您的位置:首页 > 移动开发 > Android开发

android JNI处理图片的例子

2014-03-14 14:16 281 查看
原地址:http://blog.csdn.net/xjwangliang/article/details/7065670

//下面是主程序代码

[java] view plaincopy
<pre name="code" class="java">import android.app.Activity;
import android.graphics.Bitmap;
import android.graphics.Bitmap.Config;
import android.graphics.BitmapFactory;
import android.os.Bundle;
import android.util.Log;
import android.view.View;
import android.widget.ImageView;

public class PhotoShop extends Activity {
private String tag = "IBMPhotoPhun";
private Bitmap bitmapOrig = null;
private Bitmap bitmapGray = null;
private Bitmap bitmapWip = null;
private ImageView ivDisplay = null;
// NDK STUFF
static {
try{
System.loadLibrary("haha");}
catch(Exception e) {
System.out.println("dd");
}
}
public native void convertToGray(Bitmap bitmapIn, Bitmap bitmapOut);

public native void changeBrightness(int direction, Bitmap bitmap);

public native void findEdges(Bitmap bitmapIn, Bitmap bitmapOut);

// END NDK STUFF
/** Called when the activity is first created. */
@Override
public void onCreate(Bundle savedInstanceState) {
super.onCreate(savedInstanceState);
setContentView(R.layout.main);
Log.i(tag, "before image stuff");
ivDisplay = (ImageView) findViewById(R.id.ivDisplay);
// load bitmap from resources
BitmapFactory.Options options = new BitmapFactory.Options();
// Make sure it is 24 bit color as our image processing algorithm
// expects this format
options.inPreferredConfig = Config.ARGB_8888;
bitmapOrig = BitmapFactory.decodeResource(this.getResources(),
R.drawable.sample, options);
if (bitmapOrig != null)
ivDisplay.setImageBitmap(bitmapOrig);
}

public void onResetImage(View v) {
Log.i(tag, "onResetImage");
ivDisplay.setImageBitmap(bitmapOrig);
}
// 边缘化
public void onFindEdges(View v) {
Log.i(tag, "onFindEdges");
// make sure our target bitmaps are happy
bitmapGray = Bitmap.createBitmap(bitmapOrig.getWidth(),
bitmapOrig.getHeight(), Config.ALPHA_8);
bitmapWip = Bitmap.createBitmap(bitmapOrig.getWidth(),
bitmapOrig.getHeight(), Config.ALPHA_8);
// before finding edges, we need to convert this image to gray
convertToGray(bitmapOrig, bitmapGray);
// find edges in the image
findEdges(bitmapGray, bitmapWip);
ivDisplay.setImageBitmap(bitmapWip);
}

public void onConvertToGray(View v) {
Log.i(tag, "onConvertToGray");
//创建一个和原图宽和高相等的灰色(一个像素是一个字节)图片,内容不确定或者为空(bitmapWip != null),就是说造好了格子只等填充数据
bitmapWip = Bitmap.createBitmap(bitmapOrig.getWidth(),
bitmapOrig.getHeight(), Config.ALPHA_8);
/**
* java中图片的config(相当于格式)
*  ALPHA_8     (2),
RGB_565     (4),
ARGB_4444   (5),
ARGB_8888   (6);
与android不同:
ANDROID_BITMAP_FORMAT_RGBA_8888:1
ANDROID_BITMAP_FORMAT_A_8:8
*/
convertToGray(bitmapOrig, bitmapWip);
ivDisplay.setImageBitmap(bitmapWip);
}
//先要获得bitmapWip才能处理,否则setImageBitmap(byll)虽然不会报错,但是空白的
public void onDimmer(View v) {
Log.i(tag, "onDimmer");
changeBrightness(2, bitmapWip);
ivDisplay.setImageBitmap(bitmapWip);
}
//先要获得bitmapWip才能处理,否则setImageBitmap(byll)虽然不会报错,但是空白的
public void onBrighter(View v) {
Log.i(tag, "onBrighter");
changeBrightness(1, bitmapWip);
ivDisplay.setImageBitmap(bitmapWip);
}
}

</pre><pre name="code" class="java">//下面是JNI层的代码</pre><pre name="code" class="java">//====================================================</pre><pre name="code" class="java">include <jni.h>
#include <android/log.h>
#include <android/bitmap.h>//【导入】
#define LOG_TAG "libibmphotophun"
#define LOGI(...) __android_log_print(ANDROID_LOG_INFO,LOG_TAG,__VA_ARGS__)
#define LOGE(...) __android_log_print(ANDROID_LOG_ERROR,LOG_TAG,__VA_ARGS__)
typedef struct//彩色图片数据
{
uint8_t alpha;
uint8_t red;
uint8_t green;
uint8_t blue;
} argb;
/*
convertToGray
Pixel operation
bitmapcolor:24位(实际是32)彩色图片;格式:ANDROID_BITMAP_FORMAT_RGBA_8888:
bitmapgray灰色图片(8位);格式:ANDROID_BITMAP_FORMAT_A_8
根据彩色图片填充灰色图片数据
*/
JNIEXPORT void JNICALL Java_com_msi_ibm_ndk_IBMPhotoPhun_convertToGray(JNIEnv
* env, jobject obj, jobject bitmapcolor,jobject bitmapgray)
{
AndroidBitmapInfo   infocolor;//AndroidBitmapInfo图片信息
void*               pixelscolor;//图片地址
AndroidBitmapInfo   infogray;
void*               pixelsgray;
int                 ret;
int                 y;
int                 x;
LOGI("convertToGray");

//AndroidBitmap_getInfo(env, bitmapcolor, &infocolor)得到图片信息
//AndroidBitmap_lockPixels(env, bitmapcolor, &pixelscolor)得到图片地址

if ((ret = AndroidBitmap_getInfo(env, bitmapcolor, &infocolor)) < 0) {//【AndroidBitmap_getInfo】没有图片信息(宽高等等)
LOGE("AndroidBitmap_getInfo() failed ! error=%d", ret);
return;
}
if ((ret = AndroidBitmap_getInfo(env, bitmapgray, &infogray)) < 0) {
LOGE("AndroidBitmap_getInfo() failed ! error=%d", ret);
return;
}
LOGI("color image :: width is %d; height is %d; stride is %d; format is %d;flags is%d",
infocolor.width,infocolor.height,infocolor.stride,infocolor.format,infocolor.flags);//【获得图片的宽和高、格式】

if (infocolor.format != ANDROID_BITMAP_FORMAT_RGBA_8888) {//不是24位图片
LOGE("Bitmap format is not RGBA_8888 !");
return;
}
LOGI("gray image :: width is %d; height is %d; stride is %d; format is %d;flags is
%d",infogray.width,infogray.height,infogray.stride,infogray.format,infogray.flags);

if (infogray.format != ANDROID_BITMAP_FORMAT_A_8) {//不是8位图片
LOGE("Bitmap format is not A_8 !");
return;
}
if ((ret = AndroidBitmap_lockPixels(env, bitmapcolor, &pixelscolor)) < 0) {//【锁定AndroidBitmap_lockPixels】
//锁定之后,pixelscolor指向图片的首地址 
LOGE("AndroidBitmap_lockPixels() failed ! error=%d", ret);
}
if ((ret = AndroidBitmap_lockPixels(env, bitmapgray, &pixelsgray)) < 0) {
//锁定之后,pixelscolor指向图片的首地址
LOGE("AndroidBitmap_lockPixels() failed ! error=%d", ret);
}
// modify pixels with image processing algorithm
for (y=0;y<infocolor.height;y++) {//infocolor.height获得图片的高度
//这是首次使用argb
argb * line = (argb *) pixelscolor;//每一行的首地址(刚开始是图片的首地址,即第一行的首地址,下面会换行)

uint8_t * grayline = (uint8_t *) pixelsgray;//每一行的首地址(数据类型为一个无符号字节)

for (x=0;x<infocolor.width;x++) {
grayline[x] = 0.3 * line[x].red + 0.59 * line[x].green + 0.11*line[x].blue;
}
pixelscolor = (char *)pixelscolor + infocolor.stride;   //换行:每行的首地址 + 每行的跨度 = 下一行的首地址
pixelsgray = (char *) pixelsgray + infogray.stride;     //换行:每行的首地址 + 每行的跨度 = 下一行的首地址
}
LOGI("unlocking pixels");
AndroidBitmap_unlockPixels(env, bitmapcolor);
AndroidBitmap_unlockPixels(env, bitmapgray);
}
/**
The AndroidBitmapInfo structure, defined in bitmap.h, is helpful for
learning about a Bitmap object.

The AndroidBitmap_getInfo function, found in the jnigraphics library,
obtains information about a specific Bitmap object.

a char represents a signed 8-bit value, so a char pointer (char *) allows you to reference an 8-bit
value and perform operations through that pointer. The image data is represented as uint8_t, which
means an unsigned 8-bit value, where each byte holds a value ranging from 0 to 255. A collection of
three 8-bit unsigned values represents a pixel of image data for a 24-bit image.
图象数据使用unit8_t(8位无符号,0-255),24-bit的图象就是3个8比特无符号数。

Working through an image involves working on the individual rows of data and
moving across the columns. The Bitmap structure contains a member known as the
stride. The stride represents the width, in bytes, of a row of image data. For
example, a 24-bit color plus alpha channel image has 32 bits, or 4 bytes, per pixel.
So an image with a width of 320 pixels has a stride of 320*4 or 1,280 bytes. An 8-bit
grayscale image has 8 bits, or 1 byte, per pixel. A grayscale bitmap with a width of
320 pixels has a stride of 320*1 or simply 320 bytes. With this information in mind,
let's look at the image processing algorithm for converting a color image to a
grayscale image:
Bitmap结构有一个属性stride(跨度):表示一行数据的长度。例如,24比特的图象每个像素有包含32位数据
(红绿蓝三种颜色,再加上alpha属性),所以宽度为320像素的图片的跨度(stride)为320*4字节。而rayscale
image每个像素只有8比特,就是一个字节,如果320像素,那么stride为320字节。

下面是原作者解释:
1. When the image data is "locked," the base address of the image data is
referenced by a pointer named pixelscolor for the input color image
and pixelsgray for the output grayscale image.

2. Two for-next loops allow you to iterate over the entire image.

1. First, you iterate over the height of the image, one pass per "row."
Use the infocolor.height value to get the count of the rows.

2. On each pass through the rows a pointer is set up to the memory
location corresponding to the first "column" of image data for the
row.

3. As you iterate over the columns for a particular row, you convert
each pixel of color data to a single value representing the
grayscale value.

4. When the complete row is converted you need to advance the
pointers to the next row. This is done by jumping forward in
memory by the stride value.

*/
/*
changeBrightness
Pixel Operation
改变亮度:direction=1表示增加亮度,direction=2减少亮度
bitmap只接受一个灰色图片(一个像素一个字节,格式ANDROID_BITMAP_FORMAT_A_8)
将灰色图片的像素值改变
*/
JNIEXPORT void JNICALL Java_com_msi_ibm_ndk_IBMPhotoPhun_changeBrightness(JNIEnv
* env, jobject obj, int direction,jobject bitmap)
{
AndroidBitmapInfo   infogray;
void*               pixelsgray;
int                 ret;
int                 y;
int                 x;
uint8_t         save;//好像没有用到

//=======================
if ((ret = AndroidBitmap_getInfo(env, bitmap, &infogray)) < 0) {
LOGE("AndroidBitmap_getInfo() failed ! error=%d", ret);
return;
}

LOGI("gray image :: width is %d; height is %d; stride is %d; format is %d;flags is
%d",infogray.width,infogray.height,infogray.stride,infogray.format,infogray.flags);

if (infogray.format != ANDROID_BITMAP_FORMAT_A_8) {
LOGE("Bitmap format is not A_8 !");
return;
}
if ((ret = AndroidBitmap_lockPixels(env, bitmap, &pixelsgray)) < 0) {
LOGE("AndroidBitmap_lockPixels() failed ! error=%d", ret);
}
// modify pixels with image processing algorithm
LOGI("time to modify pixels....");

for (y=0;y<infogray.height;y++) {//每个像素
uint8_t * grayline = (uint8_t *) pixelsgray;
int v;
for (x=0;x<infogray.width;x++) {
v = (int) grayline[x];
if (direction == 1)//与上面不同,灰色图片每个像素是一个字节,所以只需要改变一个字节
v -=5;
else
v += 5;

if (v >= 255) {
grayline[x] = 255;
} else if (v <= 0) {
grayline[x] = 0;
} else {
grayline[x] = (uint8_t) v;
}
}
pixelsgray = (char *) pixelsgray + infogray.stride;
}
AndroidBitmap_unlockPixels(env, bitmap);
}

/*

/*
findEdges
Matrix operation
利用bitmapgray填充bitmapedges
*/
JNIEXPORT void JNICALL Java_com_msi_ibm_ndk_IBMPhotoPhun_findEdges(JNIEnv
* env, jobject obj, jobject bitmapgray,jobject bitmapedges)
{
AndroidBitmapInfo   infogray;
void*                   pixelsgray;
AndroidBitmapInfo   infoedges;
void*                   pixelsedge;
int                 ret;
int                 y;
int                 x;
int                 sumX,sumY,sum;
int                 i,j;
int                 Gx[3][3];
int                 Gy[3][3];
uint8_t *graydata;
uint8_t *edgedata;

LOGI("findEdges running");

/**
[  -1    0    1  ]
Gx: [  -2    0    2   ]
[  -1    0    1  ]

[   1    2    1  ]
Gy: [   0    0    0   ]
[  -1   -2   -1  ]
*/
Gx[0][0] = -1;Gx[0][1] = 0;Gx[0][2] = 1;
Gx[1][0] = -2;Gx[1][1] = 0;Gx[1][2] = 2;
Gx[2][0] = -1;Gx[2][1] = 0;Gx[2][2] = 1;
Gy[0][0] = 1;Gy[0][1] = 2;Gy[0][2] = 1;
Gy[1][0] = 0;Gy[1][1] = 0;Gy[1][2] = 0;
Gy[2][0] = -1;Gy[2][1] = -2;Gy[2][2] = -1;

if ((ret = AndroidBitmap_getInfo(env, bitmapgray, &infogray)) < 0) {
LOGE("AndroidBitmap_getInfo() failed ! error=%d", ret);
return;
}
if ((ret = AndroidBitmap_getInfo(env, bitmapedges, &infoedges)) < 0) {
LOGE("AndroidBitmap_getInfo() failed ! error=%d", ret);
return;
}
LOGI("gray image :: width is %d; height is %d; stride is %d; format is %d;flags is
%d",infogray.width,infogray.height,infogray.stride,infogray.format,infogray.flags);
if (infogray.format != ANDROID_BITMAP_FORMAT_A_8) {
LOGE("Bitmap format is not A_8 !");
return;
}
LOGI("color image :: width is %d; height is %d; stride is %d; format is %d;flags is
%d",infoedges.width,infoedges.height,infoedges.stride,infoedges.format,infoedges.flags);
if (infoedges.format != ANDROID_BITMAP_FORMAT_A_8) {
LOGE("Bitmap format is not A_8 !");
return;
}
if ((ret = AndroidBitmap_lockPixels(env, bitmapgray, &pixelsgray)) < 0) {
LOGE("AndroidBitmap_lockPixels() failed ! error=%d", ret);
}
if ((ret = AndroidBitmap_lockPixels(env, bitmapedges, &pixelsedge)) < 0) {
LOGE("AndroidBitmap_lockPixels() failed ! error=%d", ret);
}
// modify pixels with image processing algorithm
LOGI("time to modify pixels....");

graydata = (uint8_t *) pixelsgray;
edgedata = (uint8_t *) pixelsedge;

for (y=0;y<=infogray.height - 1;y++) {
for (x=0;x<infogray.width -1;x++) {
sumX = 0;
sumY = 0;
// check boundaries
if (y==0 || y == infogray.height-1) {
sum = 0;
} else if (x == 0 || x == infogray.width -1) {
sum = 0;
} else {
// calc X gradient
for (i=-1;i<=1;i++) {
for (j=-1;j<=1;j++) {
sumX += (int) ( (*(graydata + x + i + (y + j)* infogray.stride)) * Gx[i+1][j+1]);
}
}
// calc Y gradient
for (i=-1;i<=1;i++) {
for (j=-1;j<=1;j++) {
sumY += (int) ( (*(graydata + x + i + (y + j)* infogray.stride)) * Gy[i+1][j+1]);
}
}
sum = abs(sumX) + abs(sumY);
}
if (sum>255) sum = 255;
if (sum<0) sum = 0;
*(edgedata + x + y*infogray.width) = 255 - (uint8_t) sum;//给边界图片bitmapedges填充数据
}
}
AndroidBitmap_unlockPixels(env, bitmapgray);
AndroidBitmap_unlockPixels(env, bitmapedges);
</pre><pre name="code" class="java">}//结束</pre><br><img src="http://hi.csdn.net/attachment/201112/13/0_1323741853c9Ch.gif" alt=""><img src="http://hi.csdn.net/attachment/201112/13/0_1323741880F49S.gif" alt=""><br>
内容来自用户分享和网络整理,不保证内容的准确性,如有侵权内容,可联系管理员处理 点击这里给我发消息
标签: