Android摄像头开发:实时摄像头视频预览帧的编码问题(二)

Android开发:实时处理摄像头预览帧视频------浅析PreviewCallback,onPreviewFrame,AsyncTask的综合应用 这里将大致框架介绍了,但很多人对onPreviewFrame()里的处理提出质疑。认为下面的转换是多余的:

            final YuvImage image = new YuvImage(mData, ImageFormat.NV21, w, h, null);
            ByteArrayOutputStream os = new ByteArrayOutputStream(mData.length);
            if(!image.compressToJpeg(new Rect(0, 0, w, h), 100, os)){
                return null;
            }
            byte[] tmp = os.toByteArray();
            Bitmap bmp = BitmapFactory.decodeByteArray(tmp, 0,tmp.length); 


因为这个mData是byte[ ]格式,转换流程是:byte[ ]---YuvImage----ByteArrayOutputStream---byte[ ]-----Bitmap。乍一看这个转换还真是多余了。看看看goolge的api:

public abstract void onPreviewFrame (byte[] data, Camera camera)

Added in API level 1
Called as preview frames are displayed. This callback is invoked on the event thread open(int) was called from.

If using the YV12 format, refer to the equations in setPreviewFormat(int) for the arrangement of the pixel data in the preview callback buffers.

Parameters
data	the contents of the preview frame in the format defined by ImageFormat, which can be queried with getPreviewFormat(). If setPreviewFormat(int) is never called, the default will be the YCbCr_420_SP (NV21) format.
camera	the Camera service object.


大致意思是:可以用getPreviewFormat()查询支持的预览帧格式。如果setPreviewFormat(INT) 从未被调用,默认将使用YCbCr_420_SP的格式(NV21)。

setPreviewFormat里,它又说:

public void setPreviewFormat (int pixel_format)

Added in API level 1
Sets the image format for preview pictures.

If this is never called, the default format will be NV21, which uses the NV21 encoding format.

Use getSupportedPreviewFormats() to get a list of the available preview formats.

It is strongly recommended that either NV21 or YV12 is used, since they are supported by all camera devices.

For YV12, the image buffer that is received is not necessarily tightly packed, as there may be padding at the end of each row of pixel data, as described in YV12. For camera callback data, it can be assumed that the stride of the Y and UV data is the smallest possible that meets the alignment requirements. That is, if the preview size is width x height, then the following equations describe the buffer index for the beginning of row y for the Y plane and row c for the U and V planes: 

 yStride   = (int) ceil(width / 16.0) * 16;
 uvStride  = (int) ceil( (yStride / 2) / 16.0) * 16;
 ySize     = yStride * height;
 uvSize    = uvStride * height / 2;
 yRowIndex = yStride * y;
 uRowIndex = ySize + uvSize + uvStride * c;
 vRowIndex = ySize + uvStride * c;
 size      = ySize + uvSize * 2;

强烈建议使用NV21格式和YV21格式,而默认情况下是NV21格式,也就是YUV420SP的。因此不经过转换,直接用BitmapFactory解析是不能成功的。事实也是如此。直接解析mData将会得到如下的错误:

 

另外下面也提到NV21是通用的。

getSupportedPreviewFormats ()

 

Added in API level 5

Gets the supported preview formats. NV21 is always supported. YV12 is always supported since API level 12.

 

 

如果嫌YuvImage进行压缩解析的慢,只能自己写转换函数了,网上常见的有三种:

一:这里只是一个编码框架

参考这里:Android 实时视频采集—Camera预览采集

// 【获取视频预览帧的接口】
   mJpegPreviewCallback = new Camera.PreviewCallback()
   {
    @Override
    public void onPreviewFrame(byte[] data, Camera camera) 
    {
     //传递进来的data,默认是YUV420SP的
 // TODO Auto-generated method stub    
     try
     {
      Log.i(TAG, "going into onPreviewFrame");
      //mYUV420sp = data;   // 获取原生的YUV420SP数据
      YUVIMGLEN = data.length;
      
      // 拷贝原生yuv420sp数据
      mYuvBufferlock.acquire();
      System.arraycopy(data, 0, mYUV420SPSendBuffer, 0, data.length);
      //System.arraycopy(data, 0, mWrtieBuffer, 0, data.length);
      mYuvBufferlock.release();
      
      // 开启编码线程,如开启PEG编码方式线程
      mSendThread1.start();
      
     } catch (Exception e)
     {
      Log.v("System.out", e.toString());
     }// endtry    
    }// endonPriview  
   };

 

二、下面是将yuv420sp转成rgb参考这里:android视频采集

private void updateIM() {
try {
// 解析YUV成RGB格式
decodeYUV420SP(byteArray, yuv420sp, width, height);
DataBuffer dataBuffer = new DataBufferByte(byteArray, numBands);
WritableRaster wr = Raster.createWritableRaster(sampleModel,
dataBuffer, new Point(0, 0));
im = new BufferedImage(cm, wr, false, null);
} catch (Exception ex) {
ex.printStackTrace();
}
}

private static void decodeYUV420SP(byte[] rgbBuf, byte[] yuv420sp,
int width, int height) {
final int frameSize = width * height;
if (rgbBuf == null)
throw new NullPointerException("buffer 'rgbBuf' is null");
if (rgbBuf.length < frameSize * 3)
throw new IllegalArgumentException("buffer 'rgbBuf' size "
+ rgbBuf.length + " < minimum " + frameSize * 3);

if (yuv420sp == null)
throw new NullPointerException("buffer 'yuv420sp' is null");

if (yuv420sp.length < frameSize * 3 / 2)
throw new IllegalArgumentException("buffer 'yuv420sp' size "
+ yuv420sp.length + " < minimum " + frameSize * 3 / 2);

int i = 0, y = 0;
int uvp = 0, u = 0, v = 0;
int y1192 = 0, r = 0, g = 0, b = 0;

for (int j = 0, yp = 0; j < height; j++) {
uvp = frameSize + (j >> 1) * width;
u = 0;
v = 0;
for (i = 0; i < width; i++, yp++) {
y = (0xff & ((int) yuv420sp[yp])) - 16;
if (y < 0)
y = 0;
if ((i & 1) == 0) {
v = (0xff & yuv420sp[uvp++]) - 128;
u = (0xff & yuv420sp[uvp++]) - 128;
}

y1192 = 1192 * y;
r = (y1192 + 1634 * v);
g = (y1192 - 833 * v - 400 * u);
b = (y1192 + 2066 * u);

if (r < 0)
r = 0;
else if (r > 262143)
r = 262143;
if (g < 0)
g = 0;
else if (g > 262143)
g = 262143;
if (b < 0)
b = 0;
else if (b > 262143)
b = 262143;

rgbBuf[yp * 3] = (byte) (r >> 10);
rgbBuf[yp * 3 + 1] = (byte) (g >> 10);
rgbBuf[yp * 3 + 2] = (byte) (b >> 10);
}
}
}

public static void main(String[] args) {
Frame f = new FlushMe();
}
}

 

三、将YUV420SP转成YUV420格式

    参考这里:Android如何实现边采集边上传

 

private byte[] changeYUV420SP2P(byte[]data,int length){

int width = 176;

int height = 144;

byte[] str = new byte[length];

System.arraycopy(data, 0, str, 0,width*height);



int strIndex = width*height;



for(int i = width*height+1; i < length ;i+=2)

{

str[strIndex++] = data[i];

}





for(int i = width*height;i<length;i+=2)

{

str[strIndex++] = data[i];

}

return str;

} 


至于怎么从YUV420SP中直接提取出Y分量进行后续检测,这个还要研究一番。有知道的大神多赐教。
  

----------------------------------------------------------------------------------------本文系原创,转载请注明作者:yanzi1225627

欢迎android爱好者加群248217350,备注:yanzi

 

发布了376 篇原创文章 · 获赞 2296 · 访问量 699万+
展开阅读全文

android byte转图片的问题

12-07

```public class MainActivity extends Activity { private ImageView iv1,iv2; @Override protected void onCreate(Bundle savedInstanceState) { super.onCreate(savedInstanceState); setContentView(R.layout.activity_main); initview(); //获取iv1的图片并将图片转成byte Bitmap bitmap = ((BitmapDrawable)iv1.getDrawable()).getBitmap(); ByteArrayOutputStream baos = new ByteArrayOutputStream(); bitmap.compress(Bitmap.CompressFormat.JPEG, 100, baos); byte[] byteArray = baos.toByteArray(); String sendString; try { //将byte转为String sendString = new String(byteArray , "UTF-8" ); try { //将String转回byte byte[] data = sendString.getBytes("UTF-8"); // 为UTF8编吗 // 把二进制图片转成位图 YuvImage yuvimage = new YuvImage(data, ImageFormat.NV21, 100, 100, null); // 20、20分别是图的宽度与高度 ByteArrayOutputStream baos2 = new ByteArrayOutputStream(); // 80--JPG图片的质量[0-100],100最高 yuvimage.compressToJpeg(new Rect(0, 0, 20, 20), 80, baos2); byte[] jdata = baos2.toByteArray(); Bitmap bitmap2 = BitmapFactory.decodeByteArray(jdata, 0, jdata.length); iv2.setImageBitmap(bitmap2); } catch (UnsupportedEncodingException e) { // TODO Auto-generated catch block e.printStackTrace(); } } catch (UnsupportedEncodingException e1) { // TODO Auto-generated catch block e1.printStackTrace(); } } private void initview() { // TODO Auto-generated method stub iv1=(ImageView) findViewById(R.id.imageView1); iv2=(ImageView) findViewById(R.id.imageView2); } } ``` ``` ![图片说明](https://img-ask.csdn.net/upload/201512/07/1449466888_815004.png) 转换完成之后图片不一样,这是怎么回事,求大神解惑 问答

没有更多推荐了,返回首页

©️2019 CSDN 皮肤主题: 大白 设计师: CSDN官方博客

分享到微信朋友圈

×

扫一扫,手机浏览