summaryrefslogtreecommitdiffstats
path: root/libavcodec
diff options
context:
space:
mode:
authorReimar Döffinger <Reimar.Doeffinger@gmx.de>2009-05-31 09:09:32 +0000
committerReimar Döffinger <Reimar.Doeffinger@gmx.de>2009-05-31 09:09:32 +0000
commit8c18e490c4f4d7c109f27b77f4502b5e612e62b4 (patch)
tree22697abb946f801a582032550eaed17821a18cba /libavcodec
parent55619690e816b378937e2c86188aa61c30b78ae8 (diff)
downloadffmpeg-streaming-8c18e490c4f4d7c109f27b77f4502b5e612e62b4.zip
ffmpeg-streaming-8c18e490c4f4d7c109f27b77f4502b5e612e62b4.tar.gz
Make lcldec produce YUV output when the input file is coded like that, instead
of having it do its own inefficient fixed-point YUV to RGB conversion. Originally committed as revision 19033 to svn://svn.ffmpeg.org/ffmpeg/trunk
Diffstat (limited to 'libavcodec')
-rw-r--r--libavcodec/lcldec.c162
1 files changed, 59 insertions, 103 deletions
diff --git a/libavcodec/lcldec.c b/libavcodec/lcldec.c
index 6c97c6b..c8a5ee6 100644
--- a/libavcodec/lcldec.c
+++ b/libavcodec/lcldec.c
@@ -71,42 +71,6 @@ typedef struct LclDecContext {
} LclDecContext;
-/*
- *
- * Helper functions
- *
- */
-static inline unsigned char fix (int pix14)
-{
- int tmp;
-
- tmp = (pix14 + 0x80000) >> 20;
- return av_clip_uint8(tmp);
-}
-
-
-
-static inline unsigned char get_b (unsigned char yq, signed char bq)
-{
- return fix((yq << 20) + bq * 1858076);
-}
-
-
-
-static inline unsigned char get_g (unsigned char yq, signed char bq, signed char rq)
-{
- return fix((yq << 20) - bq * 360857 - rq * 748830);
-}
-
-
-
-static inline unsigned char get_r (unsigned char yq, signed char rq)
-{
- return fix((yq << 20) + rq * 1470103);
-}
-
-
-
static unsigned int mszh_decomp(unsigned char * srcptr, int srclen, unsigned char * destptr, unsigned int destsize)
{
unsigned char *destptr_bak = destptr;
@@ -166,6 +130,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, AVPac
unsigned int pixel_ptr;
int row, col;
unsigned char *outptr;
+ uint8_t *y_out, *u_out, *v_out;
unsigned int width = avctx->width; // Real image width
unsigned int height = avctx->height; // Real image height
unsigned int mszh_dlen;
@@ -397,36 +362,35 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, AVPac
}
/* Convert colorspace */
+ y_out = c->pic.data[0] + (height - 1) * c->pic.linesize[0];
+ u_out = c->pic.data[1] + (height - 1) * c->pic.linesize[1];
+ v_out = c->pic.data[2] + (height - 1) * c->pic.linesize[2];
switch (c->imgtype) {
case IMGTYPE_YUV111:
- for (row = height - 1; row >= 0; row--) {
- pixel_ptr = row * c->pic.linesize[0];
+ for (row = 0; row < height; row++) {
for (col = 0; col < width; col++) {
- outptr[pixel_ptr++] = get_b(encoded[0], encoded[1]);
- outptr[pixel_ptr++] = get_g(encoded[0], encoded[1], encoded[2]);
- outptr[pixel_ptr++] = get_r(encoded[0], encoded[2]);
- encoded += 3;
+ y_out[col] = *encoded++;
+ u_out[col] = *encoded++ + 128;
+ v_out[col] = *encoded++ + 128;
}
+ y_out -= c->pic.linesize[0];
+ u_out -= c->pic.linesize[1];
+ v_out -= c->pic.linesize[2];
}
break;
case IMGTYPE_YUV422:
- for (row = height - 1; row >= 0; row--) {
- pixel_ptr = row * c->pic.linesize[0];
- for (col = 0; col < width/4; col++) {
- outptr[pixel_ptr++] = get_b(encoded[0], encoded[4]);
- outptr[pixel_ptr++] = get_g(encoded[0], encoded[4], encoded[6]);
- outptr[pixel_ptr++] = get_r(encoded[0], encoded[6]);
- outptr[pixel_ptr++] = get_b(encoded[1], encoded[4]);
- outptr[pixel_ptr++] = get_g(encoded[1], encoded[4], encoded[6]);
- outptr[pixel_ptr++] = get_r(encoded[1], encoded[6]);
- outptr[pixel_ptr++] = get_b(encoded[2], encoded[5]);
- outptr[pixel_ptr++] = get_g(encoded[2], encoded[5], encoded[7]);
- outptr[pixel_ptr++] = get_r(encoded[2], encoded[7]);
- outptr[pixel_ptr++] = get_b(encoded[3], encoded[5]);
- outptr[pixel_ptr++] = get_g(encoded[3], encoded[5], encoded[7]);
- outptr[pixel_ptr++] = get_r(encoded[3], encoded[7]);
- encoded += 8;
+ for (row = 0; row < height; row++) {
+ for (col = 0; col < width - 3; col += 4) {
+ memcpy(y_out + col, encoded, 4);
+ encoded += 4;
+ u_out[ col >> 1 ] = *encoded++ + 128;
+ u_out[(col >> 1) + 1] = *encoded++ + 128;
+ v_out[ col >> 1 ] = *encoded++ + 128;
+ v_out[(col >> 1) + 1] = *encoded++ + 128;
}
+ y_out -= c->pic.linesize[0];
+ u_out -= c->pic.linesize[1];
+ v_out -= c->pic.linesize[2];
}
break;
case IMGTYPE_RGB24:
@@ -437,58 +401,46 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, AVPac
}
break;
case IMGTYPE_YUV411:
- for (row = height - 1; row >= 0; row--) {
- pixel_ptr = row * c->pic.linesize[0];
- for (col = 0; col < width/4; col++) {
- outptr[pixel_ptr++] = get_b(encoded[0], encoded[4]);
- outptr[pixel_ptr++] = get_g(encoded[0], encoded[4], encoded[5]);
- outptr[pixel_ptr++] = get_r(encoded[0], encoded[5]);
- outptr[pixel_ptr++] = get_b(encoded[1], encoded[4]);
- outptr[pixel_ptr++] = get_g(encoded[1], encoded[4], encoded[5]);
- outptr[pixel_ptr++] = get_r(encoded[1], encoded[5]);
- outptr[pixel_ptr++] = get_b(encoded[2], encoded[4]);
- outptr[pixel_ptr++] = get_g(encoded[2], encoded[4], encoded[5]);
- outptr[pixel_ptr++] = get_r(encoded[2], encoded[5]);
- outptr[pixel_ptr++] = get_b(encoded[3], encoded[4]);
- outptr[pixel_ptr++] = get_g(encoded[3], encoded[4], encoded[5]);
- outptr[pixel_ptr++] = get_r(encoded[3], encoded[5]);
- encoded += 6;
+ for (row = 0; row < height; row++) {
+ for (col = 0; col < width - 3; col += 4) {
+ memcpy(y_out + col, encoded, 4);
+ encoded += 4;
+ u_out[col >> 2] = *encoded++ + 128;
+ v_out[col >> 2] = *encoded++ + 128;
}
+ y_out -= c->pic.linesize[0];
+ u_out -= c->pic.linesize[1];
+ v_out -= c->pic.linesize[2];
}
break;
case IMGTYPE_YUV211:
- for (row = height - 1; row >= 0; row--) {
- pixel_ptr = row * c->pic.linesize[0];
- for (col = 0; col < width/2; col++) {
- outptr[pixel_ptr++] = get_b(encoded[0], encoded[2]);
- outptr[pixel_ptr++] = get_g(encoded[0], encoded[2], encoded[3]);
- outptr[pixel_ptr++] = get_r(encoded[0], encoded[3]);
- outptr[pixel_ptr++] = get_b(encoded[1], encoded[2]);
- outptr[pixel_ptr++] = get_g(encoded[1], encoded[2], encoded[3]);
- outptr[pixel_ptr++] = get_r(encoded[1], encoded[3]);
- encoded += 4;
+ for (row = 0; row < height; row++) {
+ for (col = 0; col < width - 1; col += 2) {
+ memcpy(y_out + col, encoded, 2);
+ encoded += 2;
+ u_out[col >> 1] = *encoded++ + 128;
+ v_out[col >> 1] = *encoded++ + 128;
}
+ y_out -= c->pic.linesize[0];
+ u_out -= c->pic.linesize[1];
+ v_out -= c->pic.linesize[2];
}
break;
case IMGTYPE_YUV420:
- for (row = height / 2 - 1; row >= 0; row--) {
- pixel_ptr = 2 * row * c->pic.linesize[0];
- for (col = 0; col < width/2; col++) {
- outptr[pixel_ptr] = get_b(encoded[0], encoded[4]);
- outptr[pixel_ptr+1] = get_g(encoded[0], encoded[4], encoded[5]);
- outptr[pixel_ptr+2] = get_r(encoded[0], encoded[5]);
- outptr[pixel_ptr+3] = get_b(encoded[1], encoded[4]);
- outptr[pixel_ptr+4] = get_g(encoded[1], encoded[4], encoded[5]);
- outptr[pixel_ptr+5] = get_r(encoded[1], encoded[5]);
- outptr[pixel_ptr-c->pic.linesize[0]] = get_b(encoded[2], encoded[4]);
- outptr[pixel_ptr-c->pic.linesize[0]+1] = get_g(encoded[2], encoded[4], encoded[5]);
- outptr[pixel_ptr-c->pic.linesize[0]+2] = get_r(encoded[2], encoded[5]);
- outptr[pixel_ptr-c->pic.linesize[0]+3] = get_b(encoded[3], encoded[4]);
- outptr[pixel_ptr-c->pic.linesize[0]+4] = get_g(encoded[3], encoded[4], encoded[5]);
- outptr[pixel_ptr-c->pic.linesize[0]+5] = get_r(encoded[3], encoded[5]);
- pixel_ptr += 6;
- encoded += 6;
+ u_out = c->pic.data[1] + ((height >> 1) - 1) * c->pic.linesize[1];
+ v_out = c->pic.data[2] + ((height >> 1) - 1) * c->pic.linesize[2];
+ for (row = 0; row < height - 1; row += 2) {
+ for (col = 0; col < width - 1; col += 2) {
+ memcpy(y_out + col, encoded, 2);
+ encoded += 2;
+ memcpy(y_out + col - c->pic.linesize[0], encoded, 2);
+ encoded += 2;
+ u_out[col >> 1] = *encoded++ + 128;
+ v_out[col >> 1] = *encoded++ + 128;
}
+ y_out -= c->pic.linesize[0] << 1;
+ u_out -= c->pic.linesize[1];
+ v_out -= c->pic.linesize[2];
}
break;
default:
@@ -543,31 +495,37 @@ static av_cold int decode_init(AVCodecContext *avctx)
case IMGTYPE_YUV111:
c->decomp_size = basesize * 3;
max_decomp_size = max_basesize * 3;
+ avctx->pix_fmt = PIX_FMT_YUV444P;
av_log(avctx, AV_LOG_INFO, "Image type is YUV 1:1:1.\n");
break;
case IMGTYPE_YUV422:
c->decomp_size = basesize * 2;
max_decomp_size = max_basesize * 2;
+ avctx->pix_fmt = PIX_FMT_YUV422P;
av_log(avctx, AV_LOG_INFO, "Image type is YUV 4:2:2.\n");
break;
case IMGTYPE_RGB24:
c->decomp_size = basesize * 3;
max_decomp_size = max_basesize * 3;
+ avctx->pix_fmt = PIX_FMT_BGR24;
av_log(avctx, AV_LOG_INFO, "Image type is RGB 24.\n");
break;
case IMGTYPE_YUV411:
c->decomp_size = basesize / 2 * 3;
max_decomp_size = max_basesize / 2 * 3;
+ avctx->pix_fmt = PIX_FMT_YUV411P;
av_log(avctx, AV_LOG_INFO, "Image type is YUV 4:1:1.\n");
break;
case IMGTYPE_YUV211:
c->decomp_size = basesize * 2;
max_decomp_size = max_basesize * 2;
+ avctx->pix_fmt = PIX_FMT_YUV422P;
av_log(avctx, AV_LOG_INFO, "Image type is YUV 2:1:1.\n");
break;
case IMGTYPE_YUV420:
c->decomp_size = basesize / 2 * 3;
max_decomp_size = max_basesize / 2 * 3;
+ avctx->pix_fmt = PIX_FMT_YUV420P;
av_log(avctx, AV_LOG_INFO, "Image type is YUV 4:2:0.\n");
break;
default:
@@ -657,8 +615,6 @@ static av_cold int decode_init(AVCodecContext *avctx)
#endif
}
- avctx->pix_fmt = PIX_FMT_BGR24;
-
return 0;
}
OpenPOWER on IntegriCloud