diff options
Diffstat (limited to 'y4minput.c')
-rw-r--r-- | y4minput.c | 784 |
1 files changed, 411 insertions, 373 deletions
diff --git a/y4minput.c b/y4minput.c index ff9ffbcc4..47f005a95 100644 --- a/y4minput.c +++ b/y4minput.c @@ -14,7 +14,7 @@ #include <string.h> #include "y4minput.h" -static int y4m_parse_tags(y4m_input *_y4m,char *_tags){ +static int y4m_parse_tags(y4m_input *_y4m, char *_tags) { int got_w; int got_h; int got_fps; @@ -23,55 +23,61 @@ static int y4m_parse_tags(y4m_input *_y4m,char *_tags){ int got_chroma; char *p; char *q; - got_w=got_h=got_fps=got_interlace=got_par=got_chroma=0; - for(p=_tags;;p=q){ + got_w = got_h = got_fps = got_interlace = got_par = got_chroma = 0; + for (p = _tags;; p = q) { /*Skip any leading spaces.*/ - while(*p==' ')p++; + while (*p == ' ')p++; /*If that's all we have, stop.*/ - if(p[0]=='\0')break; + if (p[0] == '\0')break; /*Find the end of this tag.*/ - for(q=p+1;*q!='\0'&&*q!=' ';q++); + for (q = p + 1; *q != '\0' && *q != ' '; q++); /*Process the tag.*/ - switch(p[0]){ - case 'W':{ - if(sscanf(p+1,"%d",&_y4m->pic_w)!=1)return -1; - got_w=1; - }break; - case 'H':{ - if(sscanf(p+1,"%d",&_y4m->pic_h)!=1)return -1; - got_h=1; - }break; - case 'F':{ - if(sscanf(p+1,"%d:%d",&_y4m->fps_n,&_y4m->fps_d)!=2){ + switch (p[0]) { + case 'W': { + if (sscanf(p + 1, "%d", &_y4m->pic_w) != 1)return -1; + got_w = 1; + } + break; + case 'H': { + if (sscanf(p + 1, "%d", &_y4m->pic_h) != 1)return -1; + got_h = 1; + } + break; + case 'F': { + if (sscanf(p + 1, "%d:%d", &_y4m->fps_n, &_y4m->fps_d) != 2) { return -1; } - got_fps=1; - }break; - case 'I':{ - _y4m->interlace=p[1]; - got_interlace=1; - }break; - case 'A':{ - if(sscanf(p+1,"%d:%d",&_y4m->par_n,&_y4m->par_d)!=2){ + got_fps = 1; + } + break; + case 'I': { + _y4m->interlace = p[1]; + got_interlace = 1; + } + break; + case 'A': { + if (sscanf(p + 1, "%d:%d", &_y4m->par_n, &_y4m->par_d) != 2) { return -1; } - got_par=1; - }break; - case 'C':{ - if(q-p>16)return -1; - memcpy(_y4m->chroma_type,p+1,q-p-1); - _y4m->chroma_type[q-p-1]='\0'; - got_chroma=1; - }break; + got_par = 1; + } + break; + case 'C': { + if (q - p > 16)return -1; + memcpy(_y4m->chroma_type, p + 1, q - p - 1); + _y4m->chroma_type[q - p - 1] = '\0'; + got_chroma = 1; + } + break; /*Ignore unknown tags.*/ } } - if(!got_w||!got_h||!got_fps)return -1; - if(!got_interlace)_y4m->interlace='?'; - if(!got_par)_y4m->par_n=_y4m->par_d=0; + if (!got_w || !got_h || !got_fps)return -1; + if (!got_interlace)_y4m->interlace = '?'; + if (!got_par)_y4m->par_n = _y4m->par_d = 0; /*Chroma-type is not specified in older files, e.g., those generated by mplayer.*/ - if(!got_chroma)strcpy(_y4m->chroma_type,"420"); + if (!got_chroma)strcpy(_y4m->chroma_type, "420"); return 0; } @@ -145,48 +151,48 @@ static int y4m_parse_tags(y4m_input *_y4m,char *_tags){ lines, and they are vertically co-sited with the luma samples in both the mpeg2 and jpeg cases (thus requiring no vertical resampling).*/ static void y4m_42xmpeg2_42xjpeg_helper(unsigned char *_dst, - const unsigned char *_src,int _c_w,int _c_h){ + const unsigned char *_src, int _c_w, int _c_h) { int y; int x; - for(y=0;y<_c_h;y++){ + for (y = 0; y < _c_h; y++) { /*Filter: [4 -17 114 35 -9 1]/128, derived from a 6-tap Lanczos window.*/ - for(x=0;x<OC_MINI(_c_w,2);x++){ - _dst[x]=(unsigned char)OC_CLAMPI(0,(4*_src[0]-17*_src[OC_MAXI(x-1,0)]+ - 114*_src[x]+35*_src[OC_MINI(x+1,_c_w-1)]-9*_src[OC_MINI(x+2,_c_w-1)]+ - _src[OC_MINI(x+3,_c_w-1)]+64)>>7,255); + for (x = 0; x < OC_MINI(_c_w, 2); x++) { + _dst[x] = (unsigned char)OC_CLAMPI(0, (4 * _src[0] - 17 * _src[OC_MAXI(x - 1, 0)] + + 114 * _src[x] + 35 * _src[OC_MINI(x + 1, _c_w - 1)] - 9 * _src[OC_MINI(x + 2, _c_w - 1)] + + _src[OC_MINI(x + 3, _c_w - 1)] + 64) >> 7, 255); } - for(;x<_c_w-3;x++){ - _dst[x]=(unsigned char)OC_CLAMPI(0,(4*_src[x-2]-17*_src[x-1]+ - 114*_src[x]+35*_src[x+1]-9*_src[x+2]+_src[x+3]+64)>>7,255); + for (; x < _c_w - 3; x++) { + _dst[x] = (unsigned char)OC_CLAMPI(0, (4 * _src[x - 2] - 17 * _src[x - 1] + + 114 * _src[x] + 35 * _src[x + 1] - 9 * _src[x + 2] + _src[x + 3] + 64) >> 7, 255); } - for(;x<_c_w;x++){ - _dst[x]=(unsigned char)OC_CLAMPI(0,(4*_src[x-2]-17*_src[x-1]+ - 114*_src[x]+35*_src[OC_MINI(x+1,_c_w-1)]-9*_src[OC_MINI(x+2,_c_w-1)]+ - _src[_c_w-1]+64)>>7,255); + for (; x < _c_w; x++) { + _dst[x] = (unsigned char)OC_CLAMPI(0, (4 * _src[x - 2] - 17 * _src[x - 1] + + 114 * _src[x] + 35 * _src[OC_MINI(x + 1, _c_w - 1)] - 9 * _src[OC_MINI(x + 2, _c_w - 1)] + + _src[_c_w - 1] + 64) >> 7, 255); } - _dst+=_c_w; - _src+=_c_w; + _dst += _c_w; + _src += _c_w; } } /*Handles both 422 and 420mpeg2 to 422jpeg and 420jpeg, respectively.*/ -static void y4m_convert_42xmpeg2_42xjpeg(y4m_input *_y4m,unsigned char *_dst, - unsigned char *_aux){ +static void y4m_convert_42xmpeg2_42xjpeg(y4m_input *_y4m, unsigned char *_dst, + unsigned char *_aux) { int c_w; int c_h; int c_sz; int pli; /*Skip past the luma data.*/ - _dst+=_y4m->pic_w*_y4m->pic_h; + _dst += _y4m->pic_w * _y4m->pic_h; /*Compute the size of each chroma plane.*/ - c_w=(_y4m->pic_w+_y4m->dst_c_dec_h-1)/_y4m->dst_c_dec_h; - c_h=(_y4m->pic_h+_y4m->dst_c_dec_v-1)/_y4m->dst_c_dec_v; - c_sz=c_w*c_h; - for(pli=1;pli<3;pli++){ - y4m_42xmpeg2_42xjpeg_helper(_dst,_aux,c_w,c_h); - _dst+=c_sz; - _aux+=c_sz; + c_w = (_y4m->pic_w + _y4m->dst_c_dec_h - 1) / _y4m->dst_c_dec_h; + c_h = (_y4m->pic_h + _y4m->dst_c_dec_v - 1) / _y4m->dst_c_dec_v; + c_sz = c_w * c_h; + for (pli = 1; pli < 3; pli++) { + y4m_42xmpeg2_42xjpeg_helper(_dst, _aux, c_w, c_h); + _dst += c_sz; + _aux += c_sz; } } @@ -233,8 +239,8 @@ static void y4m_convert_42xmpeg2_42xjpeg(y4m_input *_y4m,unsigned char *_dst, the chroma plane's resolution) to the right. Then we use another filter to move the C_r location down one quarter pixel, and the C_b location up one quarter pixel.*/ -static void y4m_convert_42xpaldv_42xjpeg(y4m_input *_y4m,unsigned char *_dst, - unsigned char *_aux){ +static void y4m_convert_42xpaldv_42xjpeg(y4m_input *_y4m, unsigned char *_dst, + unsigned char *_aux) { unsigned char *tmp; int c_w; int c_h; @@ -243,69 +249,71 @@ static void y4m_convert_42xpaldv_42xjpeg(y4m_input *_y4m,unsigned char *_dst, int y; int x; /*Skip past the luma data.*/ - _dst+=_y4m->pic_w*_y4m->pic_h; + _dst += _y4m->pic_w * _y4m->pic_h; /*Compute the size of each chroma plane.*/ - c_w=(_y4m->pic_w+1)/2; - c_h=(_y4m->pic_h+_y4m->dst_c_dec_h-1)/_y4m->dst_c_dec_h; - c_sz=c_w*c_h; - tmp=_aux+2*c_sz; - for(pli=1;pli<3;pli++){ + c_w = (_y4m->pic_w + 1) / 2; + c_h = (_y4m->pic_h + _y4m->dst_c_dec_h - 1) / _y4m->dst_c_dec_h; + c_sz = c_w * c_h; + tmp = _aux + 2 * c_sz; + for (pli = 1; pli < 3; pli++) { /*First do the horizontal re-sampling. This is the same as the mpeg2 case, except that after the horizontal case, we need to apply a second vertical filter.*/ - y4m_42xmpeg2_42xjpeg_helper(tmp,_aux,c_w,c_h); - _aux+=c_sz; - switch(pli){ - case 1:{ + y4m_42xmpeg2_42xjpeg_helper(tmp, _aux, c_w, c_h); + _aux += c_sz; + switch (pli) { + case 1: { /*Slide C_b up a quarter-pel. This is the same filter used above, but in the other order.*/ - for(x=0;x<c_w;x++){ - for(y=0;y<OC_MINI(c_h,3);y++){ - _dst[y*c_w]=(unsigned char)OC_CLAMPI(0,(tmp[0] - -9*tmp[OC_MAXI(y-2,0)*c_w]+35*tmp[OC_MAXI(y-1,0)*c_w] - +114*tmp[y*c_w]-17*tmp[OC_MINI(y+1,c_h-1)*c_w] - +4*tmp[OC_MINI(y+2,c_h-1)*c_w]+64)>>7,255); + for (x = 0; x < c_w; x++) { + for (y = 0; y < OC_MINI(c_h, 3); y++) { + _dst[y * c_w] = (unsigned char)OC_CLAMPI(0, (tmp[0] + - 9 * tmp[OC_MAXI(y - 2, 0) * c_w] + 35 * tmp[OC_MAXI(y - 1, 0) * c_w] + + 114 * tmp[y * c_w] - 17 * tmp[OC_MINI(y + 1, c_h - 1) * c_w] + + 4 * tmp[OC_MINI(y + 2, c_h - 1) * c_w] + 64) >> 7, 255); } - for(;y<c_h-2;y++){ - _dst[y*c_w]=(unsigned char)OC_CLAMPI(0,(tmp[(y-3)*c_w] - -9*tmp[(y-2)*c_w]+35*tmp[(y-1)*c_w]+114*tmp[y*c_w] - -17*tmp[(y+1)*c_w]+4*tmp[(y+2)*c_w]+64)>>7,255); + for (; y < c_h - 2; y++) { + _dst[y * c_w] = (unsigned char)OC_CLAMPI(0, (tmp[(y - 3) * c_w] + - 9 * tmp[(y - 2) * c_w] + 35 * tmp[(y - 1) * c_w] + 114 * tmp[y * c_w] + - 17 * tmp[(y + 1) * c_w] + 4 * tmp[(y + 2) * c_w] + 64) >> 7, 255); } - for(;y<c_h;y++){ - _dst[y*c_w]=(unsigned char)OC_CLAMPI(0,(tmp[(y-3)*c_w] - -9*tmp[(y-2)*c_w]+35*tmp[(y-1)*c_w]+114*tmp[y*c_w] - -17*tmp[OC_MINI(y+1,c_h-1)*c_w]+4*tmp[(c_h-1)*c_w]+64)>>7,255); + for (; y < c_h; y++) { + _dst[y * c_w] = (unsigned char)OC_CLAMPI(0, (tmp[(y - 3) * c_w] + - 9 * tmp[(y - 2) * c_w] + 35 * tmp[(y - 1) * c_w] + 114 * tmp[y * c_w] + - 17 * tmp[OC_MINI(y + 1, c_h - 1) * c_w] + 4 * tmp[(c_h - 1) * c_w] + 64) >> 7, 255); } _dst++; tmp++; } - _dst+=c_sz-c_w; - tmp-=c_w; - }break; - case 2:{ + _dst += c_sz - c_w; + tmp -= c_w; + } + break; + case 2: { /*Slide C_r down a quarter-pel. This is the same as the horizontal filter.*/ - for(x=0;x<c_w;x++){ - for(y=0;y<OC_MINI(c_h,2);y++){ - _dst[y*c_w]=(unsigned char)OC_CLAMPI(0,(4*tmp[0] - -17*tmp[OC_MAXI(y-1,0)*c_w]+114*tmp[y*c_w] - +35*tmp[OC_MINI(y+1,c_h-1)*c_w]-9*tmp[OC_MINI(y+2,c_h-1)*c_w] - +tmp[OC_MINI(y+3,c_h-1)*c_w]+64)>>7,255); + for (x = 0; x < c_w; x++) { + for (y = 0; y < OC_MINI(c_h, 2); y++) { + _dst[y * c_w] = (unsigned char)OC_CLAMPI(0, (4 * tmp[0] + - 17 * tmp[OC_MAXI(y - 1, 0) * c_w] + 114 * tmp[y * c_w] + + 35 * tmp[OC_MINI(y + 1, c_h - 1) * c_w] - 9 * tmp[OC_MINI(y + 2, c_h - 1) * c_w] + + tmp[OC_MINI(y + 3, c_h - 1) * c_w] + 64) >> 7, 255); } - for(;y<c_h-3;y++){ - _dst[y*c_w]=(unsigned char)OC_CLAMPI(0,(4*tmp[(y-2)*c_w] - -17*tmp[(y-1)*c_w]+114*tmp[y*c_w]+35*tmp[(y+1)*c_w] - -9*tmp[(y+2)*c_w]+tmp[(y+3)*c_w]+64)>>7,255); + for (; y < c_h - 3; y++) { + _dst[y * c_w] = (unsigned char)OC_CLAMPI(0, (4 * tmp[(y - 2) * c_w] + - 17 * tmp[(y - 1) * c_w] + 114 * tmp[y * c_w] + 35 * tmp[(y + 1) * c_w] + - 9 * tmp[(y + 2) * c_w] + tmp[(y + 3) * c_w] + 64) >> 7, 255); } - for(;y<c_h;y++){ - _dst[y*c_w]=(unsigned char)OC_CLAMPI(0,(4*tmp[(y-2)*c_w] - -17*tmp[(y-1)*c_w]+114*tmp[y*c_w]+35*tmp[OC_MINI(y+1,c_h-1)*c_w] - -9*tmp[OC_MINI(y+2,c_h-1)*c_w]+tmp[(c_h-1)*c_w]+64)>>7,255); + for (; y < c_h; y++) { + _dst[y * c_w] = (unsigned char)OC_CLAMPI(0, (4 * tmp[(y - 2) * c_w] + - 17 * tmp[(y - 1) * c_w] + 114 * tmp[y * c_w] + 35 * tmp[OC_MINI(y + 1, c_h - 1) * c_w] + - 9 * tmp[OC_MINI(y + 2, c_h - 1) * c_w] + tmp[(c_h - 1) * c_w] + 64) >> 7, 255); } _dst++; tmp++; } - }break; + } + break; } /*For actual interlaced material, this would have to be done separately on each field, and the shift amounts would be different. @@ -320,27 +328,27 @@ static void y4m_convert_42xpaldv_42xjpeg(y4m_input *_y4m,unsigned char *_dst, /*Perform vertical filtering to reduce a single plane from 4:2:2 to 4:2:0. This is used as a helper by several converation routines.*/ static void y4m_422jpeg_420jpeg_helper(unsigned char *_dst, - const unsigned char *_src,int _c_w,int _c_h){ + const unsigned char *_src, int _c_w, int _c_h) { int y; int x; /*Filter: [3 -17 78 78 -17 3]/128, derived from a 6-tap Lanczos window.*/ - for(x=0;x<_c_w;x++){ - for(y=0;y<OC_MINI(_c_h,2);y+=2){ - _dst[(y>>1)*_c_w]=OC_CLAMPI(0,(64*_src[0] - +78*_src[OC_MINI(1,_c_h-1)*_c_w] - -17*_src[OC_MINI(2,_c_h-1)*_c_w] - +3*_src[OC_MINI(3,_c_h-1)*_c_w]+64)>>7,255); + for (x = 0; x < _c_w; x++) { + for (y = 0; y < OC_MINI(_c_h, 2); y += 2) { + _dst[(y >> 1)*_c_w] = OC_CLAMPI(0, (64 * _src[0] + + 78 * _src[OC_MINI(1, _c_h - 1) * _c_w] + - 17 * _src[OC_MINI(2, _c_h - 1) * _c_w] + + 3 * _src[OC_MINI(3, _c_h - 1) * _c_w] + 64) >> 7, 255); } - for(;y<_c_h-3;y+=2){ - _dst[(y>>1)*_c_w]=OC_CLAMPI(0,(3*(_src[(y-2)*_c_w]+_src[(y+3)*_c_w]) - -17*(_src[(y-1)*_c_w]+_src[(y+2)*_c_w]) - +78*(_src[y*_c_w]+_src[(y+1)*_c_w])+64)>>7,255); + for (; y < _c_h - 3; y += 2) { + _dst[(y >> 1)*_c_w] = OC_CLAMPI(0, (3 * (_src[(y - 2) * _c_w] + _src[(y + 3) * _c_w]) + - 17 * (_src[(y - 1) * _c_w] + _src[(y + 2) * _c_w]) + + 78 * (_src[y * _c_w] + _src[(y + 1) * _c_w]) + 64) >> 7, 255); } - for(;y<_c_h;y+=2){ - _dst[(y>>1)*_c_w]=OC_CLAMPI(0,(3*(_src[(y-2)*_c_w] - +_src[(_c_h-1)*_c_w])-17*(_src[(y-1)*_c_w] - +_src[OC_MINI(y+2,_c_h-1)*_c_w]) - +78*(_src[y*_c_w]+_src[OC_MINI(y+1,_c_h-1)*_c_w])+64)>>7,255); + for (; y < _c_h; y += 2) { + _dst[(y >> 1)*_c_w] = OC_CLAMPI(0, (3 * (_src[(y - 2) * _c_w] + + _src[(_c_h - 1) * _c_w]) - 17 * (_src[(y - 1) * _c_w] + + _src[OC_MINI(y + 2, _c_h - 1) * _c_w]) + + 78 * (_src[y * _c_w] + _src[OC_MINI(y + 1, _c_h - 1) * _c_w]) + 64) >> 7, 255); } _src++; _dst++; @@ -385,8 +393,8 @@ static void y4m_422jpeg_420jpeg_helper(unsigned char *_dst, We use a resampling filter to decimate the chroma planes by two in the vertical direction.*/ -static void y4m_convert_422jpeg_420jpeg(y4m_input *_y4m,unsigned char *_dst, - unsigned char *_aux){ +static void y4m_convert_422jpeg_420jpeg(y4m_input *_y4m, unsigned char *_dst, + unsigned char *_aux) { int c_w; int c_h; int c_sz; @@ -395,18 +403,18 @@ static void y4m_convert_422jpeg_420jpeg(y4m_input *_y4m,unsigned char *_dst, int dst_c_sz; int pli; /*Skip past the luma data.*/ - _dst+=_y4m->pic_w*_y4m->pic_h; + _dst += _y4m->pic_w * _y4m->pic_h; /*Compute the size of each chroma plane.*/ - c_w=(_y4m->pic_w+_y4m->src_c_dec_h-1)/_y4m->src_c_dec_h; - c_h=_y4m->pic_h; - dst_c_w=(_y4m->pic_w+_y4m->dst_c_dec_h-1)/_y4m->dst_c_dec_h; - dst_c_h=(_y4m->pic_h+_y4m->dst_c_dec_v-1)/_y4m->dst_c_dec_v; - c_sz=c_w*c_h; - dst_c_sz=dst_c_w*dst_c_h; - for(pli=1;pli<3;pli++){ - y4m_422jpeg_420jpeg_helper(_dst,_aux,c_w,c_h); - _aux+=c_sz; - _dst+=dst_c_sz; + c_w = (_y4m->pic_w + _y4m->src_c_dec_h - 1) / _y4m->src_c_dec_h; + c_h = _y4m->pic_h; + dst_c_w = (_y4m->pic_w + _y4m->dst_c_dec_h - 1) / _y4m->dst_c_dec_h; + dst_c_h = (_y4m->pic_h + _y4m->dst_c_dec_v - 1) / _y4m->dst_c_dec_v; + c_sz = c_w * c_h; + dst_c_sz = dst_c_w * dst_c_h; + for (pli = 1; pli < 3; pli++) { + y4m_422jpeg_420jpeg_helper(_dst, _aux, c_w, c_h); + _aux += c_sz; + _dst += dst_c_sz; } } @@ -450,8 +458,8 @@ static void y4m_convert_422jpeg_420jpeg(y4m_input *_y4m,unsigned char *_dst, pixel (at the original chroma resolution) to the right. Then we use a second resampling filter to decimate the chroma planes by two in the vertical direction.*/ -static void y4m_convert_422_420jpeg(y4m_input *_y4m,unsigned char *_dst, - unsigned char *_aux){ +static void y4m_convert_422_420jpeg(y4m_input *_y4m, unsigned char *_dst, + unsigned char *_aux) { unsigned char *tmp; int c_w; int c_h; @@ -460,24 +468,24 @@ static void y4m_convert_422_420jpeg(y4m_input *_y4m,unsigned char *_dst, int dst_c_sz; int pli; /*Skip past the luma data.*/ - _dst+=_y4m->pic_w*_y4m->pic_h; + _dst += _y4m->pic_w * _y4m->pic_h; /*Compute the size of each chroma plane.*/ - c_w=(_y4m->pic_w+_y4m->src_c_dec_h-1)/_y4m->src_c_dec_h; - c_h=_y4m->pic_h; - dst_c_h=(_y4m->pic_h+_y4m->dst_c_dec_v-1)/_y4m->dst_c_dec_v; - c_sz=c_w*c_h; - dst_c_sz=c_w*dst_c_h; - tmp=_aux+2*c_sz; - for(pli=1;pli<3;pli++){ + c_w = (_y4m->pic_w + _y4m->src_c_dec_h - 1) / _y4m->src_c_dec_h; + c_h = _y4m->pic_h; + dst_c_h = (_y4m->pic_h + _y4m->dst_c_dec_v - 1) / _y4m->dst_c_dec_v; + c_sz = c_w * c_h; + dst_c_sz = c_w * dst_c_h; + tmp = _aux + 2 * c_sz; + for (pli = 1; pli < 3; pli++) { /*In reality, the horizontal and vertical steps could be pipelined, for less memory consumption and better cache performance, but we do them separately for simplicity.*/ /*First do horizontal filtering (convert to 422jpeg)*/ - y4m_42xmpeg2_42xjpeg_helper(tmp,_aux,c_w,c_h); + y4m_42xmpeg2_42xjpeg_helper(tmp, _aux, c_w, c_h); /*Now do the vertical filtering.*/ - y4m_422jpeg_420jpeg_helper(_dst,tmp,c_w,c_h); - _aux+=c_sz; - _dst+=dst_c_sz; + y4m_422jpeg_420jpeg_helper(_dst, tmp, c_w, c_h); + _aux += c_sz; + _dst += dst_c_sz; } } @@ -522,8 +530,8 @@ static void y4m_convert_422_420jpeg(y4m_input *_y4m,unsigned char *_dst, right. Then we use another filter to decimate the planes by 2 in the vertical direction.*/ -static void y4m_convert_411_420jpeg(y4m_input *_y4m,unsigned char *_dst, - unsigned char *_aux){ +static void y4m_convert_411_420jpeg(y4m_input *_y4m, unsigned char *_dst, + unsigned char *_aux) { unsigned char *tmp; int c_w; int c_h; @@ -536,57 +544,57 @@ static void y4m_convert_411_420jpeg(y4m_input *_y4m,unsigned char *_dst, int y; int x; /*Skip past the luma data.*/ - _dst+=_y4m->pic_w*_y4m->pic_h; + _dst += _y4m->pic_w * _y4m->pic_h; /*Compute the size of each chroma plane.*/ - c_w=(_y4m->pic_w+_y4m->src_c_dec_h-1)/_y4m->src_c_dec_h; - c_h=_y4m->pic_h; - dst_c_w=(_y4m->pic_w+_y4m->dst_c_dec_h-1)/_y4m->dst_c_dec_h; - dst_c_h=(_y4m->pic_h+_y4m->dst_c_dec_v-1)/_y4m->dst_c_dec_v; - c_sz=c_w*c_h; - dst_c_sz=dst_c_w*dst_c_h; - tmp_sz=dst_c_w*c_h; - tmp=_aux+2*c_sz; - for(pli=1;pli<3;pli++){ + c_w = (_y4m->pic_w + _y4m->src_c_dec_h - 1) / _y4m->src_c_dec_h; + c_h = _y4m->pic_h; + dst_c_w = (_y4m->pic_w + _y4m->dst_c_dec_h - 1) / _y4m->dst_c_dec_h; + dst_c_h = (_y4m->pic_h + _y4m->dst_c_dec_v - 1) / _y4m->dst_c_dec_v; + c_sz = c_w * c_h; + dst_c_sz = dst_c_w * dst_c_h; + tmp_sz = dst_c_w * c_h; + tmp = _aux + 2 * c_sz; + for (pli = 1; pli < 3; pli++) { /*In reality, the horizontal and vertical steps could be pipelined, for less memory consumption and better cache performance, but we do them separately for simplicity.*/ /*First do horizontal filtering (convert to 422jpeg)*/ - for(y=0;y<c_h;y++){ + for (y = 0; y < c_h; y++) { /*Filters: [1 110 18 -1]/128 and [-3 50 86 -5]/128, both derived from a 4-tap Mitchell window.*/ - for(x=0;x<OC_MINI(c_w,1);x++){ - tmp[x<<1]=(unsigned char)OC_CLAMPI(0,(111*_aux[0] - +18*_aux[OC_MINI(1,c_w-1)]-_aux[OC_MINI(2,c_w-1)]+64)>>7,255); - tmp[x<<1|1]=(unsigned char)OC_CLAMPI(0,(47*_aux[0] - +86*_aux[OC_MINI(1,c_w-1)]-5*_aux[OC_MINI(2,c_w-1)]+64)>>7,255); + for (x = 0; x < OC_MINI(c_w, 1); x++) { + tmp[x << 1] = (unsigned char)OC_CLAMPI(0, (111 * _aux[0] + + 18 * _aux[OC_MINI(1, c_w - 1)] - _aux[OC_MINI(2, c_w - 1)] + 64) >> 7, 255); + tmp[x << 1 | 1] = (unsigned char)OC_CLAMPI(0, (47 * _aux[0] + + 86 * _aux[OC_MINI(1, c_w - 1)] - 5 * _aux[OC_MINI(2, c_w - 1)] + 64) >> 7, 255); } - for(;x<c_w-2;x++){ - tmp[x<<1]=(unsigned char)OC_CLAMPI(0,(_aux[x-1]+110*_aux[x] - +18*_aux[x+1]-_aux[x+2]+64)>>7,255); - tmp[x<<1|1]=(unsigned char)OC_CLAMPI(0,(-3*_aux[x-1]+50*_aux[x] - +86*_aux[x+1]-5*_aux[x+2]+64)>>7,255); + for (; x < c_w - 2; x++) { + tmp[x << 1] = (unsigned char)OC_CLAMPI(0, (_aux[x - 1] + 110 * _aux[x] + + 18 * _aux[x + 1] - _aux[x + 2] + 64) >> 7, 255); + tmp[x << 1 | 1] = (unsigned char)OC_CLAMPI(0, (-3 * _aux[x - 1] + 50 * _aux[x] + + 86 * _aux[x + 1] - 5 * _aux[x + 2] + 64) >> 7, 255); } - for(;x<c_w;x++){ - tmp[x<<1]=(unsigned char)OC_CLAMPI(0,(_aux[x-1]+110*_aux[x] - +18*_aux[OC_MINI(x+1,c_w-1)]-_aux[c_w-1]+64)>>7,255); - if((x<<1|1)<dst_c_w){ - tmp[x<<1|1]=(unsigned char)OC_CLAMPI(0,(-3*_aux[x-1]+50*_aux[x] - +86*_aux[OC_MINI(x+1,c_w-1)]-5*_aux[c_w-1]+64)>>7,255); + for (; x < c_w; x++) { + tmp[x << 1] = (unsigned char)OC_CLAMPI(0, (_aux[x - 1] + 110 * _aux[x] + + 18 * _aux[OC_MINI(x + 1, c_w - 1)] - _aux[c_w - 1] + 64) >> 7, 255); + if ((x << 1 | 1) < dst_c_w) { + tmp[x << 1 | 1] = (unsigned char)OC_CLAMPI(0, (-3 * _aux[x - 1] + 50 * _aux[x] + + 86 * _aux[OC_MINI(x + 1, c_w - 1)] - 5 * _aux[c_w - 1] + 64) >> 7, 255); } } - tmp+=dst_c_w; - _aux+=c_w; + tmp += dst_c_w; + _aux += c_w; } - tmp-=tmp_sz; + tmp -= tmp_sz; /*Now do the vertical filtering.*/ - y4m_422jpeg_420jpeg_helper(_dst,tmp,dst_c_w,c_h); - _dst+=dst_c_sz; + y4m_422jpeg_420jpeg_helper(_dst, tmp, dst_c_w, c_h); + _dst += dst_c_sz; } } /*Convert 444 to 420jpeg.*/ -static void y4m_convert_444_420jpeg(y4m_input *_y4m,unsigned char *_dst, - unsigned char *_aux){ +static void y4m_convert_444_420jpeg(y4m_input *_y4m, unsigned char *_dst, + unsigned char *_aux) { unsigned char *tmp; int c_w; int c_h; @@ -599,218 +607,247 @@ static void y4m_convert_444_420jpeg(y4m_input *_y4m,unsigned char *_dst, int y; int x; /*Skip past the luma data.*/ - _dst+=_y4m->pic_w*_y4m->pic_h; + _dst += _y4m->pic_w * _y4m->pic_h; /*Compute the size of each chroma plane.*/ - c_w=(_y4m->pic_w+_y4m->src_c_dec_h-1)/_y4m->src_c_dec_h; - c_h=_y4m->pic_h; - dst_c_w=(_y4m->pic_w+_y4m->dst_c_dec_h-1)/_y4m->dst_c_dec_h; - dst_c_h=(_y4m->pic_h+_y4m->dst_c_dec_v-1)/_y4m->dst_c_dec_v; - c_sz=c_w*c_h; - dst_c_sz=dst_c_w*dst_c_h; - tmp_sz=dst_c_w*c_h; - tmp=_aux+2*c_sz; - for(pli=1;pli<3;pli++){ + c_w = (_y4m->pic_w + _y4m->src_c_dec_h - 1) / _y4m->src_c_dec_h; + c_h = _y4m->pic_h; + dst_c_w = (_y4m->pic_w + _y4m->dst_c_dec_h - 1) / _y4m->dst_c_dec_h; + dst_c_h = (_y4m->pic_h + _y4m->dst_c_dec_v - 1) / _y4m->dst_c_dec_v; + c_sz = c_w * c_h; + dst_c_sz = dst_c_w * dst_c_h; + tmp_sz = dst_c_w * c_h; + tmp = _aux + 2 * c_sz; + for (pli = 1; pli < 3; pli++) { /*Filter: [3 -17 78 78 -17 3]/128, derived from a 6-tap Lanczos window.*/ - for(y=0;y<c_h;y++){ - for(x=0;x<OC_MINI(c_w,2);x+=2){ - tmp[x>>1]=OC_CLAMPI(0,(64*_aux[0]+78*_aux[OC_MINI(1,c_w-1)] - -17*_aux[OC_MINI(2,c_w-1)] - +3*_aux[OC_MINI(3,c_w-1)]+64)>>7,255); + for (y = 0; y < c_h; y++) { + for (x = 0; x < OC_MINI(c_w, 2); x += 2) { + tmp[x >> 1] = OC_CLAMPI(0, (64 * _aux[0] + 78 * _aux[OC_MINI(1, c_w - 1)] + - 17 * _aux[OC_MINI(2, c_w - 1)] + + 3 * _aux[OC_MINI(3, c_w - 1)] + 64) >> 7, 255); } - for(;x<c_w-3;x+=2){ - tmp[x>>1]=OC_CLAMPI(0,(3*(_aux[x-2]+_aux[x+3]) - -17*(_aux[x-1]+_aux[x+2])+78*(_aux[x]+_aux[x+1])+64)>>7,255); + for (; x < c_w - 3; x += 2) { + tmp[x >> 1] = OC_CLAMPI(0, (3 * (_aux[x - 2] + _aux[x + 3]) + - 17 * (_aux[x - 1] + _aux[x + 2]) + 78 * (_aux[x] + _aux[x + 1]) + 64) >> 7, 255); } - for(;x<c_w;x+=2){ - tmp[x>>1]=OC_CLAMPI(0,(3*(_aux[x-2]+_aux[c_w-1])- - 17*(_aux[x-1]+_aux[OC_MINI(x+2,c_w-1)])+ - 78*(_aux[x]+_aux[OC_MINI(x+1,c_w-1)])+64)>>7,255); + for (; x < c_w; x += 2) { + tmp[x >> 1] = OC_CLAMPI(0, (3 * (_aux[x - 2] + _aux[c_w - 1]) - + 17 * (_aux[x - 1] + _aux[OC_MINI(x + 2, c_w - 1)]) + + 78 * (_aux[x] + _aux[OC_MINI(x + 1, c_w - 1)]) + 64) >> 7, 255); } - tmp+=dst_c_w; - _aux+=c_w; + tmp += dst_c_w; + _aux += c_w; } - tmp-=tmp_sz; + tmp -= tmp_sz; /*Now do the vertical filtering.*/ - y4m_422jpeg_420jpeg_helper(_dst,tmp,dst_c_w,c_h); - _dst+=dst_c_sz; + y4m_422jpeg_420jpeg_helper(_dst, tmp, dst_c_w, c_h); + _dst += dst_c_sz; } } /*The image is padded with empty chroma components at 4:2:0.*/ -static void y4m_convert_mono_420jpeg(y4m_input *_y4m,unsigned char *_dst, - unsigned char *_aux){ +static void y4m_convert_mono_420jpeg(y4m_input *_y4m, unsigned char *_dst, + unsigned char *_aux) { int c_sz; - _dst+=_y4m->pic_w*_y4m->pic_h; - c_sz=((_y4m->pic_w+_y4m->dst_c_dec_h-1)/_y4m->dst_c_dec_h)* - ((_y4m->pic_h+_y4m->dst_c_dec_v-1)/_y4m->dst_c_dec_v); - memset(_dst,128,c_sz*2); + _dst += _y4m->pic_w * _y4m->pic_h; + c_sz = ((_y4m->pic_w + _y4m->dst_c_dec_h - 1) / _y4m->dst_c_dec_h) * + ((_y4m->pic_h + _y4m->dst_c_dec_v - 1) / _y4m->dst_c_dec_v); + memset(_dst, 128, c_sz * 2); } /*No conversion function needed.*/ -static void y4m_convert_null(y4m_input *_y4m,unsigned char *_dst, - unsigned char *_aux){ +static void y4m_convert_null(y4m_input *_y4m, unsigned char *_dst, + unsigned char *_aux) { } -int y4m_input_open(y4m_input *_y4m,FILE *_fin,char *_skip,int _nskip){ +int y4m_input_open(y4m_input *_y4m, FILE *_fin, char *_skip, int _nskip, + int only_420) { char buffer[80]; int ret; int i; /*Read until newline, or 80 cols, whichever happens first.*/ - for(i=0;i<79;i++){ - if(_nskip>0){ - buffer[i]=*_skip++; + for (i = 0; i < 79; i++) { + if (_nskip > 0) { + buffer[i] = *_skip++; _nskip--; + } else { + ret = (int)fread(buffer + i, 1, 1, _fin); + if (ret < 1)return -1; } - else{ - ret=(int)fread(buffer+i,1,1,_fin); - if(ret<1)return -1; - } - if(buffer[i]=='\n')break; + if (buffer[i] == '\n')break; } /*We skipped too much header data.*/ - if(_nskip>0)return -1; - if(i==79){ - fprintf(stderr,"Error parsing header; not a YUV2MPEG2 file?\n"); + if (_nskip > 0)return -1; + if (i == 79) { + fprintf(stderr, "Error parsing header; not a YUV2MPEG2 file?\n"); return -1; } - buffer[i]='\0'; - if(memcmp(buffer,"YUV4MPEG",8)){ - fprintf(stderr,"Incomplete magic for YUV4MPEG file.\n"); + buffer[i] = '\0'; + if (memcmp(buffer, "YUV4MPEG", 8)) { + fprintf(stderr, "Incomplete magic for YUV4MPEG file.\n"); return -1; } - if(buffer[8]!='2'){ - fprintf(stderr,"Incorrect YUV input file version; YUV4MPEG2 required.\n"); + if (buffer[8] != '2') { + fprintf(stderr, "Incorrect YUV input file version; YUV4MPEG2 required.\n"); } - ret=y4m_parse_tags(_y4m,buffer+5); - if(ret<0){ - fprintf(stderr,"Error parsing YUV4MPEG2 header.\n"); + ret = y4m_parse_tags(_y4m, buffer + 5); + if (ret < 0) { + fprintf(stderr, "Error parsing YUV4MPEG2 header.\n"); return ret; } - if(_y4m->interlace=='?'){ - fprintf(stderr,"Warning: Input video interlacing format unknown; " - "assuming progressive scan.\n"); - } - else if(_y4m->interlace!='p'){ - fprintf(stderr,"Input video is interlaced; " - "Only progressive scan handled.\n"); + if (_y4m->interlace == '?') { + fprintf(stderr, "Warning: Input video interlacing format unknown; " + "assuming progressive scan.\n"); + } else if (_y4m->interlace != 'p') { + fprintf(stderr, "Input video is interlaced; " + "Only progressive scan handled.\n"); return -1; } - if(strcmp(_y4m->chroma_type,"420")==0|| - strcmp(_y4m->chroma_type,"420jpeg")==0){ - _y4m->src_c_dec_h=_y4m->dst_c_dec_h=_y4m->src_c_dec_v=_y4m->dst_c_dec_v=2; - _y4m->dst_buf_read_sz=_y4m->pic_w*_y4m->pic_h - +2*((_y4m->pic_w+1)/2)*((_y4m->pic_h+1)/2); + _y4m->vpx_fmt = VPX_IMG_FMT_I420; + _y4m->vpx_bps = 12; + if (strcmp(_y4m->chroma_type, "420") == 0 || + strcmp(_y4m->chroma_type, "420jpeg") == 0) { + _y4m->src_c_dec_h = _y4m->dst_c_dec_h = _y4m->src_c_dec_v = _y4m->dst_c_dec_v = 2; + _y4m->dst_buf_read_sz = _y4m->pic_w * _y4m->pic_h + + 2 * ((_y4m->pic_w + 1) / 2) * ((_y4m->pic_h + 1) / 2); /*Natively supported: no conversion required.*/ - _y4m->aux_buf_sz=_y4m->aux_buf_read_sz=0; - _y4m->convert=y4m_convert_null; - } - else if(strcmp(_y4m->chroma_type,"420mpeg2")==0){ - _y4m->src_c_dec_h=_y4m->dst_c_dec_h=_y4m->src_c_dec_v=_y4m->dst_c_dec_v=2; - _y4m->dst_buf_read_sz=_y4m->pic_w*_y4m->pic_h; + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz = 0; + _y4m->convert = y4m_convert_null; + } else if (strcmp(_y4m->chroma_type, "420mpeg2") == 0) { + _y4m->src_c_dec_h = _y4m->dst_c_dec_h = _y4m->src_c_dec_v = _y4m->dst_c_dec_v = 2; + _y4m->dst_buf_read_sz = _y4m->pic_w * _y4m->pic_h; /*Chroma filter required: read into the aux buf first.*/ - _y4m->aux_buf_sz=_y4m->aux_buf_read_sz= - 2*((_y4m->pic_w+1)/2)*((_y4m->pic_h+1)/2); - _y4m->convert=y4m_convert_42xmpeg2_42xjpeg; - } - else if(strcmp(_y4m->chroma_type,"420paldv")==0){ - _y4m->src_c_dec_h=_y4m->dst_c_dec_h=_y4m->src_c_dec_v=_y4m->dst_c_dec_v=2; - _y4m->dst_buf_read_sz=_y4m->pic_w*_y4m->pic_h; + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz = + 2 * ((_y4m->pic_w + 1) / 2) * ((_y4m->pic_h + 1) / 2); + _y4m->convert = y4m_convert_42xmpeg2_42xjpeg; + } else if (strcmp(_y4m->chroma_type, "420paldv") == 0) { + _y4m->src_c_dec_h = _y4m->dst_c_dec_h = _y4m->src_c_dec_v = _y4m->dst_c_dec_v = 2; + _y4m->dst_buf_read_sz = _y4m->pic_w * _y4m->pic_h; /*Chroma filter required: read into the aux buf first. We need to make two filter passes, so we need some extra space in the aux buffer.*/ - _y4m->aux_buf_sz=3*((_y4m->pic_w+1)/2)*((_y4m->pic_h+1)/2); - _y4m->aux_buf_read_sz=2*((_y4m->pic_w+1)/2)*((_y4m->pic_h+1)/2); - _y4m->convert=y4m_convert_42xpaldv_42xjpeg; - } - else if(strcmp(_y4m->chroma_type,"422jpeg")==0){ - _y4m->src_c_dec_h=_y4m->dst_c_dec_h=2; - _y4m->src_c_dec_v=1; - _y4m->dst_c_dec_v=2; - _y4m->dst_buf_read_sz=_y4m->pic_w*_y4m->pic_h; + _y4m->aux_buf_sz = 3 * ((_y4m->pic_w + 1) / 2) * ((_y4m->pic_h + 1) / 2); + _y4m->aux_buf_read_sz = 2 * ((_y4m->pic_w + 1) / 2) * ((_y4m->pic_h + 1) / 2); + _y4m->convert = y4m_convert_42xpaldv_42xjpeg; + } else if (strcmp(_y4m->chroma_type, "422jpeg") == 0) { + _y4m->src_c_dec_h = _y4m->dst_c_dec_h = 2; + _y4m->src_c_dec_v = 1; + _y4m->dst_c_dec_v = 2; + _y4m->dst_buf_read_sz = _y4m->pic_w * _y4m->pic_h; /*Chroma filter required: read into the aux buf first.*/ - _y4m->aux_buf_sz=_y4m->aux_buf_read_sz=2*((_y4m->pic_w+1)/2)*_y4m->pic_h; - _y4m->convert=y4m_convert_422jpeg_420jpeg; - } - else if(strcmp(_y4m->chroma_type,"422")==0){ - _y4m->src_c_dec_h=_y4m->dst_c_dec_h=2; - _y4m->src_c_dec_v=1; - _y4m->dst_c_dec_v=2; - _y4m->dst_buf_read_sz=_y4m->pic_w*_y4m->pic_h; - /*Chroma filter required: read into the aux buf first. - We need to make two filter passes, so we need some extra space in the - aux buffer.*/ - _y4m->aux_buf_read_sz=2*((_y4m->pic_w+1)/2)*_y4m->pic_h; - _y4m->aux_buf_sz=_y4m->aux_buf_read_sz+((_y4m->pic_w+1)/2)*_y4m->pic_h; - _y4m->convert=y4m_convert_422_420jpeg; - } - else if(strcmp(_y4m->chroma_type,"411")==0){ - _y4m->src_c_dec_h=4; - _y4m->dst_c_dec_h=2; - _y4m->src_c_dec_v=1; - _y4m->dst_c_dec_v=2; - _y4m->dst_buf_read_sz=_y4m->pic_w*_y4m->pic_h; - /*Chroma filter required: read into the aux buf first. - We need to make two filter passes, so we need some extra space in the - aux buffer.*/ - _y4m->aux_buf_read_sz=2*((_y4m->pic_w+3)/4)*_y4m->pic_h; - _y4m->aux_buf_sz=_y4m->aux_buf_read_sz+((_y4m->pic_w+1)/2)*_y4m->pic_h; - _y4m->convert=y4m_convert_411_420jpeg; - } - else if(strcmp(_y4m->chroma_type,"444")==0){ - _y4m->src_c_dec_h=1; - _y4m->dst_c_dec_h=2; - _y4m->src_c_dec_v=1; - _y4m->dst_c_dec_v=2; - _y4m->dst_buf_read_sz=_y4m->pic_w*_y4m->pic_h; + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz = 2 * ((_y4m->pic_w + 1) / 2) * _y4m->pic_h; + _y4m->convert = y4m_convert_422jpeg_420jpeg; + } else if (strcmp(_y4m->chroma_type, "422") == 0) { + _y4m->src_c_dec_h = 2; + _y4m->src_c_dec_v = 1; + if (only_420) { + _y4m->dst_c_dec_h = 2; + _y4m->dst_c_dec_v = 2; + _y4m->dst_buf_read_sz = _y4m->pic_w * _y4m->pic_h; + /*Chroma filter required: read into the aux buf first. + We need to make two filter passes, so we need some extra space in the + aux buffer.*/ + _y4m->aux_buf_read_sz = 2 * ((_y4m->pic_w + 1) / 2) * _y4m->pic_h; + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz + + ((_y4m->pic_w + 1) / 2) * _y4m->pic_h; + _y4m->convert = y4m_convert_422_420jpeg; + } else { + _y4m->vpx_fmt = VPX_IMG_FMT_I422; + _y4m->vpx_bps = 16; + _y4m->dst_c_dec_h = _y4m->src_c_dec_h; + _y4m->dst_c_dec_v = _y4m->src_c_dec_v; + _y4m->dst_buf_read_sz = _y4m->pic_w * _y4m->pic_h + + 2 * ((_y4m->pic_w + 1) / 2) * _y4m->pic_h; + /*Natively supported: no conversion required.*/ + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz = 0; + _y4m->convert = y4m_convert_null; + } + } else if (strcmp(_y4m->chroma_type, "411") == 0) { + _y4m->src_c_dec_h = 4; + _y4m->dst_c_dec_h = 2; + _y4m->src_c_dec_v = 1; + _y4m->dst_c_dec_v = 2; + _y4m->dst_buf_read_sz = _y4m->pic_w * _y4m->pic_h; /*Chroma filter required: read into the aux buf first. We need to make two filter passes, so we need some extra space in the aux buffer.*/ - _y4m->aux_buf_read_sz=2*_y4m->pic_w*_y4m->pic_h; - _y4m->aux_buf_sz=_y4m->aux_buf_read_sz+((_y4m->pic_w+1)/2)*_y4m->pic_h; - _y4m->convert=y4m_convert_444_420jpeg; - } - else if(strcmp(_y4m->chroma_type,"444alpha")==0){ - _y4m->src_c_dec_h=1; - _y4m->dst_c_dec_h=2; - _y4m->src_c_dec_v=1; - _y4m->dst_c_dec_v=2; - _y4m->dst_buf_read_sz=_y4m->pic_w*_y4m->pic_h; - /*Chroma filter required: read into the aux buf first. - We need to make two filter passes, so we need some extra space in the - aux buffer. - The extra plane also gets read into the aux buf. - It will be discarded.*/ - _y4m->aux_buf_sz=_y4m->aux_buf_read_sz=3*_y4m->pic_w*_y4m->pic_h; - _y4m->convert=y4m_convert_444_420jpeg; - } - else if(strcmp(_y4m->chroma_type,"mono")==0){ - _y4m->src_c_dec_h=_y4m->src_c_dec_v=0; - _y4m->dst_c_dec_h=_y4m->dst_c_dec_v=2; - _y4m->dst_buf_read_sz=_y4m->pic_w*_y4m->pic_h; + _y4m->aux_buf_read_sz = 2 * ((_y4m->pic_w + 3) / 4) * _y4m->pic_h; + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz + ((_y4m->pic_w + 1) / 2) * _y4m->pic_h; + _y4m->convert = y4m_convert_411_420jpeg; + } else if (strcmp(_y4m->chroma_type, "444") == 0) { + _y4m->src_c_dec_h = 1; + _y4m->src_c_dec_v = 1; + if (only_420) { + _y4m->dst_c_dec_h = 2; + _y4m->dst_c_dec_v = 2; + _y4m->dst_buf_read_sz = _y4m->pic_w * _y4m->pic_h; + /*Chroma filter required: read into the aux buf first. + We need to make two filter passes, so we need some extra space in the + aux buffer.*/ + _y4m->aux_buf_read_sz = 2 * _y4m->pic_w * _y4m->pic_h; + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz + + ((_y4m->pic_w + 1) / 2) * _y4m->pic_h; + _y4m->convert = y4m_convert_444_420jpeg; + } else { + _y4m->vpx_fmt = VPX_IMG_FMT_I444; + _y4m->vpx_bps = 24; + _y4m->dst_c_dec_h = _y4m->src_c_dec_h; + _y4m->dst_c_dec_v = _y4m->src_c_dec_v; + _y4m->dst_buf_read_sz = 3 * _y4m->pic_w * _y4m->pic_h; + /*Natively supported: no conversion required.*/ + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz = 0; + _y4m->convert = y4m_convert_null; + } + } else if (strcmp(_y4m->chroma_type, "444alpha") == 0) { + _y4m->src_c_dec_h = 1; + _y4m->src_c_dec_v = 1; + if (only_420) { + _y4m->dst_c_dec_h = 2; + _y4m->dst_c_dec_v = 2; + _y4m->dst_buf_read_sz = _y4m->pic_w * _y4m->pic_h; + /*Chroma filter required: read into the aux buf first. + We need to make two filter passes, so we need some extra space in the + aux buffer. + The extra plane also gets read into the aux buf. + It will be discarded.*/ + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz = 3 * _y4m->pic_w * _y4m->pic_h; + _y4m->convert = y4m_convert_444_420jpeg; + } else { + _y4m->vpx_fmt = VPX_IMG_FMT_444A; + _y4m->vpx_bps = 32; + _y4m->dst_c_dec_h = _y4m->src_c_dec_h; + _y4m->dst_c_dec_v = _y4m->src_c_dec_v; + _y4m->dst_buf_read_sz = 4 * _y4m->pic_w * _y4m->pic_h; + /*Natively supported: no conversion required.*/ + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz = 0; + _y4m->convert = y4m_convert_null; + } + } else if (strcmp(_y4m->chroma_type, "mono") == 0) { + _y4m->src_c_dec_h = _y4m->src_c_dec_v = 0; + _y4m->dst_c_dec_h = _y4m->dst_c_dec_v = 2; + _y4m->dst_buf_read_sz = _y4m->pic_w * _y4m->pic_h; /*No extra space required, but we need to clear the chroma planes.*/ - _y4m->aux_buf_sz=_y4m->aux_buf_read_sz=0; - _y4m->convert=y4m_convert_mono_420jpeg; - } - else{ - fprintf(stderr,"Unknown chroma sampling type: %s\n",_y4m->chroma_type); + _y4m->aux_buf_sz = _y4m->aux_buf_read_sz = 0; + _y4m->convert = y4m_convert_mono_420jpeg; + } else { + fprintf(stderr, "Unknown chroma sampling type: %s\n", _y4m->chroma_type); return -1; } /*The size of the final frame buffers is always computed from the destination chroma decimation type.*/ - _y4m->dst_buf_sz=_y4m->pic_w*_y4m->pic_h - +2*((_y4m->pic_w+_y4m->dst_c_dec_h-1)/_y4m->dst_c_dec_h)* - ((_y4m->pic_h+_y4m->dst_c_dec_v-1)/_y4m->dst_c_dec_v); - _y4m->dst_buf=(unsigned char *)malloc(_y4m->dst_buf_sz); - _y4m->aux_buf=(unsigned char *)malloc(_y4m->aux_buf_sz); + _y4m->dst_buf_sz = _y4m->pic_w * _y4m->pic_h + + 2 * ((_y4m->pic_w + _y4m->dst_c_dec_h - 1) / _y4m->dst_c_dec_h) * + ((_y4m->pic_h + _y4m->dst_c_dec_v - 1) / _y4m->dst_c_dec_v); + _y4m->dst_buf = (unsigned char *)malloc(_y4m->dst_buf_sz); + _y4m->aux_buf = (unsigned char *)malloc(_y4m->aux_buf_sz); return 0; } -void y4m_input_close(y4m_input *_y4m){ +void y4m_input_close(y4m_input *_y4m) { free(_y4m->dst_buf); free(_y4m->aux_buf); } -int y4m_input_fetch_frame(y4m_input *_y4m,FILE *_fin,vpx_image_t *_img){ +int y4m_input_fetch_frame(y4m_input *_y4m, FILE *_fin, vpx_image_t *_img) { char frame[6]; int pic_sz; int c_w; @@ -818,54 +855,55 @@ int y4m_input_fetch_frame(y4m_input *_y4m,FILE *_fin,vpx_image_t *_img){ int c_sz; int ret; /*Read and skip the frame header.*/ - ret=(int)fread(frame,1,6,_fin); - if(ret<6)return 0; - if(memcmp(frame,"FRAME",5)){ - fprintf(stderr,"Loss of framing in Y4M input data\n"); + ret = (int)fread(frame, 1, 6, _fin); + if (ret < 6)return 0; + if (memcmp(frame, "FRAME", 5)) { + fprintf(stderr, "Loss of framing in Y4M input data\n"); return -1; } - if(frame[5]!='\n'){ + if (frame[5] != '\n') { char c; int j; - for(j=0;j<79&&fread(&c,1,1,_fin)&&c!='\n';j++); - if(j==79){ - fprintf(stderr,"Error parsing Y4M frame header\n"); + for (j = 0; j < 79 && fread(&c, 1, 1, _fin) && c != '\n'; j++); + if (j == 79) { + fprintf(stderr, "Error parsing Y4M frame header\n"); return -1; } } /*Read the frame data that needs no conversion.*/ - if(fread(_y4m->dst_buf,1,_y4m->dst_buf_read_sz,_fin)!=_y4m->dst_buf_read_sz){ - fprintf(stderr,"Error reading Y4M frame data.\n"); + if (fread(_y4m->dst_buf, 1, _y4m->dst_buf_read_sz, _fin) != _y4m->dst_buf_read_sz) { + fprintf(stderr, "Error reading Y4M frame data.\n"); return -1; } /*Read the frame data that does need conversion.*/ - if(fread(_y4m->aux_buf,1,_y4m->aux_buf_read_sz,_fin)!=_y4m->aux_buf_read_sz){ - fprintf(stderr,"Error reading Y4M frame data.\n"); + if (fread(_y4m->aux_buf, 1, _y4m->aux_buf_read_sz, _fin) != _y4m->aux_buf_read_sz) { + fprintf(stderr, "Error reading Y4M frame data.\n"); return -1; } /*Now convert the just read frame.*/ - (*_y4m->convert)(_y4m,_y4m->dst_buf,_y4m->aux_buf); + (*_y4m->convert)(_y4m, _y4m->dst_buf, _y4m->aux_buf); /*Fill in the frame buffer pointers. We don't use vpx_img_wrap() because it forces padding for odd picture sizes, which would require a separate fread call for every row.*/ - memset(_img,0,sizeof(*_img)); + memset(_img, 0, sizeof(*_img)); /*Y4M has the planes in Y'CbCr order, which libvpx calls Y, U, and V.*/ - _img->fmt=IMG_FMT_I420; - _img->w=_img->d_w=_y4m->pic_w; - _img->h=_img->d_h=_y4m->pic_h; - /*This is hard-coded to 4:2:0 for now, as that's all VP8 supports.*/ - _img->x_chroma_shift=1; - _img->y_chroma_shift=1; - _img->bps=12; + _img->fmt = _y4m->vpx_fmt; + _img->w = _img->d_w = _y4m->pic_w; + _img->h = _img->d_h = _y4m->pic_h; + _img->x_chroma_shift = _y4m->dst_c_dec_h >> 1; + _img->y_chroma_shift = _y4m->dst_c_dec_v >> 1; + _img->bps = _y4m->vpx_bps; + /*Set up the buffer pointers.*/ - pic_sz=_y4m->pic_w*_y4m->pic_h; - c_w=(_y4m->pic_w+_y4m->dst_c_dec_h-1)/_y4m->dst_c_dec_h; - c_h=(_y4m->pic_h+_y4m->dst_c_dec_v-1)/_y4m->dst_c_dec_v; - c_sz=c_w*c_h; - _img->stride[PLANE_Y]=_y4m->pic_w; - _img->stride[PLANE_U]=_img->stride[PLANE_V]=c_w; - _img->planes[PLANE_Y]=_y4m->dst_buf; - _img->planes[PLANE_U]=_y4m->dst_buf+pic_sz; - _img->planes[PLANE_V]=_y4m->dst_buf+pic_sz+c_sz; + pic_sz = _y4m->pic_w * _y4m->pic_h; + c_w = (_y4m->pic_w + _y4m->dst_c_dec_h - 1) / _y4m->dst_c_dec_h; + c_h = (_y4m->pic_h + _y4m->dst_c_dec_v - 1) / _y4m->dst_c_dec_v; + c_sz = c_w * c_h; + _img->stride[PLANE_Y] = _img->stride[PLANE_ALPHA] = _y4m->pic_w; + _img->stride[PLANE_U] = _img->stride[PLANE_V] = c_w; + _img->planes[PLANE_Y] = _y4m->dst_buf; + _img->planes[PLANE_U] = _y4m->dst_buf + pic_sz; + _img->planes[PLANE_V] = _y4m->dst_buf + pic_sz + c_sz; + _img->planes[PLANE_ALPHA] = _y4m->dst_buf + pic_sz + 2 * c_sz; return 1; } |