| @@ -2343,7 +2343,7 @@ static VideoState *global_video_state; | |||||
| static int decode_interrupt_cb(void *ctx) | static int decode_interrupt_cb(void *ctx) | ||||
| { | { | ||||
| return (global_video_state && global_video_state->abort_request); | |||||
| return global_video_state && global_video_state->abort_request; | |||||
| } | } | ||||
| /* this thread gets the stream from the disk or the network */ | /* this thread gets the stream from the disk or the network */ | ||||
| @@ -914,8 +914,9 @@ int check_stream_specifier(AVFormatContext *s, AVStream *st, const char *spec) | |||||
| if (*endptr++ == ':') { | if (*endptr++ == ':') { | ||||
| int stream_idx = strtol(endptr, NULL, 0); | int stream_idx = strtol(endptr, NULL, 0); | ||||
| return (stream_idx >= 0 && stream_idx < s->programs[i]->nb_stream_indexes && | |||||
| st->index == s->programs[i]->stream_index[stream_idx]); | |||||
| return stream_idx >= 0 && | |||||
| stream_idx < s->programs[i]->nb_stream_indexes && | |||||
| st->index == s->programs[i]->stream_index[stream_idx]; | |||||
| } | } | ||||
| for (j = 0; j < s->programs[i]->nb_stream_indexes; j++) | for (j = 0; j < s->programs[i]->nb_stream_indexes; j++) | ||||
| @@ -742,7 +742,7 @@ static int decodeFrame(ATRAC3Context *q, const uint8_t* databuf, | |||||
| result = decodeChannelSoundUnit(q,&q->gb, q->pUnits, out_samples[0], 0, JOINT_STEREO); | result = decodeChannelSoundUnit(q,&q->gb, q->pUnits, out_samples[0], 0, JOINT_STEREO); | ||||
| if (result != 0) | if (result != 0) | ||||
| return (result); | |||||
| return result; | |||||
| /* Framedata of the su2 in the joint-stereo mode is encoded in | /* Framedata of the su2 in the joint-stereo mode is encoded in | ||||
| * reverse byte order so we need to swap it first. */ | * reverse byte order so we need to swap it first. */ | ||||
| @@ -783,7 +783,7 @@ static int decodeFrame(ATRAC3Context *q, const uint8_t* databuf, | |||||
| /* Decode Sound Unit 2. */ | /* Decode Sound Unit 2. */ | ||||
| result = decodeChannelSoundUnit(q,&q->gb, &q->pUnits[1], out_samples[1], 1, JOINT_STEREO); | result = decodeChannelSoundUnit(q,&q->gb, &q->pUnits[1], out_samples[1], 1, JOINT_STEREO); | ||||
| if (result != 0) | if (result != 0) | ||||
| return (result); | |||||
| return result; | |||||
| /* Reconstruct the channel coefficients. */ | /* Reconstruct the channel coefficients. */ | ||||
| reverseMatrixing(out_samples[0], out_samples[1], q->matrix_coeff_index_prev, q->matrix_coeff_index_now); | reverseMatrixing(out_samples[0], out_samples[1], q->matrix_coeff_index_prev, q->matrix_coeff_index_now); | ||||
| @@ -802,7 +802,7 @@ static int decodeFrame(ATRAC3Context *q, const uint8_t* databuf, | |||||
| result = decodeChannelSoundUnit(q,&q->gb, &q->pUnits[i], out_samples[i], i, q->codingMode); | result = decodeChannelSoundUnit(q,&q->gb, &q->pUnits[i], out_samples[i], i, q->codingMode); | ||||
| if (result != 0) | if (result != 0) | ||||
| return (result); | |||||
| return result; | |||||
| } | } | ||||
| } | } | ||||
| @@ -757,7 +757,7 @@ static av_always_inline int dv_guess_dct_mode(DVVideoContext *s, uint8_t *data, | |||||
| if (ps > 0) { | if (ps > 0) { | ||||
| int is = s->ildct_cmp(NULL, data , NULL, linesize<<1, 4) + | int is = s->ildct_cmp(NULL, data , NULL, linesize<<1, 4) + | ||||
| s->ildct_cmp(NULL, data + linesize, NULL, linesize<<1, 4); | s->ildct_cmp(NULL, data + linesize, NULL, linesize<<1, 4); | ||||
| return (ps > is); | |||||
| return ps > is; | |||||
| } | } | ||||
| } | } | ||||
| @@ -148,7 +148,7 @@ static int decode_plane_subdivision(GetBitContext *gb) | |||||
| static inline int scale_tile_size(int def_size, int size_factor) | static inline int scale_tile_size(int def_size, int size_factor) | ||||
| { | { | ||||
| return (size_factor == 15 ? def_size : (size_factor + 1) << 5); | |||||
| return size_factor == 15 ? def_size : (size_factor + 1) << 5; | |||||
| } | } | ||||
| /** | /** | ||||
| @@ -195,10 +195,10 @@ typedef struct { | |||||
| /** compare some properties of two pictures */ | /** compare some properties of two pictures */ | ||||
| static inline int ivi_pic_config_cmp(IVIPicConfig *str1, IVIPicConfig *str2) | static inline int ivi_pic_config_cmp(IVIPicConfig *str1, IVIPicConfig *str2) | ||||
| { | { | ||||
| return (str1->pic_width != str2->pic_width || str1->pic_height != str2->pic_height || | |||||
| str1->chroma_width != str2->chroma_width || str1->chroma_height != str2->chroma_height || | |||||
| str1->tile_width != str2->tile_width || str1->tile_height != str2->tile_height || | |||||
| str1->luma_bands != str2->luma_bands || str1->chroma_bands != str2->chroma_bands); | |||||
| return str1->pic_width != str2->pic_width || str1->pic_height != str2->pic_height || | |||||
| str1->chroma_width != str2->chroma_width || str1->chroma_height != str2->chroma_height || | |||||
| str1->tile_width != str2->tile_width || str1->tile_height != str2->tile_height || | |||||
| str1->luma_bands != str2->luma_bands || str1->chroma_bands != str2->chroma_bands; | |||||
| } | } | ||||
| /** calculate number of tiles in a stride */ | /** calculate number of tiles in a stride */ | ||||
| @@ -105,7 +105,7 @@ static av_cold int oggvorbis_init_encoder(vorbis_info *vi, AVCodecContext *avcco | |||||
| /* How many bytes are needed for a buffer of length 'l' */ | /* How many bytes are needed for a buffer of length 'l' */ | ||||
| static int xiph_len(int l) | static int xiph_len(int l) | ||||
| { | { | ||||
| return (1 + l / 255 + l); | |||||
| return 1 + l / 255 + l; | |||||
| } | } | ||||
| static av_cold int oggvorbis_encode_init(AVCodecContext *avccontext) | static av_cold int oggvorbis_encode_init(AVCodecContext *avccontext) | ||||
| @@ -214,7 +214,7 @@ static int adaptive_cb_search(const int16_t *adapt_cb, float *work, | |||||
| ff_celp_lp_synthesis_filterf(work, coefs, exc, BLOCKSIZE, LPC_ORDER); | ff_celp_lp_synthesis_filterf(work, coefs, exc, BLOCKSIZE, LPC_ORDER); | ||||
| for (i = 0; i < BLOCKSIZE; i++) | for (i = 0; i < BLOCKSIZE; i++) | ||||
| data[i] -= best_gain * work[i]; | data[i] -= best_gain * work[i]; | ||||
| return (best_vect - BLOCKSIZE / 2 + 1); | |||||
| return best_vect - BLOCKSIZE / 2 + 1; | |||||
| } | } | ||||
| @@ -26,8 +26,8 @@ | |||||
| /** Reconstruct bitstream f_code */ | /** Reconstruct bitstream f_code */ | ||||
| static inline int mpeg2_get_f_code(MpegEncContext *s) | static inline int mpeg2_get_f_code(MpegEncContext *s) | ||||
| { | { | ||||
| return ((s->mpeg_f_code[0][0] << 12) | (s->mpeg_f_code[0][1] << 8) | | |||||
| (s->mpeg_f_code[1][0] << 4) | s->mpeg_f_code[1][1]); | |||||
| return (s->mpeg_f_code[0][0] << 12) | (s->mpeg_f_code[0][1] << 8) | | |||||
| (s->mpeg_f_code[1][0] << 4) | s->mpeg_f_code[1][1]; | |||||
| } | } | ||||
| /** Determine frame start: first field for field picture or frame picture */ | /** Determine frame start: first field for field picture or frame picture */ | ||||
| @@ -42,10 +42,10 @@ static inline int vc1_has_MVTYPEMB_bitplane(VC1Context *v) | |||||
| { | { | ||||
| if (v->mv_type_is_raw) | if (v->mv_type_is_raw) | ||||
| return 0; | return 0; | ||||
| return (v->s.pict_type == AV_PICTURE_TYPE_P && | |||||
| (v->mv_mode == MV_PMODE_MIXED_MV || | |||||
| (v->mv_mode == MV_PMODE_INTENSITY_COMP && | |||||
| v->mv_mode2 == MV_PMODE_MIXED_MV))); | |||||
| return v->s.pict_type == AV_PICTURE_TYPE_P && | |||||
| (v->mv_mode == MV_PMODE_MIXED_MV || | |||||
| (v->mv_mode == MV_PMODE_INTENSITY_COMP && | |||||
| v->mv_mode2 == MV_PMODE_MIXED_MV)); | |||||
| } | } | ||||
| /** Check whether the SKIPMB bitplane is present */ | /** Check whether the SKIPMB bitplane is present */ | ||||
| @@ -53,8 +53,8 @@ static inline int vc1_has_SKIPMB_bitplane(VC1Context *v) | |||||
| { | { | ||||
| if (v->skip_is_raw) | if (v->skip_is_raw) | ||||
| return 0; | return 0; | ||||
| return (v->s.pict_type == AV_PICTURE_TYPE_P || | |||||
| (v->s.pict_type == AV_PICTURE_TYPE_B && !v->bi_type)); | |||||
| return v->s.pict_type == AV_PICTURE_TYPE_P || | |||||
| (v->s.pict_type == AV_PICTURE_TYPE_B && !v->bi_type); | |||||
| } | } | ||||
| /** Check whether the DIRECTMB bitplane is present */ | /** Check whether the DIRECTMB bitplane is present */ | ||||
| @@ -70,9 +70,9 @@ static inline int vc1_has_ACPRED_bitplane(VC1Context *v) | |||||
| { | { | ||||
| if (v->acpred_is_raw) | if (v->acpred_is_raw) | ||||
| return 0; | return 0; | ||||
| return (v->profile == PROFILE_ADVANCED && | |||||
| (v->s.pict_type == AV_PICTURE_TYPE_I || | |||||
| (v->s.pict_type == AV_PICTURE_TYPE_B && v->bi_type))); | |||||
| return v->profile == PROFILE_ADVANCED && | |||||
| (v->s.pict_type == AV_PICTURE_TYPE_I || | |||||
| (v->s.pict_type == AV_PICTURE_TYPE_B && v->bi_type)); | |||||
| } | } | ||||
| /** Check whether the OVERFLAGS bitplane is present */ | /** Check whether the OVERFLAGS bitplane is present */ | ||||
| @@ -80,11 +80,11 @@ static inline int vc1_has_OVERFLAGS_bitplane(VC1Context *v) | |||||
| { | { | ||||
| if (v->overflg_is_raw) | if (v->overflg_is_raw) | ||||
| return 0; | return 0; | ||||
| return (v->profile == PROFILE_ADVANCED && | |||||
| (v->s.pict_type == AV_PICTURE_TYPE_I || | |||||
| (v->s.pict_type == AV_PICTURE_TYPE_B && v->bi_type)) && | |||||
| (v->overlap && v->pq <= 8) && | |||||
| v->condover == CONDOVER_SELECT); | |||||
| return v->profile == PROFILE_ADVANCED && | |||||
| (v->s.pict_type == AV_PICTURE_TYPE_I || | |||||
| (v->s.pict_type == AV_PICTURE_TYPE_B && v->bi_type)) && | |||||
| (v->overlap && v->pq <= 8) && | |||||
| v->condover == CONDOVER_SELECT; | |||||
| } | } | ||||
| /** Reconstruct bitstream PTYPE (7.1.1.4, index into Table-35) */ | /** Reconstruct bitstream PTYPE (7.1.1.4, index into Table-35) */ | ||||
| @@ -429,7 +429,7 @@ static av_cold void uninit(AVFilterContext *ctx) | |||||
| static inline int is_newline(uint32_t c) | static inline int is_newline(uint32_t c) | ||||
| { | { | ||||
| return (c == '\n' || c == '\r' || c == '\f' || c == '\v'); | |||||
| return c == '\n' || c == '\r' || c == '\f' || c == '\v'; | |||||
| } | } | ||||
| static int dtext_prepare_text(AVFilterContext *ctx) | static int dtext_prepare_text(AVFilterContext *ctx) | ||||