You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

529 lines
15 KiB

/*
* H.265 video codec.
* Copyright (c) 2013-2014 struktur AG, Dirk Farin <farin@struktur.de>
*
* This file is part of libde265.
*
* libde265 is free software: you can redistribute it and/or modify
* it under the terms of the GNU Lesser General Public License as
* published by the Free Software Foundation, either version 3 of
* the License, or (at your option) any later version.
*
* libde265 is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public License
* along with libde265. If not, see <http://www.gnu.org/licenses/>.
*/
#ifndef DE265_DECCTX_H
#define DE265_DECCTX_H
#include "libde265/vps.h"
#include "libde265/sps.h"
#include "libde265/pps.h"
#include "libde265/nal.h"
#include "libde265/slice.h"
#include "libde265/image.h"
#include "libde265/motion.h"
#include "libde265/de265.h"
#include "libde265/dpb.h"
#include "libde265/sei.h"
#include "libde265/threads.h"
#include "libde265/acceleration.h"
#include "libde265/nal-parser.h"
#include <memory>
#define DE265_MAX_VPS_SETS 16 // this is the maximum as defined in the standard
#define DE265_MAX_SPS_SETS 16 // this is the maximum as defined in the standard
#define DE265_MAX_PPS_SETS 64 // this is the maximum as defined in the standard
#define MAX_WARNINGS 20
class slice_segment_header;
class image_unit;
class slice_unit;
class decoder_context;
class thread_context
{
public:
thread_context();
int CtbAddrInRS;
int CtbAddrInTS;
int CtbX, CtbY;
// motion vectors
PBMotionCoding motion;
// prediction
// enum IntraPredMode IntraPredModeC[4]; // chroma intra-prediction mode for current CB
int ResScaleVal;
// residual data
uint8_t cu_transquant_bypass_flag;
uint8_t transform_skip_flag[3];
uint8_t explicit_rdpcm_flag;
uint8_t explicit_rdpcm_dir;
// we need 16 bytes of extra memory (8*int16) to shift the base for the
// alignment required for SSE code !
int16_t _coeffBuf[(32*32)+8];
int16_t *coeffBuf; // the base pointer for into _coeffBuf, aligned to 16 bytes
int16_t coeffList[3][32*32];
int16_t coeffPos[3][32*32];
int16_t nCoeff[3];
int32_t residual_luma[32*32]; // only used when cross-comp-prediction is enabled
// quantization
int IsCuQpDeltaCoded;
int CuQpDelta;
int IsCuChromaQpOffsetCoded;
int CuQpOffsetCb, CuQpOffsetCr;
int currentQPY;
int currentQG_x, currentQG_y;
int lastQPYinPreviousQG;
int qPYPrime, qPCbPrime, qPCrPrime;
CABAC_decoder cabac_decoder;
context_model_table ctx_model;
uint8_t StatCoeff[4];
decoder_context* decctx;
struct de265_image *img;
slice_segment_header* shdr;
image_unit* imgunit;
slice_unit* sliceunit;
thread_task* task; // executing thread_task or NULL if not multi-threaded
private:
thread_context(const thread_context&); // not allowed
const thread_context& operator=(const thread_context&); // not allowed
};
class error_queue
{
public:
error_queue();
void add_warning(de265_error warning, bool once);
de265_error get_warning();
private:
de265_error warnings[MAX_WARNINGS];
int nWarnings;
de265_error warnings_shown[MAX_WARNINGS]; // warnings that have already occurred
int nWarningsShown;
};
class slice_unit
{
public:
slice_unit(decoder_context* decctx);
~slice_unit();
NAL_unit* nal; // we are the owner
slice_segment_header* shdr; // not the owner (de265_image is owner)
bitreader reader;
image_unit* imgunit;
bool flush_reorder_buffer;
// decoding status
enum SliceDecodingProgress { Unprocessed,
InProgress,
Decoded
} state;
de265_progress_lock finished_threads;
int nThreads;
int first_decoded_CTB_RS; // TODO
int last_decoded_CTB_RS; // TODO
void allocate_thread_contexts(int n);
thread_context* get_thread_context(int n) {
assert(n < nThreadContexts);
return &thread_contexts[n];
}
int num_thread_contexts() const { return nThreadContexts; }
private:
thread_context* thread_contexts; /* NOTE: cannot use std::vector, because thread_context has
no copy constructor. */
int nThreadContexts;
public:
decoder_context* ctx;
private:
slice_unit(const slice_unit&); // not allowed
const slice_unit& operator=(const slice_unit&); // not allowed
};
class image_unit
{
public:
image_unit();
~image_unit();
de265_image* img;
de265_image sao_output; // if SAO is used, this is allocated and used as SAO output buffer
std::vector<slice_unit*> slice_units;
std::vector<sei_message> suffix_SEIs;
slice_unit* get_next_unprocessed_slice_segment() const {
for (int i=0;i<slice_units.size();i++) {
if (slice_units[i]->state == slice_unit::Unprocessed) {
return slice_units[i];
}
}
return NULL;
}
slice_unit* get_prev_slice_segment(slice_unit* s) const {
for (int i=1; i<slice_units.size(); i++) {
if (slice_units[i]==s) {
return slice_units[i-1];
}
}
return NULL;
}
slice_unit* get_next_slice_segment(slice_unit* s) const {
for (int i=0; i<slice_units.size()-1; i++) {
if (slice_units[i]==s) {
return slice_units[i+1];
}
}
return NULL;
}
void dump_slices() const {
for (int i=0; i<slice_units.size(); i++) {
printf("[%d] = %p\n",i,slice_units[i]);
}
}
bool all_slice_segments_processed() const {
if (slice_units.size()==0) return true;
if (slice_units.back()->state != slice_unit::Unprocessed) return true;
return false;
}
bool is_first_slice_segment(const slice_unit* s) const {
if (slice_units.size()==0) return false;
return (slice_units[0] == s);
}
enum { Invalid, // headers not read yet
Unknown, // SPS/PPS available
Reference, // will be used as reference
Leaf // not a reference picture
} role;
enum { Unprocessed,
InProgress,
Decoded,
Dropped // will not be decoded
} state;
std::vector<thread_task*> tasks; // we are the owner
/* Saved context models for WPP.
There is one saved model for the initialization of each CTB row.
The array is unused for non-WPP streams. */
std::vector<context_model_table> ctx_models; // TODO: move this into image ?
};
class base_context : public error_queue
{
public:
base_context();
virtual ~base_context() { }
// --- accelerated DSP functions ---
void set_acceleration_functions(enum de265_acceleration);
struct acceleration_functions acceleration; // CPU optimized functions
//virtual /* */ de265_image* get_image(int dpb_index) { return dpb.get_image(dpb_index); }
virtual const de265_image* get_image(int frame_id) const = 0;
virtual bool has_image(int frame_id) const = 0;
};
class decoder_context : public base_context {
public:
decoder_context();
~decoder_context();
de265_error start_thread_pool(int nThreads);
void stop_thread_pool();
void reset();
bool has_sps(int id) const { return (bool)sps[id]; }
bool has_pps(int id) const { return (bool)pps[id]; }
std::shared_ptr<const seq_parameter_set> get_shared_sps(int id) { return sps[id]; }
std::shared_ptr<const pic_parameter_set> get_shared_pps(int id) { return pps[id]; }
/* */ seq_parameter_set* get_sps(int id) { return sps[id].get(); }
const seq_parameter_set* get_sps(int id) const { return sps[id].get(); }
/* */ pic_parameter_set* get_pps(int id) { return pps[id].get(); }
const pic_parameter_set* get_pps(int id) const { return pps[id].get(); }
/*
const slice_segment_header* get_SliceHeader_atCtb(int ctb) {
return img->slices[img->get_SliceHeaderIndex_atIndex(ctb)];
}
*/
uint8_t get_nal_unit_type() const { return nal_unit_type; }
bool get_RapPicFlag() const { return RapPicFlag; }
de265_error decode_NAL(NAL_unit* nal);
de265_error decode(int* more);
de265_error decode_some(bool* did_work);
de265_error decode_slice_unit_sequential(image_unit* imgunit, slice_unit* sliceunit);
de265_error decode_slice_unit_parallel(image_unit* imgunit, slice_unit* sliceunit);
de265_error decode_slice_unit_WPP(image_unit* imgunit, slice_unit* sliceunit);
de265_error decode_slice_unit_tiles(image_unit* imgunit, slice_unit* sliceunit);
void process_nal_hdr(nal_header*);
bool process_slice_segment_header(slice_segment_header*,
de265_error*, de265_PTS pts,
nal_header* nal_hdr, void* user_data);
//void push_current_picture_to_output_queue();
de265_error push_picture_to_output_queue(image_unit*);
// --- parameters ---
bool param_sei_check_hash;
bool param_conceal_stream_errors;
bool param_suppress_faulty_pictures;
int param_sps_headers_fd;
int param_vps_headers_fd;
int param_pps_headers_fd;
int param_slice_headers_fd;
bool param_disable_deblocking;
bool param_disable_sao;
//bool param_disable_mc_residual_idct; // not implemented yet
//bool param_disable_intra_residual_idct; // not implemented yet
void set_image_allocation_functions(de265_image_allocation* allocfunc, void* userdata);
de265_image_allocation param_image_allocation_functions;
void* param_image_allocation_userdata;
// --- input stream data ---
NAL_Parser nal_parser;
int get_num_worker_threads() const { return num_worker_threads; }
/* */ de265_image* get_image(int dpb_index) { return dpb.get_image(dpb_index); }
const de265_image* get_image(int dpb_index) const { return dpb.get_image(dpb_index); }
bool has_image(int dpb_index) const { return dpb_index>=0 && dpb_index<dpb.size(); }
de265_image* get_next_picture_in_output_queue() { return dpb.get_next_picture_in_output_queue(); }
int num_pictures_in_output_queue() const { return dpb.num_pictures_in_output_queue(); }
void pop_next_picture_in_output_queue() { dpb.pop_next_picture_in_output_queue(); }
private:
de265_error read_vps_NAL(bitreader&);
de265_error read_sps_NAL(bitreader&);
de265_error read_pps_NAL(bitreader&);
de265_error read_sei_NAL(bitreader& reader, bool suffix);
de265_error read_eos_NAL(bitreader& reader);
de265_error read_slice_NAL(bitreader&, NAL_unit* nal, nal_header& nal_hdr);
private:
// --- internal data ---
std::shared_ptr<video_parameter_set> vps[ DE265_MAX_VPS_SETS ];
std::shared_ptr<seq_parameter_set> sps[ DE265_MAX_SPS_SETS ];
std::shared_ptr<pic_parameter_set> pps[ DE265_MAX_PPS_SETS ];
std::shared_ptr<video_parameter_set> current_vps;
std::shared_ptr<seq_parameter_set> current_sps;
std::shared_ptr<pic_parameter_set> current_pps;
public:
thread_pool thread_pool_;
private:
int num_worker_threads;
public:
// --- frame dropping ---
void set_limit_TID(int tid);
int get_highest_TID() const;
int get_current_TID() const { return current_HighestTid; }
int change_framerate(int more_vs_less); // 1: more, -1: less
void set_framerate_ratio(int percent);
private:
// input parameters
int limit_HighestTid; // never switch to a layer above this one
int framerate_ratio;
// current control parameters
int goal_HighestTid; // this is the layer we want to decode at
int layer_framerate_ratio; // ratio of frames to keep in the current layer
int current_HighestTid; // the layer which we are currently decoding
struct {
int8_t tid;
int8_t ratio;
} framedrop_tab[100+1];
int framedrop_tid_index[6+1];
void compute_framedrop_table();
void calc_tid_and_framerate_ratio();
private:
// --- decoded picture buffer ---
decoded_picture_buffer dpb;
int current_image_poc_lsb;
bool first_decoded_picture;
bool NoRaslOutputFlag;
bool HandleCraAsBlaFlag;
bool FirstAfterEndOfSequenceNAL;
int PicOrderCntMsb;
int prevPicOrderCntLsb; // at precTid0Pic
int prevPicOrderCntMsb; // at precTid0Pic
de265_image* img;
public:
const slice_segment_header* previous_slice_header; /* Remember the last slice for a successive
dependent slice. */
// --- motion compensation ---
public:
int PocLsbLt[MAX_NUM_REF_PICS];
int UsedByCurrPicLt[MAX_NUM_REF_PICS];
int DeltaPocMsbCycleLt[MAX_NUM_REF_PICS];
private:
int CurrDeltaPocMsbPresentFlag[MAX_NUM_REF_PICS];
int FollDeltaPocMsbPresentFlag[MAX_NUM_REF_PICS];
// The number of entries in the lists below.
int NumPocStCurrBefore;
int NumPocStCurrAfter;
int NumPocStFoll;
int NumPocLtCurr;
int NumPocLtFoll;
// These lists contain absolute POC values.
int PocStCurrBefore[MAX_NUM_REF_PICS]; // used for reference in current picture, smaller POC
int PocStCurrAfter[MAX_NUM_REF_PICS]; // used for reference in current picture, larger POC
int PocStFoll[MAX_NUM_REF_PICS]; // not used for reference in current picture, but in future picture
int PocLtCurr[MAX_NUM_REF_PICS]; // used in current picture
int PocLtFoll[MAX_NUM_REF_PICS]; // used in some future picture
// These lists contain indices into the DPB.
int RefPicSetStCurrBefore[MAX_NUM_REF_PICS];
int RefPicSetStCurrAfter[MAX_NUM_REF_PICS];
int RefPicSetStFoll[MAX_NUM_REF_PICS];
int RefPicSetLtCurr[MAX_NUM_REF_PICS];
int RefPicSetLtFoll[MAX_NUM_REF_PICS];
// --- parameters derived from parameter sets ---
// NAL
uint8_t nal_unit_type;
char IdrPicFlag;
char RapPicFlag;
// --- image unit queue ---
std::vector<image_unit*> image_units;
bool flush_reorder_buffer_at_this_frame;
private:
void init_thread_context(thread_context* tctx);
void add_task_decode_CTB_row(thread_context* tctx, bool firstSliceSubstream, int ctbRow);
void add_task_decode_slice_segment(thread_context* tctx, bool firstSliceSubstream,
int ctbX,int ctbY);
void mark_whole_slice_as_processed(image_unit* imgunit,
slice_unit* sliceunit,
int progress);
void process_picture_order_count(slice_segment_header* hdr);
int generate_unavailable_reference_picture(const seq_parameter_set* sps,
int POC, bool longTerm);
void process_reference_picture_set(slice_segment_header* hdr);
bool construct_reference_picture_lists(slice_segment_header* hdr);
void remove_images_from_dpb(const std::vector<int>& removeImageList);
void run_postprocessing_filters_sequential(struct de265_image* img);
void run_postprocessing_filters_parallel(image_unit* img);
};
#endif