本文整理汇总了C++中PacketPtr::getStreamIndex方法的典型用法代码示例。如果您正苦于以下问题:C++ PacketPtr::getStreamIndex方法的具体用法?C++ PacketPtr::getStreamIndex怎么用?C++ PacketPtr::getStreamIndex使用的例子?那么, 这里精选的方法代码示例或许可以为您提供帮助。您也可以进一步了解该方法所在类PacketPtr
的用法示例。
在下文中一共展示了PacketPtr::getStreamIndex方法的4个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的C++代码示例。
示例1: putPacket
bool Packetizer::putPacket(PacketPtr ptr) {
bool result = false;
//in case of stream is not mapped through stream_data element, discard packet
if (_streams.find(ptr->getStreamIndex()) == _streams.end()) {
return false;
}
if (_codec_overlap.find(_streams[ptr->getStreamIndex()].decoder->getCodecId()) == _codec_overlap.end()) {
_codec_overlap[_streams[ptr->getStreamIndex()].decoder->getCodecId()] = 0;
}
result = processPacket(ptr);
return result;
}
示例2: formatWriter
void formatWriter(const ContainerPtr& container, const PacketPtr& packet)
{
packetSync.doTimeSync(packet);
//if (packet->getPts() != av::NoPts)
// packet->setDts(packet->getPts());
clog << "Write FRAME: " << packet->getStreamIndex()
<< ", PTS: " << packet->getPts()
<< ", DTS: " << packet->getDts()
<< ", timeBase: " << packet->getTimeBase()
<< ", time: " << packet->getTimeBase().getDouble() * packet->getPts()
<< endl;
container->writePacket(packet, true);
}
示例3: u
boost::shared_ptr<org::esb::hive::job::ProcessUnit> ProcessUnitBuilder::build(org::esb::av::PacketListPtr list) {
boost::shared_ptr<org::esb::hive::job::ProcessUnit> u(new org::esb::hive::job::ProcessUnit());
PacketPtr p = list.front();
if(list.size()==0)return u;
int idx = p->getStreamIndex();
u->_source_stream = _map_data[idx].instream;
u->_target_stream = _map_data[idx].outstream;
u->_deinterlace = _map_data[idx].deinterlace;
u->_decoder = _map_data[idx].decoder;
u->_encoder = _map_data[idx].encoder;
u->_2passdecoder = _map_data[idx].pass2decoder;
u->_2passencoder = _map_data[idx].pass2encoder;
u->_input_packets = std::list<boost::shared_ptr<Packet> >(list.begin(), list.end());
u->_gop_size = list.size(); //- _map_data[idx].b_frame_offset;
int cou = u->_gop_size;
u->_frameRateCompensateBase = _map_data[idx].frameRateCompensateBase;
if (u->_decoder->getCodecType() == AVMEDIA_TYPE_VIDEO) {
AVRational input_framerate = u->_decoder->getFrameRate();
AVRational output_framerate = u->_encoder->getFrameRate();
//
double in = (((double) u->_gop_size) / input_framerate.num * input_framerate.den * output_framerate.num / output_framerate.den);
//double in = ((((((double) u->_gop_size)*u->_decoder->getFrameRate().den)/u->_decoder->getFrameRate().num)/u->_encoder->getFrameRate().den)*u->_encoder->getFrameRate().num);
in += _map_data[idx].frameRateCompensateBase;
/*this against some round issues*/
in += 0.000000001;
LOGDEBUG("Decimal input:"<<in);
/*spliting the double value into Integral and Fractional parts*/
org::esb::util::Decimal dec(in);
org::esb::util::Decimal::MantissaType i;
org::esb::util::Decimal::MantissaType f;
org::esb::util::Decimal::ExponentType exp = 0;
dec.getIntegralFractionalExponent<org::esb::util::Decimal::MantissaType > (i, f, exp, dec.getExponent());
/*in case to BFrames the resulting frame count is -1
* because there 1 I-Frame to much at the end
*
* e.g. IBBPBBPBBPBBPBBIBBP
* |
* this i-frame is to much
* it is always the last i-frame in a ProcessUnit
*
* this only happend when the decoder has B-Frames
*/
if (u->_decoder->getCodecId() == CODEC_ID_MPEG2VIDEO && u->_decoder->getCodecOption("has_b_frames") == "1")
i -= 1;
u->_expected_frame_count = static_cast<int> (i);
_map_data[idx].frameRateCompensateBase = f / pow((double) 10, dec.getExponent()*-1);
}
if (u->_decoder->getCodecType() == AVMEDIA_TYPE_AUDIO) {
AVRational input_timebase = u->_decoder->getTimeBase();
AVRational output_timebase = u->_encoder->getTimeBase();
int64_t input_samplerate = u->_decoder->getSampleRate();
int64_t output_samplerate = u->_encoder->getSampleRate();
int64_t output_framesize = u->_encoder->getFrameBytes();
/*
* the decoder did not returns a correct frame size
* calculation on the base of the packet->duration and packet->timebase
*/
int64_t input_framesize = av_rescale_q(p->getDuration(), p->getTimeBase(), input_timebase);
int osize = av_get_bytes_per_sample(u->_decoder->getSampleFormat());
input_framesize *= osize;
input_framesize *= u->_decoder->ctx->request_channel_layout;
int64_t raw_in_samples = input_framesize * list.size();
int64_t raw_out_samples = av_rescale_q(raw_in_samples, input_timebase, output_timebase);
raw_out_samples += _map_data[idx].frameRateCompensateBase;
int rest = raw_out_samples % (output_framesize > 0 ? output_framesize : 1);
double out;
double delta = modf(raw_out_samples / (output_framesize > 0 ? output_framesize : 1), &out);
u->_expected_frame_count = -1; //static_cast<int> (out);
_map_data[idx].frameRateCompensateBase = rest;
// LOGDEBUG(rest);
}
return u;
}
示例4: processPacket
/***
* this function packetizes the GOPs in chunks between the IFrames
* i normal cases the order of the Pictures are
* here will be each group of Picture be chunked, all of them cutted in the position of an I-Frame
* IPPPPPPPPIPPPPPPPPIPPPPPPPP
* IPPPPPPPP
* IPPPPPPPP
* IPPPPPPPP
*
* in a MPeg Stream there is a reorder with BFrames
* each group will be also chunked in the position of the I-Frame, but in the case of the delay from the Decoder
* we will append the next packets until a P-Frame arrived.
* IPBBPBBPBBIBBPBBPBBIBBPBBPBBIBBP
* IPBBPBBPBBIBB
* I PBBPBBIBB
* I PBBPBBIBB
*
*
*/
bool Packetizer::processPacket(PacketPtr ptr) {
bool result = false;
int stream_idx = ptr->getStreamIndex();
if (_streams[stream_idx].state == STATE_NOP && (ptr->isKeyFrame() || _streams[stream_idx].decoder->getCodecType() == AVMEDIA_TYPE_AUDIO)) {
_streams[stream_idx].state = STATE_START_I_FRAME;
}
if (_streams[stream_idx].state == STATE_START_I_FRAME && _streams[stream_idx].packets.size() >= _streams[stream_idx].min_packet_count && ptr->isKeyFrame()) {
_streams[stream_idx].state = STATE_END_I_FRAME;
}
if (_streams[stream_idx].state == STATE_START_I_FRAME) {
_streams[stream_idx].packets.push_back(ptr);
}
if (_streams[stream_idx].state == STATE_END_I_FRAME) {
_overlap_queue[stream_idx].push_back(ptr);
}
/**
* case handling for mpeg2 video packets(formaly streams with b frames)
* */
if (_streams[stream_idx].state == STATE_END_I_FRAME && (_streams[stream_idx].decoder->getCodecId() == CODEC_ID_MPEG2VIDEO && ptr->_pict_type == AV_PICTURE_TYPE_P)) {
/*in the first roundtrip the stream packets look like this, the first B Frames are not removed*/
/*_streams[stream_idx].packets =IBBPBBPBBPBB*/
/*in the following roundtrip the stream packets look like this*/
/*_streams[stream_idx].packets =IPBBPBBPBB*/
/*_overlap_queue[stream_idx] =IBBP*/
_streams[stream_idx].state = STATE_START_I_FRAME;
/**
* appending the next IBB from the IBBP order to the actual packet_list
* */
_streams[stream_idx].packets.insert(_streams[stream_idx].packets.end(), _overlap_queue[stream_idx].begin(), _overlap_queue[stream_idx].end() - 1);
/*_streams[stream_idx].packets =IBBPBBPBBPBBIBB*/
/*_overlap_queue[stream_idx] =IBBP*/
_packet_list.push_back(_streams[stream_idx].packets);
_streams[stream_idx].packets.clear();
/*_streams[stream_idx].packets = */
/*_overlap_queue[stream_idx] =IBBP*/
/**
* appending the IP frames from the IBBP order to the actual packet_list
* that are the first and the last entries in the overlap queue
* */
_streams[stream_idx].packets.insert(_streams[stream_idx].packets.end(), _overlap_queue[stream_idx].begin(), _overlap_queue[stream_idx].begin() + 1);
_streams[stream_idx].packets.insert(_streams[stream_idx].packets.end(), _overlap_queue[stream_idx].end() - 1, _overlap_queue[stream_idx].end());
// _streams[stream_idx].packets.insert(_streams[stream_idx].packets.end(), _overlap_queue[stream_idx].begin(), _overlap_queue[stream_idx].end());
/*_streams[stream_idx].packets =IP */
/*_overlap_queue[stream_idx] =IBBP*/
_overlap_queue[stream_idx].clear();
/*_streams[stream_idx].packets =IP */
/*_overlap_queue[stream_idx] = */
result = true;
} else if (_streams[stream_idx].state == STATE_END_I_FRAME && _streams[stream_idx].decoder->getCodecType() == AVMEDIA_TYPE_AUDIO) {
/**************************************
* this is used for all audio streams
**************************************/
_streams[stream_idx].state = STATE_START_I_FRAME;
/**copying all Packets into the actual ProcessUnit*/
_packet_list.push_back(_streams[stream_idx].packets);
/**clear out the PacketList, because they are in the ProcessUnit*/
_streams[stream_idx].packets.clear();
/**bring in the first I-Frame for the next Process Unit*/
_streams[stream_idx].packets.insert(_streams[stream_idx].packets.end(), _overlap_queue[stream_idx].begin(), _overlap_queue[stream_idx].end());
/**clear out the overlap queue, because there is only an I-Frame packet and this is now in the next ProcessUnit*/
_overlap_queue[stream_idx].clear();
result = true;
} else if (_streams[stream_idx].state == STATE_END_I_FRAME && _streams[stream_idx].decoder->getCodecId() != CODEC_ID_MPEG2VIDEO) {
/**********************************************************************
* this is used for all video stream types except MPEG2 Video Streams
**********************************************************************/
_streams[stream_idx].state = STATE_START_I_FRAME;
/**copying all Packets into the actual ProcessUnit*/
_packet_list.push_back(_streams[stream_idx].packets);
/**clear out the PacketList, because they are in the ProcessUnit*/
//.........这里部分代码省略.........