• SRS之接收推流线程:recv


    SrsPublishRecvThread、SrsRecvThread、SrsReusableThread2、SrsThread 之间的关系图

    image

    1. recv 线程函数:SrsThread::thread_fun

    void *SrsThread::thread_fun(void *arg)
    {
        SrsThread* obj = (SrsThread*)arg;
        srs_assert(obj);
        
        /* 进入线程循环 */
        obj->thread_cycle();
        
        // for valgrind to detect.
        SrsThreadContext* ctx = dynamic_cast<SrsThreadContext*>(_srs_context);
        if (ctx) {
            ctx->clear_cid();
        }
        
        st_thread_exit(NULL);
        
        return NULL;
    }
    

    1.1 SrsThread::thread_cycle

    void SrsThread::thread_cycle()
    {
        int ret = ERROR_SUCCESS;
        
        /* 生成 recv 线程的一个上下文 id */
        _srs_context->generate_id();
        srs_info("thread %s cycle start", _name);
        
        /* 将生成的 recv 上下文 id 赋给 _cid,以便 recv 的父线程醒来后
         * 可以继续往下执行,此时父线程会设置 can_run 为 true */
        _cid = _srs_context->get_id();
        
        srs_assert(handler);
        /* 调用 SrsReusableThread2 实现的 on_thread_start 函数 */
        handler->on_thread_start();
        
        // thread is running now.
        really_terminated = false;
        
        /* 上面生成好 recv 线程的上下文 id 后,这里会陷入休眠,接着会
         * 调度到 recv 的父线程执行,父线程检测到 _cid 准备好后,
         * 即会设置 can_run 为 true,表示 recv 线程可以继续往下执行了 */
        // wait for cid to ready, for parent thread to get the cid.
        while (!can_run && loop) {
            st_usleep(10 * 1000);
        }
        
        while (loop) {
            /* 该函数没有具体做任何事,忽略 */
            if ((ret = handler->on_before_cycle()) != ERROR_SUCCESS) {
                srs_warn("thread %s on before cycle failed, ignored and retry, ret=%d", 
                         _name, ret);
                goto failed;
            }
            srs_info("thread %s on before cycle success", _name);
            
            /* 调用 SrsReusableThread2 实现的 cycle 函数 */
            if ((ret = handler->cycle()) != ERROR_SUCCESS) {
                if (!srs_is_client_gracefully_close(ret) && !srs_is_system_control_error(ret)) 
                {
                    srs_warn("thread %s cycle failed, ignored and retry, ret=%d", _name, ret);
                }
                goto failed;
            }
            srs_info("thread %s cycle success", _name);
            
            if ((ret = handler->on_end_cycle()) != ERROR_SUCCESS) {
                srs_warn("thread %s on end cycle failed, ignored and retry, ret=%d", 
                         _name, ret);
                goto failed;
            }
            srs_info("thread %s on end cycle success", _name);
            
        failed:
            if (!loop) {
                break;
            }
            
            // to improve performance, donot sleep when interval is zero.
            // @see: https://github.com/ossrs/srs/issues/237
            if (cycle_interval_us != 0) {
                st_usleep(cycle_interval_us);
            }
        }
        
        // readly terminated now.
        really_terminated = true;
        
        handler->on_thread_stop();
        srs_info("thread %s cycle finished", _name);
    }
    

    1.2 SrsReusableThread2::on_thread_start

    void SrsReusableThread2::on_thread_start()
    {
        handler->on_thread_start();
    }
    

    该函数中接着调用 SrsRecvThread 实现的 on_thread_start 函数。

    1.2.1 SrsRecvThread::on_thread_start

    void SrsRecvThread::on_thread_start()
    {
        // the multiple messages writev improve performance large,
        // but the timeout recv will cause 33% sys call performance,
        // to use isolate thread to recv, can improve about 33% performance.
        // @see https://github.com/ossrs/srs/issues/194
        // @see: https://github.com/ossrs/srs/issues/217
        rtmp->set_recv_timeout(ST_UTIME_NO_TIMEOUT);
        
        handler->on_thread_start();
    }
    

    函数先设置 recv 的超时时间为 -1,然后接着调用 SrsPublishRecvThread 实现的 on_thread_start 函数。

    1.2.2 SrsPublishRecvThread::on_thread_start

    void SrsPublishRecvThread::on_thread_start()
    {
        // we donot set the auto response to false,
        // for the main thread never send message.
        
        /* 若配置文件中没有设置 mr 配置项,则默认没有启用该功能,可忽略 */
    #ifdef SRS_PERF_MERGED_READ
        if (mr) {
            // set underlayer buffer size
            set_socket_buffer(mr_sleep);
            
            // disable the merge read
            // @see https://github.com/ossrs/srs/issues/241
            rtmp->set_merge_read(true, this);
        }
    #endif
    }
    

    1.3 SrsReusableThread2::cycle

    int SrsReusableThread2::cycle()
    {
        return handler->cycle();
    }
    

    接着调用 SrsRecvThread 实现的 cycle 函数,该函数才开始真正接收客户端推流的数据。

    2. 接收推流数据:SrsRecvThread::cycle

    int SrsRecvThread::cycle()
    {
        int ret = ERROR_SUCCESS;
        
        /* 若当前没有被中断的情况下,进入循环开始接收客户端的消息 */
        while (!trd->interrupted()) {
            /* 调用 SrsPublishRecvThread 实现的 can_handle 函数,该函数
             * 默认返回 true,即默认推流线程总是可以处理消息 */
            if (!handler->can_handler()) {
                st_usleep(timeout * 1000);
                continue;
            }
            
            SrsCommonMessage* msg = NULL;
            
            // recv and handle message
            ret = rtmp->recv_message(&msg);
            if (ret == ERROR_SUCCESS) {
                /* 调用 SrsPublishRecvThread 实现的 handle 函数处理接收到的消息 */
                ret = handler->handle(msg);
            }
            
            /* 若发生错误,则中断当前 recv 线程 */
            if (ret != ERROR_SUCCESS) {
                if (!srs_is_client_gracefully_close(ret) && !srs_is_system_control_error(ret)) 
                {
                    srs_error("thread process message failed. ret=%d", ret);
                }
                
                // we use no timeout to recv, should never got any error.
                trd->interrupt();
                
                // notice the handler got a recv error
                handler->on_recv_error(ret);
                
                return ret;
            }
            srs_verbose("thread loop recv message. ret=%d", ret);
        }
        
        return ret;
    }
    

    3. 接收推流数据:SrsRtmpServer::recv_message

    int SrsRtmpServer::recv_message(SrsCommonMessage** pmsg)
    {
        return protocol->recv_message(pmsg);
    }
    

    该函数接着调用 SrsProtocol 实现的 recv_message 函数。

    3.1 SrsProtocol::recv_message

    int SrsProtocol::recv_message(SrsCommonMessage** pmsg)
    {
        *pmsg = NULL;
        
        int ret = ERROR_SUCCESS;
        
        while (true) {
            SrsCommonMessage* msg = NULL;
            
            if ((ret = recv_interlaced_message(&msg)) != ERROR_SUCCESS) {
                if (ret != ERROR_SOCKET_TIMEOUT && !srs_is_client_gracefully_close(ret)) {
                    srs_error("recv interlaced message failed. ret=%d", ret);
                }
                srs_freep(msg);
                return ret;
            }
            srs_verbose("entire msg received");
            
            /* 若获取到一个空消息,则继续获取下一个消息 */
            if (!msg) {
                srs_info("got empty message without error.");
                continue;
            }
            
            if (msg->size <= 0 || msg->header.payload_length <= 0) {
                srs_trace("ignore empty message(type=%d, size=%d, time=%"PRId64", sid=%d).",
                    msg->header.message_type, msg->header.payload_length,
                    msg->header.timestamp, msg->header.stream_id);
                srs_freep(msg);
                continue;
            }
            
            /* 该函数首先检测当前接收到的字节数是否已经达到当前窗口大小,若是,则回应客户端窗口消息
             * 然后接着对接收到的若为 应答窗口大小(5)、设置块大小(1)、用户控制消息(4) 则会进行解码,
             * 并根据解析后的内容更新当前 rtmp 服务器的上下文信息 */
            if ((ret = on_recv_message(msg)) != ERROR_SUCCESS) {
                srs_error("hook the received msg failed. ret=%d", ret);
                srs_freep(msg);
                return ret;
            }
            
            srs_verbose("got a msg, cid=%d, type=%d, size=%d, time=%"PRId64, 
                msg->header.perfer_cid, msg->header.message_type, msg->header.payload_length, 
                msg->header.timestamp);
            *pmsg = msg;
            break;
        }
    }
    

    3.2 SrsProtocol::recv_interlaced_message

    int SrsProtocol::recv_interlaced_message(SrsCommonMessage** pmsg)
    {
        int ret = ERROR_SUCCESS;
        
        // chunk stream basic header.
        char fmt = 0;
        int cid = 0;
        /* 读取 chunk 的基本头 */
        if ((ret = read_basic_header(fmt, cid)) != ERROR_SUCCESS) {
            if (ret != ERROR_SOCKET_TIMEOUT && !srs_is_client_gracefully_close(ret)) {
                srs_error("read basic header failed. ret=%d", ret);
            }
            return ret;
        }
        srs_verbose("read basic header success. fmt=%d, cid=%d", fmt, cid);
        
        // the cid must not negative.
        srs_assert(cid >= 0);
        
        // get the cached chunk stream.
        SrsChunkStream* chunk = NULL;
        
        /* 一个消息客户端可能会分成几个 chunk 发送,因此需要把每次读取
         * 的 chunk 的信息和负载缓存起来 */
        // use chunk stream cache to get the chunk info.
        // @see https://github.com/ossrs/srs/issues/249
        if (cid < SRS_PERF_CHUNK_STREAM_CACHE) {
            // chunk stream cache hit.
            srs_verbose("cs-cache hit, cid=%d", cid);
            // already init, use it direclty
            chunk = cs_cache[cid];
            srs_verbose("cached chunk stream: fmt=%d, cid=%d, size=%d, "
                        "message(type=%d, size=%d, time=%"PRId64", sid=%d)",
                chunk->fmt, chunk->cid, (chunk->msg? chunk->msg->size : 0), 
                chunk->header.message_type, chunk->header.payload_length,
                chunk->header.timestamp, chunk->header.stream_id);
        } else {
            // chunk stream cache miss, use map.
            if (chunk_streams.find(cid) == chunk_streams.end()) {
                chunk = chunk_streams[cid] = new SrsChunkStream(cid);
                // set the perfer cid of chunk,
                // which will copy to the message received.
                chunk->header.perfer_cid = cid;
                srs_verbose("cache new chunk stream: fmt=%d, cid=%d", fmt, cid);
            } else {
                chunk = chunk_streams[cid];
                srs_verbose("cached chunk stream: fmt=%d, cid=%d, size=%d, "
                            "message(type=%d, size=%d, time=%"PRId64", sid=%d)",
                    chunk->fmt, chunk->cid, (chunk->msg? chunk->msg->size : 0), 
                    chunk->header.message_type, chunk->header.payload_length,
                    chunk->header.timestamp, chunk->header.stream_id);
            }
        }
    
        // chunk stream message header
        if ((ret = read_message_header(chunk, fmt)) != ERROR_SUCCESS) {
            if (ret != ERROR_SOCKET_TIMEOUT && !srs_is_client_gracefully_close(ret)) {
                srs_error("read message header failed. ret=%d", ret);
            }
            return ret;
        }
        srs_verbose("read message header success. fmt=%d, ext_time=%d, size=%d, "
                "message(type=%d, size=%d, time=%"PRId64", sid=%d)", 
                fmt, chunk->extended_timestamp, (chunk->msg? chunk->msg->size : 0), 
                chunk->header.message_type, chunk->header.payload_length, 
                chunk->header.timestamp, chunk->header.stream_id);
        
        // read msg payload from chunk stream.
        SrsCommonMessage* msg = NULL;
        if ((ret = read_message_payload(chunk, &msg)) != ERROR_SUCCESS) {
            if (ret != ERROR_SOCKET_TIMEOUT && !srs_is_client_gracefully_close(ret)) {
                srs_error("read message payload failed. ret=%d", ret);
            }
            return ret;
        }
        
        // not got an entire RTMP message, try next chunk.
        if (!msg) {
            srs_verbose("get partial message success. size=%d, "
                        "message(type=%d, size=%d, time=%"PRId64", sid=%d)",
                    (msg? msg->size : (chunk->msg? chunk->msg->size : 0)), 
                    chunk->header.message_type, chunk->header.payload_length,
                    chunk->header.timestamp, chunk->header.stream_id);
            return ret;
        }
        
        /* 获取到完整的消息 */
        *pmsg = msg;
        srs_info("get entire message success. size=%d, "
                 "message(type=%d, size=%d, time=%"PRId64", sid=%d)",
                (msg? msg->size : (chunk->msg? chunk->msg->size : 0)), 
                chunk->header.message_type, chunk->header.payload_length,
                chunk->header.timestamp, chunk->header.stream_id);
                
        return ret;
    }
    

    4. 处理推流消息:SrsPublishRecvThread::handle

    int SrsPublishRecvThread::handle(SrsCommonMessage* msg)
    {
        int ret = ERROR_SUCCESS;
        
        // when cid changed, change it.
        if (ncid != cid) {
            _srs_context->set_id(ncid);
            cid = ncid;
        }
        
        /* 每接收到一个消息,该将该消息计数值加 1 */
        _nb_msgs++;
        
        /* 若当前消息为视频,则视频帧数加 1 */
        if (msg->header.is_video()) {
            video_frames++;
        }
    
        /* log to show the time of recv thread. */    
        srs_verbose("recv thread now=%"PRId64"us, got msg time=%"PRId64"ms, size=%d",
            srs_update_system_time_ms(), msg->header.timestamp, msg->size);
    
        /* the rtmp connection will handle this message. */
        ret = _conn->handle_publish_message(_source, msg, _is_fmle, _is_edge);
        
        /* must always free it,
         * the source will copy it if need to use. */
        srs_freep(msg);
        
        return ret;
    }
    

    该函数接着主要调用 SrsRtmpConn 实现的 handle_publish_message 函数。

    4.1 SrsRtmpConn::handle_publish_message

    int SrsRtmpConn::handle_publish_message(SrsSource* source, SrsCommonMessage* msg, 
        bool is_fmle, bool vhost_is_edge)
    {
        int ret = ERROR_SUCCESS;
        
        /* process publish event. */
        if (msg->header.is_amf0_command() || msg->header.is_amf3_command()) {
            SrsPacket* pkt = NULL;
            if ((ret = rtmp->decode_message(msg, &pkt)) != ERROR_SUCCESS) {
                srs_error("fmle decode unpublish message failed. ret=%d", ret);
                return ret;
            }
            SrsAutoFree(SrsPacket, pkt);
            
            /* for flash, any packet is republish. */
            if (!is_fmle) {
                /* flash unpublish.
                 * TODO: maybe need to support republish. */
                srs_trace("flash flash publish finished.");
                return ERROR_CONTROL_REPUBLISH;
            }
            
            /* for fmle, drop others except the fmle start packet. */
            if (dynamic_cast<SrsFMLEStartPacket*>(pkt)) {
                SrsFMLEStartPacket* unpublish = dynamic_cast<SrsFMLEStartPacket*>(pkt);
                if ((ret = rtmp->fmle_unpublish(res->stream_id, unpublish->transaction_id)) 
                    != ERROR_SUCCESS) {
                    return ret;
                }
                return ERROR_CONTROL_REPUBLISH;
            }
            
            srs_trace("fmle ignore AMF0/AMF3 command message.");
            return ret;
        }
        
        /* video, audio, data message */
        if ((ret = process_publish_message(source, msg, vhost_is_edge)) != ERROR_SUCCESS) {
            srs_error("fmle process publish message failed. ret=%d", ret);
            return ret;
        }
        
        return ret;
    }
    

    这里暂先不分析接收到 unpublish 的情况,而对于接收到 video、audio 和 data message 等消息情况下,直接调用 SrsRtmpConn 实现的 process_publish_message 进行处理。

    5. 媒体数据的处理:SrsRtmpConn::process_publish_message

    int SrsRtmpConn::process_publish_message(SrsSource* source, SrsCommonMessage* msg, 
        bool vhost_is_edge)
    {
        int ret = ERROR_SUCCESS;
        
        // for edge, directly proxy message to origin.
        if (vhost_is_edge) {
            if ((ret = source->on_edge_proxy_publish(msg)) != ERROR_SUCCESS) {
                srs_error("edge publish proxy msg failed. ret=%d", ret);
                return ret;
            }
            return ret;
        }
        
        // process audio packet
        if (msg->header.is_audio()) {
            if ((ret = source->on_audio(msg)) != ERROR_SUCCESS) {
                srs_error("source process audio message failed. ret=%d", ret);
                return ret;
            }
            return ret;
        } 
        // process video packet
        if (msg->header.is_video()) {
            if ((ret = source->on_video(msg)) != ERROR_SUCCESS) {
                srs_error("source process video message failed. ret=%d", ret);
                return ret;
            }
            return ret;
        }
        
        // process aggregate packet
        if (msg->header.is_aggregate()) {
            if ((ret = source->on_aggregate(msg)) != ERROR_SUCCESS) {
                srs_error("source process aggregate message failed. ret=%d", ret);
                return ret;
            }
            return ret;
        }
        
        // process onMetadata
        if (msg->header.is_amf0_data() || msg->header.is_amf3_data()) {
            SrsPacket* pkt = NULL;
            /* 解析元数据 */
            if ((ret = rtmp->decode_message(msg, &pkt)) != ERROR_SUCCESS) {
                srs_error("decode onMetaData message failed. ret=%d", ret);
                return ret;
            }
            SrsAutoFree(SrsPacket, pkt);
            
            if (dynamic_cast<SrsOnMetaDataPacket*>(pkt)) {
                SrsOnMetaDataPacket* metadata = dynamic_cast<SrsOnMetaDataPacket*>(pkt);
                if ((ret = source->on_meta_data(msg, metadata)) != ERROR_SUCCESS) {
                    srs_error("source process onMetaData message failed. ret=%d", ret);
                    return ret;
                }
                srs_info("process onMetaData message success.");
                return ret;
            }
            
            srs_info("ignore AMF0/AMF3 data message.");
            return ret;
        }
        
        return ret;
    }
    

    5.1 onMetaData

    通常接收到的第一个媒体数据包一般为 onMetaData,抓包图如下图所示。
    image
    接收到 onMetaData 数据包后,需要调用 SrsRtmpServer 实现的 decode_message 函数对该包进行解码。

    5.1.1 SrsRtmpServer::decode_message

    int SrsRtmpServer::decode_message(SrsCommonMessage* msg, SrsPacket** ppacket)
    {
        return protocol->decode_message(msg, ppacket);
    }
    

    该函数接着调用 SrsProtocol 实现的 decode_message 函数。

    5.1.2 SrsProtocol::decode_message

    int SrsProtocol::decode_message(SrsCommonMessage* msg, SrsPacket** packet)
    {
        *ppacket = NULL;
        
        int ret = ERROR_SUCCESS;
        
        srs_assert(msg != NULL);
        srs_assert(msg->payload != NULL);
        srs_assert(msg->size > 0);
        
        SrsStream stream;
        
        // initialize the decode stream for all message,
        // it's ok for the initialize if fast and without memory copy.
        if ((ret = stream.initialize(msg->payload, msg->size)) != ERROR_SUCCESS) {
            srs_error("initialize stream failed. ret=%d", ret);
            return ret;
        }
        srs_verbose("decode stream initialized success");
        
        // decode the packet.
        SrsPacket* packet = NULL;
        if ((ret = do_decode_message(msg->header, &stream, &packet)) != ERROR_SUCCESS) {
            srs_freep(packet);
            return ret;
        }
        
        // set to output ppacket only when success.
        *ppacket = packet;
        
        return ret;
    }
    

    该函数将消息的负载转化为一个字节流,便于调用 SrsProtocol 实现的 do_decode_message 函数对负载数据进行解码。

    5.1.3 SrsProtocol::do_decode_message

    int SrsProtocol::do_decode_message(SrsMessageHeader& header, 
        SrsStream* stream, SrsPacket* ppacket)
    {
        int ret = ERROR_SUCCESS;
        
        SrsPacket* packet = NULL;
        
        // decode specified packet type
        if (header.is_amf0_command() || header.is_amf3_command() || 
            header.is_amf0_data()    || header.is_amf3_data(0)
        {
            srs_verbose("start to decode AMF0/AMF3 command message.");
            
            // skip 1bytes to decode the amf3 command.
            if (header.is_amf3_command() && stream->require(1)) {
                srs_verbose("skip 1bytes to decode AMF3 command");
                stream->skip(1);
            }
            
            // amf0 command message.
            // need to read the command name.
            std::string command;
            if ((ret = srs_amf0_read_string(stream, command)) != ERROR_SUCCESS) {
                srs_error("decode AMF0/AMF3 command name failed. ret=%d", ret);
                return ret;
            }
            srs_verbose("AMF0/AMF3 command message, command_name=%s", command.c_str());
            
            // result/error packet
            if (command == RTMP_AMF0_COMMAND_RESULT || command == RTMP_AMF0_COMMAND_ERROR) {
                /* 这里先忽略,仅考虑对 amf0_data 类型的解码 */
                ...
            }
            
            // reset to zero(amf3 to 1) to restart decode.
            stream->skip(-1 * stream->pos());
            if (header.is_amf3_command()) {
                stream->skip(1);
            }
            
            // decode command object.
            if (command == RTMP_AMF0_COMMAND_CONNECT)
            {
                ...
            }
            ...
            /* "@setDataFrame" or "onMetaData" */
            else if (command == SRS_CONSTS_RTMP_SET_DATAFRAME || 
                     command == SRS_CONSTS_RTMP_ON_METADATA) {
                srs_info("decode the AMF0/AMF3 data(onMetaData message).");
                *ppacket = packet = new SrsOnMetaDataPacket();
                /* 调用 SrsOnMetaDataPacket 类实现的 decode 函数 */
                return packet->decode(stream);
            }
            ...
            
            // default packet to drop message.
            srs_info("drop the AMF0/AMF3 command message, command_name=%s", command.c_str());
            *ppacket = packet = new SrsPacket();
            return ret;
        } else if (header.is_user_control_message()) {
            ...
        } else if
        ...
        
        return ret;
    }
    

    对于接收到的 amf_data 类型的数据,统一构造一个 SrsOnMetaDataPacket 类,然后调用该类实现的 decode 函数进行解码。

    5.1.4 SrsOnMetaDataPacket 构造函数

    /**
     * the stream metadata.
     * FMLE: @setDataFrame
     * others: onMetaData
     */
    SrsOnMetaDataPacket::SrsOnMetaDataPacket()
    {
        name = SRS_CONSTS_RTMP_ON_METADATA;
        /**
         * Metadata of stream.
         * @remark, never be NULL, an AMF0 object instance.
         */
        metadata = SrsAmf0Any::object();
    }
    

    若为 FMLE(Flash Media Live Encoder) 软件,则发送的 amf0_data 消息名为 "@setDataFrame",其他的则为 "onMetaData"。

    5.1.5 SrsOnMetaDataPacket::decode

    int SrsOnMetaDataPacket::decode(SrsStream* stream)
    {
        int ret = ERROR_SUCCESS;
        
        if ((ret = srs_amf0_read_string(stream, name)) != ERROR_SUCCESS) {
            srs_error("decode metadata name failed. ret=%d", ret);
            return ret;
        }
        
        // ignore the @setDataFrame
        if (name == SRS_CONSTS_RTMP_SET_DATAFRAME) {
            /* 名称以 "onMetaData" 为准 */
            if ((ret = srs_amf0_read_string(stream, name)) != ERROR_SUCCESS) {
                srs_error("decode metadata name failed. ret=%d", ret);
                return ret;
            }
        }
        
        srs_verbose("decode metadata name success. name=%s", name.c_str());
        
        // the metadata mayby object or ecma array
        SrsAmf0Any* any = NULL;
        /* 由上图知,该 metadata 的数据类型是 ecma array */
        if ((ret = srs_amf0_read_any(stream, &any)) != ERROR_SUCCESS) {
            srs_error("decode metadata metadata failed. ret=%d", ret);
            return ret;
        }
        
        srs_assert(any);
        if (any_is_object()) {
            srs_freep(metadata);
            metadata = any->to_object();
            srs_info("decode metadata object success");
            return ret;
        }
        
        SrsAutoFree(SrsAmf0Any, any);
        
        if (any->is_ecma_array()) {
            SrsAmf0EcmaArray* arr = any->to_ecma_array();
        
            // if ecma array, copy to object.
            for (int i = 0; i < arr->count(); i++) {
                /* 将解析出来的数据拷贝到 metadata 的 properties 中,
                 * metadata 是指向 SrsAmf0Object 对象的指针 */
                metadata->set(arr->key_at(i), arr->value_at(i)->copy());
            }
            
            srs_info("decode metadata array success");
        }
        
        return ret;
    }
    

    该函数主要是解析 metadata 数据,然后将其保存在 SrsOnMetaDataPacket 类的成员 metadata 中。

    5.1.6 srs_amf0_read_any

    int srs_amf0_read_any(SrsStream* stream, SrsAmf0Any** ppvalue)
    {
        int ret = ERROR_SUCCESS;
        
        /* 读取 marker,发现是 ecma array 类型,则会构造一个 SrsAmf0EcmaArray 对象,
         * 通过 ppvalue 返回该对象 */
        if ((ret = SrsAmf0Any::discovery(stream, ppvalue)) != ERROR_SUCCESS) {
            srs_error("amf0 discovery any elem failed. ret=%d", ret);
            return ret;
        }
        
        srs_assert(*ppvalue);
        
        /* 调用 SrsAmf0EcmaArray 类实现的 read 函数读取metadata携带的各项property */
        if ((ret = (*ppvalue)->read(stream)) != ERROR_SUCCESS) {
            srs_error("amf0 parse elem failed. ret=%d", ret);
            srs_freep(*ppvalue);
            return ret;
        }
        
        return ret;
    }
    

    5.1.7 SrsAmf0EcmaArray::read

    int SrsAmf0EcmaArray::read(SrsStream* stream)
    {
        int ret = ERROR_SUCCESS;
        
        // marker
        if (!stream->require(1)) {
            ret = ERROR_RTMP_AMF0_DECODE;
            srs_error("amf0 read ecma_array marker failed. ret=%d", ret);
            return ret;
        }
        
        /* 读取 AMF0 type:ECMA array 为 0x08 */
        char marker = stream->read_1bytes();
        if (marker != RTMP_AMF0_EcmaArray) {
            ret = ERROR_RTMP_AMF0_DECODE;
            srs_error("amf0 check ecma_array marker failed. "
                "marker=%#x, required=%#x, ret=%d", marker, RTMP_AMF0_EcmaArray, ret);
            return ret;
        }
        srs_verbose("amf0 read ecma_array marker success");
    
        // count
        if (!stream->require(4)) {
            ret = ERROR_RTMP_AMF0_DECODE;
            srs_error("amf0 read ecma_array count failed. ret=%d", ret);
            return ret;
        }
    
        /* 读取该 ECMA array 中有多少个 property */    
        int32_t count = stream->read_4bytes();
        srs_verbose("amf0 read ecma_array count success. count=%d", count);
        
        // value
        this->_count = count;
    
        while (!stream->empty()) {
            // detect whether is eof.
            if (srs_amf0_is_object_eof(stream)) {
                SrsAmf0ObjectEOF pbj_eof;
                /* ECMA array 类型同样以 0x00 0x00 0x09 结尾,与 object 一样 */
                if ((ret = pbj_eof.read(stream)) != ERROR_SUCCESS) {
                    srs_error("amf0 ecma_array read eof failed. ret=%d", ret);
                    return ret;
                }
                srs_info("amf0 read ecma_array EOF.");
                break;
            }
            
            // property-name: utf8 string
            std::string property_name;
            /* 读取 property 的名称 */
            if ((ret =srs_amf0_read_utf8(stream, property_name)) != ERROR_SUCCESS) {
                srs_error("amf0 ecma_array read property name failed. ret=%d", ret);
                return ret;
            }
            /* 读取 property 的值:number or string or boolean */
            // property-value: any
            SrsAmf0Any* property_value = NULL;
            if ((ret = srs_amf0_read_any(stream, &property_value)) != ERROR_SUCCESS) {
                srs_error("amf0 ecma_array read property_value failed. "
                    "name=%s, ret=%d", property_name.c_str(), ret);
                return ret;
            }
            
            /* 将获取到的每一个 property 以该 property 的名称为 key,保存到 SrsAmf0EcmaArray 类的
             * 成员 properties 中,该 properties 是一个指向 SrsUnSortedHashtable 类的指针,该类的
             * 成员 properties 维护了一个 std::vector<SrsAmf0ObjectPropertyType> 容器,该容器用于
             * 存放所有获取到的 property 项 */
            // add property
            this->set(property_name, property_value);
        }
        
        return ret;
    }
    

    解析 metadata 数据成功后,接下来是调用 SrsSource 实现的 on_meta_data 函数对解析后的 metadata 做进一步的处理。

    5.1.8 SrsSource::on_meta_data

    int SrsSource::on_meta_data(SrsCommonMessage* msg, SrsOnMetaDataPacket* metadata)
    {
        int ret = ERROR_SUCCESS;
        
        /* hls 和 dvr 的暂时忽略 */
    #ifdef SRS_AUTO_HLS
        if (metadata && (ret = hls->on_meta_data(metadata->metadata)) != ERROR_SUCCESS) {
            srs_error("hls process onMetaData message failed. ret=%d", ret);
            return ret;
        }
    #endif
        
    #ifdef SRS_AUTO_DVR
        if (metadata && (ret = dvr->on_meta_data(metadata)) != ERROR_SUCCESS) {
            srs_error("dvr process onMetaData message failed. ret=%d", ret);
            return ret;
        }
    #endif
    
        SrsAmf0Any* prop = NULL;
        
        // when exists the duration, remove it to make ExoPlayer happy.
        if (metadata->metadata->get_property("duration") != NULL) {
            metadata->metadata->remove("duration");
        }
        
        // generate metadata info to print
        std::stringstream ss;
        if ((prop = metadata->metadata->ensure_property_number("width")) != NULL) {
            ss << ", width=" << (int)prop->to_number();
        }
        if ((prop = metadata->metadata->ensure_property_number("height")) != NULL) {
            ss << ", height=" << (int)prop->to_number();
        }
        if ((prop = metadata->metadata->ensure_property_number("videocodecid")) != NULL) {
            ss << ", vcodec=" << (int)prop->to_number();
        }
        if ((prop = metadata->metadata->ensure_property_number("audiocodecid")) != NULL) {
            ss << ", acodec=" << (int)prop->to_number();
        }
        srs_trace("got metadata%s", ss.str().c_str());
        
        // add server info to metadata.
        metadata->metadata->set("server", SrsAmf0Any::str(RTMP_SIG_SRS_SERVER));
        metadata->metadata->set("srs_primary", SrsAmf0Any::str(RTMP_SIG_SRS_PRIMARY));
        metadata->metadata->set("srs_authors", SrsAmf0Any::str(RTMP_SIG_SRS_AUTHROS));
        
        // version, for example, 1.0.0
        // add version to metadata, please donot remove it, for debug.
        metadata->metadata->set("server_version", SrsAmf0Any::str(RTMP_SIG_SRS_VERSION));
        
        // if allow atc_auto and bravo-atc detected, open atc for vhost.
        atc = _srs_config->get_atc(_req->vhost);
        if (_srs_config->get_atc_auto(_req->vhost)) {
            if ((prop = metadata->metadata->get_property("bravo_atc")) != NULL) {
                if (prop->is_string() && prop->to_str() == "true") {
                    atc = true;
                }
            }
        }
        
        // encode the metadata to payload
        int size = 0;
        char* payload = NULL;
        /* 调用继承自父类 SrsPacket 的函数 encode 将 metadata 中的元数据信息编码成
         * payload */
        if ((ret = metadata->encode(size, payload)) != ERROR_SUCCESS) {
            srs_error("encode metadata error. ret=%d", ret);
            srs_freep(payload);
            return ret;
        }
        srs_verbose("encode metadata success.");
        
        if (size <= 0) {
            srs_warn("ignore the invalid metadata. size=%d", size);
            return ret;
        }
        
        // when already got metadata, drop when reduce sequence header.
        bool drop_for_reduce = false;
        if (cache_metadata && _srs_config->get_reduce_sequence_header(_req->vhost)) {
            drop_for_reduce = true;
            srs_warn("drop for reduce sh metadata, size=%d", msg->size);
        }
        
        // create a shared ptr message.
        srs_freep(cache_metadata);
        cache_metadata = new SrsSharedPtrMessage();
        
        // dump message to shared ptr message.
        // the payload/size managed by cache_metadata, user should not free it.
        if ((ret = cache_metadata->create(&msg->header, payload, size)) != ERROR_SUCCESS) {
            srs_error("initialize the cache metadata failed. ret=%d", ret);
            return ret;
        }
        srs_verbose("initialize shared ptr metadata success.");
        
        // copy to all consumer
        if (!drop_for_reduce) {
            /* 若有其他客户端订阅了该直播流,则通知这些客户端 */
            std::vector<SrsConsumer*>::iterator it;
            for (it = consumer.begin(); it != consumers.end(); ++it) {
                SrsConsumer* consumer = *it;
                if ((ret = consumer->enqueue(cache_metadata, atc, jitter_algorithm)) 
                    != ERROR_SUCCESS) {
                    srs_error("dispatch the metadata failed. ret=%d", ret);
                    return ret;
                }
            }
        }
        
        // copy to all forwarders
        if (true) {
            std::vector<SrsForwarder*>::iterator it;
            for (it = forwarders.begin(); it != forwarders.end(); ++it) {
                SrsForwarder* forwarder = *it;
                if ((ret = forwarder->on_meta_data(cache_metadata)) != ERROR_SUCCESS) {
                    srs_error("forwarder process onMetaData message failed. ret=%d", ret);
                    return ret;
                }
            }
        }
        
        return ret;
    }
    

    5.1.9 SrsPacket::encode

    /*
     * the subpacket can override this encode,
     * for example, video and audio will directly set the payload without memory copy,
     * other packet which need to serialize/encode to bytes by override the
     * get_size and encode_packet.
     */
    int SrsPacket::encode(int& psize, char*& ppayload) 
    {
        int ret = ERROR_SUCCESS;
        
        int size = get_size();
        char* payload = NULL;
        
        SrsStream stream;
        
        if (size > 0) {
            payload = new char[size];
            
            if ((ret = stream.initialize(payload, sizse)) != ERROR_SUCCESS) {
                srs_error("initialize the stream failed. ret=%d", ret);
                srs_freepa(payload);
                return ret;
            }
        }
        
        /* 调用 SrsOnMetaDataPacket 类实现的 encode_packet 函数 */
        if ((ret = encode_packet(&stream)) != ERROR_SUCCESS) {
            srs_error("encode the packet failed. ret=%d", ret);
            srs_freepa(payload);
            return ret;
        }
        
        psize = size;
        ppayload = payload;
        srs_verbose("encode the packet success. size=%d", size);
        
        return ret;
    }
    

    5.1.10 SrsOnMetaDataPacket::encode_packet

    int SrsOnMetaDataPacket::encode_packet(SrsStream* stream)
    {
        int ret = ERROR_SUCCESS;
        
        if ((ret = srs_amf0_write_string(stream, name)) != ERROR_SUCCESS) {
            srs_error("encode name failed. ret=%d", ret);
            return ret;
        }
        srs_verbose("encode name success.");
        
        /* 调用 SrsAmf0Object 类实现的 write 函数 */
        if ((ret = metadata->write(stream)) != ERROR_SUCCESS) {
            srs_error("encode metadata failed. ret=%d", ret);
            return ret;
        }
        srs_verbose("encode metadata success.");
        
        srs_info("encode onMetaData packet success.");
        return ret;
    }
    

    5.1.11 SrsAmf0Object::write

    int SrsAmf0Object::write(SrsStream* stream)
    {
        int ret = ERROR_SUCCESS;
        
        // marker
        if (!stream->require(1)) {
            ret = ERROR_RTMP_AMF0_ENCODE;
            srs_error("amf0 write object marker failed. ret=%d", ret);
            return ret;
        }
        
        /* 写入 1 字节的 amf 类型 */
        stream->write_1bytes(RTMP_AMF0_Object);
        srs_verbose("amf0 write object marker success");
        
        // value
        for (int i = 0; i < properties->count(); i++) {
            std::string name = this->key_at(i);
            SrsAmf0Any* any = this->value_at(i);
            
            if ((ret = srs_amf0_write_utf8(stream, name)) != ERROR_SUCCESS) {
                srs_error("write object property name failed. ret=%d", ret);
                return ret;
            }
            
            if ((ret = srs_amf0_write_any(stream, any)) != ERROR_SUCCESS) {
                srs_error("write object property value failed. ret=%d", ret);
                return ret;
            }
            
            srs_verbose("write amf0 property success. name=%s", name.c_str());
        }
        
        /* 写入结束标志 0x00 0x00 0x09 */
        if ((ret = eof->write(stream)) != ERROR_SUCCESS) {
            srs_error("write object eof failed. ret=%d", ret);
            return ret;
        }
        
        srs_verbose("write amf0 object success.");
        
        return ret;
    }
    

    5.1.12 SrsSharedPtrMessage 构造函数

    /*
     * shared ptr message.
     * for audio/video/data message that need less memory copy.
     * and only for output.
     *
     * create first object by constructor and create(),
     * use copy if need reference count message.
     */
    SrsSharedPtrMessage::SrsSharedPtrMessage()
    {
        ptr = NULL;
    }
    

    5.1.13 SrsSharedPtrMessage::create

    /*
     * create shared ptr message,
     * from the header and payload.
     * @remark user should never free the payload.
     * @param pheader, the header to copy to the message. NULL to ignore.
     */
    int SrsSharedPtrMessage::create(SrsMessageHeader* pheader, char* payload, int size)
    {
        int ret = ERROR_SUCCESS;
        
        if (ptr) {
            ret = ERROR_SYSTEM_ASSERT_FAILED;
            srs_error("should not set the payload twice. ret=%d", ret);
            srs_assert(false);
            
            return ret;
        }
        
        /* 构造 SrsSharedPtrPayload */
        ptr = new SrsSharedPtrPayload();
        
        /* 将解析自 metadata 消息的消息头内容赋给 ptr->header 的相应成员 */
        // direct attach the data.
        if (pheader) {
            /* amf0_data: 0x12 */
            ptr->header.message_type = pheader->message_type;
            /* 负载的大小 */
            ptr->header.payload_length = size;
            ptr->header.perfer_cid = pheader->perfer_cid;
            this->timestamp = pheader->timestamp;
            this->stream_id = pheader->stream_id;
        }
        
        ptr->payload = payload;
        ptr->size = size;
        
        // message can access it.
        /* 
         * payload:
         * the payload of message, the SrsCommonMessage never know about the detail of payload,
         * user must use SrsProtocol.decode_message to get concrete packet.
         * @remark, not all message payload can be decoded to packet. for example,
         *       video/audio packet use raw bytes, no video/audio packet.
         */
        this->payload = ptr->payload;
        /*
         * current message parsed size,
         *     size <= header.payload_length
         * for the payload maybe sent in multiple chunks.
         */
        this->size = ptr->size;
        
        return ret;
    }
    

    5.1.14 SrsSharedPtrPayload 构造函数

    SrsSharedPtrMessage::SrsSharedPtrPayload::SrsSharedPtrPayload()
    {
        /* actual shared payload. */
        payload = NULL;
        /* size of payload. */
        size = 0;
        /* the reference count */
        shared_count = 0;
    }
    

    5.1.15 通知消费者:SrsConsumer::enqueue

    /**
     * enqueue an shared ptr message.
     * @param shared_msg, directly ptr, copy it if need to save it.
     * @param whether atc, donot use jitter correct if true.
     * @param ag the algorithm of time jitter.
     */
    int SrsConsumer::enqueue(SrsSharedPtrMessage* shared_msg, bool atc, 
        SrsRtmpJitterAlgorithm ag)
    {
        int ret = ERROR_SUCCESS;
        
        /* 拷贝一个副本返回给 msg */
        SrsSharedPtrMessage* msg = shared_msg->copy();
        
        /* 若 atc 为 false,则使用 jitter 进行校正 */
        if (!atc) {
            if ((ret = jitter->correct(msg, ag)) != ERROR_SUCCESS) {
                srs_freep(msg);
                return ret;
            }
        }
        
        if ((ret = queue->enqueue(msg, NULL)) != ERROR_SUCCESS) {
            return ret;
        }
        
    #ifdef SRS_PERF_QUEUE_COND_WAIT
        srs_verbose("enqueue msg, time=%"PRId64", size=%d, "
                    "duration=%d, waiting=%d, min_msg=%d", 
            msg->timestamp, msg->size, queue->duration(), mw_waiting, mw_min_msgs);
            
        // fire the mw when msgs is enough.
        /* 若有消费者,即播放客户端正在等待 msg 准备好,即上面的 queue->enqueue 成功返回 */
        if (mw_waiting) {
            int duration_ms = queue->duration();
            bool match_min_msgs = queue->size() > mw_min_msgs;
            
            // For ATC, maybe the SH timestamp bigger than A/V packet,
            // when encoder republish or overflow.
            // @see https://github.com/ossrs/srs/pull/749
            if (atc && duration_ms < 0) {
                st_cond_signal(mw_wait);
                mw_waiting = false;
                return ret;
            }
            
            // when duration ok, signal to flush.
            if (match_min_msgs && duration_ms > mw_duration) {
                st_cond_signal(mw_wait);
                mw_waitting = false;
                return ret;
            }
        }
    #endif
    
        return ret;
    }
    

    5.1.16 SrsRtmpJitter::correct

    int SrsRtmpJitter::correct(SrsSharedPtrMessage* msg, SrsRtmpJitterAlgorithm ag)
    {
        int ret = ERROR_SUCCESS;
        
        // for performance issue
        if (ag != SrsRtmpJitterAlgorithmFULL) {
            // all jitter correct features is disabled, ignore.
            if (ag == SrsRtmpJitterAlgorithmOFF) {
                return ret;
            }
            
            // start at zero, but donot ensure monotonically increasing.
            if (ag == SrsRtmpJitterAlgothmZERO) {
                // for the first time, last_pkt_corrent_time is -1.
                if (last_pkt_correct_time == -1) {
                    last_pkt_correct_time = msg->timestamp;
                }
                msg->timestamp -= last_pkt_correct_time;
                return ret;
            }
            
            // other algorithm, ignore.
            return ret;
        }
        
        // full jitter algorithm, do jitter correct.
        // set to 0 for metadata.
        if (!msg->is_av()) {
            msg->timestamp = 0;
            return ret;
        }
        
        /**
         * we use a very simple time jitter detect/correct algorithm:
         * 1. delta: ensure the delta is positive and valid,
         *     we set the delta to DEFAULT_FRAME_TIME_MS,
         *     if the delta of time is nagative or greater than CONST_MAX_JITTER_MS.
         * 2. last_pkt_time: specifies the original packet time,
         *     is used to detect next jitter.
         * 3. last_pkt_correct_time: simply add the positive delta,
         *     and enforce the time monotonically.
         */
        int64_t time = msg->timestamp;
        int64_t delta = time - last_pkt_time;
        
        // if jitter detected, reset the delta.
        if (delta < CONST_MAX_JITTER_MS_NED || delta > CONST_MAX_JITTER_MS) {
            // use default 10ms to notice the problem of stream.
            // @see https://github.com/ossrs/srs/issues/425
            delta = DEFAULT_FRAME_TIME_MS;
            
            srs_info("jitter detected, last_pts=%"PRId64", pts=%"PRId64", "
                     "diff=%"PRId64", last_time=%"PRId64", time=%"PRId64", diff=%"PRId64"",
                last_pkt_time, time, time - last_pkt_time, last_pkt_correct_time, 
                last_pkt_correct_time + delta, delta);
        } else {
            srs_verbose("timestamp no jitter. time=%"PRId64", "
                        "last_pkt=%"PRId64", correct_to=%"PRId64"", 
                time, last_pkt_time, last_pkt_correct_time + delta);
        }
        
        last_pkt_correct_time = srs_max(0, last_pkt_correct_time + delta);
        
        msg->timestamp = last_pkt_correct_time;
        last_pkt_time = time;
        
        return ret;
    }
    

    若传入的第二个参数为 SrsRtmpJitterAlgorithmOFF,则禁止所有的 jitter 校正,构造 SrsSource 的时候默认初始化为 SrsRtmpJitterAlgorithmOFF。

    5.1.17 SrsMessageQueue::enqueue

    /*
     * enqueue the message, the timestamp always monotonically.
     * @param msg, the msg to enqueue, user never free it whatever the return code.
     * @param is_overflow, whether overflow and shrinked. NULL to ignore.
     */
    int SrsMessageQueue::enqueue(SrsSharedPtrMessage* msg, bool* is_overflow)
    {
        int ret = ERROR_SUCCESS;
        
        if (msg->is_av()) {
            if (av_start_time == -1) {
                av_start_time = msg->timestamp;
            }
            
            av_end_time = msg->timestamp;
        }
        
        /* 若声明了 SRS_PERF_QUEUE_FAST_VECTOR 宏,则调用 SrsFastVector 类
         * 实现的 push_back 函数 */
        msgs.push_back(msg);
        
        /* 检测 msgs 队列是否溢出 */
        while (av_end_time - av_start_time > queue_size_ms) {
            // notice the caller queue already overflow and sharinked.
            if (is_overflow) {
                *is_overflow = true;
            }
            
            /* 满溢的情况下,移除一个 gop */
            sharink();
        }
        
        return ret;
    }
    

    5.1.18 SrsFastVector::push_back

    void SrsFastVector::push_back(SrsSharedPtrMessage* msg)
    {
        // increase vector.
        if (count >= nb_msgs) {
            int size = nb_msgs * 2;
            SrsSharedPtrMessage** buf = new SrsSharedPtrMessage*[size];
            for (int i = 0; i < nb_msgs; i++) {
                buf[i] = msgs[i];
            }
            srs_warn("fast vector incrase %d=>%d", nb_msgs, size);
            
            // use new array.
            srs_freep(msgs);
            msgs = buf;
            nb_msgs = size;
        }
        
        /* msgs 指向一个 SrsSharedPtrMessage 类型的二级数组,该数组的成员是
         * SrsSharedPtrMessage* */
        msgs[count++] = msg;
    } 
    

    该函数是直接将 msg 放入到 SrsFastVector 类的成员 msgs 数组中(若该数组大小足够的话)。

    5.1.19 SrsMessageQueue::shrink

    /*
     * remove a gop from the front.
     * if no iframe found, clear it.
     */
    void SrsMessageQueue::shrink()
    {
        SrsSharedPtrMessage* video_sh = NULL;
        SrsSharedPtrMessage* audio_sh = NULL;
        int msgs_size = (int)msgs.size();
        
        // remove all msg
        // ignore the sequence header
        for (int i = 0; i < (int)msgs.size(); i++) {
            SrsSharedPtrMessage* msg = msgs.at(i);
            
            if (msg->is_video() && 
                SrsFlvCodec::video_is_sequence_header(msg->payload, msg->size)) 
            {
                srs_freep(video_sh);
                video_sh = msg;
                continue;
            }
            else if (msg->is_audio() && 
                     SrsFlvCodec::audio_is_sequence_header(msg->payload, msg->size)) 
            {
                srs_freep(audio_sh);
                audio_sh = msg;
                continue;
            }
            
            srs_freep(msg);
        }
        msgs.clear();
        
        // update av_start_time
        av_start_time = av_end_time;
        // push_back sequence header and update timestamp
        if (video_sh) {
            video_sh->timestamp = av_end_time;
            msgs.push_back(video_sh);
        }
        if (audio_sh) {
            audio_sh->timestamp = av_end_time;
            msgs.push_back(audio_sh);
        }
        
        if (_ignore_shrink) {
            srs_info("shrink the cache queue, size=%d, removed=%d, max=%.2f", 
                (int)msgs.size(), msgs_size - (int)msgs.size(), queue_size_ms / 1000.0);
        } else {
            srs_trace("shrink the cache queue, size=%d, removed=%d, max=%.2f", 
                (int)msgs.size(), msgs_size - (int)msgs.size(), queue_size_ms / 1000.0);
        }
    }
    

    5.1.20 SrsMessageQueue::duration

    /* get the duration of queue. */
    int SrsMessageQueue::duration()
    {
        return (int)(av_end_time - av_start_time);
    }
    

    计算当前消息队列中所有消息的总 duration。

    5.1.21 st_cond_signal

    int st_cond_signal(_st_cond_t *cvar)
    {
      return _st_cond_signal(cvar, 0);
    }
    

    5.1.22 _st_cond_signal

    static int _st_cond_signal(_st_cond_t *cvar, int broadcast)
    {
        _st_thread_t *thread;
        _st_clist_t *q;
        
        for (q = cvar->wait_q.next; q != &cvar->wait_q; q = q->next) {
            thread = _ST_THREAD_WAITQ_PTR(q);
            if (thread->state == _ST_ST_COND_WAIT) {
                if (thread->flags & _ST_ST_ON_SLEEPQ) 
                    _ST_DEL_SLEEPQ(thread);
                
                /* Make thread runnable */
                thread->state = _ST_ST_RUNNABLE;
                _ST_ADD_RUNQ(thread);
                if (!broadcast)
                    break;s
            }
        }
        
        return 0;
    }
    

    5.2 Audio

    假设接收到的第一个音频包如下图。

    对于接收到的音频包,在 SrsRtmpConn::process_publish_message 函数中直接调用 SrsSource 类实现的 on_audio 函数进行处理。

    5.2.1 SrsSource::on_audio

    int SrsSource::on_audio(SrsCommonMessage* shared_audio)
    {
        int ret = ERROR_SUCCESS;
        
        // monotically increate detect.
        if (!mix_correct && is_monotonically_increase) {
            if (last_packet_time > 0 && shared_audio->header.timestamp < last_packet_time) {
                is_monotonically_increate = false;
                srs_warn("AUDIO: stream not monotonically increase, please open mix_correct.");
            }
        }
        /* 记录接收到的 audio 帧的时间戳 */
        last_packet_time = shared_audio->header.timestamp;
        
        // convert shared_audio to msg, user should not use shared_audio again.
        // the payload is transfer to msg, and set to NULL in shared_audio.
        SrsSharedPtrMessage msg;
        /* 构造一个 SrsSharedPtrMessage 类,并将接收到的音频消息的消息头还有负载
         * 拷贝到该新构建的 SrsSharedPtrMessage 类中 */
        if ((ret = msg.create(shared_audio)) != ERROR_SUCCESS) {
            srs_error("initialize the audio failed. ret=%d", ret);
            return ret;
        }
        srs_info("Audio dts=%"PRId64", size=%d", msg.timestamp, msg.size);
        
        /* 若没有开启 mix_correct 校正,则直接处理音频数据 */
        // directly process the audio message.
        if (!mix_correct) {
            return on_audio_imp(&msg);
        }
        
        // insert msg to the queue.
        mix_queue->push(msg.copy());
        
        // fetch someone from mix_queue.
        SrsSharedPtrMessage* m = mix_queue->pop();
        if (!m) {
            return ret;
        }
        
        // comsumer the monotonically increase message.
        if (m->is_audio()) {
            ret = on_audio_imp(m);
        } else {
            ret = on_video_imp(m);
        }
        srs_freep(m);
        
        return ret;
    }
    

    5.2.2 SrsSource::on_audio_imp

    int SrsSource::on_audio_imp(SrsSharedPtrMessage* msg)
    {
        int ret = ERROR_SUCCESS;
        
        srs_info("Audio dts=%"PRId64", size=%d", msg->timestamp, msg->size);
        bool is_aac_sequence_header = 
             SrsFlvCodec::audio_is_sequence_header(msg->payload, msg->size);
        bool is_sequence_header = is_aac_sequence_header;
        
        // whether consumer should drop for the duplicated sequnece header.
        bool drop_for_reduce = false;
        if (is_sequence_header && cache_sh_audio && 
            _srs_config->get_reduce_sequence_header(_req->vhost)) {
            if (cache_sh_audio->size == msg->size) {
                drop_for_reduce = srs_bytes_equals(cache_sh_audio->payload, 
                                                   msg->payload, msg->size);
                srs_warn("drop for reduce sh audio, size=%d", msg->size);
            }
        }
        
        /* 若是 AAC sequence header,则将该音频包的数据保存到 aac_extra_data 中 */
        // cache the sequence header if aac
        // donot cache the sequence header to gop_cache, return here.
        if (is_aac_sequence_header) {
            // parse detail audio codec
            SrsAvcAacCodec codec;
            SrsCodecSample sample;
            /* 对接收到的音频数据进行解析 */
            if ((ret = codec.audio_aac_demux(msg->payload, msg->size, &sample)) 
                != ERROR_SUCCESS) {
                srs_error("source codec demux audio failed. ret=%d", ret);
                return ret;
            }
            
            static int flv_sample_sizes[] = {8, 16, 0};
            static int flv_shound_types[] = {1, 2, 0};
            
            // when got audio stream info.
            SrsStatistic* stat = SrsStatistic::instance();
            if ((ret = stat->on_audio_info(_req, SrsCodecAudioAAC, sample.sound_rate, 
                       sample.sound_type, codec.aac_object)) != ERROR_SUCCESS) {
                return ret;
            }
            
            srs_trace("%dB audio sh, codec(%d, profile=%s, %dchannels, %dkbps, %dHZ), "
                "flv(%dbits, %dchannels, %dHZ)",
                msg->size, codec.audio_codec_id,
                srs_codec_aac_object2str(codec.aac_object).c_str(), codec.aac_channels,
                codec.audio_data_rate / 1000, aac_sample_rates[codec.aac_sample_rate],
                flv_sample_sizes[sample.sound_size], flv_sound_types[sample.sound_type],
                flv_sample_rates[sample.sound_rate]);
        }
        
    #ifdef SRS_AUTO_HLS
        if ((ret = hls->on_audio(msg)) != ERROR_SUCCESS) {
            // apply the error strategy for hls.
            // @see https://github.com/ossrs/srs/issues/264
            std::string hls_error_strategy = _srs_config->get_hls_on_error(_req->vhost);
            if (srs_config_hls_is_on_error_ignore(hls_error_strategy)) {
                srs_warn("hls process audio message failed, ignore and disable hls. ret=%d", 
                         ret);
                
                // unpublish, ignore ret.
                hls->on_unpublish();
                
                // ignore.
                ret = ERROR_SUCCESS;
            } else if (srs_config_hls_is_on_error_continue(hls_error_strategy)) {
                if (srs_hls_can_continue(ret, cache_sh_audio, msg)) {
                    ret = ERROR_SUCCESS;
                } else {
                    srs_warn("hls continue audio failed. ret=%d", ret);
                    return ret;
                }
            } else {
                srs_warn("hls disconnect publisher for audio error. ret=%d", ret);
                return ret;
            }
        }
    #endif
    
    #ifdef SRS_AUTO_DVR
        if ((ret = dvr->on_audio(msg)) != ERROR_SUCCESS) {
            srs_warn("dvr process audio message failed, ignore and disable dvr. ret=%d", ret);
            
            // unpublish, ignore ret.
            dvr->on_unpublish();
            
            // ignore.
            ret = ERROR_SUCCESS;
        }
    #endif
    
    #ifdef SRS_AUTO_HDS
        if ((ret = hds->on_audio(msg)) != ERROR_SUCCESS) {
            srs_warn("hds process audio message failed, ignore and disable dvr. ret=%d", ret);
            
            // unpublish, ignore ret.
            hds->on_unpublish();
            // ignore.
            ret = ERROR_SUCCESS;
        }
    #endif
    
        /* 将接收到的 audio message 放入到 consumer 所持有的 queue 队列中 */
        // copy to all consumer
        if (!drop_for_reduce) {
            for (int i = 0; i < (int)consumers.size(); i++) {
                SrsConsumer* consumer = consumers.at(i);
                if ((ret = consumer->enqueue(msg, atc, jitter_algorithm)) != ERROR_SUCCESS) {
                    srs_error("dispatch the audio failed. ret=%d", ret);
                    return ret;
                }
            }
            srs_info("dispatch audio success.");
        }
        
        // copy to all forwarders.
        if (true) {
            std::vector<SrsForwarder*>::iterator it;
            for (it = forwarders.begin(); it != forwarders.end(); ++it) {
                SrsForwarder* forwarder = *it;
                if ((ret = forwarder->on_audio(msg)) != ERROR_SUCCESS) {
                    srs_error("forwarder process audio message failed. ret=%d", ret);
                    return ret;
                }
            }
        }
        
        // cache the sequence header of aac, or first packet of mp3.
        // for example, the mp3 is used for hls to write the "right" audio codec.
        // TODO: FIXME: to refine the stream info system.
        if (is_aac_sequence_header || !cache_sh_audio) {
            srs_freep(cache_sh_audio);
            cache_sh_audio = msg->copy();
        }
        
        // when sequence header, donot push to gop cache and adjust the timestamp.
        if (is_sequence_header) {
            return ret;
        }
        
        // cache the last gop packets
        if ((ret = gop_cache->cache(msg)) != ERROR_SUCCESS) {
            srs_error("shrink gop cache failed. ret=%d", ret);
            return ret;
        }
        srs_verbose("cache gop success.");
        
        // if aac, update the sequence header to abs time.
        if (atc) {
            if (cache_sh_audio) {
                cache_sh_audio->timestamp = msg->timestamp;
            }
            if (cache_metadata) {
                cache_metadata->timestamp = msg->timestamp;
            }
        }
        
        return ret;
    }
    

    调用 SrsAvcAacCodec 类实现的 audio_aac_demux 对接收到的 aac 数据进行解析之前,需要构造两个类对象:SrsAvcAacCodec 和 SrsCodecSample。

    5.2.3 SrsAvcAacCodec 构造函数

    /*
     * the h264/avc and aac codec, for media stream.
     *
     * to demux the FLV/RTMP video/audio packet to sample,
     * add each NALUs of h.264 as a sample unit to sample,
     * while the entire aac raw data as a sample unit.
     *
     * for sequence header,
     * demux it and save it in the avc_extra_data and aac_extra_data。
     *
     * for the codec info, such as audio sample rate,
     * decode from FLV/RTMP header, then use codec info in sequence 
     * header to override it.
     */
    SrsAvcAacCodec::SrsAvcAacCodec()
    {
        /* for sequence header, whether parse the h.264 sps. */
        avc_parse_sps               = true;
        
        width                       = 0;
        height                      = 0;
        duration                    = 0;
        /* lengthSizeMinusOne, H.264-AVC-ISO_IEC_14496-15.pdf, page 16 */
        NAL_unit_length             = 0;
        frame_rate                  = 0;
    
        video_data_rate             = 0;
        video_codec_id              = 0;
    
        audio_data_rate             = 0;
        audio_codec_id              = 0;
    
        /* profile_idc, H.264-AVC-ISO_IEC_14496-10.pdf, page 45. */
        avc_profile                 = SrsAvcProfileReserved;
        /* level_idc, H.264-AVC-ISO_IEC_14496-10.pdf, page 45. */
        avc_level                   = SrsAvcLevelReserved;
        /**
         * audio specified
         * audioObjectType, in 1.6.2.1 AudioSpecificConfig, page 33,
         * 1.5.1.1 Audio object type definition, page 23,
         *           in aac-mp4a-format-ISO_IEC_14496-3+2001.pdf.
         */
        aac_object                  = SrsAacObjectTypeReserved;
        /* samplingFrequencyIndex */
        aac_sample_rate             = SRS_AAC_SAMPLE_RATE_UNSET; // sample rate ignored
        /* channelConfiguration */
        aac_channels                = 0;
        /**
         * the avc extra data, the AVC sequence header,
         * without the flv codec header,
         * @see: ffmpeg, AVCodecContext::extradata
         */
        avc_extra_size              = 0;
        avc_extra_data              = NULL;
        /**
         * the aac extra data, the AAC sequence header,
         * without the flv codec header,
         * @see: ffmpeg, AVCodecContext::extradata
         */
        aac_extra_size              = 0;
        aac_extra_data              = NULL;
    
        sequenceParameterSetLength  = 0;
        sequenceParameterSetNALUnit = NULL;
        pictureParameterSetLength   = 0;
        pictureParameterSetNALUnit  = NULL;
    
        /* the avc payload format. */
        payload_format = SrsAvcPayloadFormatGuess;
        stream = new SrsStream();
    }
    

    5.2.4 SrsCodecSample 构造函数

    /*
     * the samples in the flv audio/video packet.
     * the sample used to analysis a video/audio packet,
     * split the h.264 NALUs to buffers, or aac raw data to a buffer,
     * and decode the video/audio specified infos.
     *
     * the sample unit:
     *     a video packet codec in h.264 contains many NALUs, each is a sample unit.
     *     a audio packet codec in aac is a sample unit.
     * @remark, the video/audio sequence header is not sample unit,
     *     all sequence header stores as extra data,
     *     @see SrsAvcAacCodec.avc_extra_data and SrsAvcAacCodec.aac_extra_data
     * @remark, user must clear all samples before decode a new video/audio packet.
     */
    SrsCodecSample::SrsCodecSample()
    {
        /* 复位所有保存的数据 */
        clear();
    }
    
    /*
     * clear all samples.
     * the sample units never copy the bytes, it directly use the ptr,
     * so when video/audio packet is destroyed, the sample must be clear.
     * in a word, user must clear sample before demux it.
     * @remark demux sample use SrsAvcAacCodec.audio_aac_demux or video_avc_demux.
     */
    void SrsCodecSample::clear()
    {
        is_video = false;
        nb_sample_units = 0;
    
        cts = 0;
        frame_type = SrsCodecVideoAVCFrameReserved;
        avc_packet_type = SrsCodecVideoAVCTypeReserved;
        has_sps_pps = has_aud = has_idr = false;
        first_nalu_type = SrsAvcNaluTypeReserved;
        
        acodec = SrsCodecAudioReserved1;
        sound_rate = SrsCodecAudioSampleRateReserved;
        sound_size = SrsCodecAudioSampleSizeReserved;
        sound_type = SrsCodecAudioSoundTypeReserved;
        aac_packet_type = SrsCodecAudioTypeReserved;
    }
    

    5.2.5 SrsAvcAacCodec::audio_aac_demux

    /*
     * demux the audio packet in aac codec.
     * the packet mux in FLV/RTMP format defined in flv specification.
     * demux the audio specified data(sound format, sound_size, ...) to sample.
     * demux the aac specified data(aac_profile, ...) to codec from sequence header.
     * demux the aac raw sample units.
     */
    int SrsAvcAacCodec::audio_aac_demux(char* data, int size, SrsCodecSample* sample) 
    {
        int ret = ERROR_SUCCESS;
        
        sample->is_video = false;
        
        if (!data || size <= 0) {
            srs_trace("no audio present, ignore it.");
            return ret;
        }
        
        if ((ret = stream->initialize(data, size)) != ERROR_SUCCESS) {
            return ret;
        }
        
        // audio decode 
        if (!stream->require(1)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("aac decode sound_format failed. ret=%d", ret);
            return ret;
        }
        
        // @see: E.4.2 Audio Tags, video_file_format_spec_v10_1.pdf, page 76
        int8_t sound_format = stream->read_1bytes();
        
        /* 音频类型,即声道 */
        int8_t sound_type = sound_format & 0x01;
        /* 音频采样精度 */
        int8_t sound_size = (sound_format >> 1) & 0x01;
        /* 音频采样率 */
        int8_t sound_rate = (sound_format >> 2) & 0x03;
        /* 音频格式,对于 AAC,为 10 */
        sound_format = (sound_format >> 4) & 0x0f;
        
        audio_codec_id = sound_format;
        sample->acodec = (SrsCodecAudio)audio_codec_id;
        
        sample->sound_type = (SrsCodecAudioSoundType)sound_type;
        sample->sound_rate = (SrsCodecAudioSampleRate)sound_rate;
        sample->sound_size = (SrsCodecAudioSampleSize)sound_size;
        
        // supoort h.264+mp3 for hls.
        if (audio_codec_id == SrsCodecAudioMP3) {
            return ERROR_HLS_TRY_MP3;
        }
        
        // only support aac
        if (audio_codec_id != SrsCodecAudioAAC) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("aac only support mp3/aac codec. actual=%d, ret=%d", 
                      audio_codec_id, ret);
            return ret;
        }
        
        if (!stream->require(1)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("aac decode aac_packet_type failed. ret=%d", ret);
            return ret;
        }
        
        /* 读取 1 字节的 AAC Packet Type,检测该 Audio data 为 AAC Sequence Header,还是 AAC Raw */
        int8_t aac_packet_type = stream->read_1bytes();
        sample->aac_packet_type = (SrsCodecAudioType)aac_packet_type;
        
        /* 若为 AAC Sequence Header */
        if (aac_packet_type == SrsCodecAudioTypeSequenceHeader) {
            // AudioSpecificConfig
            // 1.6.2.1 AudioSpecificConfig, 
            // in aac-mp4a-format-ISO_IEC_14496-3+2001.pdf, page 33.
            aac_extra_size = stream->size() - stream->pos();
            if (aac_extra_size > 0) {
                srs_freepa(aac_extra_data);
                aac_extra_data = new char[aac_extra_size];
                memcpy(aac_extra_data, stream->data() + stream->pos(), aac_extra_size);
                
                // demux the sequence header.
                if ((ret = audio_aac_sequence_header_demux(aac_extra_data, aac_extra_size)) 
                    != ERROR_SUCCESS) {
                    return ret;
                }
            }
            
        /* 否则为 AAC Raw */
        } else if (aac_packet_type == SrsCodecAudioTypeRawData) {
            // ensure the sequence header demuxed
            if (!is_aac_codec_ok()) {
                srs_warn("aac ignore type=%d for no sequence header. ret=%d", 
                         aac_packet_type, ret);
                return ret;
            }
            
            // Raw AAC frame data in UI8 []
            // 6.3 Raw Data, aac-iso-13818-7.pdf, page 28
            if ((ret = sample->add_sample_unit(stream->data() + stream->pos(), 
                       stream->size() - stream->pos())) != ERROR_SUCCESS) {
                srs_error("aac add sample failed. ret=%d", ret);
                return ret;
            }
        } else {
            // ignored
        }
        
        // reset the sample rate by sequence header
        if (aac_sample_rate != SRS_AAC_SAMPLE_RATE_UNSET) {
            static  int aac_sample_rates[] = {
                96000, 88200, 64000, 48000,
                44100, 32000, 24000, 22050,
                16000, 12000, 11025,  8000,
                7350,     0,     0,    0
            };
            switch (aac_sample_rates[aac_sample_rate]) {
                case 11025:
                    sample->sound_rate = SrsCodecAudioSampleRate11025;
                    break;
                case 22050:
                    sample->sound_rate = SrsCodecAudioSampleRate22050;
                    break;
                case 44100:
                    sample->sound_rate = SrsCodecAudioSampleRate44100;
                    break;
                default:
                    break;
            }
        }
        
        srs_info("aac decoded, type=%d, codec=%d, asize=%d, rate=%d, format=%d, size=%d",
            sound_type, audio_codec_id, sound_size, sound_rate, sound_format, size);
        
        return ret;
    }
    

    5.2.6 SrsAvcAacCodec::audio_aac_sequence_header_demux

    /* directly demux the sequence header, without RTMP packet header. */
    int SrsAvcAacCodec::audio_aac_sequence_header_demux(char* data, int size)
    {
        int ret = ERROR_SUCCESS;
        
        if ((ret = stream->initialize(data, size)) != ERROR_SUCCESS) {
            return ret;
        }
        
        /* only need to decode the first 2bytes:
         *     audioObjectType, aac_profile, 5bits.
         *     samplingFrequencyIndex, aac_sample_rate, 4bites.
         *     channelConfiguration, aac_channels, 4bits */
        if (!stream->require(2)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("audio codec decode aac sequence header failed. ret=%d", ret);
            return ret;
        }
        u_int8_t profile_ObjectType = stream->read_1bytes();
        u_int8_t samplingFrequencyIndex = stream->read_1bytes();
        
        aac_channels = (samplingFrequencyIndex >> 3) & 0x0f;
        samplingFrequencyIndex = ((profile_ObjectType << 1) & 0x0e) | 
                                 ((samplingFrequencyIndex >> 7) & 0x01);
        profile_ObjectType = (profile_ObjectType >> 3) & 0x1f;
        
        // set the aac sample rate.
        aac_sampel_rate = samplingFrequencyIndex;
        
        // convert the object tyep in sequence header to aac profile of ADTS.
        aac_object = (SrsAacObjectType)profile_ObjectType;
        if (aac_object == SrsAacObjectTypeReserved) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("audio codec decode aac sequence header failed, "
                "adts object=%d invalid. ret=%d", profile_ObjectType, ret);
            return ret;
        }
        
        // TODO: FIXME: to support aac he/he-v2, see: ngx_rtmp_codec_parse_aac_header
        // 
        // donot force to LC, @see: https://github.com/ossrs/srs/issues/81
        // the source will print the sequence header info.
        //if (aac_profile > 3) {
            // Mark all extended profiles as LC
            // to make Android as happy as possible.
            // @see: ngx_rtmp_hls_parse_aac_header
            //aac_profile = 1;
        //}
        
        return ret;
    }
    

    该函数主要是解析 AAC Sequence header。

    5.2.7 SrsCodecSample::add_sample_unit

    /*
     * add the a sample unit, it's a h.264 NALU or aac raw data.
     * the sample unit directly use the ptr of packet bytes,
     * so user must never use sample unit when packet is destroyed.
     * in a word, user must clear sample before demux it.
     */
    int SrsCodecSample::add_sample_unit(char* bytes, int size)
    {
        int ret = ERROR_SUCCESS;
        
        if (nb_sample_units >= SRS_SRS_MAX_CODEC_SAMPLE) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("hls decode samples error, "
                "exceed the max count: %d, ret=%d", SRS_SRS_MAX_CODEC_SAMPLE, ret);
            return ret;
        }
        
        SrsCodecSampleUnit* sample_uint = &sample_units[nb_sample_units++];
        sample_uint->bytes = bytes;
        sample_uint->size = size;
        
        // for video, parse the nalu type, set the IDR flag.
        if (is_video) {
            SrsAvcNaluType nal_uint_type = (SrsAvcNaluType)(bytes[0] & 0x1f);
            
            if (nal_unit_type == SrsAvcNaluTypeIDR) {
                has_idr = true;
            } else if (nal_unit_type == SrsAvcNaluTypeSPS || 
                       nal_unit_type == SrsAvcNaluTypePPS) {
                has_sps_pps = true;
            } else if (nal_unit_type == SrsAvcNaluTypeAccessUnitDelimiter) {
                has_aud = true;
            }
            
            if (first_nalu_type == SrsAvcNaluTypeReserved) {
                first_nalu_type = nal_unit_type;
            }
        }
        
        return ret;
    }
    

    5.2.8 SrsGopCache::cache

    /* only for h264 codec
     * 1. cache the gop when got h264 video packet.
     * 2. clear gop when got keyframe. 
     * @param shared_msg, directly ptr, copy it if need to save it. */
    int SrsGopCache::cache(SrsSharedPtrMessage* shared_msg)
    {
        int ret = ERROR_SUCCESS;
        
        if (!enable_gop_cache) {
            srs_verbose("gop cache is disabled.");
            return ret;
        }
        
        // the gop cache know when to gop it.
        SrsSharedPtrMessage* msg = shared_msg;
        
        // got video, update the video count if acceptable
        if (msg->is_video()) {
            // drop video when not h.264
            if (!SrsFlvCodec::video_is_h264(msg->payload, msg->size)) {
                srs_info("gop cache drop video for none h.264");
                return ret;
            }
        
            /* the video frame count, avoid cache for pure audio stream. */    
            cached_video_count++;
            /*
             * when user disabled video when publishing, and gop cache enabled,
             * we will cache the audio/video for we already got video, but we never
             * know when to clear the gop cache, for there is no video in future,
             * so we must guess whether user disabled the video.
             * when we got some audios after laster video, for instance, 600 audio packets,
             * about 3s(26ms per packet) 115 audio packets, clear gop cache.
             *
             * @remark, it is ok for performance, for when we clear the gop cache,
             *     gop cache is disabled for pure audio stream.
             * @see: https://github.com/ossrs/srs/issues/124
             */
            audio_after_last_video_count = 0;
        }
        
        // no acceptable video or pure audio, disable the cache.
        if (pure_audio()) {
            srs_verbose("ignore any frame util got a h264 video frame.");
            return ret;
        }
        
        // ok, gop cache enabled, and got an audio.
        if (msg->is_audio()) {
            audio_after_last_video_count++;
        }
        
        // clear gop cache when pure audio count overflow
        if (audio_after_last_video_count > SRS_PURE_AUDIO_GUESS_COUNT) {
            srs_warn("clear gop cache for guess pure audio overflow");
            clear();
            return ret;
        }
        
        // clear gop cache when got key frame
        if (msg->is_video() && SrsFlvCodec::video_is_keyframe(msg->payload, msg->size)) {
            srs_info("clear gop cache when got keyframe. vcount=%d, count=%d", 
                cached_video_count, (int)gop_cache.size());
            
            clear();
            
            // current msg is video frame, so we set to 1.
            cached_video_count = 1;
        }
        
        // cache the frame.
        gop_cache.push_back(msg->copy());
        
        return ret;
    }
    

    该函数是将接收到的音频/视频包保存到 gop_cache 中,但是有两点需要注意:

    1. 当接收到一个 video packet,且为 keyframe 时,需要将 gop_cache 清空,从新开始缓存,并设当前缓存的 video 计数值 cached_video_count 为 1;
    2. 用 audio_after_last_video_count 记录自上一次获得 video 后,到下一次再次接收到 video 时,当前已经接收到的 audio 个数,若该值超过 115(26ms per packet, 大概 3s),则假设客户端已经禁止发送 video 了,此时需要清空 gop_cache,不再缓存 audio。

    5.3 Video

    假设接收到的第一个 video 如下图:

    参考 多媒体文件格式之FLV 中关于 Video Tag 的分析,可知上图为 AVC Sequence Header,即 sps,pps 数据。

    在 SrsRtmpConn::process_publish_message 函数中,若接收到 video,则调用 SrsSource::on_video 函数进行处理。

    5.3.1 SrsSource::on_video

    int SrsSource::on_video(SrsCommonMessage* shared_video)
    {
        int ret = ERROR_SUCCESS;
        
        // monotically increate detect.
        if (!mix_correct && is_monotically_increate) {
            if (last_packet_time > 0 && shared_video->header.timestamp < last_packet_time) {
                is_monotonically_increase = false;
                srs_warn("VIDEO: stream not monotonically increase, please open mix_correct.");
            }
        }
        last_packet_time = shared_video->header.timestamp;
        
        // drop any unknown header video.
        // @see https://github.com/ossrs/srs/issues/421
        if (!SrsFlvCodec::video_is_acceptable(shared_video->payload, shared_video->size)) {
            char b0 = 0x00;
            if (shared_video->size > 0) {
                b0 = shared_video->payload[0];
            }
            
            srs_warn("drop unknown header video, size=%d, bytes[0]=%#x", 
                     shared_video->size, b0);
            return ret;
        }
        
        // convert shared_video to msg, user should not use shared_video again.
        // the payload is transfer to msg, and set to NULL in shared_video.
        SrsSharedPtrMessage msg;
        /* 将 shared_video 中的数据拷贝到 msg 中 */
        if ((ret = msg.create(shared_video)) != ERROR_SUCCESS) {
            srs_error("initialize the video failed. ret=%d", ret);
            return ret;
        }
        srs_info("Video dts=%"PRId64", size=%d", msg.timestamp, msg.size);
        
        // directly process the audio message.
        if (!mix_correct) {
            return on_video_imp(&msg);
        }
        
        // insert msg to the queue.
        mix_queue->push(msg.copy());
        
        // fetch someone from mix queue.
        SrsSharedPtrMessage* m = mix_queue->pop();
        if (!m) {
            return ret;
        }
        
        // consume the monotonically increase message.
        if (m->is_audio()) {
            ret = on_audio_imp(m);
        } else {
            ret = on_video_imp(m);
        }
        srs_freep(m);
        
        return m;
    }
    

    该函数中先是检测接收到 video 是正确的消息后,然后构造一个 SrsSharedPtrMessage 类的对象,并将接收到的 video 数据拷贝到该对象中,然后调用 SrsSource::on_video_imp 进行处理.

    5.3.2 SrsSource::on_video_imp

    int SrsSource::on_video_imp(SrsSharedPtrMessage* msg) 
    {
        int ret = ERROR_SUCCESS;
        
        srs_info("Video dts=%"PRId64", size=%d", msg->timestamp, msg->size);
        
        /* 检测是否是关键帧,并且 videodatatype 为 sequence header,即 sps,pps 数据 */
        bool is_sequence_header = 
             SrsFlvCodec::video_is_sequence_header(msg->payload, msg->size);
        
        // whether consumer should drop for the duplicated sequence header.
        bool drop_for_reduce = false;
        /* 若配置文件中使能了 reduce_sequence_header 配置项 */
        if (is_sequence_header && cache_sh_video && 
            _srs_config->get_reduce_sequence_header(_req->vhost)) {
            if (cache_sh_video->size == msg->size) {
                drop_for_reduce = srs_bytes_equals(cache_sh_video->payload, 
                                  msg->payload, msg->size);
                srs_warn("drop for reduce sh video, size=%d", msg->size);
            }
        }
        
        // cache the sequence header if h264
        // donot cache the sequence header to gop_cache, return here.
        if (is_sequence_header) {
            srs_freep(cache_sh_video);
            /* the cached video sequence header. */
            cache_sh_video = msg->copy();
            
            // parse detail audio codec
            SrsAvcAacCodec codec;
            
            /* 默认使能解析 sps */
            // user can disable the sps parse to workaround when parse sps failed.
            // @see https://github.com/ossrs/srs/issues/474
            codec.avc_parse_sps = _srs_config->get_parse_sps(_req->vhost);
            
            SrsCodecSample sample;
            if ((ret = codec.video_avc_demux(msg->payload, msg->size, &sample)) 
                != ERROR_SUCCESS) {
                srs_error("source codec demux video failed. ret=%d", ret);
                return ret;
            }
            
            // when got video stream info.
            SrsStatistic* stat = SrsStatistic::instance();
            if ((ret = stat->on_video_info(_req, SrsCodecVideoAVC, 
                       codec.avc_profile, codec.avc_level)) 
                != ERROR_SUCCESS) {
                return ret;
            }
            
            srs_trace("%dB video sh,  "
                      "codec(%d, profile=%s, level=%s, %dx%d, %dkbps, %dfps, %ds)",
                msg->size, codec.video_codec_id,
                srs_codec_avc_profile2str(codec.avc_profile).c_str(),
                srs_codec_avc_level2str(codec.avc_level).c_str(), codec.width, codec.height,
                codec.video_data_rate / 1000, codec.frame_rate, codec.duration);
        }
        
    #ifdef SRS_AUTO_HLS
        ...
    #endif
        
    #ifdef SRS_AUTO_DVR
        ...
    #endif
    
    #ifdef SRS_AUTO_HDS
        ...
    #endif
    
        // copy to all consumer
        if (!drop_for_reduce) {
            for (int i = 0; i < (int)consumers.size(); i++) {
                SrsConsumer* consumer = consumers.at(i);
                if ((ret = consumer->enqueue(msg, atc, jitter_algorithm)) != ERROR_SUCCESS) {
                    srs_error("dispatch the video failed. ret=%d", ret);
                    return ret;
                }
            }
            srs_info("dispatch video success.");
        }
        
        // copy to all forwarders.
        if (!forwarders.empty()) {
            std::vector<SrsForwarder*>::iterator it;
            for (it = forwarders.begin(); it != forwarders.end(); ++it) {
                SrsForwarder* forwarder = *it;
                if ((ret = forwarder->on_video(msg)) != ERROR_SUCCESS) {
                    srs_error("forwarder process video message failed. ret=%d", ret);
                    return ret;
                }
            }
        }
        
        // when sequence heeader, donot push to gop cache and adjust the timestamp.
        if (is_sequence_header) {
            return ret;
        }
        
        // cache the last gop packets
        if ((ret = gop_cache->cache(msg)) != ERROR_SUCCESS) {
            srs_error("gop cache msg failed. ret=%d", ret);
            return ret;
        }
        srs_verbose("cache gop success.");
        
        // if atc, update the sequence header to abs time.
        if (atc) {
            if (cache_sh_video) {
                cache_sh_video->timestamp = msg->timestamp;
            }
            if (cache_metadata) {
                cache_metadata->timestamp = msg->timestamp;
            }
        }
        
        return ret;
    }
    

    5.3.3 SrsAvcAacCodec::video_avc_demux

    /* demux the video packet in h.264 codec.
     * the packet mux in FLV/RTMP format defined in flv specification. 
     * demux the video specified data(frame_type, codec_id, ...) to sample.
     * demux the h.264 specified data(avc_profile, ...) to codec from sequence header.
     * demux the h.264 MALUs to sample units. */
    int SrsAvcAacCodec::video_avc_demux(char* data, int size, SrsCodecSample* sample) 
    {
        int ret = ERROR_SUCCESS;
        
        sample->is_video = true;
        
        if (!data || size <= 0) {
            srs_trace("no video present, ignore it.");
            return ret;
        }
        
        if ((ret = stream->initialize(data, size)) != ERROR_SUCCESS) {
            return ret;
        }
        
        // video decode
        if (!stream->require(1)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode frame_type failed. ret=%d", ret);
            return ret;
        }
        
        // @see: E.4.3 Video Tags, video_file_format_spec_v10_1.pdf, page 78
        int8_t frame_type = stream->read_1bytes();
        /* 获取编码 id,一般为 7,即 AVC 编码,SRS 仅支持 AVC 编码 */
        int8_t codec_id = frame_type & 0x0f;
        /* 获取帧类型,为 1 则为 keyframe */
        frame_type = (frame_type >> 4) & 0x0f;
        
        sample->frame_type = (SrsCodecVideoAVCFrame)frame_type;
        
        // ignore info frame without error,
        // @see https://github.com/ossrs/srs/issues/288#issuecomment-69863909
        if (sample->frame_type == SrsCodecVideoAVCFrameVideoInfoFrame) {
            srs_warn("avc igone the info frame, ret=%d", ret);
            return ret;
        }
        
        // only support h.264/avc
        if (codec_id != SrsCodecVideoAVC) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc only support video h.264/avc codec. actual=%d, ret=%d", 
                      codec_id, ret);
            return ret;
        }
        video_codec_id = codec_id;
        
        if (!stream->require(4)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode avc_packet_type failed. ret=%d", ret);
            return ret;
        }
        /* 获取 AVC packet 类型,0:AVC Sequence Header,1: AVC NALU Units */
        int8_t avc_packet_type = stream->read_1bytes();
        /* 获取 cts,如果 avc_packet_type 为 1,则为 cts 偏移,否则为 0 则为 0 */
        int32_t composition_time = stream->read_3bytes();
        
        // pts = dts + ctx.
        sample->ctx = composition_time;
        sample->avc_packet_type = (SrsCodecVideoAVCType)avc_packet_type;
        
        /* 若为 AVC Sequence Header,则解码 sps,pps */
        if (avc_packet_type == SrsCodecVideoAVCTypeSequenceHeader) {
            if ((ret = avc_demux_sps_pps(stream)) != ERROR_SUCCESS) {
                return ret;
            }
            
        /* 若为 AVC NALU Unit,则解码 H.264 NALU */
        } else if (avc_packet_type == SrsCodecVideoAVCTypeNALU) {
            if ((ret = video_nalu_demux(stream, sample)) != ERROR_SUCCESS) {
                return ret;
            }
        } else {
            / ignored.
        }
        
        srs_info("avc decoded, type=%d, codec=%d, avc=%d, cts=%d, size=%d",
            frame_type, video_codec_id, avc_packet_type, composition_time, size);
        
        return ret;
    }
    

    5.3.4 SrsAvcAacCodec::avc_demux_sps_pps

    int SrsAvcAacCodec::avc_demux_sps_pps(SrsStream* stream) 
    {
        int ret = ERROR_SUCCESS;
        
        // AVCDecoderConfigurationRecord
        // 5.2.4.1.1 Syntax, H.264-AVC-ISO_IEC_14496-15.pdf, page 16
        avc_extra_size = stream->size() - stream->pos();
        if (avc_extra_size > 0) {
            srs_freepa(avc_extra_data);
            avc_extra_data = new char[avc_extra_size];
            memcpy(avc_extra_data, stream->data() + stream->pos(), avc_extra_size);
        }
        
        if (!stream->require(6)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode sequenc header failed. ret=%d", ret);
            return ret;
        }
        // int8_t configurationVersion = stream->read_1bytes();
        stream->read_1bytes(); // configurationVersion
        // int8_t AVCProfileIndication = stream->read_1bytes();
        /*
         * the profile for avc/h.264.
         * @see Annex A Profiles and levels, H.264-AVC-ISO_IEC_14496-10.pdf, page 205.
         */
        avc_profile = (SrsAvcProfile)stream->read_1bytes(); // AVCProfileIndication
        // int8_t profile_compatibility = stream->read_1bytes();
        stream->read_1bytes();  // profile_compatibility
        // int8_t AVCLevelIndication = stream->read_1bytes(); 
        avc_level = (SrsAvcLevel)stream->read_1bytes(); // AVCLevelIndication
        
        // parse the NALU size.
        int8_t lengthSizeMinusOne = stream->read_1bytes();
        lengthSizeMinusOne &= 0x03;
        /* lengthSizeMinusOne,一般为 3, H.264-AVC-ISO_IEC_14496-15.pdf, page 16 */
        NAL_unit_length = lengthSizeMinusOne;
        
        // 5.3.4.2.1 Syntax, H.264-AVC-ISO_IEC_14496-15.pdf, page 16
        // 5.2.4.1 AVC decoder configuration record
        // 5.2.4.1.2 Semantics
        // The value of this field shall be one of 0, 1, or 3 corresponding to a
        // length encoded with 1, 2, or 4 bytes, respectively.
        if (NAL_unit_length == 2) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("sps lengthSizeMinusOne should never be 2. ret=%d", ret);
            return ret;
        }
        
        // 1 sps, 7.3.2.1 Sequence parameter set RBSP syntax
        // H.264-AVC-ISO_IEC_14496-10.pdf, page 45.
        if (!stream->require(1)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode sequenc header sps failed. ret=%d", ret);
            return ret;
        }
        int8_t numOfSequenceParameterSets = stream->read_1bytes();
        /* 获取 SPS NALU 的个数,一般为 1 */
        numOfSequenceParameterSets &= 0x1f;
        if (numOfSequenceParameterSets != 1) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode sequenc header sps failed. ret=%d", ret);
            return ret;
        }
        if (!stream->require(2)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode sequenc header sps size failed. ret=%d", ret);
            return ret;
        }
        /* 获取该 SPS NALU 的大小 */
        sequenceParameterSetLength = stream->read_2bytes();
        if (!stream->require(sequenceParameterSetLength)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode sequenc header sps data failed. ret=%d", ret);
            return ret;
        }
        if (sequenceParameterSetLength > 0) {
            srs_freepa(sequenceParameterSetNALUnit);
            sequenceParameterSetNALUnit = new char[sequenceParameterSetLength];
            /* 将 SPS NALU 的数据拷贝到 sequenceParameterSetNALUnit 中 */
            stream->read_bytes(sequenceParameterSetNALUnit, sequenceParameterSetLength);
        }
        // 1 pps
        if (!stream->require(1)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode sequenc header pps failed. ret=%d", ret);
            return ret;
        }
        int8_t numOfPictureParameterSets = stream->read_1bytes();
        /* 获取 PPS NALU 的个数,一般为 1 */
        numOfPictureParameterSets &= 0x1f;
        if (numOfPictureParameterSets != 1) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode sequenc header pps failed. ret=%d", ret);
            return ret;
        }
        if (!stream->require(2)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode sequenc header pps size failed. ret=%d", ret);
            return ret;
        }
        pictureParameterSetLength = stream->read_2bytes();
        if (!stream->require(pictureParameterSetLength)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode sequenc header pps data failed. ret=%d", ret);
            return ret;
        }
        if (pictureParameterSetLength > 0) {
            srs_freepa(pictureParameterSetNALUnit);
            pictureParameterSetNALUnit = new char[pictureParameterSetLength];
            /* 将 PPS NALU 的数据拷贝到 pictureParameterSetNALUnit 中 */
            stream->read_bytes(pictureParameterSetNALUnit, pictureParameterSetLength);
        }
        
        /* 解析 SPS NALU */
        return avc_demux_sps();
    }
    

    该函数分别将 SPS 和 PPS 提取保存到 sequenceParameterSetNALUnit 和 pictureParameterSetNALUnit,最后调用 avc_demux_sps 函数解析 SPS。

    5.3.5 SrsAvcAacCodec::avc_demux_sps

    /* decode the sps rbsp stream. */
    int SrsAvcAacCodec::avc_demux_sps()
    {
        int ret = ERROR_SUCCESS;
        
        if (!sequenceParameterSetLength) {
            return ret;
        }
        
        SrsStream stream;
        if ((ret = stream.initialize(sequenceParameterSetNALUnit, sequenceParameterSetLength)) 
            != ERROR_SUCCESS) {
            return ret;
        }
        
        // for NALU, 7.3.1 NAL unit syntax
        // H.264-AVC-ISO_IEC_14496-10-2012.pdf, page 61.
        if (!stream.require(1)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("avc decode sps failed. ret=%d", ret);
            return ret;
        }
        int8_t nutv = stream.read_1bytes();
        
        /* 禁止位 */
        // forbidden_zero_bit shall be equal to 0.
        int8_t forbidden_zero_bit = (nutv >> 7) & 0x01;
        if (forbidden_zero_bit) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("forbidden_zero_bit shall be equal to 0. ret=%d", ret);
            return ret;
        }
        
        /* 重要性指示位 */
        // nal_ref_idc not equal to 0 specifies that the content 
        // of the NAL nuit contains a sequence parameter set or a picture
        // parameter set or a slice of reference picture 
        // or a slice data partition of a reference picture.
        int8_t nal_ref_idc = (nutv >> 5) & 0x03;
        if (!nal_ref_idc) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("for sps, nal_ref_idc shall be not be equal to 0. ret=%d", ret);
            return ret;
        }
        
        /* NAL 单元类型,这里必须为 7,即 SPS */
        // 7.4.1 NAL unit semantics
        // H.264-AVC-ISO_IEC_14496-10-2012.pdf, page 61.
        // nal_unit_type specifies the type of RBSP data structure contained 
        // in the NAL unit as specified in Table 7-1.
        SrsAvcNaluType nal_unit_type = (SrsAvcNaluType)(nutv & 0x1f);
        if (nal_unit_type != 7) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("for sps, nal_unit_type shall be equal to 7. ret=%d", ret);
            return ret;
        }
        
        // decode the rbsp from sps.
        // rbsp[ i ] a raw byte sequence payload is specified as an ordered sequence of bytes.
        int8_t* rbsp = new int8_t[sequenceParameterSetLength];
        SrsAutoFreeA(int8_t, rbsp);
        
        /* 遍历 EBSP 数据,找到并丢弃 0x03 字节,也即将 EBSP 转为 RBSP */
        int nb_rbsp = 0;
        while (!stream.empty()) {
            rbsp[nb_rbsp] = stream.read_1bytes();
            
            // XX 00 00 03 XX, the 03 byte should be drop.
            if (nb_rbsp > 2 && rbsp[nb_rbsp - 2] == 0 && rbsp[nb_rbsp - 1] == 0 && 
                rbsp[nb_rbsp] == 3) {
                // read 1byte more.
                if (stream.empty()) {
                    break;
                }
                rbsp[nb_rbsp] = stream.read_1bytes();
                nb_rbsp++;
                
                continue;
            }
            
            nb_rbsp++;
        }
        
        /* 对提取到的 RBSP 数据进行解析 */
        return avc_demux_sps_rbsp((char*)rbsp, nb_rbsp);
    }
    

    5.3.6 SrsAvcAacCodec::avc_demux_sps_rbsp

    int SrsAvcAacCodec::avc_demux_sps_rbsp(char* rbsp, int nb_rbsp)
    {
        int ret = ERROR_SUCCESS;
        
        /* 若配置文件中配置禁止解析 sps,则直接返回 */
        // we donot parse the detail of sps.
        // @see https://github.com/ossrs/srs/issues/474
        if (!avc_parse_sps) {
            return ret;
        }
        
        // reparse the rbsp.
        SrsStream stream;
        if ((ret = stream.initialize(rbsp, nb_rbsp)) != ERROR_SUCCESS) {
            return ret;
        }
        
        // for SPS, 7.3.2.1.1 Sequence parameter set data syntax
        // H.264-AVC-ISO_IEC_14496-10-2012.pdf, page 62.
        if (!stream.require(3)) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("sps shall atleast 3bytes. ret=%d", ret);
            return ret;
        }
        /* 本视频编码时遵循的 profile,profile 分为 Baseline,Main,Extended 等,
         * 主要用来规定编码时是否采用某些特性,比如说 Baseline Profile 就规定了
         * 只能使用 I、P slice 进行编码,关于 profile 的说明可以去查看标准的 
         * Annex A 
         */
        u_int8_t profile_idc = stream.read_1bytes();
        if (!profile_idc) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("sps the profile_idc invalid. ret=%d", ret);
            return ret;
        }
        
        int8_t flags = stream.read_1bytes();
        if (flags & 0x03) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("sps the flags invalid. ret=%d", ret);
            return ret;
        }
        
        /* 本视频遵循的 level,level 主要规定了每秒最多能处理多少个宏块,
         * 最大的帧大小,最大的解码缓存,最大的比特率等这些性能相关的东西,
         * 如果是硬解码,则比较容易出现由于视频 level 太高而不能解码的
         * 情况。 
         */
        u_int8_t level_idc = stream.read_1bytes();
        if (!level_idc) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("sps the level_idc invalid. ret=%d", ret);
            return ret;
        }
        
        
        SrsBitStream bs;
        if ((ret = bs.initialize(&stream)) != ERROR_SUCCESS) {
            return ret;
        }
        
        /* 本 SPS 的 ID,这个 ID 主要是给 PPS 用的 */
        int32_t seq_parameter_set_id = -1;
        if ((ret = srs_avc_nalu_read_uev(&bs, seq_parameter_set_id)) != ERROR_SUCCESS) {
            return ret;
        }
        if (seq_parameter_set_id < 0) {
            ret = ERROR_HLS_DECODE_ERROR;
            srs_error("sps the seq_parameter_set_id invalid. ret=%d", ret);
            return ret;
        }
        srs_info("sps parse profile=%d, level=%d, sps_id=%d", profile_idc, 
                 level_idc, seq_parameter_set_id);
        
        int32_t chroma_format_idc = -1;
        if (profile_idc == 100 || profile_idc == 110 || profile_idc == 122 
            || profile_idc == 244 || profile_idc == 44 || profile_idc == 83 
            || profile_idc == 86 || profile_idc == 118 || profile_idc == 128
        ) {
            if ((ret = srs_avc_nalu_read_uev(&bs, chroma_format_idc)) != ERROR_SUCCESS) {
                return ret;
            }
            if (chroma_format_idc == 3) {
                int8_t separate_colour_plane_flag = -1;
                if ((ret = srs_avc_nalu_read_bit(&bs, separate_colour_plane_flag)) 
                    != ERROR_SUCCESS) {
                    return ret;
                }
            }
            
            int32_t bit_depth_luma_minus8 = -1;
            if ((ret = srs_avc_nalu_read_uev(&bs, bit_depth_luma_minus8)) != ERROR_SUCCESS) {
                return ret;
            }
            
            int32_t bit_depth_chroma_minus8 = -1;
            if ((ret = srs_avc_nalu_read_uev(&bs, bit_depth_chroma_minus8)) != ERROR_SUCCESS) 
            {
                return ret;
            }
            
            int8_t qpprime_y_zero_transform_bypass_flag = -1;
            if ((ret = srs_avc_nalu_read_bit(&bs, qpprime_y_zero_transform_bypass_flag)) 
                != ERROR_SUCCESS) {
                return ret;
            }
            
            int8_t seq_scaling_matrix_present_flag = -1;
            if ((ret = srs_avc_nalu_read_bit(&bs, seq_scaling_matrix_present_flag)) 
                != ERROR_SUCCESS) {
                return ret;
            }
            if (seq_scaling_matrix_present_flag) {
                int nb_scmpfs = ((chroma_format_idc != 3)? 8:12);
                for (int i = 0; i < nb_scmpfs; i++) {
                    int8_t seq_scaling_matrix_present_flag_i = -1;
                    if ((ret = srs_avc_nalu_read_bit(&bs, seq_scaling_matrix_present_flag_i)) 
                        != ERROR_SUCCESS) {
                        return ret;
                    }
                }
            }
        }
        
        int32_t log2_max_frame_num_minus4 = -1;
        if ((ret = srs_avc_nalu_read_uev(&bs, log2_max_frame_num_minus4)) != ERROR_SUCCESS) {
            return ret;
        }
        
        int32_t pic_order_cnt_type = -1;
        if ((ret = srs_avc_nalu_read_uev(&bs, pic_order_cnt_type)) != ERROR_SUCCESS) {
            return ret;
        }
        
        if (pic_order_cnt_type == 0) {
            int32_t log2_max_pic_order_cnt_lsb_minus4 = -1;
            if ((ret = srs_avc_nalu_read_uev(&bs, log2_max_pic_order_cnt_lsb_minus4)) 
                != ERROR_SUCCESS) {
                return ret;
            }
        } else if (pic_order_cnt_type == 1) {
            int8_t delta_pic_order_always_zero_flag = -1;
            if ((ret = srs_avc_nalu_read_bit(&bs, delta_pic_order_always_zero_flag)) 
                != ERROR_SUCCESS) {
                return ret;
            }
            
            int32_t offset_for_non_ref_pic = -1;
            if ((ret = srs_avc_nalu_read_uev(&bs, offset_for_non_ref_pic)) 
                != ERROR_SUCCESS) {
                return ret;
            }
            
            int32_t offset_for_top_to_bottom_field = -1;
            if ((ret = srs_avc_nalu_read_uev(&bs, offset_for_top_to_bottom_field)) 
                != ERROR_SUCCESS) {
                return ret;
            }
            
            int32_t num_ref_frames_in_pic_order_cnt_cycle = -1;
            if ((ret = srs_avc_nalu_read_uev(&bs, num_ref_frames_in_pic_order_cnt_cycle)) 
                != ERROR_SUCCESS) {
                return ret;
            }
            if (num_ref_frames_in_pic_order_cnt_cycle < 0) {
                ret = ERROR_HLS_DECODE_ERROR;
                srs_error("sps the num_ref_frames_in_pic_order_cnt_cycle invalid. ret=%d", 
                          ret);
                return ret;
            }
            for (int i = 0; i < num_ref_frames_in_pic_order_cnt_cycle; i++) {
                int32_t offset_for_ref_frame_i = -1;
                if ((ret = srs_avc_nalu_read_uev(&bs, offset_for_ref_frame_i)) 
                    != ERROR_SUCCESS) {
                    return ret;
                }
            }
        }
        
        /* 参考帧最多能有多少个 */
        int32_t max_num_ref_frames = -1;
        if ((ret = srs_avc_nalu_read_uev(&bs, max_num_ref_frames)) != ERROR_SUCCESS) {
            return ret;
        }
        
        /* 由于码流在传输过程中可能出现丢包的情况,从而导致中间有帧缺失,
         * 如果制定了这个标记,则会在解码时对帧丢失的情况进行调整,否则
         * 就当做意外丢失处理 */
        int8_t gaps_in_frame_num_value_allowed_flag = -1;
        if ((ret = srs_avc_nalu_read_bit(&bs, gaps_in_frame_num_value_allowed_flag)) 
            != ERROR_SUCCESS) {
            return ret;
        }
        
        /* 图片宽度(宏块为单位)- 1 */
        int32_t pic_width_in_mbs_minus1 = -1;
        if ((ret = srs_avc_nalu_read_uev(&bs, pic_width_in_mbs_minus1)) 
            != ERROR_SUCCESS) {
            return ret;
        }
        
        /* 图片高度(宏块为单位)- 1 */
        int32_t pic_height_in_map_units_minus1 = -1;
        if ((ret = srs_avc_nalu_read_uev(&bs, pic_height_in_map_units_minus1)) 
            != ERROR_SUCCESS) {
            return ret;
        }
        
        width = (int)(pic_width_in_mbs_minus1 + 1) * 16;
        height = (int)(pic_height_in_map_units_minus1 + 1) * 16;
        
        return ret;
    }
    
  • 相关阅读:
    .Net开发笔记(二十一) 反射在.net中的应用
    .Net开发笔记(二十)创建一个需要授权的第三方组件
    .Net开发笔记(十九) 创建一个可以可视化设计的对象
    .net开发笔记(十八) winform中的等待框
    .Net开发笔记(十七) 应用程序扩展
    java连接https时禁用证书验证.
    How to disable SSL certificate checking with Spring RestTemplate?(使用resttemplate访问https时禁用证书检查)
    webpack打包调试react并使用babel编译jsx配置方法
    动态改变spring定时任务执行频率
    在java代码中,用xslt处理xml文件
  • 原文地址:https://www.cnblogs.com/jimodetiantang/p/9098003.html
Copyright © 2020-2023  润新知