首页> 外文学位 >Multiplexing H.264 video with AAC audio bit streams, demultiplexing and achieving lip synchronization during playback.
【24h】

Multiplexing H.264 video with AAC audio bit streams, demultiplexing and achieving lip synchronization during playback.

机译:将H.264视频与AAC音频比特流多路复用,在播放过程中多路分解并实现口型同步。

获取原文
获取原文并翻译 | 示例

摘要

H.264, MPEG-4 part-10 or AVC [5], is the latest digital video codec standard which has proven to be superior than earlier standards in terms of compression ratio, quality, bit rates and error resilience. However, the standard just defines a video codec and has no mention of any audio compression. In order to have a meaningful delivery of the video to the end user, it is necessary to associate an audio stream along with it. AAC (advanced audio coding) [1] is the latest digital audio codec standard defined in MPEG-2 and later in MPEG-4 with few changes. The audio quality of an AAC stream is observed to be better than both MP3 and AC3, which were widely used as the audio coding standard in various applications, at lower bit rates. Adopting H.264 as video codec and AAC as the audio codec, for transmission of digital multimedia through air (ATSC, DVB) or through the internet (video streaming, IPTV), facilitates the users to take advantage of the leading technologies in both audio and video. However, for these applications, treatment of video and audio as separate streams requires multiplexing the two in order to create a single bit stream for transmission. The objective of the thesis is to propose a method for effectively multiplexing the audio and video coded streams for transmission followed by demultiplexing the streams at the receiving end and achieve lip sync between the audio and video during playback. The proposed method takes advantage of the frame wise arrangement of data in both audio and video codecs. The audio and video frames are used as the first layer of packetization. The frame numbers of the audio and video data blocks are used as the reference for aligning the streams in order to achieve lip sync. The synchronizing information is embedded in the headers of the first layer of packetization. Then second layer of packetization is carried out from the first layer in order to meet the various requirements of transmission channels. Proposed method uses playback time as the criteria for allocating data packets during multiplexing in order to prevent buffer overflow or underflow at the demultiplexer end. More information is embedded into the headers to ensure an effective and fast demultiplexing process, to detect errors and correct them. Advantages and limitations of the proposed method are discussed in detail.
机译:H.264,MPEG-4 part-10或AVC [5]是最新的数字视频编解码器标准,已被证明在压缩率,质量,比特率和错误恢复能力方面优于早期标准。但是,该标准仅定义了视频编解码器,没有提及任何音频压缩。为了将视频有意义地传递给最终用户,有必要将音频流与其关联。 AAC(高级音频编码)[1]是在MPEG-2和MPEG-4中定义的最新数字音频编解码器标准,几乎没有更改。可以观察到AAC流的音频质量要好于MP3和AC3,后者在较低的比特率下已广泛用作各种应用中的音频编码标准。采用H.264作为视频编解码器,并采用AAC作为音频编解码器,以通过空中(ATSC,DVB)或通过互联网(视频流,IPTV)传输数字多媒体,从而使用户能够利用音频中的领先技术和视频。但是,对于这些应用程序,将视频和音频视为单独的流需要将两者复用,以便创建用于传输的单个位流。本文的目的是提出一种有效地复用音频和视频编码流以进行传输,然后在接收端对流进行解复用并在回放期间在音频和视频之间实现口型同步的方法。所提出的方法利用了音频和视频编解码器中数据的逐帧安排。音频和视频帧用作打包的第一层。音频和视频数据块的帧号用作对齐流以实现口形同步的参考。同步信息嵌入在打包的第一层的头中。然后,从第一层开始执行第二层打包,以满足传输通道的各种要求。为了防止缓冲器在解复用器端上溢或下溢,所提出的方法使用回放时间作为在复用期间分配数据分组的标准。标头中嵌入了更多信息,以确保有效且快速的多路分解过程,检测错误并进行纠正。详细讨论了所提出方法的优点和局限性。

著录项

  • 作者

    Murugan, Harishankar.;

  • 作者单位

    The University of Texas at Arlington.;

  • 授予单位 The University of Texas at Arlington.;
  • 学科 Engineering Electronics and Electrical.
  • 学位 M.S.
  • 年度 2007
  • 页码 68 p.
  • 总页数 68
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 无线电电子学、电信技术;
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号