We propose an application that utilizes audio and video data dependencies to achieve additional video compression in low-bit rate encoding systems such as: H.263+ video coding and G.723.1 audio coding standards. The joint correlation of synchronized audio and motion parameters has been proved to exist. A joint performance of Principal Component Analysis (PCA) by Karhunen-Loeve expansions (KL) and Tree-Structured Vector Quantization algorithms (TSVQ) based on LindeBuzo-Gray (LBG) and Competitive Learning (CL) techniques achieve as much as 60% bit reduction for the motion in the mouth region (1% of the overall output bit rate of a P frame) and provide the same motion-compensated image quality in high picture formats. We show performance evaluations that determine the optimal audio parameters, such as Linear Predictive Coefficients (LPC) or Line Spectrum Fairs (LSP), and determine the nature of the motion parameter in each macroblock of the mouth region when using Advanced Prediction Mode (APM) video coding.
展开▼