May 2008
Volume 8, Issue 6
Vision Sciences Society Annual Meeting Abstract  |   May 2008
Visual attention guided video compression
Author Affiliations
  • Zhicheng Li
    School of Automation Science and Electrical Engineering, Beihang University, and Department of Computer Science, University of Southern California
  • Laurent Itti
    Department of Computer Science, University of Southern California
Journal of Vision May 2008, Vol.8, 772. doi:
  • Views
  • Share
  • Tools
    • Alerts
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Zhicheng Li, Laurent Itti; Visual attention guided video compression. Journal of Vision 2008;8(6):772.

      Download citation file:

      © ARVO (1962-2015); The Authors (2016-present)

  • Supplements

Human visual characteristics show promising future for applications to video coding. Here, we propose, implement, and test a universal visual attention based video coding platform (VAVC). This platform includes two main parts: the visual attention module and the video coding module. The visual attention module is used to generate saliency maps (or other maps which can represent human visual characteristics) according to the human visual system (HVS) while the video coding module is used to compress the raw video sequence according to the results of the first module. Using this platform, a saliency-based video coding algorithm is implemented. The bottom-up methods proposed in Itti et al. (1998) are adopted to get the saliency map. Then we transform the saliency map into the quantization map used in the latest video coding standard H.264 to guide the residual quantization. For the salient regions, we decrease the quantization step to reduce the artifacts, and for the non-salient regions, we increase the quantization step to increase the compression ratio. In our experiment, 18 natural video sequences are adopted for encoding with different methods while 6 subjects to evaluate these encoded results. Subjects were asked to subjectively rate on a 1–5 scale the perceptual quality of 3 variants of the clips: standard H.264, our VAVC (yielding on average 17.37% smaller file sizes), and rate controlled H.264 to match the smaller size of the VAVC encoded clips. The experiment results show that, for 64.9% samples, the subjective quality of VAVC-encoded clips is equal or better than traditional H.264-encoded clips. For 87.04% samples, the subjective quality of the proposed VAVC method is equal or better than the rate-controlled H.264 method for equal file size. Our results suggest that exploiting human visual characteristics can lead to better video compression without degrading perceptual quality.

Li, Z. Itti, L. (2008). Visual attention guided video compression [Abstract]. Journal of Vision, 8(6):772, 772a,, doi:10.1167/8.6.772. [CrossRef]
 The authors gratefully acknowledge the contribution of NSF, HFSP, NGA, DARPA and Chinese Scholarship Council.

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.