Compression is one of the most important criteria in determining quality and cost of ownership in a digital video system. Advances in processing power enable more sophisticated compression techniques. Better compression lowers the total cost of ownership (see chart on page 62) and enables remote access to quality video over the Internet and wide area networks. Most digital video products allow an installer to configure the quality to match the bandwidth and storage capabilities of the system.
Why compress video?
Uncompressed digital video is simply cost prohibitive to store and impossible to transmit over most networks. Raw video generates over 1,000 Kbytes (Kilobytes) for each frame of video. This translates to a whopping 80 megabytes per second ( Mbs ) for a 10 frame per second (fps) video stream at a resolution of 320x240 pixels. Some 20 Mbs are required for a 10 fps stream at VGA resolution (640x480).
Compression reduces the amount of data required to represent a video frame. The process works by applying mathematical algorithms to the data at the video source--digital video recorder (DVR), encoder or network camera. The process must be reversed to decode the compression video for viewing on a monitor.
All video compression techniques are “ lossy ” meaning that some information is lost during the compression process. Fortunately, the human eye is forgiving and can tolerate some loss of information without impacting the perceived quality of the image.
Video compression standards fall into two basic categories: spatial (image) and temporal compression.
Spatial compression standards individually compress each frame of video and create an image stream by sequentially concatenating the frames in sequence. As more frames are created, more bandwidth and storage space is required. The relationship between recorded frames and storage is very linear.
Two popular spatial compression techniques are Wavelet and Motion JPEG or MJPEG. Wavelet is the grandfather of compression for digital video surveillance. An entire generation of DVRs have been built on wavelet compression. Most current network cameras have been built to support MJPEG. Since JPEG images pervade most Web pages, any PC can decode MJPEG video without a special decoder.
Temporal compression techniques like MPEG-4 (Part 2) do not compress each individual video frame in its entirety. MPEG-4 only compresses the differences between consecutive frames. The process works by periodically compressing a full frame called a reference or Intra frame (I-frame). The process then compresses only the subsequent differences in the following frames: called difference or P-frames. The result of this process is a significant reduction in the size of the P-frames and overall data stream. This data reduction is particularly evident when there are periods of low motion in the video scene and therefore fewer differences from frame-to-frame. (See related graphic on this page.)
H.264 is the next step in the compression evolution. The codec resulted from efforts by two standards groups and is therefore known by several names. MPEG-4 Part 10, H.264 and Advanced Video Coding (AVC) all refer to the same compression method. H.264 couples the same general motion estimation techniques found in MPEG-4 with more advanced mathematics to further reduce data rates while preserving video quality.
Nicol Verheem , chief executive officer of technology provider OpenVideo of Orange County, California expects H.264 to become the de facto compression standard for video surveillance. Verheem said H.264 can provide a data rate reduction of about 25 percent versus equivalent MPEG-4 video. Verheem noted that unlike MPEG-4 (Part 2), H.264 provides full support for high definition (megapixel) video. Relative to MPEG-4, Verheem added that H.264 will have better interoperability between manufacturers since the standard is more fully defined and leaves less room for variations in implementation. H.264 is starting to appear in digital video products such as Sony's IPela camera line.