Chapter 8 - Video
Chapter 8 - Video
Principles
Chapter 8
Video
TMH Chapter - 8 1
Video
Introduction
Motion video (or simply video) is a combination of image and
audio. It consists of a set of still images called frames displayed
to the user one after another at a specific speed, known as the
frame rate measured in number of frames per second (fps).
TMH Chapter - 8 2
Video
Introduction
TMH Chapter - 8 3
Video
Introduction
Motion picture is recorded on celluloid film and displayed in cinema
theaters by projecting on a screen, whereas motion video is represented
in the form of electrical signals as an output from video cameras.
It was only after the advent of the Pentium-II processor coupled with
cost reduction of video compression hardware, that full screen digital
video finally became a reality.
TMH Chapter - 8 4
Video
Video Camera
Analog video cameras are used to record a succession of still
images and then convert the brightness and color information of
the images into electrical signals.
TMH Chapter - 8 5
Video
Video Camera
TMH Chapter - 8 6
Video
Video Camera
The electrons migrate towards a positive potential applied to the
lens side of the target. This positive potential is applied to a thin
layer of conductive but transparent material.
TMH Chapter - 8 7
Video
Video Camera
The charge pattern is sampled point-by-point by a moving beam of
electrons which originates in an electron gun in the tube.
The beam scans the charge pattern in the same way a raster is
produced in a monitor but approach the target at a very low velocity.
The beam deposits just enough carriers to neutralize the charge pattern
formed by the holes. Excess electrons are turned back towards the
source.
TMH Chapter - 8 8
Video
Video Camera
A color TV camera essentially consists of three camera tubes in which
each tube receives selectively filtered primary colors.
Light from the scene is processed by the objective lens system. The
image formed by the lens is split into three images by glass prisms.
Thus red, green and blue images are formed. These pass through color
filters which provide highly precise primary color images which are
converted into video signals by the camera tubes. This generates the
three color signals R, G and B.
TMH Chapter - 8 9
Video
Video Camera
TMH Chapter - 8 10
Video
Video Transmission
When the displaying monitor is at a far off distance from the
camera, for example in the case of TV transmission, technical
difficulties are encountered in transmitting these signals in the
original R, G, B format.
TMH Chapter - 8 11
Video
Video Transmission
Thirdly, for TV signals, the transmission scheme had to adapt to
the existing monochrome TV transmission set up i.e. the same
signals would need to produce a monochrome image in a B/W
TV set and a color image in a color TV set.
TMH Chapter - 8 12
Video
Video Transmission
According to the theory formulated by Helmholtz the light
sensitive cells are of two types – rods and cones.
The cones that are sensitive to color are broadly divided in three
different groups. One set of cones detect the presence of blue
color, the second set perceives red color and the third is
sensitive to the green shade.
TMH Chapter - 8 13
Video
Video Transmission
The sensitivity of the human eye is greatest for the green-yellow
range decreasing towards both the red and blue ends of the
spectrum. This is the human color perception curve.
TMH Chapter - 8 14
Video
Video Transmission
Based on the spectral response curve and extensive tests with a
large number of observers, the relative intensities of the primary
colors for color transmission e.g. for color television, has been
standardized.
TMH Chapter - 8 15
Video
Video Transmission
The luminance component, describes the variation of perceived
brightness by the HVS in different portions of the image without regard
to any color information e.g. an image with only the luminance
information would be a grayscale image similar to that seen on a
monochrome TV set. The luminance component is denoted by Y.
TMH Chapter - 8 16
Video
Video Transmission
TMH Chapter - 8 17
Video
Video Transmission
The RGB output signals from a video camera are transformed to
YC format using electronic circuitry before being transmitted.
TMH Chapter - 8 18
Video
Video Transmission
TMH Chapter - 8 19
Video
Video Transmission
To generate the YC signals from the RGB signals they have to be
defined quantitatively i.e. how Y and C are related to R, G and B.
TMH Chapter - 8 20
Video
Video Transmission
The C sub-components i.e. H and S, are quantitatively defined in
terms of color difference signals referred to as blue chrominance
(Cb) and red chrominance (Cr). These are defined as shown
below:
Cb = B – Y
Cr = R – Y
TMH Chapter - 8 22
Video
Video Transmission
The color difference signals are generated by inverting Y and
adding the inverted signal separately to R and B to obtain (R-Y)
or Cr and (B-Y) or Cb
TMH Chapter - 8 23
Video
Video Transmission
Conversion of RGB signals into YC format also has another
important advantage of utilizing less bandwidth through the use
of chroma subsampling.
TMH Chapter - 8 24
Video
Video Transmission
There are different schemes of chroma sub-sampling. In the
4:2:2 scheme for every 4 units (pixels) of Y information, only 2
units each of Cb and Cr will be transmitted.
TMH Chapter - 8 26
Video
Video Signal Formats
Component video
TMH Chapter - 8 28
Video
Video Signal Formats
S-video
TMH Chapter - 8 30
Video
TV Broadcasting Standards
NTSC
NTSC is short for National Television Systems Committee. It is a
broadcast standard used in USA and Japan which uses 525 horizontal
lines at 30 (29.97) frames / sec.
The fundamental difference between the SECAM system and the NTSC
and PAL system is that the latter transmit and receive two color signals
simultaneously while in the SECAM system only one of the two
difference signals is transmitted at a time. It also uses 625 horizontal
lines at 25 frames/sec.
Here the color difference signals are denoted by DR and DB and both
occupies 1.5 MHz each. They are given by the relations:
Y = 0.3R + 0.59G + 0.11B
DR = -1.9(R-Y) = -1.33R +1.121G + 0.209B
TMH DB = 1.5(B-Y) = -0.45R –Chapter
0.885G - 8 + 1.335B 33
Video
TV Broadcasting Standards
The pictures are only slightly more detailed than NTSC images
because the signal does not contain any new information.
TMH Chapter - 8 34
Video
TV Broadcasting Standards
The international body for television standards, the International
Telecommunications Union – Radiocommunications Branch (ITU-R)
formerly known as the Consultative Committee for International
Radiocommunications (CCIR), defined a standard for digitization of
video signals known as CCIR-601 Recommendations.
A sampling rate of 13.5 MHz for the luminance signal and 6.75
MHz for the chrominance signal, for 4:2:2 sub-sampled signal,
with 8-bits per sample, is selected for both NTSC and PAL
TMH Chapter - 8 36
Video
TV Broadcasting Standards
Common Intermediate Format (CIF) is a non-interlaced format. Its
luminance resolution has 360 × 288 pixels/frame at 30 frames/second
and the chrominance has half the luminance resolution in both
horizontal and vertical directions.
Since its line value 288 represents half the active lines in the PAL
television signal, and its picture rate 30 frames/second is the same as
the NTSC system, it is a common intermediate format for both PAL and
NTSC systems.
TMH Chapter - 8 38
Video
TV Broadcasting Standards
High Definition (HD) video is a new standard for digital video for
improving picture quality compared to the standard NTSC or PAL
formats. They require a high definition monitor or TV screen (HDTV) to
be viewed and have been defined as the ITU-R recommendations
formerly known as CCIR
There are two alternate formats one relating to the standard 4:3 aspect
ratio screens with 1440 × 1152 pixels (1080 visible), and the other to the
16:9 aspect ratio wide screens with 1920 × 1152 pixels (1080 visible).
Both use either 4:2:2 sub-sampling scheme for studio applications with
50/60 Hz frame refresh rate, or 4:2:0 scheme for broadcast applications
with 25/30 Hz refresh rate.
TMH Chapter - 8 39
Video
TV Broadcasting Standards
Additionally there are three other standards based on MPEG-2 coding
The first standard called Advanced Television (ATV) in North America,
formulated by a large number of TV manufacturers known as the Grand
Alliance.
TMH Chapter - 8 41
Video
Digital Video
Displaying video on a PC monitor requires powerful video
adapter card, system processor and a high speed bus
connecting the two.
For a screen size of 640 × 480, 24-bit color and a frame rate of
25 frames per second, the bit rate necessary to display smooth
motion without jerking and stoppages amounts to 22500 KB/sec.
TMH Chapter - 8 42
Video
Digital Video
Displaying full screen video only became a reality after the
advent of the Pentium-II processor together with fast disks
capable of delivering the required output.
TMH Chapter - 8 43
Video
Digital Video Camera
Light passing through the lens of the camera is focused on a chip called
CCD.
TMH Chapter - 8 44
Video
Digital Video Camera
The DSP adjusts the contrast and brightness of the image, and
compresses the data before sending it to the camera’s storage
medium.
TMH Chapter - 8 45
Video
Digital Video Camera
TMH Chapter - 8 46
Video
Digitization
Two main components : Source and Source device Capture
device
Source device must use PAL or NTSC playback and must have
Composite video or S-video output ports.
TMH Chapter - 8 47
Video
Digitization
The source and source device can be the following :
TMH Chapter - 8 48
Video
Digitization
TMH Chapter - 8 49
Video
Digitization
A video capture card is an expansion board that can handle a variety of
different audio and video input signals and convert them from analog to
digital or vice versa.
TMH Chapter - 8 50
Video
Digitization
(2) Video compression-decompression hardware for video data.
(3) Audio compression-decompression hardware for audio data.
(4)A/D converter to convert the analog input video signals to digital form.
(5)Video OUTPUT port to feed output video signals to camera and VCR.
(6) D/A converter to convert the digital video data to analog signals for
feeding to output analog devices.
(7) Audio INPUT/OUTPUT ports for audio input and output functions.
TMH Chapter - 8 51
Video
Digitization
The main components of a video capture card also called video frame
grabber include the following :
ADC - The analog to digital converter (ADC) reads the input analog
video signal, from an analog video camera or VCP, and digitizes it using
standard procedures of sampling and quantization. The parameters for
digitization include the sampling rate for the visual and audio portions,
the color depth and the frame rate.
TMH Chapter - 8 52
Video
Digitization
Image processing parameters - include specifying the brightness,
contrast, color, audio volume etc. which are specified using the video
capture software.
These parameters are changed using a lookup table which converts the
value of an input pixel or audio sample in a pre-defined way and writes it
to an appropriate frame buffer of the capture card before being written
as a file on the hard disk.
CODEC - The video capture card often contains a chip for hardware
compression and decompression of video data in real time. There can
be multiple standards like MPEG-1, MPEG-2, H.261/263, which would
require a programmable CODEC on the card.
TMH Chapter - 8 53
Video
Digitization
TMH Chapter - 8 54
Video
Digitization
The following capabilities might be provided by a video capture
software, often bundled with a capture card.
TMH Chapter - 8 55
Video
Digitization
MPEG Capture - Certain cards allow the user to capture video directly in
the MPEG format. Here analog video data is captured, digitized and
compressed at the same time before being written to the disk. This is
suitable for capturing large volumes of video data.
TMH Chapter - 8 56
Video
File Formats
AVI - The native file format for Windows platform. Within a single file
video and audio data are kept in small chunks and arranged close
together on the disk.
TMH Chapter - 8 57
Video
File Formats
Quicktime - Developed by Apple for both Mac and Windows
platforms.
TMH Chapter - 8 58
Video
File Formats
MPEG - Lossy compression format based on intra-frame and inter-
frame redundancy.
H.263 was an improvement over H.261 supporting very low bit rates
below 64 Kbps.
TMH Chapter - 8 60
Video
File Formats
Real Video - The RV file format was developed by Real
Networks for playing video files from web pages.
TMH Chapter - 8 61
Video
File Formats
Indeo Video Interactive - This CODEC by Intel is used for video
distributed over the Internet for computers with MMX or Pentium II
processors. This CODEC includes features like flexible keyframe
control, chroma keying, on-the-fly cropping that reduces data load. Also
this CODEC employs a progressive download feature that adapts to
different network bandwidths.
TMH Chapter - 8 62
Video
File Formats
Sorenson Video - This CODEC supported by the Quicktime
architecture is useful for compressing 24-bit video intended for
CD-ROM applications and downloadable WWW video files.
Similar to Cinepak, the newer CODEC is designed for high
quality at data rates under 200 KBps.
TMH Chapter - 8 63
Video
File Formats
DivX - DivX is a video CODEC created by DivX Inc. (formerly
DivXNetworks Inc.) known for its ability to compress lengthy video
segments into small sizes. DivX 4 was released in July 2001. DivX, Inc.
has applied for patents on parts of the latest DivX codec, which is fully
MPEG-4-Advanced Simple Profile compliant.
TMH Chapter - 8 64
Video
Video Editing
Online editing is a practice of doing all editing (including the
rough cut) on the same computer that will produce the final cut.
TMH Chapter - 8 65
Video
Video Editing
SMPTE Timecode
Timecode defines how frames within a video are numbered.
When NTSC color system was developed the frame rate was changed
by a small amount from 30 to 29.97 to eliminate crosstalk between color
and audio information.
Because of this SMPTE time and real time (what the clock displays)
might be different over periods of time.
TMH Chapter - 8 66
Video
Video Editing
SMPTE Timecode
This has led to the development of two methods : Drop and Non-drop
However since the video actually plays at 29.97 frames per sec rather
than 30 frames per sec, SMPTE time will increment at a slower rate
than real world time.
TMH Chapter - 8 67
Video
Video Editing
SMPTE Timecode
Thus after a while the clock on the wall would be ahead than the
SMPTE time displayed in the video.
Since 29.97 frames are played back in 1 sec. of real time, playing 30
frames would actually take 1.001 sec of real time.
TMH Chapter - 8 68
Video
Video Editing
SMPTE Timecode
SMPTE Drop Method tries to compensate for this discrepancy by
accelerating the frame counting by “dropping frames”.
This means that occasionally SMPTE time will jump forward at a faster
speed to catch up with real time.
TMH Chapter - 8 69
Video
Video Editing
SMPTE Timecode
The time is adjusted by two frames on every minute boundary i.e. after
00:01:59:29 the next SMPTE Drop Frame time code should be
00:02:00:00 but is actually 00:02:00:02
Since frame numbers are increased by 2 every minute, in one hour the
total increase would be 120 frames.
TMH Chapter - 8 70
Video
Video Editing
SMPTE Timecode
Since the requirement is for 108 frames only, the incrementing is not
done on the following minute boundaries : 00, 10, 20, 30, 40, 50.
TMH Chapter - 8 71
Video
Video Editing
SMPTE Timecode
In SMPTE Drop Frame method no information is lost, since the method
does not actually drop frames, only frame numbers.
TMH Chapter - 8 72
Video
Video Editing
Timebase
Timebase specifies the time divisions that the editing software
uses to calculate the time position of an editing change.
For best results, the timebase should match the final frame rate
of the exported video.
TMH Chapter - 8 73
Video
Video Editing
Timebase
Suppose you are asked to combine motion-picture captured at 24 fps
with video source clips captured at 30 fps and export the final video for
an application supporting 15 fps.
TMH Chapter - 8 74
Video
Video Editing
Timebase
For the 30 fps clip, using a timebase of 30 displays all frames as
expected (half second shown below).
TMH Chapter - 8 75
Video
Video Editing
Timebase
For a 24 fps to be adjusted within a timebase of 30, frames 1, 5 and 9
need to be repeated.
Clips in an editing software with timebase of 30 would need to discard
every alternate frames for an exported frame rate of 15 fps.
TMH Chapter - 8 76
Video
Video Editing
Timebase
On the other hand, if timebase within the editing software is 24 and final
video is to be exported at 15 fps, then frames 3, 6, 8 and 11 need to be
discarded.
TMH Chapter - 8 77
Video
Video Editing
Edit Decision List (EDL)
An EDL is used in offline editing for recording the edit points.
It contains the names of the original clips, the In and Out points, and
other editing information.
TMH Chapter - 8 78
Video
Video Editing
Edit Decision List (EDL)
A standard EDL contains the following columns:
(a) Header – Contains title and type of timecode (drop-frame or
nondrop-frame) (b) Source Reel ID – Identifies the name or number of
the videotape containing the source clips
(c) Edit Mode – Indicates whether edits take place on video track, audio
track or both.
(d) Transition type – Describes the type of transition e.g. wipe, cut etc.
(e) Source In and Source Out – Lists timecodes of first and last frames
of clips
On a high-end system the EDL is accepted by an edit controller which
applies the editing changes to the high-quality clips.
TMH Chapter - 8 79