Multimedia Unit V

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 15

Multimedia Unit - 5

UNIT-V

Basics of Video- Anlog and Digital Video, How to use video on PC, introduction to graphics accelerator
cards, directX, introduction to ad/dv and IEE1394 cards, digitization of analog video to digital video,
interlacing and non interlacing, brief note on various video standards-NTSC, PAL, SECAM, HDTV.
Introduction to video capturing media and instrument, videodisk, DVCAM, camcorder. Introduction to
digital video compression techniques and various file formats= avi, mpeg, mov, real video.

Brief introduction to video editing and movie making tools- quick time, video for windows and adobe
premier.

Analog Video

When light reflected from an object passes through a video camera lens that light is converted into
electronic signals by a special sensor called charged couple device (CCD). Top-quality broadcast cameras
may have as many as three CCDs (one for each color of red, green, and blue) to enhance the resolution
of the camera. The output of the CCD is processed by the camera into a signal containing three channels
of color information and synchronization pulses (sync): There are several video standards for managing
CCD output, each dealing with the amount of separation between the components of the signal. The
more separation of the color information in the signal results in higher quality of the image (and the
more expensive the equipment).

If each channel of color information is transmitted as a separate signal on its own conductor, the signal
output is called RGB (red, green, and blue), which is the preferred method for higher quality and
professional video work. Output can also be split into two separate color channels which results in poor
quality of video.

Digital Video
Analog video has been used for years in recording / editing studios and television broadcasting. For the
purpose of incorporating video content in multimedia production video needs to be converted into the
digital format.

It has already been mentioned that processing digital video on personal computers was very difficult
initially, firstly because of the huge file sizes required, and secondly of the large bit rate and processing
power required. Full screen video only became a reality after advent of the Pentium-II processor
together with fast disks capable of delivering the required output. Even with these powerful resources
delivering video files was difficult until the reduction in prices of compression hardware and software.
Compression helped to reduce the size of video files to a great extent which required a lower bit-rate to
transfer them over communication buses. Nowadays video is rarely viewed in; the uncompressed form
unless there is specific reason for doing so, e.g. to maintain the high quality, as for medical analysis.

Digitizing video in general requires a, video capture card and associated recording software. The capture
card is usually installed at the PC end which accepts analog video from a source device and converts it
into a digital file using recording software. Alternatively the capture card can be inside a digital video
camera which is capable of producing a digital video output and recording it onto a tape. The digital
output from a digital video camera can also be fed to a PC after necessary format conversion.

Basics of Video

Of all the multimedia elements, video places the highest performance demand on your computer and its
memory and storage. Consider that a high-quality color still image on a computer screen could require
as much as a megabyte of storage memory. Multiply this by 30-the number of times per second that still
picture is replaced to provide the appearance of motion-and you would need 30 megabytes of storage
to play your video for one second, or 1.8 gigabytes of storage for a minute. Just moving these entire
pictures from computer memory to the screen at that rate would challenge the processing capability of
a super- computer. Multimedia technologies and re-search efforts today deal with compressing digital
video image data into manageable streams of information so that a massive amount of image can be
squeezed into a comparatively small data file that still delivers a good viewing experience on the
intended viewing platform during playback.

Carefully planned, well-executed video clips can make a dramatic difference in a multimedia project.

Using Video On PC

Analog video needs to be converted to the digital format before it can be displayed on a PC screen. The
procedure for conversion involves two types of devices- source devices and capture devices.

The source and source device can be one of the following:


camcorder with pre-recorded video tape

VCP with pre-recorded video tape

Video camera with live footage.

We need Video capture card to convert analog signal to digital signal along with video capture Software
such as AVI capture, AVI to MPEG Converter, MPEG capture, DAT to MPEG Converter or MPEG Editor.

DirectX

Microsoft changed the entire multimedia standards game with its DirectX standard in Windows 95.

The idea was that DirectX offered a load of commands, also known as APIs, which did things like "make a
sound on the left" or "draw a sphere in front". Games would then simply make DirectX calls and the
hardware manufacturers would have to ensure their sound and graphics card drivers understood them.

The audio portion of DirectX 1 and 2 was called DirectSound, and this offered basic stereo left and right
panning effects. As with other DirectX components, this enabled software developers to write directly to
any DirectX-compatible sound card with multiple audio streams, while utilizing 3D audio effects. Each
audio channel can be treated individually, supporting multiple sampling rates and the ability to add
software-based effects. DirectSound itself acts as a sound-mixing engine, using system RAM to hold the
different audio streams in play for the few milliseconds they must wait before being mixed and sent on
to the sound card. Under ideal conditions, DirectSound can mix and output the requested sounds in as
little as 20 milliseconds.

DirectX 3 introduced DirectSound3D (DS3D) which offered a range of commands to place a sound
anywhere in 3D space. This was known as positional audio, and required significant processing power.
Sadly we had to wait for DirectX 5 before Microsoft allowed DS3D to be accelerated by third-party
hardware, reducing the stress on the main system CPU. DirectX 6 supported DirectMusic, offering
increased versatility in composing music for games and other applications.

DS3D positional audio is one of the features supported by the latest generation of PCI sound cards.
Simply put, positional audio manipulates the characteristics of sounds to make them seem to come from
a specific direction, such as from behind or from far to the left. DirectSound3D gives game developers a
set of API commands they can use to position audio elements. Furthermore, as with much of DirectX,
DirectSound3D is scaleable: if an application asks for positional effects and no hardware support for such
effects is found, then DirectSound3D will provide the necessary software to offer the positional effect,
using the CPU for processing.

DS3D may have supported positional audio, but it didn't offer much support for adding reverb, let alone
considering individual reflections, to simulate different environments. Fortunately DS3D does support
extensions to the API, and this need was soon met by a couple of new sound standards which have
gained widespread support from games developers: Aureal's A3D technology and Creative Technology's
Environmental Audio Extensions (EAX).
Broadcast video standards

Three analog broadcast video standards are commonly in use around the world: NTSC, PAL, and SECAM.
In the United States, the NTSC standard is being phased out, replaced by ATSC digital television
standard. Because these standards and formats are not easily interchangeable, it is important to know
where your multimedia project will be used. A video cassette recorded in USA which uses NTSC will not
play on a television set in any European country (which uses either PAL or SECAM), even thought the
recording method and style of the cassette is “VHS”. Likewise, tapes recorded in European PAL or
SECAM formats will not play back on an NTSC video cassette recorder. Each system is based on a
different standard that defines the way information is encoded to produce the electronics signal that
ultimately creates a television picture. Multi-format VCRs can play back all three standards but typically
cannot dub from one standard to another; dubbing between standards still require high-end specialized
equipment.

National Television Standard Committee (NTSC)

The United States, Canada, Mexico, Japan, and many other countries use a system for broadcasting and
displaying video that is based upon the specifications set forth by the 1952 National Television
Standards Committee. These standards define a method for encoding information into the electronic
signal that ultimately creates a television picture. As specified by the NTSC standard, a single frame of
video is made up of 525 horizontal scan lines drawn onto the inside face of a phosphor-coated picture
tube every 30th of a second by a fast-moving electron beam. The drawing occurs so fast that your eye
perceives the image as stable. The electron beam actually makes two passes as it draws a single video
frame, first laying down all the odd-numbered lines, then all the even-numbered lines. Each of these
passes (which happen at a rate of 60 per second, or 60 Hz) paints a field, and the two fields are
combined to create a single frame at a rate of 30 frames per second (fps). (Technically, the speed is
actually 29.97 Hz.) This process of building a single frame from two fields is called interlacing, a
technique that helps to prevent flicker on television screens. Computer monitors use a different
progressive-scan technology, and draw the lines of an entire frame in a single pass, without interlacing
them and without flicker.

Phase Alternate Line (PAL)

The Phase Alternate Line (PAL) system is used in the United Kingdom, Western Europe, Australia, South
Africa, China, and South America. PAL increases the screen resolution to 625 horizontal lines, but slows
the scan rate to 25 frames per second. As with NTSC, the even and odd lines are interlaced, each field
taking 1/50th of a second to draw (50 Hz).
Sequential Color and Memory (SECAM)

The Sequential Color and Memory is used in France, Eastern Europe, the former USSR, and a few other
countries. Although SECAM is a 625-line, 50 Hz system, it differs greatly from both the NTSC and the PAL
color systems in its basic technology and broadcast method. Often, however, TV sets sold in Europe
utilize dual components and can handle both PAL and SECAM systems.

Advanced Television Standard Committee (ATSC) and Digital Television (DTV)

What started as the High Definition Television (HDTV) initiative of the Federal Communications
Commission in the 1980s, changed first to the Advanced Television (ATV) initiative and then finished as
the Digital Television (DTV) initiative by the time the FCC announced the change in 1996. This standard,
slightly modified from the Digital Television Standard and Digital Audio Compression Standard, moves
U.S. television from an analog to digital standard and provides TV stations with sufficient bandwidth to
present four or five Standard Television signals (STV, providing the NTSC's resolution of 525 lines with a
3:4 aspect ratio, but in a digital signal) or one HDTV signal (providing 1,080 lines of resolution with a
movie screen's 16:9 aspect ratio). More significantly for multimedia producers, this emerging standard
allows for transmission of data to computers and for new ATV interactive services. As of May 2003,
1,587 TV stations in the United States (94 percent) had been granted a DTV construction permit or
license. Among those, 1,081 stations were actually broadcasting a DTV signal, almost all simultaneously-
casting their regular TV signal. According to the current schedule, all the stations are to cease
broadcasting on their analog channel and completely switch to a digital signal by 2006.

High Definition Television (HDTV)

It provides high resolution in 16: 9 aspect ratios. This aspect ratio allows the viewing of Cinemascope
and Panavision movies. There is contention between the broadcast and computer industries about
whether to use interlacing or progressive-scan technologies. The broadcast industry has promulgate an
ultra-high-resolution, 1920xl080 interlaced format to become the cornerstone of a new generation of
high-end entertainment centers, but the computer industry would like to settle on a 1280x720
progressive-scan system for HDTV. While the 1920xl080 format provides more pixel than the 1280x720
standard, the refresh rates are quite different. This higher-resolution interlaced format delivers only half
the picture every 1/60 of a second, and because of the interlacing, on highly detailed images there is a
great deal of screen flicker at 30 Hz. The computer people argue that the picture quality at 1280x720 is
superior and steady. Both formats have been included in the HDTV standard by the Advanced Television
System!

Today's multimedia monitors typically use a screen pixel ratio of 4:3 (800x600), but the new HDTV
standard specifies a ratio of 16:9 (1280x720), much wider than tall. There is no easy way to stretch and
shrink existing graphics material to this new aspect ratio, so new multimedia design and interface
principles will need to be developed for HDTV presentations.

Digitization of Analog Video to Digital Video

Video, like sound, is usually recorded and played as an analog signal. It must therefore be digitized in
order to be incorporated into a multimedia title. A video source, such as a video camera, VCR, TV, or
videodisc, is connected to a video capture card in a computer. As the video source is played, the analog
signal is sent to the video card and converted into a digital file that is stored on the hard drive. At the
same time, the sound from the video source is also digitized.

One of the advantages of digitized video is that it can be easily edited. Analog video, such as a videotape
is linear; there is a beginning, middle, and end. If you want to edit it, you need to continually rewind,
pause, and fast-forward the tape to display the desired frames. Digitized video, on the other hand,
allows random access to any part of the Video, and editing can be as easy as the cut-and-paste process
in a word processing program. In addition, adding special effects such as fly-in titles and transitions is
relatively simple.

Introduction to Video Compression Technique

Because of the large sizes associated with video files, video compression, decompression programs,
known as codecs, have been developed. These programs can substantially reduce the size of video files,
which means that more video can fit on a single CD and that the speed of transferring video from a CD
to the computer can be increased. There are two types of compression: loss less and lossy. Lossy
compression actually eliminates some of the data in the image and therefore provides greater
compression ratios than lossless compression. When the compression ratio is made high the quality of
decompressed image becomes poor. Thus, the trade-off is file size versus image quality. Lossy
compression is applied to video because some drop in the quality is not noticeable in moving images.

Certain standards have been established for compression programs, including JPEG (Joint Photographic
Experts Groups) and MPEG (Motion Picture Experts Group). Both of these programs reduce the file size
of graphic images by eliminating redundant information. Often areas of an image (especially
backgrounds) contain similar information. JPEG compression identifies these areas and stores them as
blocks of pixels instead of pixel by pixel, thus reducing the amount of, information needed to store the
image. Compression rations of 20: 1 can be achieved without substantially affecting image quality. A 20:
1 compression ratio would reduce a 1 MB file to only 50 KB.
MPEG adds another process to the still-image compression when working with video. MPEG looks for
the changes in the image from frame to frame. Key frames are identified every few frames, and the
changes that occur from key frame to key frame are recorded.

MPEG can provide greater compression ratios than JPEG, but it requires hardware (a card inserted in the
computer) that is not needed for JPEG compression. This limits the use of MPEG compression for
multimedia titles, because MPEG cards are not standard on the typical multimedia playback system.

Two widely used video compression software programs are Apple's QuickTime (and Quicklime for
Windows) and Microsoft's Video for Windows. Quicklime is popular because it runs on both, Apple and
Windows-based computers. It uses lossy compression coding and can achieve ratios of 5: 1 to 25: 1.
Video for Windows uses a format called Audio Video Interleave (AVI) which, like QuickTime,
synchronizes the sound and, motion of a video file. In this example, the unchanging background is stored
only once every 15 frames; only the moving spaceship is recorded frame by frame.

Video Formats

The AVI Format

The AVI (Audio Video Interleave) format was developed by Microsoft.The AVI format is supported by all
computers running Windows, and by all the most popular web browsers. It is a very common format on
the Internet, but not always possible to play on non-Windows computers. Videos stored in the AVI
format have the extension .avi.

The Windows Media Format

The Windows Media format is developed by Microsoft. Windows Media is a common format on the
Internet, but Windows Media movies cannot be played on non-Windows computer without an extra
(free) component installed. Some later Windows Media movies cannot play at all on non-Windows
computers because no player is available. Videos stored in the Windows Media format have the
extension .wmv.

The MPEG Format

The MPEG (Moving Pictures Expert Group) format is the most popular format on the Internet. It is cross-
platform, and supported by all the most popular web browsers. Videos stored in the MPEG format have
the extension .mpg or .mpeg.

The QuickTime Format

The QuickTime format is developed by Apple. QuickTime is a common format on the Internet, but
QuickTime movies cannot be played on a Windows computer without an extra (free) component
installed. Videos stored in the QuickTime format have the extension .mov.

The RealVideo Format


The RealVideo format was developed for the Internet by Real Media. The format allows streaming of
video (on-line video, Internet TV) with low bandwidths. Because of the low bandwidth priority, quality is
often reduced.

Videos stored in the RealVideo format have the extension .rm or .ram.

The Shockwave (Flash) Format

The Shockwave format was developed by Macromedia. The Shockwave format requires an extra
component to play. This component comes preinstalled with the latest versions of Netscape and
Internet Explorer.

Videos stored in the Shockwave format have the extension .swf.

Software for Capturing and Editing Videos

Several steps are needed to prepare video to be incorporated into a multi- media title. These include
capturing and digitizing the video from some video source, such as a video camera, VCR, TV, or
videodisc; editing the digitized video; and compressing the video. Some software programs specialize in
one or the other of these steps, and other programs, such as Adobe Premiere can perform all of them.
Although capturing and compressing are necessary, it is editing that receives the most attention. Editing
digitize video is similar to editing analog video, except that it is easier. For one this it is much quicker to
access frames in digital form than in analog. For example, with analog video, a lot of time is spent fast-
forwarding a rewinding the videotape to locate the desired frames; whereas with digital editing you can
quickly jump from the first frame to the last-or anywhere in between. Removing frames or moving them
to another location is as easy as the cut-and-paste process in a word processing program. The following
are some other features that may be included in editing software programs:

Incorporating transitions such as dissolves, wipes, and spins

Superimposing titles and animating them, such as a fly-in logo

Applying special effects to various images, such as twisting, zooming, rotating, and distorting

Synchronizing sound with the video

Applying filters that control color balance, brightness and contrast, blurring, distortions, and morphing
Introduction to Video Capture Media and Instruments:

Digital video cameras come in two different image capture formats: interlaced and progressive scan.
Interlaced cameras record the image in alternating sets of lines: the odd-numbered lines are scanned,
and then the even-numbered lines are scanned, then the odd-numbered lines are scanned again, and so
on. One set of odd or even lines is referred to as a "field", and a consecutive pairing of two fields of
opposite parity is called a frame.

A progressive scanning digital video camera records each frame as distinct, with both fields being
identical. Thus, interlaced video captures twice as many fields per second as progressive video does
when both operate at the same number of frames per second.

Progressive scan camcorders are generally more desirable because of the similarities they share with
film. They both record frames progressively, which results in a crisper image. They can both shoot at 24
frames per second, which results in motion strobing (blurring of the subject when fast movement
occurs). Thus, progressive scanning video cameras tend to be more expensive than their interlaced
counterparts

Standard film stocks such as 16 mm and 35 mm record at 24 frames per second. For video, there are two
frame rate standards: NTSC, and PAL, which shoot at 30/1.001 (about 29.97) frames per second and 25
frames per second, respectively.

Digital video can be copied with no degradation in quality. No matter how many generations a digital
source is copied, it will be as clear as the original first generation of digital footage.

Digital video can be processed and edited on an NLE, or non-linear editing station, a device built
exclusively to edit video and audio. These frequently can import from analog as well as digital sources,
but are not intended to do anything other than edit videos. Digital video can also be edited on a
personal computer which has the proper hardware and software. Using an NLE station, digital video can
be manipulated to follow an order, or sequence, of video clips.

Digital video is used outside of movie making. Digital television (including higher quality HDTV) started
to spread in most developed countries in early 2000s. Digital video is also used in modern mobile
phones and video conferencing systems. Digital video is also used for Internet distribution of media,
including streaming video.

Many types of video compression exist for serving digital video over the internet, and onto DVDs.
Although digital technique allows for a wide variety of edit effects, most common is the hard cut and an
editable video format like DV-video allows repeated cutting without loss of quality, because any
compression across frames is lossless. While DV video is not compressed beyond its own codec while
editing, the file sizes that result are not practical for delivery onto optical discs or over the internet, with
codecs such as the Windows Media format, MPEG2, MPEG4, Real Media, the more recent H.264, and
the Sorenson media codec. Probably the most widely used formats for delivering video over the internet
are MPEG4 and Windows Media, while MPEG2 is used almost exclusively for DVDs, providing an
exceptional image in minimal size but resulting in a high level of CPU consumption to decompress.

In analog systems, the video signal from the camera is delivered to the Video In connector(s) of a VCR,
where it is recorded on magnetic video tape. A camcorder combines both camera and tape recorder in a
single device. One or two channels of sound may also be recorded on the video tape (mono or stereo).
The video signal is written to tape by a spinning recording head that changes the local magnetic
properties of the tape's surface in a series of long diagonal stripes. Because the head is tilted at a slight
angle compared with the path of the tape, it follows a helical (spiral) path, which is called helical scan
recording. Each stripe represents information for one field of a video frame. A single video frame is
made up of two fields that are interlaced. Audio is recorded 0n a separate straight- line track at the top
of the videotape, although with some recording systems, sound is recorded helically between the video
tracks. At the bottom of the tape is a control track containing the pulses used to regulate speed.
Tracking is fine adjustment of the tape so that the tracks are properly aligned as the tape moves across
the playback head. This is how your VCR works.

In digital systems, the video signal from the camera is first digitized as a single frame, and the data is
compressed before it is written to the tape in one of several proprietary and competing formats: DV,
DVCPRO, or DVCAM.

Video-capture capability is not confined to camcorders. Cellphones, digital single lens reflex and
compact digicams, laptops, and personal media players frequently offer some form of video-capture
capability. In general, these multipurpose-devices offer less functionality for video-capture, than a
traditional camcorder. The absence of manual adjustments, external-audio input, and even basic
usability functions (such as autofocus and lens-zoom) are common limitations. More importantly, few
can capture to standard TV-video formats (480p60, 720p60, 1080i30), and instead record in either non-
TV resolutions (320x240, 640x480, etc.) or slower frame-rates (15fps, 30fps.)

Different type of storage media for Video

Some recent camcorders record video on flash memory devices, Microdrives, small hard disks, and size-
reduced DVD-RAM or DVD-Rs using MPEG-1, MPEG-2 or MPEG-4 formats. Most other digital consumer
camcorders record in DV or HDV format on tape and transfer content over FireWire.

Camcorders are often classified by their storage device: VHS, VHS-C, Betamax, Video8 are examples of
older, videotape-based camcorders which record video in analog form. Newer camcorders include
Digital8, MiniDV, DVD, Hard Disk and solid-state (flash) semiconductor memory, which all record video
in digital form. (Please see the digital video page for details.) In older digital camcorders, the imager-
chip, the CCD was considered an analog component, so the digital namesake is in reference to the
camcorder's processing and recording of the video. Many next generation camcorders use a CMOS
imager, which register photons as binary data as soon as the photons hit the imager
The IEEE 1394 interface

It is a serial bus interface standard for high-speed communications and isochronous real-time data
transfer, frequently used by personal computers, as well as in digital audio, digital video, automotive,
and aeronautics applications. The interface is also known by the brand names of FireWire (Apple), i.LINK
(Sony), and Lynx (Texas Instruments). IEEE 1394 replaced parallel SCSI in many applications, because of
lower implementation costs and a simplified, more adaptable cabling system.

IEEE 1394 was adopted as the High-Definition Audio-Video Network Alliance (HANA) standard
connection interface for A/V (audio/visual) component communication and control.FireWire is also
available in wireless, fiber optic, and coaxial versions using the isochronous protocols.

Nearly all digital camcorders have included a four-circuit 1394 interface, though, except for premium
models, such inclusion is becoming less common. It remains the primary transfer mechanism for high
end professional audio and video equipment. Since 2003 many computers intended for home or
professional audio/video use have built-in FireWire/i.LINK ports, especially prevalent with Sony and
Apple's computers. The legacy (alpha) 1394 port is also available on premium retail motherboards.

A Camcorder

A camcorder (video CAMera reCORDER) is an electronic device that combines a video camera and a
video recorder into one unit. Equipment manufacturers do not seem to have strict guidelines for the
term usage. Marketing materials may present a video recording device as a camcorder, but the delivery
package would identify content as video camera recorder.

In order to differentiate a camcorder from other devices that are capable of recording video, like cell
phones and compact digital cameras, a camcorder is generally identified as a portable device having
video capture and recording as its primary function.

The earliest camcorders employed analog recording onto videotape. Since the 1990s digital recording
has become the norm, but tape remained the primary recording media. Starting from early 2000s tape
as storage media is being gradually replaced with tapeless solutions like optical disks, hard disk drives
and flash memory.

All tape-based camcorders use removable media in form of video cassettes. Camcorders that do not use
magnetic tape are often called tapeless camcorders and may use optical discs (removable), solid-state
flash memory (removable or built-in) or a hard disk drive (removable or built-in).

Camcorders that permit using more than one type of media, like built-in hard disk drive and memory
card, are often called hybrid camcorders

Video cameras originally designed for television broadcast were large and heavy, mounted on special
pedestals, and wired to remote recorders located in separate rooms.
As technology advanced, out-of-studio video recording was made possible by means of compact video
cameras and portable video recorders. The recording unit could be detached from the camera and
carried to a shooting location. While the camera itself could be quite compact, the fact that a separate
recorder had to be carried along made on-location shooting a two-man job. Specialized video cassette
recorders were introduced by both JVC (VHS) and Sony (Umatic & Betamax) to be used for mobile work.

In 1982 Sony released the Betacam system. A part of this system was a single camera-recorder unit,
which eliminated the cable between camera and recorder and dramatically improved the freedom of a
cameraman. Betacam quickly became the standard for both news-gathering and in-studio video editing.

In 1983 Sony released the first consumer camcorder - the Betamovie BMC-100P. It used a Betamax
cassette and could not be held with one hand, so it was typically resting on a shoulder. In the same year
JVC released the first camcorder based on VHS-C format. In 1985 Sony came up with its own compact
video cassette format — Video8. Both formats had their benefits and drawbacks, and neither won the
format war.

In 1985, Panasonic, RCA, and Hitachi began producing camcorders that recorded to full-sized VHS
cassette and offered up to 3 hours of record time. These shoulders mount camcorders found use for
industrial videographers, and college TV studios. Super VHS full-sized camcorders were released in 1987
which exceeded broadcast quality and provided an inexpensive way to collect news segments or
videographies.

In 1986 Sony introduced the first digital video format, D1. Video was recorded in uncompressed form
and required enormous bandwidth for its time. In 1992 Ampex used D1 form-factor to create DCT, the
first digital video format that utilized data compression. The compression utilized discrete cosine
transform algorithm, which is used in most modern commercial digital video formats.

In 1995 Sony, JVC, Panasonic and other video camera manufacturers launched DV. Its variant using a
smaller MiniDV cassette quickly became a de-facto standard for home and semi-professional video
production, for independent filmmaking and for citizen journalism.

In 2000 Panasonic launched DVCPRO HD, expanding DV codec to support high definition. The format
was intended for use in professional camcorders and used full-size DVCPRO cassettes. In 2003 Sony, JVC,
Canon and Sharp introduced HDV, the first truly affordable high definition video format, which used
inexpensive MiniDV cassettes.

In 2003 Sony pioneered XDCAM, the first tapeless video format, which uses Professional Disc as
recording media. Panasonic followed next year, offering P2 solid state memory cards as recording
medium for DVCPRO HD video.

In 2006 Panasonic and Sony introduced AVCHD as an inexpensive consumer-grade tapeless high
definition video format. Presently AVCHD camcorders are manufactured by Sony, Panasonic, Canon, JVC
and Hitachi.
In 2007 Sony introduced XDCAM EX, which offers similar recording modes to XDCAM HD, but records on
SxS memory cards.

With proliferation of file-based digital formats the relationship between recording media and recording
format became weaker than ever: the same video can be recorded onto different media. With tapeless
formats, recording media has become a storage device for digital files, signifying convergence of video
and computer industries.

DVCAM

Sony's DVCAM is a professional variant of the DV standard that uses the same cassettes as DV and
MiniDV, but transports the tape 50% faster having 50% wider track, 15 micrometres instead of 10
micrometres. This variant uses the same codec as regular DV, however, the wider track lowers the
chances of dropout errors. The LP mode of consumer DV is not supported. All DVCAM recorders and
cameras can play back DV material, but DVCPRO support was only recently added to some models like
DSR-1800, DSR-2000, DSR-1600. DVCAM tapes (or DV tapes recorded in DVCAM mode) have their
recording time reduced by one third.

Because of the wider track, DVCAM has the ability to do a frame accurate insert tape edit. DV will vary
by a few frames on each edit compared to the preview. Another feature of DVCAM is locked audio. If
several generations of copies are made on DV, the audio sync may drift. On DVCam this does not happe

Video Capture Card

Video Capture Card is essentially an expansion board that can handle a variety of different audio and
video input signals and convert them into analog to digital or vice versa. Rendering support for a variety
of TV signal formats, e.g. NTSC, PAL, SECAM, imposes a level of complexity in the design together with
recently introduced standards for HDTV. A typical circuit board consists of the following components:

Video INPUT port to accept the video input signals from NTSC/PAL/SECAM broadcast signals, video
camera or VCR The input port may conform to the composite-video or S-video standards.

Video compression-decompression hardware for video data.

Audio compression-decompression hardware for audio data.

A/D converter to convert the analog input video signals to digital form.

What is DV?

As you can guess, DV stands for "Digital Video". It is the new high resolution digital video standard.
DV is compressed at the camera, on the tape itself. The camcorder has the DV "codec" built in.

The DV spec is a 720x480 image size with a 5:1 compression. DV video information is carried in a
nominal 25 megabit per second data stream. The color information is sampled at 4:1:1 for NTSC, and
4:2:0 for PAL.

Unlike MJPEG compressed video, DV video can't be scaled. You can't lower the screen size, change the
screen size or data rate.

DV format is typically reckoned to be equal to or slightly better than Betacam SP or MII in terms of
picture quality. Two types of DV camcorders, DVCAM and DVCPRO, are widely used in TV industry today.

However, for most of us, DV often refers to MiniDV actually. MiniDV is just the home level DV format. It
is compressed to a constant throughput of 3,600 kilobytes per second. The video quality is not as good
as Betacam, but much better than S-video.

What is FireWire?

Technically, it is the high speed, short distance data transfer protocol IEEE1394. Apple didn’t like the
numbers and so called it "FireWire". Sony didn’t like it either, and so they called it "iLink". And they are
all the same thing.

When the FireWire concept was first announced a few years ago, it was envisioned that it would
become a new standard that would replace SCSI and link all our consumer electronics equipment and
computers together. Now, the dust has settled and the hype has died down. The only application for
FireWire that has actually come to fruition is for transferring digital video (DV) information directly from
a camcorder (or VCR) to your hard drive.

What's the difference between DV and FireWire?

DV is the actual format of the video.

FireWire is the port and protocol that lets you transfer the DV data to your computer. The full FireWire
spec includes frame accurate device control and the ability to read and write the digital video.

When the video goes through the 1394 cable, into the capture card, and onto the hard drive, nothing is
done to the video. It is a digital copy. It's identical to the original. And this is really nice.

How's the quality of DV?

The DV (MiniDV) spec is a 720x480 image size, at roughly a 5:1 compression. More accurately, it is
compressed at a constant throughput of 3600 kilobytes per second which averages out to 5:1
compression.

The images are crisp, bright and have excellent depth and contrast. In general, it's acceptable even in TV
stations.
Best of all, the information is stored on the video tape in digital form, so it can be copied over and over
without any loss.

You might also like