36
Unit IV: TEXT Types of text, text compression: Huffman coding, LZ and LZW, text file formats: TXT, DOC, RTF, PDF, PS. Video: Video signal formats, video transmission standards: EDTV, CCER, CIF, SIF, HDTV, Digitization of video, video recording systems: VHS, video compact cassette, DVCAN, Camcorder, laser disk, VCD, DVD-Video, Micro-MV, Video formats: MOV, Real video, H-261, H-263, Cinepack, Nerodigital, Video editing, DVD formats

Unit iv

Embed Size (px)

Citation preview

Unit IV: TEXT

Types of text, text compression: Huffman coding, LZ and LZW, text file formats: TXT, DOC, RTF, PDF, PS.

Video: Video signal formats, video transmission standards: EDTV, CCER, CIF, SIF, HDTV, Digitization of video, video recording systems: VHS, video compact cassette, DVCAN, Camcorder, laser disk, VCD, DVD-Video, Micro-MV,

Video formats: MOV, Real video, H-261, H-263, Cinepack, Nerodigital, Video editing, DVD formats

Types of text

There are 3 types of text that can be used to produce pages of a document : unformatted text, formatted text and hypertext.

Unformatted Text – Plaintext. Comprises of fixed sized characters from a limited character set (ASCII table). Characters – alphabetic, numeric and punctuation also called printable characters and control characters – backspace, linefeed, carriage return, space, delete, escape.

Formatted Text – changes the appearance of characters ie bold, underline, italics, varying shapes, sizes and colours.

Hypertext – sequential navigation, cross references which lead to multiple searches by hyperlinks

Architecture of Hypertext Document

There are 3 layers.

Presentation Layer : An HTML document displayed in the web browser.

Hypertext Abstract Machine : It is placed between the presentation and storage layers. For example interpreting the <a>

Storage Layer : the specific properties of the different discrete and continuous media need to be considered. Storage of HTML file along with graphic and audio files can be used to generate a display of a page containing the graphic and a background music.

Text Compression

Compression algorithms are used to reduce the size of the text file during storage. An algorithm must be applied to decompress the file before its contents can be displayed on the screen; without changing the text or character.

There are 2 types of text compression methods.

1. Huffman coding

2. LZ and LZW

Huffman Coding

This type of coding is intended for applications in which the text to be compressed has known characteristics in terms of the characters used and their relative frequencies of occurrences. Instead of using fixed length code words, an optimum set of variable length codewords is derived such that the shortest codeword is used to represent the most frequently occurring characters.

Method : Step 1: Analyze frequencies of the occurrences of each character.Step 2: Generate codewords.

Huffman Coding Example

A document contains letters A, B, C, D, E, F with frequencies as indicated :

A : 0.25B : 0.10C : 0.20D : 0.15E : 0.26F : 0.04

Solution : Step 1 – store the values in descending order of the frequency. The last 2 characters (leaf nodes) at the base are assigned values 1 and 0.

Huffman coding example

E : 0.26A : 0.25C : 0.20D : 0.15B : 0.10 (1) = 0.14F : 0.04 (0)

Step 2 : Replace the leaf nodes by a branch node whose weight is the sum of the weights of the leaf nodes. Form a new column to the right containing the new branch node combined with the remaining nodes from the first column arranged in their correct weight order.

Repeat step 1 to 2 till there are only 2 nodes remaining.

Huffman coding example...E : 0.26A : 0.25C : 0.20D : 0.15 (1) = 0.29 : 0.14 (0)

: 0.29E : 0.26A : 0.25 (1) = 0.45C : 0.20 (0)

: 0.45 : 0.29 (1)E : 0.26 (0)

: 0.55 (1) : 0.45 (0)

Huffman coding example

To derive the resulting codeword for each character, start from the first column to the left and proceed to the right, concatenating the values assigned to the character at each column.

A : (1) (0) ---> 01B : (1) (0) (1) (1) ---> 1101C : (0) (0)D : () () ()E : () ()F : () () () ()

Text File Formats

TXT : Unformatted text document created by an editor like Notepad on Windows platform. Encoding – ASCII, Unicode (UTF-8, UTF-16)

DOC : Stores documents created by the MS Word package. Not considered a document exchange format as it requires propriety software.

RTF : Developed for cross platform document exchanges. The control codes are human readable similar to HTML code.

PDF : Developed by Adobe Systems for cross platform exchange of documents. In addition to text, it also supports images and graphics. It is an open standard.

PS : PostScript is a page description language (to describe the contents of a page such that it can be accurately displayed on output devices usually a printer) used mainly for desktop publishing.

Video

Video Signal Formats :

1. Component Video – A video signal which is stored or transmitted as three separate component signals. The simplest form is the collection of R, G and B signals. Three separate wires and connectors are usually used to convey such signals from the camera to another device for storage or playback. In situations where the same set of signals are required to drive both a color monitor as well as a B/W monitor, RGB signals are replaced by Y, Cb and Cr signals, also delivered along three separate wires.

Video signal formats

2. Composite Video – Composite video is the most common type of video interface for sending or receiving an analog video signal to or from a television set. A composite video interface might connect a VHS tape player, DVD player or game console to a television.

Composite video is a yellow, RCA jack, normally found next to two audio jacks, one red, the other white. The three jacks together provide an interface for audiovisual connections. A video stream is composed of a Y signal for luminescence or black and white values and a C signal for chrominance or color. The Y signal provides brightness and contrast, allowing for deep rich blacks and startling bright whites.

Video Signal Format

3. S-Video – Super Video is an analog signal format where the luminance and chrominance portions are transmitted separately using multiple wires instead of the same wire as for composite video. The picture quality is better than that of composite video because of reduced interference but the cable is more expensive and is usually found in high end VCRs and capture cards.

1. Ground2. Ground3. Luminance (Y)4. Chrominance (C)

Chrominance

Chrominance (chroma for short) is the signal used in video systems to convey the color information of the picture, separately from the accompanying luma signal. Chrominance is usually represented as two color-difference components: U = B' − Y' (blue − luma) and V = R' − Y' (red − luma). Each of these difference components may have scale factors and offsets applied to it, as specified by the applicable video standard.

4. SCART Connector – (Syndicat des Constructeurs d'Appareils Radiorecepteurs et Televiseurs) is a French standard of a 21 pin audio and video connector. It can be used to connect VCRs, DVD players, set top boxes, game systems and computers to television sets. SCART attempts to provide a standardized connector containing all the signals for audio video applications.

NTSC

National Television System Committee (NTSC) is a television broadcasting system used in a number of countries. It specifies a standard using:- 525 horizontal lines, - 30 frames/sec, - an aspect ratio (image width : image height) of 4:3, - 2 interlaced fields /frame (also called 2:1 interlacing), - 262.5 lines per field, - 60 fields per second and - frequency modulation for the audio signal.

In video, a field is one of the many still images which are displayed sequentially to create the impression of motion on the screen. Two fields comprise one video frame.

NTSCColour information was added to the black and white image by adding a

colour sub carrier to the video signal.

Of the 525 horizontal lines, only 480 are actively used for generating the picture on the screen, the rest being used for synchronization and vertical retrace.

In colour transmission the luminance signal takes the place of the original monochrome signal used for B/W TVs. The chrominance part of the signal is represented by two sub components known as I(in-phase) and Q(Quadrature).

A television channel occupies 6 MHz of bandwidth; majority of which (0-4 MHz) is occupied by the luminance signal Y (brightness information), 1.24 MHz for video carrier frequency, 1.5 MHz for colour information.

NTSC has some colour problems because of which professional also call it “Never the same colour”

PAL

Phase Alternation (or Alternating) Lines (PAL) is a TV broadcasting standard used in a number of countries. The PAL system uses 625 horizontal lines (gives better quality picture) at 25 frames / sec, 2:1 interlacing with 2 fields per frame, 4:3 aspect ratio and a chroma sub sampling scheme of 4:2:2.

PAL also uses a colour sub carrier signal with 2 amplitude modulated Q components, however its location is different compared to NTSC.

PAL is also called “Perfect At Last”. Out of the 625 lines, only 576 lines are actively used to display the image on the screen, the remaining lines being used for synchronization and vertical retrace.

video transmission standards

1. EDTV – Enhanced Definition Television SystemsThey are conventional systems modified to offer improved vertical and horizontal resolutions. An emerging system is known as Improved Definition Television (IDTV) which improves NTSC (National Television Systems Committee) image. By using digital memory to double the scanning lines from 525 to 1050. The pictures are only slightly more detailed than NTSC images because the signal does not contain any new information. By separating the chrominance and luminance parts of the video signal, IDTV prevents cross interference between the two.

video transmission standards

2. CCIR – (ITU-R) RecommendationsThe International Telecommunications Union –

Radiocommunications Branch (ITU-R) formerly known as the Consultative Committee for International Radiocommunications (CCIR), defined a standard for digitization of video signals known as CCIR 601 Recommendations. A colour signal has 3 components – a luminance component and two chrominance components. The CCIR format has 2 options : one for NTSC TV and another for PAL TV. (Both of them being interlaced formats)

CCIR ...

Each TV channel occupies a maximum bandwidth of 6 MHz. To digitize the entire information, a sampling rate of 12 MHz needs to be chosen a per Nyquist's postulate, though a higher rate of 13.5 MHz has been proposed by CCIR in order to take care of the imperfections of the filter due to which some high frequency component always percolates through the filter.

The frame size for digital NTSC is fixed at 720 by 480 pixels whereas the size for digital PAL is fixed at 720 by 576 pixels (based on specifications).

CCIR proposes using two chroma sub sampling schemas – 4:2:0 and 4:2:2

Video trans. std....3. CIF – (common intermediate format) it is non interlaced format.

Its luminance resolution has 360 x 288 pixels / frame at 30 frames / sec and the chrominance has half the luminance resolution in both horizontal and vertical directions. It is a common intermediate format for both PAL and NTSC systems. In the NTSC system only a conversion of line number is needed while in the PAL system only a picture rate conversion is required.

4. SIF – Source Input Format has luminance resolution of 360 x 240 pixels / frame at 30 frames / sec or 360 x 288 pixels / frame at 25 frames / sec. It is usually used for video conferencing. For both cases, the resolution of the chrominance components is half of the luminance resolution in both horizontal and vertical dimensions.

5. HDTV – High definition TV is a new standard for digital video for improving picture quality compared to the standard NTSC or PAL formats. They require a high definition monitor or TV screen to be viewed and have been defined as the ITU-R recommendations. There are 2 alternate formats one relating to the standard 4:3 aspect ratio screen with 1440 x 1152 (1080 visible) pixels and the other to the 16:9 aspect ratio wide screen with 1920 x 1152 pixels (1080 visible).

Both use either 4:2:2 sub-sampling scheme for studio applications with 50 / 60 Hz frame refresh rate or 4:2:0 scheme for broadcast applications with 25/30 Hz refresh rate.

Video trans. std....

Analog video has been used for years in recording / editing studios and television broadcasting which needs to be converted to digital format for incorporating it in multimedia production.

Digitizing reduces the file size due to compression.

Digitizing video in general requires a video capture card and associated recording software. The capture card is usually installed at the PC end, which accepts analog video from a source device and converts it into a digital file using a recording software. Alternatively the capture card can be inside a digital video camera which is capable of producing a digital video output and recording it onto a tape.

Digitization of video

video recording systems

VHS – The Video Home System is a recording and playing standard for Video Cassette Recorders. Early VHS machines could rewind and fast forward the tape considerably faster. It contains a 12.7 mm (½ inch) wide magnetic tape

wound between two spools allowing it to be slowly passed over the various playback and recording heads of the video cassette recorder.

Super VHS and VHS C (compact) and S-VHS-C and D-VHS were some of the improved versions of VHS.

video recording systems

video compact cassette – VCC was the first practical home video cassette recorder system. The system used ½ inch tapes coated with chromium dioxide available in three versions : 30, 45 and 60 minutes. To prevent cross talk between adjacent

video tracks it used an unrecorded guard band which is essentially a small space between tracks to prevent interference when reading data.

DVCAM – it is a variant of the Digital Video (DV) format which uses the same DV cassettes but uses greater track widths in SP mode (standard play), EP (extended play) is not supported. Greater track width makes the tape more robust.

video recording systems

Camcorder – it is a generic term for a portable device for recording of audio and video onto a storage device within it. The term is a combination of camera and recorder in one unit. Prior to the introduction of the camcorder, the video camera needed to be connected to a separate device the VCR

for recording video footage. Sony introduced its model known as handycam. Initially camcorders used the VHS or Betamax tapes and were quite bulky, but later with the introduction of 8 mm tapes, they reduced in Size. Now a days, digital camcorders use MiniDV or DVD for storage.

video recording systems

laser disk – it was the first commercial optical disc storage medium and was used primarily for the presentation of movies. Originally was called Reflective Optical Videodisc System. Video was stored on LD as an analog

signal, while audio could be stored in either analog or digital format and in a variety of surround sound formats. Like on a CD, the surface of the disc is an aluminium foil covered by pits and lands, but on a audio CD or DVD the pits and lands will signify binary codes, on an LD, the pits are created using frequency modulation of an analog signal.

video recording systems VCD – with a display resolution equal to 352 x 240 (NTSC) pixels

or 352 x 288 (PAL), a VCD holds about 74 minutes of compressed video. It is in MPEG-1 format; audio is encoded as MPEG layer 2; video is stored at 1150 kilobits per sec, audio at 224 kilobits per sec. Overall bit rate of VCD is approx equal to the bit rate of ordinary audio CD, the length of video that can be stored is similar to the length of audio in an audio CD.

The following folder structure is used on a VCD : MPEGAV (contains main movie), SEGMENT (contains optional still pictures), EXT (contains optional list of I frames), VCD (contains album and disc identification information), CDDA (contains optional audio files).

SVCD (Super VCD) is a format used for storing video on standard CDs. Picture and audio quality is higher than VCD but lower compared to DVD.

video recording systems

DVD-Video – DVD video was the first DVD format to be launched. It uses the MPEG2 data compression algorithm to encode the video data. It uses a display size of 720 x 480 pixels and instead of encoding RGB components uses a YcrCb representation. It may include one or more subtitle tracks in various languages including those made especially for the hearing impaired.

DVD has 4 complementary system designed to restrict the DVD user in various ways : Macrovision (analog anti copying mechanism), Content Scrambling System – CSS (discourages people from making perfect digital copies), region codes (denoting the areas of the world in which distribution and playback are intended), and disabled user options (UOP) – specifies whether the user may perform any operations like selecting menu, forwarding etc on remote control.

video recording systems

Micro-MV – it was a proprietary videotape format introduced in 2001 by Sony. This cassette is physically smaller than a Digital8 or DV cassette. In fact, MicroMV is the smallest videotape format — 70% smaller than MiniDV or about the size of two US quarter coins. Each cassette can hold up to 60 minutes of video.

Video formatsMOV – Quick Time Movie requires a program called Movie Player

for playback which is freely downloadable.

Real video – the RM file format was developed by Real Networks for playing video files from web pages. It supports straming which means that the video file starts playing even before they are fully downloaded from the Internet.

H-261 – developed by ITU-T for videotelephony and video conferencing applications in a ISDN environment.

H-263 – it was designed for very low bit rate applications like videotelephony, videoconferencing, interactive games, surveillance video.

Cinepack – its greatest strength is its extremely low CPU requirements.

Video formats ...

Nero digital – is a software product that uses MPEG 4 conforming compression technology with support for two MPEG 4 video coding algorithms and two forms of MPEG 4 audio.

Video editing

Online and Offline editing

SMPTE (society of motion pictures and television engineers) Time Code – time code defines how frames in a movie are counted and affects the way you view and edit a clip. Frames are counted differently when editing video for TV than when editing for motion picture film.

Time Base – when editing a video, several source clips may need to be imported to create the output clip. The source frame rates of these source clips determine how many frame rates are displayed per second within these clips.

Edit Decision List – when an offline edit is completed on a low end system, a table of scene sequences called an edit decision list is created.

DVD formatsDVD – video : requires a DVD drive.

DVD – Audio : describes a high quality audio storage format. They employ a robust copy prevention mechanism called Content Protection for Prerecorded Media (CPPM).

DVD – R : is composed of 2 0.6 mm thick polycarbonate discs bonded with an adhesive to each other. It is approved by DVD forum. It is supported by 90% DVD players.

DVD + R : not approved by DVD forum. It is a writable disc with 4.7 GB capacity.

DVD - RW : is a rewritable optical disc with capacity like DVD-R ie 4.7 GB. Main advantage is to erase and rewrite. It is approved by DVD Forum.

DVD formatsDVD + RW : is a rewritable optical disc with 4.7 GB capacity. It is

not approved by DVD Forum. They are commonly used for volatile data such as backups or collections of files.

DVD – RAM : they are used in computers as well as camcorders. It is more closely related to hard disk technology as it has concentric tracks instead of one long spiral track.

DVD+R DL (dual layer) : also called DVD+R9. It uses discs with two layers of writable polycarbonate thereby doubling the storage capactiy.

MiniDVD : it was developed for use in camcorders and holds 30 min of video.

DualDisc : is a combination of CD & DVD with one format on one side of the disc and the other on the opposite.