Video

Last updated

A one-minute animated video showing an example of a media production process

Video is an electronic medium for the recording, copying, playback, broadcasting, and display of moving visual media. [1] Video was first developed for mechanical television systems, which were quickly replaced by cathode-ray tube (CRT) systems, which, in turn, were replaced by flat-panel displays of several types.

Contents

Video systems vary in display resolution, aspect ratio, refresh rate, color capabilities, and other qualities. Analog and digital variants exist and can be carried on a variety of media, including radio broadcasts, magnetic tape, optical discs, computer files, and network streaming.

Etymology

The word video comes from the Latin video (I see). [2]

History

Analog video

NTSC composite video signal (analog) NTSC Signal.png
NTSC composite video signal (analog)

Video developed from facsimile systems developed in the mid-19th century. Early mechanical video scanners, such as the Nipkow disk, were patented as early as 1884, however, it took several decades before practical video systems could be developed, many decades after film. Film records using a sequence of miniature photographic images visible to the eye when the film is physically examined. Video, by contrast, encodes images electronically, turning the images into analog or digital electronic signals for transmission or recording. [3]

Video technology was first developed for mechanical television systems, which were quickly replaced by cathode-ray tube (CRT) television systems. Video was originally exclusively live technology. Live video cameras used an electron beam, which would scan a photoconductive plate with the desired image and produce a voltage signal proportional to the brightness in each part of the image. The signal could then be sent to televisions, where another beam would receive and display the image. [4] Charles Ginsburg led an Ampex research team to develop one of the first practical video tape recorders (VTR). In 1951, the first VTR captured live images from television cameras by writing the camera's electrical signal onto magnetic videotape.

Video recorders were sold for $50,000 in 1956, and videotapes cost US$300 per one-hour reel. [5] However, prices gradually dropped over the years; in 1971, Sony began selling videocassette recorder (VCR) decks and tapes into the consumer market. [6]

Digital video

Digital video is capable of higher quality and, eventually, a much lower cost than earlier analog technology. After the commercial introduction of the DVD in 1997 and later the Blu-ray Disc in 2006, sales of videotape and recording equipment plummeted. Advances in computer technology allow even inexpensive personal computers and smartphones to capture, store, edit, and transmit digital video, further reducing the cost of video production and allowing programmers and broadcasters to move to tapeless production. The advent of digital broadcasting and the subsequent digital television transition are in the process of relegating analog video to the status of a legacy technology in most parts of the world. The development of high-resolution video cameras with improved dynamic range and color gamuts, along with the introduction of high-dynamic-range digital intermediate data formats with improved color depth, has caused digital video technology to converge with film technology. Since 2013, the use of digital cameras in Hollywood has surpassed the use of film cameras. [7]

Characteristics of video streams

Number of frames per second

Frame rate , the number of still pictures per unit of time of video, ranges from six or eight frames per second (frame/s) for old mechanical cameras to 120 or more frames per second for new professional cameras. PAL standards (Europe, Asia, Australia, etc.) and SECAM (France, Russia, parts of Africa, etc.) specify 25 frame/s, while NTSC standards (United States, Canada, Japan, etc.) specify 29.97 frame/s. [8] Film is shot at a slower frame rate of 24 frames per second, which slightly complicates the process of transferring a cinematic motion picture to video. The minimum frame rate to achieve a comfortable illusion of a moving image is about sixteen frames per second. [9]

Interlaced vs. progressive

Video can be interlaced or progressive. In progressive scan systems, each refresh period updates all scan lines in each frame in sequence. When displaying a natively progressive broadcast or recorded signal, the result is the optimum spatial resolution of both the stationary and moving parts of the image. Interlacing was invented as a way to reduce flicker in early mechanical and CRT video displays without increasing the number of complete frames per second. Interlacing retains detail while requiring lower bandwidth compared to progressive scanning. [10] [11]

In interlaced video, the horizontal scan lines of each complete frame are treated as if numbered consecutively and captured as two fields: an odd field (upper field) consisting of the odd-numbered lines and an even field (lower field) consisting of the even-numbered lines. Analog display devices reproduce each frame, effectively doubling the frame rate as far as perceptible overall flicker is concerned. When the image capture device acquires the fields one at a time, rather than dividing up a complete frame after it is captured, the frame rate for motion is effectively doubled as well, resulting in smoother, more lifelike reproduction of rapidly moving parts of the image when viewed on an interlaced CRT display. [10] [11]

NTSC, PAL, and SECAM are interlaced formats. Abbreviated video resolution specifications often include an i to indicate interlacing. For example, PAL video format is often described as 576i50, where 576 indicates the total number of horizontal scan lines, i indicates interlacing, and 50 indicates 50 fields (half-frames) per second. [11] [12]

When displaying a natively interlaced signal on a progressive scan device, the overall spatial resolution is degraded by simple line doubling—artifacts, such as flickering or "comb" effects in moving parts of the image that appear unless special signal processing eliminates them. [10] [13] A procedure known as deinterlacing can optimize the display of an interlaced video signal from an analog, DVD, or satellite source on a progressive scan device such as an LCD television, digital video projector, or plasma panel. Deinterlacing cannot, however, produce video quality that is equivalent to true progressive scan source material. [11] [12] [13]

Aspect ratio

Comparison of common cinematography and traditional television (green) aspect ratios Aspect ratios.svg
Comparison of common cinematography and traditional television (green) aspect ratios

Aspect ratio describes the proportional relationship between the width and height of video screens and video picture elements. All popular video formats are rectangular, and this can be described by a ratio between width and height. The ratio of width to height for a traditional television screen is 4:3, or about 1.33:1. High-definition televisions use an aspect ratio of 16:9, or about 1.78:1. The aspect ratio of a full 35 mm film frame with soundtrack (also known as the Academy ratio) is 1.375:1. [14] [15]

Pixels on computer monitors are usually square, but pixels used in digital video often have non-square aspect ratios, such as those used in the PAL and NTSC variants of the CCIR 601 digital video standard and the corresponding anamorphic widescreen formats. The 720 by 480 pixel raster uses thin pixels on a 4:3 aspect ratio display and fat pixels on a 16:9 display. [14] [15]

The popularity of viewing video on mobile phones has led to the growth of vertical video. Mary Meeker, a partner at Silicon Valley venture capital firm Kleiner Perkins Caufield & Byers, highlighted the growth of vertical video viewing in her 2015 Internet Trends Report growing from 5% of video viewing in 2010 to 29% in 2015. Vertical video ads like Snapchat's are watched in their entirety nine times more frequently than landscape video ads. [16]

Color model and depth

Example of U-V color plane, Y value=0.5 YUV UV plane.svg
Example of U-V color plane, Y value=0.5

The color model uses the video color representation and maps encoded color values to visible colors reproduced by the system. There are several such representations in common use: typically, YIQ is used in NTSC television, YUV is used in PAL television, YDbDr is used by SECAM television, and YCbCr is used for digital video. [17] [18]

The number of distinct colors a pixel can represent depends on the color depth expressed in the number of bits per pixel. A common way to reduce the amount of data required in digital video is by chroma subsampling (e.g., 4:4:4, 4:2:2, etc.). Because the human eye is less sensitive to details in color than brightness, the luminance data for all pixels is maintained, while the chrominance data is averaged for a number of pixels in a block, and the same value is used for all of them. For example, this results in a 50% reduction in chrominance data using 2-pixel blocks (4:2:2) or 75% using 4-pixel blocks (4:2:0). This process does not reduce the number of possible color values that can be displayed, but it reduces the number of distinct points at which the color changes. [12] [17] [18]

Video quality

Video quality can be measured with formal metrics like peak signal-to-noise ratio (PSNR) or through subjective video quality assessment using expert observation. Many subjective video quality methods are described in the ITU-T recommendation BT.500. One of the standardized methods is the Double Stimulus Impairment Scale (DSIS). In DSIS, each expert views an unimpaired reference video, followed by an impaired version of the same video. The expert then rates the impaired video using a scale ranging from "impairments are imperceptible" to "impairments are very annoying."

Video compression method (digital only)

Uncompressed video delivers maximum quality, but at a very high data rate. A variety of methods are used to compress video streams, with the most effective ones using a group of pictures (GOP) to reduce spatial and temporal redundancy. Broadly speaking, spatial redundancy is reduced by registering differences between parts of a single frame; this task is known as intraframe compression and is closely related to image compression. Likewise, temporal redundancy can be reduced by registering differences between frames; this task is known as interframe compression, including motion compensation and other techniques. The most common modern compression standards are MPEG-2, used for DVD, Blu-ray, and satellite television, and MPEG-4, used for AVCHD, mobile phones (3GP), and the Internet. [19] [20]

Stereoscopic

Stereoscopic video for 3D film and other applications can be displayed using several different methods: [21] [22]

Formats

Different layers of video transmission and storage each provide their own set of formats to choose from.

For transmission, there is a physical connector and signal protocol (see List of video connectors). A given physical link can carry certain display standards that specify a particular refresh rate, display resolution, and color space.

Many analog and digital recording formats are in use, and digital video clips can also be stored on a computer file system as files, which have their own formats. In addition to the physical format used by the data storage device or transmission medium, the stream of ones and zeros that is sent must be in a particular digital video coding format, for which a number is available.

Analog video

Analog video is a video signal represented by one or more analog signals. Analog color video signals include luminance (Y) and chrominance (C). When combined into one channel, as is the case among others with NTSC, PAL, and SECAM, it is called composite video. Analog video may be carried in separate channels, as in two-channel S-Video (YC) and multi-channel component video formats.

Analog video is used in both consumer and professional television production applications.

Digital video

Digital video signal formats have been adopted, including serial digital interface (SDI), Digital Visual Interface (DVI), High-Definition Multimedia Interface (HDMI) and DisplayPort Interface.

Transport medium

Video can be transmitted or transported in a variety of ways including wireless terrestrial television as an analog or digital signal, coaxial cable in a closed-circuit system as an analog signal. Broadcast or studio cameras use a single or dual coaxial cable system using serial digital interface (SDI). See List of video connectors for information about physical connectors and related signal standards.

Video may be transported over networks and other shared digital communications links using, for instance, MPEG transport stream, SMPTE 2022 and SMPTE 2110.

Display standards

Digital television

Digital television broadcasts use the MPEG-2 and other video coding formats and include:

Analog television

Analog television broadcast standards include:

An analog video format consists of more information than the visible content of the frame. Preceding and following the image are lines and pixels containing metadata and synchronization information. This surrounding margin is known as a blanking interval or blanking region; the horizontal and vertical front porch and back porch are the building blocks of the blanking interval.

Computer displays

Computer display standards specify a combination of aspect ratio, display size, display resolution, color depth, and refresh rate. A list of common resolutions is available.

Recording

A VHS video cassette tape VHS-Video-Tape-Top-Flat.jpg
A VHS video cassette tape

Early television was almost exclusively a live medium, with some programs recorded to film for historical purposes using Kinescope. The analog video tape recorder was commercially introduced in 1951. The following list is in rough chronological order. All formats listed were sold to and used by broadcasters, video producers, or consumers; or were important historically. [23] [24]

Digital video tape recorders offered improved quality compared to analog recorders. [24] [26]

Optical storage mediums offered an alternative, especially in consumer applications, to bulky tape formats. [23] [27]

Digital encoding formats

A video codec is software or hardware that compresses and decompresses digital video. In the context of video compression, codec is a portmanteau of encoder and decoder, while a device that only compresses is typically called an encoder , and one that only decompresses is a decoder. The compressed data format usually conforms to a standard video coding format. The compression is typically lossy, meaning that the compressed video lacks some information present in the original video. A consequence of this is that decompressed video has lower quality than the original, uncompressed video because there is insufficient information to accurately reconstruct the original video. [28]

See also

General
Video format
Video usage
Video screen recording software

Related Research Articles

<span class="mw-page-title-main">Digital video</span> Digital electronic representation of moving visual images

Digital video is an electronic representation of moving visual images (video) in the form of encoded digital data. This is in contrast to analog video, which represents moving visual images in the form of analog signals. Digital video comprises a series of digital images displayed in rapid succession, usually at 24, 25, 30, or 60 frames per second. Digital video has many advantages such as easy copying, multicasting, sharing and storage.

<span class="mw-page-title-main">NTSC</span> Analog television system

NTSC is the first American standard for analog television, published and adopted in 1941. In 1961, it was assigned the designation System M. It is also known as EIA standard 170.

<span class="mw-page-title-main">Interlaced video</span> Technique for doubling the perceived frame rate of a video display

Interlaced video is a technique for doubling the perceived frame rate of a video display without consuming extra bandwidth. The interlaced signal contains two fields of a video frame captured consecutively. This enhances motion perception to the viewer, and reduces flicker by taking advantage of the characteristics of the human visual system.

<span class="mw-page-title-main">Chroma subsampling</span> Practice of encoding color images

Chroma subsampling is the practice of encoding images by implementing less resolution for chroma information than for luma information, taking advantage of the human visual system's lower acuity for color differences than for luminance.

<span class="mw-page-title-main">Rec. 601</span> Standard from the International Telecommunication Union

ITU-R Recommendation BT.601, more commonly known by the abbreviations Rec. 601 or BT.601, is a standard originally issued in 1982 by the CCIR for encoding interlaced analog video signals in digital video form. It includes methods of encoding 525-line 60 Hz and 625-line 50 Hz signals, both with an active region covering 720 luminance samples and 360 chrominance samples per line. The color encoding system is known as YCbCr 4:2:2.

<span class="mw-page-title-main">ATSC standards</span> Standards for digital television in the US

Advanced Television Systems Committee (ATSC) standards are an international set of standards for broadcast and digital television transmission over terrestrial, cable and satellite networks. It is largely a replacement for the analog NTSC standard and, like that standard, is used mostly in the United States, Mexico, Canada, South Korea and Trinidad & Tobago. Several former NTSC users, such as Japan, have not used ATSC during their digital television transition, because they adopted other systems such as ISDB developed by Japan, and DVB developed in Europe, for example.

<span class="mw-page-title-main">Camcorder</span> Video camera with built-in video recorder

A camcorder is a self-contained portable electronic device with video and recording as its primary function. It is typically equipped with an articulating screen mounted on the left side, a belt to facilitate holding on the right side, hot-swappable battery facing towards the user, hot-swappable recording media, and an internally contained quiet optical zoom lens.

Broadcasttelevision systems are the encoding or formatting systems for the transmission and reception of terrestrial television signals.

<span class="mw-page-title-main">Common Intermediate Format</span>

CIF, also known as FCIF, is a standardized format for the picture resolution, frame rate, color space, and color subsampling of digital video sequences used in video teleconferencing systems. It was first defined in the H.261 standard in 1988.

<span class="mw-page-title-main">Display resolution</span> Width and height of a display in pixels

The display resolution or display modes of a digital television, computer monitor, or other display device is the number of distinct pixels in each dimension that can be displayed. It can be an ambiguous term especially as the displayed resolution is controlled by different factors in cathode-ray tube (CRT) displays, flat-panel displays and projection displays using fixed picture-element (pixel) arrays.

1080i is a term used in high-definition television (HDTV) and video display technology. It means a video mode with 1080 lines of vertical resolution. The "i" stands for interlaced scanning method. This format was once a standard in HDTV. It was particularly used for broadcast television. This is because it can deliver high-resolution images without needing excessive bandwidth. This format is used in the SMPTE 292M standard.

<span class="mw-page-title-main">480i</span> Standard-definition video mode

480i is the video mode used for standard-definition digital video in the Caribbean, Japan, South Korea, Taiwan, Philippines, Myanmar, Western Sahara, and most of the Americas. The other common standard definition digital standard, used in the rest of the world, is 576i.

Progressive segmented Frame is a scheme designed to acquire, store, modify, and distribute progressive scan video using interlaced equipment.

<span class="mw-page-title-main">PCM adaptor</span> Encodes digital audio as video

A PCM adaptor is a device that encodes digital audio as video for recording on a videocassette recorder. The adapter also has the ability to decode a video signal back to digital audio for playback. This digital audio system was used for mastering early compact discs.

Source Input Format (SIF) defined in MPEG-1, is a video format that was developed to allow the storage and transmission of digital video.

Analog high-definition television has referred to a variety of analog video broadcast television systems with various display resolutions throughout history.

MUSE, commercially known as Hi-Vision was a Japanese analog high-definition television system, with design efforts going back to 1979.

Television standards conversion is the process of changing a television transmission or recording from one video system to another. Converting video between different numbers of lines, frame rates, and color models in video pictures is a complex technical problem. However, the international exchange of television programming makes standards conversion necessary so that video may be viewed in another nation with a differing standard. Typically video is fed into video standards converter which produces a copy according to a different video standard. One of the most common conversions is between the NTSC and PAL standards.

Broadcast-safe video is a term used in the broadcast industry to define video and audio compliant with the technical or regulatory broadcast requirements of the target area or region the feed might be broadcasting to. In the United States, the Federal Communications Commission (FCC) is the regulatory authority; in most of Europe, standards are set by the European Broadcasting Union (EBU).

High-definition television (HDTV) describes a television or video system which provides a substantially higher image resolution than the previous generation of technologies. The term has been used since at least 1933; in more recent times, it refers to the generation following standard-definition television (SDTV). It is the standard video format used in most broadcasts: terrestrial broadcast television, cable television, satellite television.

References

  1. "Video – HiDef Audio and Video". hidefnj.com. Archived from the original on May 14, 2017. Retrieved March 30, 2017.
  2. "video", Online Etymology Dictionary
  3. Amidon, Audrey (June 25, 2013). "Film Preservation 101: What's the Difference Between a Film and a Video?". The Unwritten Record. US National Archives.
  4. "Vocademy - Learn for Free - Electronics Technology - Analog Circuits - Analog Television". vocademy.net. Retrieved 2024-06-29.
  5. Elen, Richard. "TV Technology 10. Roll VTR". Archived from the original on October 27, 2011.
  6. "Vintage Umatic VCR – Sony VO-1600. The worlds first VCR. 1971". Rewind Museum. Archived from the original on February 22, 2014. Retrieved February 21, 2014.
  7. Follows, Stephen (February 11, 2019). "The use of digital vs celluloid film on Hollywood movies". Archived from the original on April 11, 2022. Retrieved February 19, 2022.
  8. Soseman, Ned. "What's the difference between 59.94fps and 60fps?". Archived from the original on June 29, 2017. Retrieved July 12, 2017.
  9. Watson, Andrew B. (1986). "Temporal Sensitivity" (PDF). Sensory Processes and Perception. Archived from the original (PDF) on March 8, 2016.
  10. 1 2 3 Bovik, Alan C. (2005). Handbook of image and video processing (2nd ed.). Amsterdam: Elsevier Academic Press. pp. 14–21. ISBN   978-0-08-053361-2. OCLC   190789775. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  11. 1 2 3 4 Wright, Steve (2002). Digital compositing for film and video. Boston: Focal Press. ISBN   978-0-08-050436-0. OCLC   499054489. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  12. 1 2 3 Brown, Blain (2013). Cinematography: Theory and Practice: Image Making for Cinematographers and Directors. Taylor & Francis. pp. 159–166. ISBN   9781136047381.
  13. 1 2 Parker, Michael (2013). Digital Video Processing for Engineers : a Foundation for Embedded Systems Design. Suhel Dhanani. Amsterdam. ISBN   978-0-12-415761-3. OCLC   815408915. Archived from the original on August 25, 2022. Retrieved August 25, 2022.{{cite book}}: CS1 maint: location missing publisher (link)
  14. 1 2 Bing, Benny (2010). 3D and HD broadband video networking. Boston: Artech House. pp. 57–70. ISBN   978-1-60807-052-7. OCLC   672322796. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  15. 1 2 Stump, David (2022). Digital cinematography : fundamentals, tools, techniques, and workflows (2nd ed.). New York, NY: Routledge. pp. 125–139. ISBN   978-0-429-46885-8. OCLC   1233023513. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  16. Constine, Josh (May 27, 2015). "The Most Important Insights From Mary Meeker's 2015 Internet Trends Report". TechCrunch. Archived from the original on August 4, 2015. Retrieved August 6, 2015.
  17. 1 2 Li, Ze-Nian; Drew, Mark S.; Liu, Jiangchun (2021). Fundamentals of multimedia (3rd ed.). Cham, Switzerland: Springer. pp. 108–117. ISBN   978-3-030-62124-7. OCLC   1243420273. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  18. 1 2 Banerjee, Sreeparna (2019). "Video in Multimedia". Elements of multimedia. Boca Raton: CRC Press. ISBN   978-0-429-43320-7. OCLC   1098279086. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  19. Andy Beach (2008). Real World Video Compression. Peachpit Press. ISBN   978-0-13-208951-7. OCLC   1302274863. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  20. Sanz, Jorge L. C. (1996). Image Technology : Advances in Image Processing, Multimedia and Machine Vision. Berlin, Heidelberg: Springer Berlin Heidelberg. ISBN   978-3-642-58288-2. OCLC   840292528. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  21. Ekmekcioglu, Erhan; Fernando, Anil; Worrall, Stewart (2013). 3DTV : processing and transmission of 3D video signals. Chichester, West Sussex, United Kingdom: Wiley & Sons. ISBN   978-1-118-70573-5. OCLC   844775006. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  22. Block, Bruce A.; McNally, Phillip (2013). 3D storytelling : how stereoscopic 3D works and how to use it. Burlington, MA: Taylor & Francis. ISBN   978-1-136-03881-5. OCLC   858027807. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  23. 1 2 Tozer, E.P.J. (2013). Broadcast engineer's reference book (1st ed.). New York. pp. 470–476. ISBN   978-1-136-02417-7. OCLC   1300579454. Archived from the original on August 25, 2022. Retrieved August 25, 2022.{{cite book}}: CS1 maint: location missing publisher (link)
  24. 1 2 Pizzi, Skip; Jones, Graham (2014). A Broadcast Engineering Tutorial for Non-Engineers (4th ed.). Hoboken: Taylor and Francis. pp. 145–152. ISBN   978-1-317-90683-4. OCLC   879025861. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  25. 1 2 "Sony HD Formats Guide (2008)" (PDF). pro.sony.com. Archived (PDF) from the original on March 6, 2015. Retrieved November 16, 2014.
  26. Ward, Peter (2015). "Video Recording Formats". Multiskilling for television production. Alan Bermingham, Chris Wherry. New York: Focal Press. ISBN   978-0-08-051230-3. OCLC   958102392. Archived from the original on August 25, 2022. Retrieved August 25, 2022.
  27. Merskin, Debra L., ed. (2020). The Sage international encyclopedia of mass media and society. Thousand Oaks, California. ISBN   978-1-4833-7551-9. OCLC   1130315057. Archived from the original on June 3, 2020. Retrieved August 25, 2022.{{cite book}}: CS1 maint: location missing publisher (link)
  28. Ghanbari, Mohammed (2003). Standard Codecs: Image Compression to Advanced Video Coding. Institution of Engineering and Technology. pp. 1–12. ISBN   9780852967102. Archived from the original on August 8, 2019. Retrieved November 27, 2019.