Multiple sub-Nyquist sampling encoding

Last updated

MUSE (Multiple sub-Nyquist Sampling Encoding), [1] commercially known as Hi-Vision (a contraction of HIgh-definition teleVISION) [1] was a Japanese analog high-definition television system, with design efforts going back to 1979. [2]

Contents

It used dot-interlacing and digital video compression to deliver 1125 line, 60 field-per-second (1125i60) [2] signals to the home. The system was standardized as ITU-R recommendation BO.786 [3] and specified by SMPTE 260M, [4] using a colorimetry matrix specified by SMPTE 240M. [5] As with other analog systems, not all lines carry visible information. On MUSE there are 1035 active interlaced lines, therefore this system is sometimes also mentioned as 1035i. [6] It employed 2-dimensional filtering, dot-interlacing, motion-vector compensation and line-sequential color encoding with time compression to "fold" an original 20 MHz bandwidth source signal into just 8.1 MHz.

Japan began broadcasting wideband analog HDTV signals in December 1988, [7] initially with an aspect ratio of 2:1. The Sony HDVS high-definition video system was used to create content for the MUSE system. [2] By the time of its commercial launch in 1991, digital HDTV was already under development in the United States. Hi-Vision was mainly broadcast by NHK through their BShi satellite TV channel.

On May 20, 1994, Panasonic released the first MUSE LaserDisc player. [8] There were also a number of players available from other brands like Pioneer and Sony

Hi-Vision continued broadcasting in analog until 2007.

History

MUSE was developed by NHK Science & Technology Research Laboratories in the 1980s as a compression system for Hi-Vision HDTV signals.

Modulation research

Technical specifications

MUSE's "1125 lines" are an analog measurement, which includes non-video scan lines taking place while a CRT's electron beam returns to the top of the screen to begin scanning the next field. Only 1035 lines have picture information. Digital signals count only the lines (rows of pixels) that have actual detail, so NTSC's 525 lines become 486i (rounded to 480 to be MPEG compatible), PAL's 625 lines become 576i, and MUSE would be 1035i. To convert the bandwidth of Hi-Vision MUSE into "conventional" lines-of-horizontal resolution (as is used in the NTSC world), multiply 29.9 lines per MHz of bandwidth. (NTSC and PAL/SECAM are 79.9 lines per MHz) - this calculation of 29.9 lines works for all current HD systems including Blu-ray and HD-DVD. So, for MUSE, during a still picture, the lines of resolution would be: 598-lines of luminance resolution per-picture-height. The chroma resolution is: 209-lines. The horizontal luminance measurement approximately matches the vertical resolution of a 1080 interlaced image when the Kell factor and interlace factor are taken into account.

Key features of the MUSE system:

Colorimetry

The MUSE luminance signal encodes , specified as the following mix of the original RGB color channels: [3]

The chrominance signal encodes and difference signals. By using these three signals (, and ), a MUSE receiver can retrieve the original RGB color components using the following matrix: [3]

The system used a colorimetry matrix specified by SMPTE 240M [5] [15] [16] (with coefficients corresponding to the SMPTE RP 145 primaries, also known as SMPTE-C, in use at the time the standard was created). [17] The chromaticity of the primary colors and white point are: [16] [5]

MUSE colorimetry (SMPTE 240M / SMPTE "C")
PrimariesCIE 1931 coordinates
xy
Red0.6300.340
Green0.3100.595
Blue0.1550.070
White Point D65 0.31270.3290

The luma () function is specified as: [5]

The blue color difference () is amplitude-scaled (), according to: [5]

The red color difference () is amplitude-scaled (), according to: [5]

Signal and Transmission

MUSE is a 1125 line system (1035 visible), and is not pulse and sync compatible with the digital 1080 line system used by modern HDTV. Originally, it was a 1125 line, interlaced, 60 Hz, system with a 5/3 (1.66:1) aspect ratio and an optimal viewing distance of roughly 3.3H.

For terrestrial MUSE transmission a bandwidth limited FM system was devised. A satellite transmission system uses uncompressed FM.

The pre-compression bandwidth for is 20 MHz, and the pre-compression bandwidth for chrominance is a 7.425 MHz carrier.

The Japanese initially explored the idea of frequency modulation of a conventionally constructed composite signal. This would create a signal similar in structure to the composite video NTSC signal - with the ( luminance ) at the lower frequencies and the ( chrominance ) above. Approximately 3 kW of power would be required, in order to get 40 dB of signal to noise ratio for a composite FM signal in the 22 GHz band. This was incompatible with satellite broadcast techniques and bandwidth.

To overcome this limitation, it was decided to use a separate transmission of and . This reduces the effective frequency range and lowers the required power. Approximately 570 W (360 for and 210 for ) would be needed in order to get a 40 dB of signal to noise ratio for a separate FM signal in the 22 GHz satellite band. This was feasible.

There is one more power saving that appears from the character of the human eye. The lack of visual response to low frequency noise allows significant reduction in transponder power if the higher video frequencies are emphasized prior to modulation at the transmitter and then de-emphasized at the receiver. This method was adopted, with crossover frequencies for the emphasis/de-emphasis at 5.2 MHz for and 1.6 MHz for . With this in place, the power requirements drop to 260 W of power (190 for and 69 for ).

Sampling systems and ratios

The subsampling in a video system is usually expressed as a three part ratio. The three terms of the ratio are: the number of brightness (luma) samples, followed by the number of samples of the two color (chroma) components and , for each complete sample area. Traditionally the value for brightness is always 4, with the rest of the values scaled accordingly.

A sampling of 4:4:4 indicates that all three components are fully sampled. A sampling of 4:2:0, for example, indicated that the two chroma components are sampled at half the horizontal sample rate of luma - the horizontal chroma resolution is halved. This reduces the bandwidth of an uncompressed video signal by one-third.

MUSE implements a similar system as a means of reducing bandwidth, but instead of static sampling, the actual ratio varies according to the amount of motion on the screen. In practice, MUSE sampling will vary from approximately 4:2:1 to 4:0.5:0.25, depending on the amount of movement. Thus the red-green chroma component has between one-half and one-eighth the sampling resolution of the luma component , and the blue-yellow chroma has half the resolution of red-green.

Audio subsystem

MUSE had a discrete 2- or 4-channel digital audio system called "DANCE", which stood for Digital Audio Near-instantaneous Compression and Expansion.

It used differential audio transmission (differential pulse-code modulation) that was not psychoacoustics-based like MPEG-1 Layer II. It used a fixed transmission rate of 1350 kbp/s. Like the PAL NICAM stereo system, it used near-instantaneous companding (as opposed to Syllabic-companding like the dbx system uses) and non-linear 13-bit digital encoding at a 32 kHz sample rate.

It could also operate in a 48 kHz 16-bit mode. The DANCE system was well documented in numerous NHK technical papers and in a NHK-published book issued in the USA called Hi-Vision Technology. [18]

The DANCE audio codec was superseded by Dolby AC-3 (a.k.a. Dolby Digital), DTS Coherent Acoustics (a.k.a. DTS Zeta 6x20 or ARTEC), MPEG-1 Layer III (a.k.a. MP3), MPEG-2 Layer I, MPEG-4 AAC and many other audio coders. The methods of this codec are described in the IEEE paper: [19]

Real world performance issues

MUSE had a four-field dot-interlacing cycle, meaning it took four fields to complete a single MUSE frame. Thus, only stationary images were transmitted at full resolution. However, as MUSE lowers the horizontal and vertical resolution of material that varies greatly from frame to frame, moving images were blurred. Because MUSE used motion-compensation, whole camera pans maintained full resolution, but individual moving elements could be reduced to only a quarter of the full frame resolution. Because the mix between motion and non-motion was encoded on a pixel-by-pixel basis, it wasn't as visible as most would think. Later, NHK came up with backwards compatible methods of MUSE encoding/decoding that greatly increased resolution in moving areas of the image as well as increasing the chroma resolution during motion. This so-called MUSE-III system was used for broadcasts starting in 1995 and a very few of the last Hi-Vision MUSE LaserDiscs used it ( A River Runs Through It is one Hi-Vision LD that used it). During early demonstrations of the MUSE system, complaints were common about the decoder's large size, which led to the creation of a miniaturized decoder. [1]

Shadows and multipath still plague this analog frequency modulated transmission mode.

Japan has since switched to a digital HDTV system based on ISDB, but the original MUSE-based BS Satellite channel 9 (NHK BS Hi-vision) was broadcast until September 30, 2007.

Cultural and geopolitical impacts

Internal reasons inside Japan that led to the creation of Hi-Vision

MUSE, as the US public came to know it, was initially covered in the magazine Popular Science in the mid-1980s. The US television networks did not provide much coverage of MUSE until the late 1980s, as there were few public demonstrations of the system outside Japan.

Because Japan had its own domestic frequency allocation tables (that were more open to the deployment of MUSE) it became possible for this television system to be transmitted by Ku Band satellite technology by the end of the 1980s.

The US FCC in the late 1980s began to issue directives that would allow MUSE to be tested in the US, providing it could be fit into a 6 MHz System-M channel.

The Europeans (in the form of the European Broadcasting Union (EBU)) were impressed with MUSE, but could never adopt it because it is a 60 Hz TV system, not a 50 Hz system that is standard in Europe and the rest of the world (outside the Americas and Japan).

The EBU development and deployment of B-MAC, D-MAC and much later on HD-MAC were made possible by Hi-Vision's technical success. In many ways MAC transmission systems are better than MUSE because of the total separation of colour from brightness in the time domain within the MAC signal structure.

Like Hi-Vision, HD-MAC could not be transmitted in 8 MHz channels without substantial modification and a severe loss of quality and frame rate. A 6 MHz version Hi-Vision was experimented with in the US, [7] but it too had severe quality problems so the FCC never fully sanctioned its use as a domestic terrestrial television transmission standard.

The US ATSC working group that had led to the creation of NTSC in the 1950s was reactivated in the early 1990s because of Hi-Vision's success. Many aspects of the DVB standard are based on work done by the ATSC working group, however most of the impact is in support for 60 Hz (as well as 24 Hz for film transmission) and uniform sampling rates and interoperable screen sizes.

Device support for Hi-Vision

Hi-Vision LaserDiscs

On May 20, 1994, Panasonic released the first MUSE LaserDisc player. [8] There were a number of MUSE LaserDisc players available in Japan: Pioneer HLD-XØ, HLD-X9, HLD-1000, HLD-V500, HLD-V700; Sony HIL-1000, HIL-C1 and HIL-C2EX; the last two of which have OEM versions made by Panasonic, LX-HD10 and LX-HD20. Players also supported standard NTSC LaserDiscs. Hi-Vision LaserDiscs are extremely rare and expensive. [7]

The HDL-5800 Video Disc Recorder recorded both high definition still images and continuous video onto an optical disc and was part of the early analog wideband Sony HDVS high-definition video system which supported the MUSE system. Capable of recording HD still images and video onto either the WHD-3AL0 or the WHD-33A0 optical disc; WHD-3Al0 for CLV mode (up to 10 minute video or 18,000 still frames per side); WHD-33A0 for CAV mode (up to 3 minute video or 5400 still frames per side).

The HDL-2000 was a full band high definition video disc player. [7]

Video cassettes

W-VHS allowed home recording of Hi-Vision programmes.

See also

The analog TV systems these systems were meant to replace:

Related standards:

Related Research Articles

<span class="mw-page-title-main">NTSC</span> Analog television system

NTSC is the first American standard for analog television, published in 1941. In 1961, it was assigned the designation System M. It is also known as EIA standard 170.

<span class="mw-page-title-main">PAL</span> Colour encoding system for analogue television

Phase Alternating Line (PAL) is a colour encoding system for analogue television. It was one of three major analogue colour television standards, the others being NTSC and SECAM. In most countries it was broadcast at 625 lines, 50 fields per second, and associated with CCIR analogue broadcast television systems B, D, G, H, I or K. The articles on analog broadcast television systems further describe frame rates, image resolution, and audio modulation.

<span class="mw-page-title-main">SECAM</span> French analog color television system

SECAM, also written SÉCAM, is an analog color television system that was used in France, Russia and some other countries or territories of Europe and Africa. It was one of three major analog color television standards, the others being PAL and NTSC. Like PAL, a SECAM picture is also made up of 625 interlaced lines and is displayed at a rate of 25 frames per second. However, due to the way SECAM processes color information, it is not compatible with the German PAL video format standard. This page primarily discusses the SECAM colour encoding system. The articles on broadcast television systems and analog television further describe frame rates, image resolution, and audio modulation. SECAM video is composite video because the luminance and chrominance are transmitted together as one signal.

<span class="mw-page-title-main">Interlaced video</span> Technique for doubling the perceived frame rate of a video display

Interlaced video is a technique for doubling the perceived frame rate of a video display without consuming extra bandwidth. The interlaced signal contains two fields of a video frame captured consecutively. This enhances motion perception to the viewer, and reduces flicker by taking advantage of the phi phenomenon.

<span class="mw-page-title-main">Y′UV</span> Mathematical color model

Y′UV, also written YUV, is the color model found in the PAL analogue color TV standard. A color is described as a Y′ component (luma) and two chroma components U and V. The prime symbol (') denotes that the luma is calculated from gamma-corrected RGB input and that it is different from true luminance. Today, the term YUV is commonly used in the computer industry to describe colorspaces that are encoded using YCbCr.

<span class="mw-page-title-main">Chroma subsampling</span> Practice of encoding images

Chroma subsampling is the practice of encoding images by implementing less resolution for chroma information than for luma information, taking advantage of the human visual system's lower acuity for color differences than for luminance.

<span class="mw-page-title-main">Rec. 601</span> Standard from the International Telecommunication Union

ITU-R Recommendation BT.601, more commonly known by the abbreviations Rec. 601 or BT.601, is a standard originally issued in 1982 by the CCIR for encoding interlaced analog video signals in digital video form. It includes methods of encoding 525-line 60 Hz and 625-line 50 Hz signals, both with an active region covering 720 luminance samples and 360 chrominance samples per line. The color encoding system is known as YCbCr 4:2:2.

<span class="mw-page-title-main">D-1 (Sony)</span> Magnetic tape-based videocassette format

D-1 or 4:2:2 Component Digital is an SMPTE digital recording video standard, introduced in 1986 through efforts by SMPTE engineering committees. It started as a Sony and Bosch – BTS product and was the first major professional digital video format. SMPTE standardized the format within ITU-R 601, also known as Rec. 601, which was derived from SMPTE 125M and EBU 3246-E standards.

<span class="mw-page-title-main">YIQ</span> Color space

YIQ is the color space used by the analog NTSC color TV system. I stands for in-phase, while Q stands for quadrature, referring to the components used in quadrature amplitude modulation. Other TV systems used different color spaces, such as YUV for PAL or YDbDr for SECAM. Later digital standards use the YCbCr color space. These color spaces are all broadly related, and work based on the principle of adding a color component named chrominance, to a black and white image named luma.

<span class="mw-page-title-main">Serial digital interface</span> Family of digital video interfaces

Serial digital interface (SDI) is a family of digital video interfaces first standardized by SMPTE in 1989. For example, ITU-R BT.656 and SMPTE 259M define digital video interfaces used for broadcast-grade video. A related standard, known as high-definition serial digital interface (HD-SDI), is standardized in SMPTE 292M; this provides a nominal data rate of 1.485 Gbit/s.

<span class="mw-page-title-main">YCbCr</span> Family of digital colour spaces

YCbCr, Y′CbCr, or Y Pb/Cb Pr/Cr, also written as YCBCR or Y′CBCR, is a family of color spaces used as a part of the color image pipeline in video and digital photography systems. Y′ is the luma component and CB and CR are the blue-difference and red-difference chroma components. Y′ is distinguished from Y, which is luminance, meaning that light intensity is nonlinearly encoded based on gamma corrected RGB primaries.

HD-MAC was a broadcast television standard proposed by the European Commission in 1986, as part of Eureka 95 project. It belongs to the MAC - Multiplexed Analogue Components standard family. It is an early attempt by the EEC to provide High-definition television (HDTV) in Europe. It is a complex mix of analogue signal, multiplexed with digital sound, and assistance data for decoding (DATV). The video signal was encoded with a modified D2-MAC encoder.

Optical resolution describes the ability of an imaging system to resolve detail, in the object that is being imaged. An imaging system may have many individual components, including one or more lenses, and/or recording and display components. Each of these contributes to the optical resolution of the system; the environment in which the imaging is done often is a further important factor.

<span class="mw-page-title-main">Multiplexed Analogue Components</span> 1980s analog television standard

Multiplexed Analogue Components (MAC) was an analog television standard where luminance and chrominance components were transmitted separately. This was an evolution from older color TV systems where there was interference between chrominance and luminance.

Analog high-definition television has referred to a variety of analog video broadcast television systems with various display resolutions throughout history.

In video, luma represents the brightness in an image. Luma is typically paired with chrominance. Luma represents the achromatic image, while the chroma components represent the color information. Converting R′G′B′ sources into luma and chroma allows for chroma subsampling: because human vision has finer spatial sensitivity to luminance differences than chromatic differences, video systems can store and transmit chromatic information at lower resolution, optimizing perceived detail at a particular bandwidth.

<span class="mw-page-title-main">Rec. 709</span> Standard for HDTV image encoding and signal characteristics

Rec. 709, also known as Rec.709, BT.709, and ITU 709, is a standard developed by ITU-R for image encoding and signal characteristics of high-definition television.

High-definition television (HDTV) describes a television or video system which provides a substantially higher image resolution than the previous generation of technologies. The term has been used since at least 1933; in more recent times, it refers to the generation following standard-definition television (SDTV). It is the current de facto standard video format used in most broadcasts: terrestrial broadcast television, cable television, satellite television.

<span class="mw-page-title-main">Rec. 2020</span> ITU-R recommendation

ITU-R Recommendation BT.2020, more commonly known by the abbreviations Rec. 2020 or BT.2020, defines various aspects of ultra-high-definition television (UHDTV) with standard dynamic range (SDR) and wide color gamut (WCG), including picture resolutions, frame rates with progressive scan, bit depths, color primaries, RGB and luma-chroma color representations, chroma subsamplings, and an opto-electronic transfer function. The first version of Rec. 2020 was posted on the International Telecommunication Union (ITU) website on August 23, 2012, and two further editions have been published since then.

Clear-Vision is a Japanese EDTV television system introduced in the 1990s, that improves audio and video quality while remaining compatible with the existing broadcast standard. Developed to improve analog NTSC, it adds features like progressive scan, ghost cancellation and widescreen image format. A similar system named PALPlus was develop in Europe with the goal of improving analog PAL broadcasts.

References

  1. 1 2 3 4 "DBNSTJ : Realization of High-Definition Television by MUSE System". dbnst.nii.ac.jp.
  2. 1 2 3 Cianci, Philip J. (January 10, 2014). High Definition Television: The Creation, Development and Implementation of HDTV Technology. McFarland. ISBN   9780786487974 via Google Books.
  3. 1 2 3 "MUSE system for HDTV broadcasting-satellite services" (PDF). International Telecommunication Union. 1992. ITU-R BO.786.
  4. "ST 240:1999 - SMPTE Standard - For Television — 1125-Line High-Definition Production Systems — Signal Parameters". St 240:1999: 1–7. November 30, 1999. doi:10.5594/SMPTE.ST240.1999. ISBN   978-1-61482-389-6 via IEEE Xplore.
  5. 1 2 3 4 5 6 7 8 9 10 11 ANSI/SMPTE 240M-1995 - Signal Parameters 1125-Line High-Definition Production Systems (PDF). SMPTE. 1995.
  6. Poynton, Charles (January 3, 2003). Digital Video and HD: Algorithms and Interfaces. Elsevier. ISBN   9780080504308 via Google Books.
  7. 1 2 3 4 "MUSE LaserDisc". ura.caldc.com. Retrieved 2022-10-19.
  8. 1 2 "MUSE HI-DEF LaserDisc Players". LaserDisc UK Web Site. Archived from the original on 30 April 2016. Retrieved 10 October 2021.
  9. Jun-ichi, Ishida; Ninomiya, Yuichi (December 19, 1982). "3. Signal and Transmission Equipment for High-Definition TV". The Journal of the Institute of Television Engineers of Japan. 36 (10): 882–888. doi: 10.3169/itej1978.36.10_882 via CiNii.
  10. 1 2 Fujio, Takashi (December 19, 1980). "High-Definition Television System for Future : Desirable Standard, Signal Form and Broadcasting System". ITE Technical Report. 4 (28): 19–24. doi:10.11485/tvtr.4.28_19 via CiNii.
  11. Fujio, Takashi (December 19, 1981). "High Definitional Television". The Journal of the Institute of Television Engineers of Japan. 35 (12): 1016–1023. doi: 10.3169/itej1978.35.1016 via CiNii.
  12. Komoto, Taro; Ishida, Junichi; Hata, Masaji; Yasunaga, Keiichi (December 19, 1979). "YC Separate Transmission of high Definition Television Signal by BSE". ITE Technical Report. 3 (26): 61–66. doi:10.11485/tvtr.3.26_61 via CiNii.
  13. FUJIO, Takashi (December 19, 1984). "High-Definition Television System". ITE Technical Report. 8 (1): 33–39. doi:10.11485/tvtr.8.1_33 via CiNii.
  14. FUJIO, Takashi (August 19, 2006). "Rowing a Boat to the HDTV New World". The Journal of the Institute of Electronics, Information and Communication Engineers. 89 (8): 728–734 via CiNii.
  15. "SMPTE-240M Y'PbPr". www5.in.tum.de.
  16. 1 2 "Detailed Colorspace Descriptions". www.linuxtv.org.
  17. Charles A. Poynton, Digital Video and HDTV: Algorithms and Interfaces, Morgan–Kaufmann, 2003. online
  18. NHK (1993). High Definition Television - Hi Vision Technology. ISBN   0-442-00798-1.
  19. Naganawa, K.; Hori, Y.; Yanase, S.; Itoh, N.; Asano, Y. (August 19, 1991). "A single-chip audio signal processor for HDTV receiver". IEEE Transactions on Consumer Electronics. 37 (3): 677–683. doi:10.1109/30.85585. S2CID   62603128.