IXML

Last updated

iXML is an open standard for the inclusion of location sound metadata in Broadcast Wave audio files, video files and also IP video and audio streams. This includes things like Scene, Take and Notes information. [1]

Contents

It is the result of extended discussions between the various manufacturers of Field recorders and editing systems. It is designed to standardise the exchange of metadata between these systems.

The iXML specification describes an WAV RIFF chunk in BWF files which contains standard XML data following the iXML specification. [1] It also introduces the concept of using iXML in IP video streams such as NDI

Prior to the development of the iXML specification, the film and TV industry relied on the BWF bext description chunk which was used differently by many vendors to roughly encode some small metadata, but was invariably undefined, with too little space for full information. Whilst many systems tried to read what they could from the bext data, because of no specification and limited space, bext usefulness was limited.

History

The iXML concept was born during a meeting of various vendors, including manufacturers of field recorders, NLEs and DAWs, hosted by the Institute of Broadcast Sound, in London, on 8 July 2004. The "i" in iXML recognises the part the IBS played in bringing together such a diverse blend of normally competitive manufacturers to collectively solve the increasingly difficult problem of metadata interchange, with an elegant, capable and completely public specification.

Following initial discussions between Mark Gilbert of Gallery, John Ellwood of SynchroArts and J.P Beauviala of Aaton at the IBS meeting, the iXML 1.0 Specification was developed, drawn up and published by Mark Gilbert of Gallery UK at http://www.ixml.info. The iXML Specification is maintained by Gallery UK, and changes made are based on discussions by the iXML committee.

Soon after the IBS meeting, Gallery shipped Metacorder which was the first iXML compatible device. Mark Gilbert continued to promote the iXML format all over the world during 2004 and 2005 and gradually products emerged from other vendors. Other early adopters of iXML included SynchroArt's TITAN utility, and HHB's Portadrive field recorder.

In 2019 Gallery introduced the notion of using iXML in NDI based video/audio streams and the iXML 3.0 specification was announced which will define this.

In Use

Today, dozens of industry standard audio products support iXML with 100% interchange of metadata between systems. This includes all the current location field recorders (for which the spec was originally designed), several DAW applications (including Digidesign Pro Tools, Sony Vegas, Cockos Reaper and Steinberg Cubase/Nuendo) and various utilities. In late 2007, Apple Inc. added iXML support to their Final Cut Pro nonlinear video editor, which is used by more than 800,000 users worldwide. iXML support was introduced with version 6.02 of Final Cut Pro.

Related Research Articles

<span class="mw-page-title-main">Audio file format</span> Computer format for digital audio

An audio file format is a file format for storing digital audio data on a computer system. The bit layout of the audio data is called the audio coding format and can be uncompressed, or compressed to reduce the file size, often using lossy compression. The data can be a raw bitstream in an audio coding format, but it is usually embedded in a container format or an audio data format with defined storage layer.

Waveform Audio File Format is an audio file format standard for storing an audio bitstream on personal computers. The format was developed and published for the first time in 1991 by IBM and Microsoft. It is the main format used on Microsoft Windows systems for uncompressed audio. The usual bitstream encoding is the linear pulse-code modulation (LPCM) format.

Audio Video Interleave is a proprietary multimedia container format and Windows standard introduced by Microsoft in November 1992 as part of its Video for Windows software. AVI files can contain both audio and video data in a file container that allows synchronous audio-with-video playback. Like the DVD video format, AVI files support multiple streaming audio and video, although these features are seldom used.

Resource Interchange File Format (RIFF) is a generic file container format for storing data in tagged chunks. It is primarily used for audio and video, though it can be used for arbitrary data.

Material Exchange Format (MXF) is a container format for professional digital video and audio media defined by a set of SMPTE standards. A typical example of its use is for delivering advertisements to TV stations and tapeless archiving of broadcast TV programs. It is also used as part of the Digital Cinema Package for delivering movies to commercial theaters.

The Advanced Authoring Format (AAF) is a file format for professional cross-platform data interchange, designed for the video post-production and authoring environment. It was created by the Advanced Media Workflow Association (AMWA), and is now being standardized through the Society of Motion Picture and Television Engineers (SMPTE).

A FourCC is a sequence of four bytes used to uniquely identify data formats. It originated from the OSType or ResType metadata system used in classic Mac OS and was adopted for the Amiga/Electronic Arts Interchange File Format and derivatives. The idea was later reused to identify compressed data types in QuickTime and DirectShow.

<span class="mw-page-title-main">Program and System Information Protocol</span> Video and audio industry protocol

The Program and System Information Protocol (PSIP) is the MPEG and privately defined program-specific information originally defined by General Instrument for the DigiCipher 2 system and later extended for the ATSC digital television system for carrying metadata about each channel in the broadcast MPEG transport stream of a television station and for publishing information about television programs so that viewers can select what to watch by title and description. Its FM radio equivalent is Radio Data System (RDS).

A container format or metafile is a file format that allows multiple data streams to be embedded into a single file, usually along with metadata for identifying and further detailing those streams. Notable examples of container formats include archive files and formats used for multimedia playback. Among the earliest cross-platform container formats were Distinguished Encoding Rules and the 1985 Interchange File Format.

The Extensible Metadata Platform (XMP) is an ISO standard, originally created by Adobe Systems Inc., for the creation, processing and interchange of standardized and custom metadata for digital documents and data sets.

Broadcast Wave Format (BWF) is an extension of the popular Microsoft WAV audio format and is the recording format of most file-based non-linear digital recorders used for motion picture, radio and television production. It was first specified by the European Broadcasting Union in 1997, and updated in 2001 and 2003. It has been accepted as the ITU recommendation ITU-R BS.1352-3, Annex 1.

Flash Video is a container file format used to deliver digital video content over the Internet using Adobe Flash Player version 6 and newer. Flash Video content may also be embedded within SWF files. There are two different Flash Video file formats: FLV and F4V. The audio and video data within FLV files are encoded in the same way as SWF files. The F4V file format is based on the ISO base media file format, starting with Flash Player 9 update 3. Both formats are supported in Adobe Flash Player and developed by Adobe Systems. FLV was originally developed by Macromedia. In the early 2000s, Flash Video was the de facto standard for web-based streaming video. Users include Hulu, VEVO, Yahoo! Video, metacafe, Reuters.com, and many other news providers.

The Institute of Professional Sound, previously the Institute of Broadcast Sound, is an organisation for audio professionals. The organisation provides opportunities for training and conferencing to assist in maintaining high standards in all areas of professional audio operations. The organisation is based in the UK.

RF64 is a BWF-compatible multichannel audio file format enabling file sizes to exceed 4 GiB. It has been specified by the European Broadcasting Union. It has been accepted as the ITU recommendation ITU-R BS.2088.

RSS-TV is an XML-based navigation protocol for Internet media services based on the RSS standard.

A file format is a standard way that information is encoded for storage in a computer file. It specifies how bits are used to encode information in a digital storage medium. File formats may be either proprietary or free.

HTML audio is a subject of the HTML specification, incorporating audio input, playback, and synthesis, as well as speech to text, all in the browser.

The Media Object Server (MOS) protocol allows newsroom computer systems (NRCS) to communicate using a standard protocol with video servers, audio servers, still stores, and character generators for broadcast production.

The Physical Security Interoperability Alliance (PSIA) is a global consortium of more than 65 physical security manufacturers and systems integrators focused on promoting interoperability of IP-enabled security devices and systems across the physical security ecosystem as well as enterprise and building automation systems.

Network Device Interface (NDI) is a software specification developed by the technology company NewTek. It enables high-definition video to be transmitted, received, and communicated over a computer network with low latency and high quality. This royalty-free specification supports frame-accurate switching, making it suitable for live production environments.

References

  1. 1 2 "iXML". fcp.cafe. Retrieved 21 August 2024.