Pop music automation

Last updated

Pop music automation is a field of study among musicians and computer scientists with a goal of producing successful pop music algorithmically. It is often based on the premise that pop music is especially formulaic, unchanging, and easy to compose. The idea of automating pop music composition is related to many ideas in algorithmic music, Artificial Intelligence (AI) and computational creativity.

Contents

History of automation in music

Algorithms (or, at the very least, formal sets of rules) have been used to compose music for centuries; the procedures used to plot voice-leading in counterpoint, for example, can often be reduced to algorithmic determinant. Now the term is usually reserved, however, for the use of formal procedures to make music without human intervention.

Classical music automation software exists that generates music in the style of Mozart and Bach and jazz. Most notably, David Cope [1] has written a software system called "Experiments in Musical Intelligence" (or "EMI") that is capable of analyzing and generalizing from existing music by a human composer to generate novel musical compositions in the same style. EMI's output is convincing enough to persuade human listeners that its music is human-generated to a high level of competence.

Creativity research in jazz has focused on the process of improvisation and the cognitive demands that this places on a musical agent: reasoning about time, remembering and conceptualizing what has already been played, and planning ahead for what might be played next.

Inevitably associated with Pop music automation is Pop music analysis.

Projects in Pop music automation may include, but are not limited to, ideas in melody creation and song development, vocal generation or improvement, automatic accompaniment and lyric composition.

Automatic accompaniment

Some systems exist that automatically choose chords to accompany a vocal melody in real-time. A user with no musical experience can create a song with instrumental accompaniment just by singing into a microphone. An example is a Microsoft Research project called Songsmith, [2] which trains a Hidden Markov model using a music database and uses that model to select chords for new melodies.

Melody generation

Automatic melody generation is often done with a Markov chain, the states of the system become note or pitch values, and a probability vector for each note is constructed, completing a transition probability matrix (see below). An algorithm is constructed to produce an output note values based on the transition matrix weightings, which could be MIDI note values, frequency (Hz), or any other desirable metric.

1st-order matrix
NoteAC#Eb
A0.10.60.3
C#0.250.050.7
Eb0.70.30
2nd-order matrix
NoteADG
AA0.180.60.22
AD0.50.50
AG0.150.750.1
DD001
DA0.2500.75
DG0.90.10
GG0.40.40.2
GA0.50.250.25
GD100

A second-order Markov chain can be introduced by considering the current state and also the previous state, as indicated in the second table. Higher, nth-order chains tend to "group" particular notes together, while 'breaking off' into other patterns and sequences occasionally. These higher-order chains tend to generate results with a sense of phrasal structure, rather than the 'aimless wandering' produced by a first-order system. [3]

Lyric composition

Automated lyric creating software may take forms such as:

The Tra-la-Lyrics system [4] produces song lyrics, in Portuguese, for a given melody. This not only involves matching each word syllable with a note in the melody, but also matching the word's stress with the strong beats of the melody.

This involves natural language processing. Pablo Gervás [5] has developed a noteworthy system called ASPERA that employs a case-based reasoning (CBR) approach to generating poetic formulations of a given input text via a composition of poetic fragments that are retrieved from a case-base of existing poems. Each poem fragment in the ASPERA case-base is annotated with a prose string that expresses the meaning of the fragment, and this prose string is used as the retrieval key for each fragment. Metrical rules are then used to combine these fragments into a well-formed poetic structure.

Programs like TALE-SPIN [6] and The MINSTREL [7] system represent a complex elaboration of this basis approach, distinguishing a range of character-level goals in the story from a range of author-level goals for the story. Systems like Bringsjord's BRUTUS [8] can create stories with complex interpersonal themes like betrayal. On-line metaphor generation systems like 'Sardonicus' or 'Aristotle' [9] can suggest lexical metaphors for a given descriptive goal (e.g., to describe a supermodel as skinny, the source terms “pencil”, “whip”, “whippet”, “rope”, “stick-insect” and “snake” are suggested).

Using a language database (such as wordnet) one can create musings on a subject that may be weak grammatically but are still sensical. See such projects as the Flowerewolf automatic poetry generator or the Dada engine.

Software

More or less free

Commercial

See also

Related Research Articles

Computer music is the application of computing technology in music composition, to help human composers create new music or to have computers independently create music, such as with algorithmic composition programs. It includes the theory and application of new and existing computer software technologies and basic aspects of music, such as sound synthesis, digital signal processing, sound design, sonic diffusion, acoustics, electrical engineering, and psychoacoustics. The field of computer music can trace its roots back to the origins of electronic music, and the first experiments and innovations with electronic instruments at the turn of the 20th century.

Sheet music is a handwritten or printed form of musical notation that uses musical symbols to indicate the pitches, rhythms, or chords of a song or instrumental musical piece. Like its analogs – printed books or pamphlets in English, Arabic, or other languages – the medium of sheet music typically is paper. However, access to musical notation since the 1980s has included the presentation of musical notation on computer screens and the development of scorewriter computer programs that can notate a song or piece electronically, and, in some cases, "play back" the notated music using a synthesizer or virtual instruments.

A music sequencer is a device or application software that can record, edit, or play back music, by handling note and performance information in several forms, typically CV/Gate, MIDI, or Open Sound Control, and possibly audio and automation data for digital audio workstations (DAWs) and plug-ins.

<span class="mw-page-title-main">Generative art</span> Art created by a set of rules, often using computers

Generative art refers to art that in whole or in part has been created with the use of an autonomous system. An autonomous system in this context is generally one that is non-human and can independently determine features of an artwork that would otherwise require decisions made directly by the artist. In some cases the human creator may claim that the generative system represents their own artistic idea, and in others that the system takes on the role of the creator.

<span class="mw-page-title-main">Max Mathews</span> American pioneer in computer music

Max Vernon Mathews was an American pioneer of computer music.

Algorithmic composition is the technique of using algorithms to create music.

Generative music is a term popularized by Brian Eno to describe music that is ever-different and changing, and that is created by a system.

David Cope is an American author, composer, scientist, and Dickerson Emeriti Professor of Music at UC Santa Cruz. His primary area of research involves artificial intelligence and music; he writes programs and algorithms that can analyze existing music and create new compositions in the style of the original input music. He taught the groundbreaking summer workshop in Workshop in Algorithmic Computer Music (WACM) that was open to the public as well as a general education course entitled Artificial Intelligence and Music for enrolled UCSC students. Cope is also co-founder and CTO Emeritus of Recombinant Inc., a music technology company.

<span class="mw-page-title-main">Transcription (music)</span>

In music, transcription is the practice of notating a piece or a sound which was previously unnotated and/or unpopular as a written music, for example, a jazz improvisation or a video game soundtrack. When a musician is tasked with creating sheet music from a recording and they write down the notes that make up the piece in music notation, it is said that they created a musical transcription of that recording. Transcription may also mean rewriting a piece of music, either solo or ensemble, for another instrument or other instruments than which it was originally intended. The Beethoven Symphonies transcribed for solo piano by Franz Liszt are an example. Transcription in this sense is sometimes called arrangement, although strictly speaking transcriptions are faithful adaptations, whereas arrangements change significant aspects of the original piece.

<span class="mw-page-title-main">Eduardo Reck Miranda</span> Musical artist

Eduardo Reck Miranda is a Brazilian composer of chamber and electroacoustic pieces but is most notable in the United Kingdom for his scientific research into computer music, particularly in the field of human-machine interfaces where brain waves will replace keyboards and voice commands to permit the disabled to express themselves musically.

<span class="mw-page-title-main">Algorithmic art</span> Art genre

Algorithmic art or algorithm art is art, mostly visual art, in which the design is generated by an algorithm. Algorithmic artists are sometimes called algorists.

Evolutionary music is the audio counterpart to evolutionary art, whereby algorithmic music is created using an evolutionary algorithm. The process begins with a population of individuals which by some means or other produce audio, which is either initialized randomly or based on human-generated music. Then through the repeated application of computational steps analogous to biological selection, recombination and mutation the aim is for the produced audio to become more musical. Evolutionary sound synthesis is a related technique for generating sounds or synthesizer instruments. Evolutionary music is typically generated using an interactive evolutionary algorithm where the fitness function is the user or audience, as it is difficult to capture the aesthetic qualities of music computationally. However, research into automated measures of musical quality is also active. Evolutionary computation techniques have also been applied to harmonization and accompaniment tasks. The most commonly used evolutionary computation techniques are genetic algorithms and genetic programming.

<span class="mw-page-title-main">Impro-Visor</span> Scorewriter

Impro-Visor is an educational tool for creating and playing a lead sheet, with a particular orientation toward representing jazz solos.

<span class="mw-page-title-main">Computational creativity</span> Multidisciplinary endeavour

Computational creativity is a multidisciplinary endeavour that is located at the intersection of the fields of artificial intelligence, cognitive psychology, philosophy, and the arts.

Artificial intelligence and music (AIM) is a common subject in the International Computer Music Conference, the Computing Society Conference and the International Joint Conference on Artificial Intelligence. The first International Computer Music Conference (ICMC) was held in 1974 at Michigan State University. Current research includes the application of AI in music composition, performance, theory and digital sound processing.

A story generator or plot generator is a tool that generates basic narratives or plot ideas. The generator could be in the form of a computer program, a chart with multiple columns, a book composed of panels that flip independently of one another, or a set of several adjacent reels that spin independently of one another, allowing a user to select elements of a narrative plot. The tool may allow the user to select elements for the narrative, or it may combine them randomly, a specific variation known as a random plot generator. Such tools can be created for virtually any genre, although they tend to produce formulaic and hackneyed situations.

<span class="mw-page-title-main">François Pachet</span> French computer scientist (born 1964)

François Pachet is a French scientist, composer and director of the Spotify Creator Technology Research Lab. Before joining Spotify he led Sony Computer Science Laboratory in Paris. He is one of the pioneers of computer music closely linked to artificial intelligence, especially in the field of machine improvisation and style modelling. He has been elected ECCAI Fellow in 2014.

<span class="mw-page-title-main">Outline of machine learning</span> Overview of and topical guide to machine learning

The following outline is provided as an overview of and topical guide to machine learning. Machine learning is a subfield of soft computing within computer science that evolved from the study of pattern recognition and computational learning theory in artificial intelligence. In 1959, Arthur Samuel defined machine learning as a "field of study that gives computers the ability to learn without being explicitly programmed". Machine learning explores the study and construction of algorithms that can learn from and make predictions on data. Such algorithms operate by building a model from an example training set of input observations in order to make data-driven predictions or decisions expressed as outputs, rather than following strictly static program instructions.

Margareta (Maya) Ackerman is a Russian-American computer scientist known for her research in cluster analysis and algorithmic composition of music. She is an assistant professor of computer science and engineering at Santa Clara University, and the founder and CEO of algorithmic music firm WaveAI.

Jukedeck was a British technology company founded in 2012. It built a website that let users create royalty-free music using artificial intelligence.

References

  1. Cope, David (2006), Computer Models of Musical Creativity, Cambridge, MA: MIT Press
  2. and
  3. Curtis Roads, ed. (1996), The Computer Music Tutorial, MIT Press, ISBN   0-262-18158-4
  4. Gonçalo Oliveira, Hugo; et al. (2007), Tra-la-lyrics: an approach to generate text based on rhythm, Proceedings of the 4th International Joint Workshop on Computational Creativity, pp. 47–55, London, UK, (June 2007)
  5. Gervás, Pablo (2001), An expert system for the composition of formal Spanish poetry, Journal of Knowledge-Based Systems 14(3-4) pp 181188
  6. Meehan, James (1981), TALE-SPIN, Shank, R. C. and Riesbeck, C. K., (eds.), Inside Computer Understanding: Five Programs plus Miniatures. Hillsdale, NJ: Lawrence Erlbaum Associates
  7. Turner, S.R. (1994), The Creative Process: A Computer Model of Storytelling, Hillsdale, NJ: Lawrence Erlbaum Associates
  8. Bringsjord, S., Ferrucci, D. A. (2000), Artificial Intelligence and Literary Creativity. Inside the Mind of BRUTUS, a Storytelling Machine., Hillsdale NJ: Lawrence Erlbaum Associates{{citation}}: CS1 maint: multiple names: authors list (link)
  9. Veale, Tony, Hao, Yanfen (2007), Comprehending and Generating Apt Metaphors: A Web-driven, Case-based Approach to Figurative Language, Proceedings of AAAI 2007, the 22nd AAAI Conference on Artificial Intelligence. Vancouver, Canada{{citation}}: CS1 maint: multiple names: authors list (link)