fbpx
Wikipedia

Music and artificial intelligence

Music and artificial intelligence is the development of music software programs which use AI to generate music.[1] As with applications in other fields, AI in music also simulates mental tasks. A prominent feature is the capability of an AI algorithm to learn based on past data, such as in computer accompaniment technology, wherein the AI is capable of listening to a human performer and performing accompaniment.[2] Artificial intelligence also drives interactive composition technology, wherein a computer composes music in response to a live performance. There are other AI applications in music that cover not only music composition, production, and performance but also how music is marketed and consumed. Several music player programs have also been developed to use voice recognition and natural language processing technology for music voice control. Current research includes the application of AI in music composition, performance, theory and digital sound processing.

Erwin Panofksy proposed that in all art, there existed 3 levels of meaning: primary meaning, or the natural subject; secondary meaning, or the conventional subject; and tertiary meaning, the intrinsic content of the subject.[3][4] AI music explores the foremost of these, creating music without the "intention" which is usually behind it, leaving composers who listen to machine-generated pieces feeling unsettled by the lack of apparent meaning.[5]

History edit

Artificial intelligence finds its beginnings in music with the transcription problem: accurately recording a performance into musical notation as it is played. Père Engramelle's schematic of a "piano roll," a mode of automatically recording note timing and duration in a way which could be easily transcribed to proper musical notation by hand, was first implemented by German engineers J.F. Unger and J. Hohlfield in 1752.[6]

In 1957, the ILLIAC I (Illinois Automatic Computer) produced the "Illiac Suite for String Quartet," a completely computer-generated piece of music. The computer was programmed to accomplish this by composer Lejaren Hiller and mathematician Leonard Isaacson.[citation needed]

In 1960, Russian researcher Rudolf Zaripov published worldwide first paper on algorithmic music composing using the "Ural-1" computer.[7]

In 1965, inventor Ray Kurzweil developed software capable of recognizing musical patterns and synthesizing new compositions from them. The computer first appeared on the quiz show I've Got a Secret.[citation needed]

By 1983, Yamaha Corporation's Kansei Music System had gained momentum, and a paper was published on its development in 1989. The software utilized music information processing and artificial intelligence techniques to essentially solve the transcription problem for simpler melodies, although higher-level melodies and musical complexities are regarded even today as difficult deep learning tasks, and near-perfect transcription is still a subject of research.[6][8]

In 1997, an artificial intelligence program named Experiments in Musical Intelligence (EMI) appeared to outperform a human composer at the task of composing a piece of music to imitate the style of Bach.[9] EMI would later become the basis for a more sophisticated algorithm called Emily Howell, named for its creator.

In 2002, the music research team at the Sony Computer Science Laboratory Paris, led by French composer and scientist François Pachet, designed the Continuator, an algorithm uniquely capable of resuming a composition after a live musician stopped.[citation needed]

Emily Howell would continue to make advancements in musical artificial intelligence, publishing its first album "From Darkness, Light" in 2009, and its second "Breathless" by 2012. Since then, many more pieces by artificial intelligence and various groups have been published.[citation needed]

In 2010, Iamus became the first AI to produce a fragment of original contemporary classical music, in its own style: "Iamus' Opus 1." Located at the Universidad de Malága (Malága University) in Spain, the computer can generate a fully original piece in a variety of musical styles in the span of eight minutes.[citation needed]

Software applications edit

Interactive scores edit

Multimedia Scenarios in interactive scores are represented by temporal objects, temporal relations, and interactive objects. Examples of temporal objects are sounds, videos and light controls. Temporal objects can be triggered by interactive objects (usually launched by the user) and several temporal objects can be executed simultaneously. A temporal object may contain other temporal objects: this hierarchy allows us to control the start or end of a temporal object by controlling the start or end of its parent. Hierarchy is ever-present in all kinds of music: music pieces are often characterized by movements, parts, motives, and measures, among other segments.[10][11]

Computer Accompaniment (Carnegie Mellon University) edit

The Computer Music Project at Carnegie Mellon University develops computer music and interactive performance technology to enhance human musical experience and creativity. This interdisciplinary effort draws on music theory, cognitive science, artificial intelligence and machine learning, human computer interaction, real-time systems, computer graphics and animation, multimedia, programming languages, and signal processing.[12]

ChucK edit

Developed at Princeton University by Ge Wang and Perry Cook, ChucK is a text-based, cross-platform language.[13] By extracting and classifying the theoretical techniques it finds in musical pieces, the software is able to synthesize entirely new pieces from the techniques it has learned.[14] The technology is used by SLOrk (Stanford Laptop Orchestra)[15] and PLOrk (Princeton Laptop Orchestra).

Jukedeck edit

Jukedeck was a website that let people use artificial intelligence to generate original, royalty-free music for use in videos.[16][17] The team started building the music generation technology in 2010,[18] formed a company around it in 2012,[19] and launched the website publicly in 2015.[17] The technology used was originally a rule-based algorithmic composition system,[20] which was later replaced with artificial neural networks.[16] The website was used to create over 1 million pieces of music, and brands that used it included Coca-Cola, Google, UKTV, and the Natural History Museum, London.[21] In 2019, the company was acquired by ByteDance.[22][23][24]

MorpheuS edit

MorpheuS[25] is a research project by Dorien Herremans and Elaine Chew at Queen Mary University of London, funded by a Marie Skłodowská-Curie EU project. The system uses an optimization approach based on a variable neighborhood search algorithm to morph existing template pieces into novel pieces with a set level of tonal tension that changes dynamically throughout the piece. This optimization approach allows for the integration of a pattern detection technique in order to enforce long term structure and recurring themes in the generated music. Pieces composed by MorpheuS have been performed at concerts in both Stanford and London.

AIVA edit

Created in February 2016, in Luxembourg, AIVA is a program that produces soundtracks for any type of media. The algorithms behind AIVA are based on deep learning architectures[26] AIVA has also been used to compose a Rock track called On the Edge,[27] as well as a pop tune Love Sick[28] in collaboration with singer Taryn Southern,[29] for the creation of her 2018 album "I am AI".

Google Magenta edit

20-second music clip generated by MusicLM using the prompt "hypnotic ambient electronic music"

Google's Magenta team has published several AI music applications and technical papers since their launch in 2016.[30] In 2017 they released the NSynth algorithm and dataset,[31] and an open source hardware musical instrument, designed to facilitate musicians in using the algorithm.[32] The instrument was used by notable artists such as Grimes and YACHT in their albums.[33][34] In 2018, they released a piano improvisation app called Piano Genie. This was later followed by Magenta Studio, a suite of 5 MIDI plugins that allow music producers to elaborate on existing music in their DAW.[35] In 2023, their machine learning team published a technical paper on GitHub that described MusicLM, a private text-to-music generator which they'd developed.[36][37]

Riffusion edit

 
Generated spectrogram from the prompt "bossa nova with electric guitar" (top), and the resulting audio after conversion (bottom)

Riffusion is a neural network, designed by Seth Forsgren and Hayk Martiros, that generates music using images of sound rather than audio.[38] It was created as a fine-tuning of Stable Diffusion, an existing open-source model for generating images from text prompts, on spectrograms.[38] This results in a model which uses text prompts to generate image files, which can be put through an inverse Fourier transform and converted into audio files.[39] While these files are only several seconds long, the model can also use latent space between outputs to interpolate different files together.[38][40] This is accomplished using a functionality of the Stable Diffusion model known as img2img.[41]

The resulting music has been described as "de otro mundo" (otherworldly),[42] although unlikely to replace man-made music.[42] The model was made available on December 15, 2022, with the code also freely available on GitHub.[39] It is one of many models derived from Stable Diffusion.[41]

Riffusion is classified within a subset of AI text-to-music generators. In December 2022, Mubert[43] similarly used Stable Diffusion to turn descriptive text into music loops. In January 2023, Google published a paper on their own text-to-music generator called MusicLM.[44][45]

Musical Applications edit

Artificial Intelligence has the opportunity to impact how producers create music by giving reiterations of a track that follow a prompt given by the creator. These prompts allow the AI to follow a certain style that the artist is trying to go for.[46]

AI has also been seen in musical analysis where it has been used for feature extraction, pattern recognition, and musical recommendations.[47]

Composition edit

Artificial intelligence has had major impacts in the composition sector as it has influenced the ideas of composers/producers and has the potential to make the industry more accessible to newcomers.[48] With its development in music, it has already been seen to be used in collaboration with producers. Artists use these software to help generate ideas and bring out musical styles by prompting the AI to follow specific requirements that fit their needs.[48] Software such as ChatGPT have been used by producers  to do these tasks, while other software such as Ozone11 have been used to automate time consuming and complex activities such as mastering. [49] Future compositional impacts by the technology include style emulation and fusion, and revision and refinement. Development of these types of software can give ease of access to newcomers to the music industry.[48]

Copyright edit

In the United States, the current legal framework tends to apply traditional copyright laws to AI, despite its differences with the human creative process.[50] However, music outputs solely generated by AI are not granted copyright protection. In the compendium of the U.S. Copyright Office Practices, the Copyright Office has stated that it would not grant copyrights to “works that lack human authorship” and “the Office will not register works produced by a machine or mere mechanical process that operates randomly or automatically without any creative input or intervention from a human author.”[51] In February 2022, the Copyright Review Board rejected an application to copyright AI-generated artwork on the basis that it "lacked the required human authorship necessary to sustain a claim in copyright."[52]

The situation in the European Union (EU) is similar to the US, because its legal framework also emphasizes the role of human involvement in a copyright-protected work.[53] According to the European Union Intellectual Property Office and the recent jurisprudence of the Court of Justice of the European Union, the originality criterion requires the work to be the author’s own intellectual creation, reflecting the personality of the author evidenced by the creative choices made during its production, requires distinct level of human involvement.[53] The reCreating Europe project, funded by the European Union’s Horizon 2020 research and innovation program, delves into the challenges posed by AI-generated contents including music, suggesting legal certainty and balanced protection that encourages innovation while respecting copyright norms.[53] The recognition of AIVA marks a significant departure from traditional views on authorship and copyrights in the realm of music composition, allowing AI artists capable of releasing music and earning royalties. This acceptance marks AIVA as a pioneering instance where an AI has been formally acknowledged within the music production.[54]

The recent advancements in artificial intelligence made by groups such as Stability AI, OpenAI, and Google has incurred an enormous sum of copyright claims leveled against generative technology, including AI music. Should these lawsuits succeed, the machine learning models behind these technologies would have their datasets restricted to the public domain.[55]

Musical deepfakes edit

A more nascent development of AI in music is the application of audio deepfakes to cast the lyrics or musical style of a preexisting song to the voice or style of another artist. This has raised many concerns regarding the legality of technology, as well as the ethics of employing it, particularly in the context of artistic identity.[56] Furthermore, it has also raised the question of to whom the authorship of these works is attributed. As AI cannot hold authorship of its own, current speculation suggests that there will be no clear answer until further rulings are made regarding machine learning technologies as a whole.[57] Most recent preventative measures have started to be developed by Google and Universal Music group who have taken into royalties and credit attribution to allow producers to replicated the voices and styles of artists.[58]

Heart on My Sleeve edit

In 2023 an artist known as “ghostwriter977” created a musical deepfake called Heart on My Sleeve that copied the voices of Drake and The Weeknd  by prompting an AI to create the track for them[59] .The track was submitted for Grammy consideration for the best rap song and song of the year [60].  It went viral and gained traction on TikTok and received a positive response from the audience leading to its official release on Apple Music, Spotify, and YouTube in April of 2023.[61] Many believed the track was fully composed by an AI software, but the producer claimed the songwriting, production, and voice were still done him.[59] It would later be rescinded from any Grammy considerations due to it not following the guidelines necessary to be considered for a grammy award[61].  The track would end up being removed from all music platforms by Universal Music Group.[61]  

See also edit

References edit

  1. ^ D. Herremans; C.H.; Chuan, E. Chew (2017). "A Functional Taxonomy of Music Generation Systems". ACM Computing Surveys. 50 (5): 69:1–30. arXiv:1812.04186. doi:10.1145/3108242. S2CID 3483927.
  2. ^ Dannenberg, Roger. (PDF). Semantic Scholar. S2CID 17787070. Archived from the original (PDF) on August 23, 2018. Retrieved August 23, 2018.
  3. ^ Erwin Panofsky, Studies in Iconology: Humanistic Themes in the Art of the Renaissance. Oxford 1939.
  4. ^ Dilly, Heinrich (2020), Arnold, Heinz Ludwig (ed.), "Panofsky, Erwin: Zum Problem der Beschreibung und Inhaltsdeutung von Werken der bildenden Kunst", Kindlers Literatur Lexikon (KLL) (in German), Stuttgart: J.B. Metzler, pp. 1–2, doi:10.1007/978-3-476-05728-0_16027-1, ISBN 978-3-476-05728-0, retrieved 2024-03-03
  5. ^ Miranda, Eduardo Reck, ed. (2021). "Handbook of Artificial Intelligence for Music" (PDF). SpringerLink. doi:10.1007/978-3-030-72116-9. ISBN 978-3-030-72115-2.
  6. ^ a b Roads, Curtis (1985). "Research in music and artificial intelligence". ACM Computing Surveys. 17 (2): 163–190. doi:10.1145/4468.4469. Retrieved 2024-03-06.
  7. ^ Zaripov, Rudolf (1960). "Об алгоритмическом описании процесса сочинения музыки (On algorithmic description of process of music composition)". Proceedings of the USSR Academy of Sciences. 132 (6).
  8. ^ Katayose, Haruhiro; Inokuchi, Seiji (1989). "The Kansei Music System". Computer Music Journal. 13 (4): 72–77. doi:10.2307/3679555. ISSN 0148-9267. JSTOR 3679555.
  9. ^ Johnson, George (11 November 1997). "Undiscovered Bach? No, a Computer Wrote It". The New York Times. Retrieved 29 April 2020. Dr. Larson was hurt when the audience concluded that his piece -- a simple, engaging form called a two-part invention -- was written by the computer. But he felt somewhat mollified when the listeners went on to decide that the invention composed by EMI (pronounced Emmy) was genuine Bach.
  10. ^ Mauricio Toro, Myriam Desainte-Catherine, Camilo Rueda. Formal semantics for interactive music scores: a framework to design, specify properties and execute interactive scenarios. Journal of Mathematics and Music 8 (1)
  11. ^ "Open Software System for Interactive Applications". Retrieved 23 January 2018.
  12. ^ Computer Music Group. 2.cs.cmu.edu. Retrieved on 2010-12-22.
  13. ^ ChucK => Strongly-timed, On-the-fly Audio Programming Language. Chuck.cs.princeton.edu. Retrieved on 2010-12-22.
  14. ^ Foundations of On-the-fly Learning in the ChucK Programming Language
  15. ^ Driver, Dustin. (1999-03-26) Pro - Profiles - Stanford Laptop Orchestra (SLOrk), pg. 1. Apple. Retrieved on 2010-12-22.
  16. ^ a b "From Jingles to Pop Hits, A.I. Is Music to Some Ears". The New York Times. 22 January 2017. Retrieved 2023-01-03.
  17. ^ a b "Need Music For A Video? Jukedeck's AI Composer Makes Cheap, Custom Soundtracks". techcrunch.com. 7 December 2015. Retrieved 2023-01-03.
  18. ^ "What Will Happen When Machines Write Songs Just as Well as Your Favorite Musician?". motherjones.com. Retrieved 2023-01-03.
  19. ^ Cookson, Robert (7 December 2015). "Jukedeck's computer composes music at touch of a button". Financial Times. Retrieved 2023-01-03.
  20. ^ "Jukedeck: the software that writes music by itself, note by note". Wired UK. Retrieved 2023-01-03.
  21. ^ "Robot rock: how AI singstars use machine learning to write harmonies". standard.co.uk. March 2018. Retrieved 2023-01-03.
  22. ^ "TIKTOK OWNER BYTEDANCE BUYS AI MUSIC COMPANY JUKEDECK". musicbusinessworldwide.com. 23 July 2019. Retrieved 2023-01-03.
  23. ^ "As TikTok's Music Licensing Reportedly Expires, Owner ByteDance Purchases AI Music Creation Startup JukeDeck". digitalmusicnews.com. 23 July 2019. Retrieved 2023-01-03.
  24. ^ "An AI-generated music app is now part of the TikTok group". sea.mashable.com. 24 July 2019. Retrieved 2023-01-03.
  25. ^ D. Herremans; E. Chew (2016). "MorpheuS: Automatic music generation with recurrent pattern constraints and tension profiles". IEEE Transactions on Affective Computing. PP(1). arXiv:1812.04832. doi:10.1109/TAFFC.2017.2737984. S2CID 54475410.
  26. ^ [1]. AIVA 2016
  27. ^ [2] AI-generated Rock Music: the Making Of
  28. ^ [3] Love Sick | Composed with Artificial Intelligence - Official Video with Lyrics | Taryn Southern
  29. ^ [4] Algo-Rhythms: the future of album collaboration
  30. ^ [5] Welcome to Magenta. Douglas Eck. Published June 1, 2016.
  31. ^ Engel, Jesse; Resnick, Cinjon; Roberts, Adam; Dieleman, Sander; Eck, Douglas; Simonyan, Karen; Norouzi, Mohammad (2017). "Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders". arXiv:1704.01279. {{cite journal}}: Cite journal requires |journal= (help)
  32. ^ Open NSynth Super, Google Creative Lab, 2023-02-13, retrieved 2023-02-14
  33. ^ "Cover Story: Grimes is ready to play the villain". Crack Magazine. Retrieved 2023-02-14.
  34. ^ "What Machine-Learning Taught the Band YACHT About Themselves". Los Angeleno. 2019-09-18. Retrieved 2023-02-14.
  35. ^ [6] Magenta Studio
  36. ^ [7] MusicLM on Github. Authored by Andrea Agostinelli, Timo I. Denk, Zalán Borsos, Jesse Engel, Mauro Verzetti, Antoine Caillon, Qingqing Huang, Aren Jansen, Adam Roberts, Marco Tagliasacchi, Matt Sharifi, Neil Zeghidour, Christian Frank. Published January 26, 2023.
  37. ^ [8] Understanding What Makes MusicLM Unique. Published January 27, 2023.
  38. ^ a b c Coldewey, Devin (December 15, 2022). "Try 'Riffusion,' an AI model that composes music by visualizing it".
  39. ^ a b Nasi, Michele (December 15, 2022). "Riffusion: creare tracce audio con l'intelligenza artificiale". IlSoftware.it.
  40. ^ "Essayez "Riffusion", un modèle d'IA qui compose de la musique en la visualisant". December 15, 2022.
  41. ^ a b "文章に沿った楽曲を自動生成してくれるAI「Riffusion」登場、画像生成AI「Stable Diffusion」ベースで誰でも自由に利用可能". GIGAZINE.
  42. ^ a b Llano, Eutropio (December 15, 2022). "El generador de imágenes AI también puede producir música (con resultados de otro mundo)".
  43. ^ "Mubert launches Text-to-Music interface – a completely new way to generate music from a single text prompt". December 21, 2022.
  44. ^ "MusicLM: Generating Music From Text". January 26, 2023.
  45. ^ "5 Reasons Google's MusicLM AI Text-to-Music App is Different". January 27, 2023.
  46. ^ Miranda, Eduardo Reck, ed. (2021). Shibboleth Authentication Request. doi:10.1007/978-3-030-72116-9. ISBN 978-3-030-72115-2. Retrieved 2024-04-03. {{cite book}}: |website= ignored (help)
  47. ^ Zhang, Yifei (December 2023). "Utilizing Computational Music Analysis and AI for Enhanced Music Composition: Exploring Pre- and Post-Analysis". Journal of Advanced Zoology. 44 (S-6): 1377–1390. doi:10.17762/jaz.v44is6.2470. S2CID 265936281.
  48. ^ a b c Miranda, Eduardo Reck, ed. (2021). Handbook of Artificial Intelligence for Music: Foundations, Advanced Approaches, and Developments for Creativity. Oxford handbooks (1st ed. 2021 ed.). Cham: Springer International Publishing. ISBN 978-3-030-72115-2.
  49. ^ Sunkel, Cameron (2023-12-16). "New Research Reveals Top AI Tools Utilized by Music Producers". EDM.com - The Latest Electronic Dance Music News, Reviews & Artists. Retrieved 2024-04-03.
  50. ^ "Art created by AI cannot be copyrighted, says US officials – what does this mean for music?". MusicTech. Retrieved 2022-10-27.
  51. ^ "Can (and should) AI-generated works be protected by copyright?". Hypebot. 2022-02-28. Retrieved 2022-10-27.
  52. ^ Re: Second Request for Reconsideration for Refusal to Register A Recent Entrance to Paradise (Correspondence ID 1-3ZPC6C3; SR # 1-7100387071) (PDF) (Report). Copyright Review Board, United States Copyright Office. 2022-02-14.
  53. ^ a b c Bulayenko, Oleksandr; Quintais, João Pedro; Gervais, Daniel J.; Poort, Joost (February 28, 2022). "AI Music Outputs: Challenges to the Copyright Legal Framework". reCreating Europe Report. Retrieved 2024-04-03.
  54. ^ Ahuja, Virendra (June 11, 2021). "Artificial Intelligence and Copyright: Issues and Challenges". ILI Law Review Winter Issue 2020. Retrieved 2024-04-03.
  55. ^ Samuelson, Pamela (2023-07-14). "Generative AI meets copyright". Science. 381 (6654): 158–161. doi:10.1126/science.adi0656. ISSN 0036-8075. PMID 37440639.
  56. ^ DeepDrake ft. BTS-GAN and TayloRVC: An Exploratory Analysis of Musical Deepfakes and Hosting Platforms
  57. ^ AI and Deepfake Voice Cloning: Innovation, Copyright and Artists’ Rights
  58. ^ "Google and Universal Music negotiate deal over AI 'deepfakes'". www.ft.com. Retrieved 2024-04-03.
  59. ^ a b Robinson, Kristin (2023-10-11). "Ghostwriter, the Mastermind Behind the Viral Drake AI Song, Speaks For the First Time". Billboard. Retrieved 2024-04-03.
  60. ^ "Drake/The Weeknd deepfake song "Heart on My Sleeve" submitted to Grammys". The FADER. Retrieved 2024-04-03.
  61. ^ a b c "The AI deepfake of Drake and The Weeknd will not be eligible for a GRAMMY". Mixmag. Retrieved 2024-04-03.

Further reading edit

  • Understanding Music with AI: Perspectives on Music Cognition 2021-01-10 at the Wayback Machine. Edited by Mira Balaban, Kemal Ebcioglu, and Otto Laske. AAAI Press.
  • Proceedings of a Workshop held as part of AI-ED 93, World Conference on Artificial Intelligence in Education on Music Education: An Artificial Intelligence Approach
  • Tanguiane (Tangian), Andranick (1993). Artificial Perception and Music Recognition. Lecture Notes in Artificial Intelligence. Vol. 746. Berlin-Heidelberg: Springer. ISBN 978-3-540-57394-4.

External links edit

  • The Music Informatics Research Group
  • Mixdevil - Is AI Good Gor Music Producers
  • OpenDream

music, artificial, intelligence, development, music, software, programs, which, generate, music, with, applications, other, fields, music, also, simulates, mental, tasks, prominent, feature, capability, algorithm, learn, based, past, data, such, computer, acco. Music and artificial intelligence is the development of music software programs which use AI to generate music 1 As with applications in other fields AI in music also simulates mental tasks A prominent feature is the capability of an AI algorithm to learn based on past data such as in computer accompaniment technology wherein the AI is capable of listening to a human performer and performing accompaniment 2 Artificial intelligence also drives interactive composition technology wherein a computer composes music in response to a live performance There are other AI applications in music that cover not only music composition production and performance but also how music is marketed and consumed Several music player programs have also been developed to use voice recognition and natural language processing technology for music voice control Current research includes the application of AI in music composition performance theory and digital sound processing Erwin Panofksy proposed that in all art there existed 3 levels of meaning primary meaning or the natural subject secondary meaning or the conventional subject and tertiary meaning the intrinsic content of the subject 3 4 AI music explores the foremost of these creating music without the intention which is usually behind it leaving composers who listen to machine generated pieces feeling unsettled by the lack of apparent meaning 5 Contents 1 History 2 Software applications 2 1 Interactive scores 2 2 Computer Accompaniment Carnegie Mellon University 2 3 ChucK 2 4 Jukedeck 2 5 MorpheuS 2 6 AIVA 2 7 Google Magenta 2 8 Riffusion 3 Musical Applications 3 1 Composition 4 Copyright 5 Musical deepfakes 5 1 Heart on My Sleeve 6 See also 7 References 8 Further reading 9 External linksHistory editArtificial intelligence finds its beginnings in music with the transcription problem accurately recording a performance into musical notation as it is played Pere Engramelle s schematic of a piano roll a mode of automatically recording note timing and duration in a way which could be easily transcribed to proper musical notation by hand was first implemented by German engineers J F Unger and J Hohlfield in 1752 6 In 1957 the ILLIAC I Illinois Automatic Computer produced the Illiac Suite for String Quartet a completely computer generated piece of music The computer was programmed to accomplish this by composer Lejaren Hiller and mathematician Leonard Isaacson citation needed In 1960 Russian researcher Rudolf Zaripov published worldwide first paper on algorithmic music composing using the Ural 1 computer 7 In 1965 inventor Ray Kurzweil developed software capable of recognizing musical patterns and synthesizing new compositions from them The computer first appeared on the quiz show I ve Got a Secret citation needed By 1983 Yamaha Corporation s Kansei Music System had gained momentum and a paper was published on its development in 1989 The software utilized music information processing and artificial intelligence techniques to essentially solve the transcription problem for simpler melodies although higher level melodies and musical complexities are regarded even today as difficult deep learning tasks and near perfect transcription is still a subject of research 6 8 In 1997 an artificial intelligence program named Experiments in Musical Intelligence EMI appeared to outperform a human composer at the task of composing a piece of music to imitate the style of Bach 9 EMI would later become the basis for a more sophisticated algorithm called Emily Howell named for its creator In 2002 the music research team at the Sony Computer Science Laboratory Paris led by French composer and scientist Francois Pachet designed the Continuator an algorithm uniquely capable of resuming a composition after a live musician stopped citation needed Emily Howell would continue to make advancements in musical artificial intelligence publishing its first album From Darkness Light in 2009 and its second Breathless by 2012 Since then many more pieces by artificial intelligence and various groups have been published citation needed In 2010 Iamus became the first AI to produce a fragment of original contemporary classical music in its own style Iamus Opus 1 Located at the Universidad de Malaga Malaga University in Spain the computer can generate a fully original piece in a variety of musical styles in the span of eight minutes citation needed Software applications editInteractive scores edit Multimedia Scenarios in interactive scores are represented by temporal objects temporal relations and interactive objects Examples of temporal objects are sounds videos and light controls Temporal objects can be triggered by interactive objects usually launched by the user and several temporal objects can be executed simultaneously A temporal object may contain other temporal objects this hierarchy allows us to control the start or end of a temporal object by controlling the start or end of its parent Hierarchy is ever present in all kinds of music music pieces are often characterized by movements parts motives and measures among other segments 10 11 Computer Accompaniment Carnegie Mellon University edit The Computer Music Project at Carnegie Mellon University develops computer music and interactive performance technology to enhance human musical experience and creativity This interdisciplinary effort draws on music theory cognitive science artificial intelligence and machine learning human computer interaction real time systems computer graphics and animation multimedia programming languages and signal processing 12 ChucK edit Main article ChucK Developed at Princeton University by Ge Wang and Perry Cook ChucK is a text based cross platform language 13 By extracting and classifying the theoretical techniques it finds in musical pieces the software is able to synthesize entirely new pieces from the techniques it has learned 14 The technology is used by SLOrk Stanford Laptop Orchestra 15 and PLOrk Princeton Laptop Orchestra Jukedeck edit Main article Jukedeck Jukedeck was a website that let people use artificial intelligence to generate original royalty free music for use in videos 16 17 The team started building the music generation technology in 2010 18 formed a company around it in 2012 19 and launched the website publicly in 2015 17 The technology used was originally a rule based algorithmic composition system 20 which was later replaced with artificial neural networks 16 The website was used to create over 1 million pieces of music and brands that used it included Coca Cola Google UKTV and the Natural History Museum London 21 In 2019 the company was acquired by ByteDance 22 23 24 MorpheuS edit MorpheuS 25 is a research project by Dorien Herremans and Elaine Chew at Queen Mary University of London funded by a Marie Sklodowska Curie EU project The system uses an optimization approach based on a variable neighborhood search algorithm to morph existing template pieces into novel pieces with a set level of tonal tension that changes dynamically throughout the piece This optimization approach allows for the integration of a pattern detection technique in order to enforce long term structure and recurring themes in the generated music Pieces composed by MorpheuS have been performed at concerts in both Stanford and London AIVA edit Main article AIVA Created in February 2016 in Luxembourg AIVA is a program that produces soundtracks for any type of media The algorithms behind AIVA are based on deep learning architectures 26 AIVA has also been used to compose a Rock track called On the Edge 27 as well as a pop tune Love Sick 28 in collaboration with singer Taryn Southern 29 for the creation of her 2018 album I am AI Google Magenta edit source source 20 second music clip generated by MusicLM using the prompt hypnotic ambient electronic music Google s Magenta team has published several AI music applications and technical papers since their launch in 2016 30 In 2017 they released the NSynth algorithm and dataset 31 and an open source hardware musical instrument designed to facilitate musicians in using the algorithm 32 The instrument was used by notable artists such as Grimes and YACHT in their albums 33 34 In 2018 they released a piano improvisation app called Piano Genie This was later followed by Magenta Studio a suite of 5 MIDI plugins that allow music producers to elaborate on existing music in their DAW 35 In 2023 their machine learning team published a technical paper on GitHub that described MusicLM a private text to music generator which they d developed 36 37 Riffusion edit This section is an excerpt from Riffusion edit nbsp source source Generated spectrogram from the prompt bossa nova with electric guitar top and the resulting audio after conversion bottom Riffusion is a neural network designed by Seth Forsgren and Hayk Martiros that generates music using images of sound rather than audio 38 It was created as a fine tuning of Stable Diffusion an existing open source model for generating images from text prompts on spectrograms 38 This results in a model which uses text prompts to generate image files which can be put through an inverse Fourier transform and converted into audio files 39 While these files are only several seconds long the model can also use latent space between outputs to interpolate different files together 38 40 This is accomplished using a functionality of the Stable Diffusion model known as img2img 41 The resulting music has been described as de otro mundo otherworldly 42 although unlikely to replace man made music 42 The model was made available on December 15 2022 with the code also freely available on GitHub 39 It is one of many models derived from Stable Diffusion 41 Riffusion is classified within a subset of AI text to music generators In December 2022 Mubert 43 similarly used Stable Diffusion to turn descriptive text into music loops In January 2023 Google published a paper on their own text to music generator called MusicLM 44 45 Musical Applications editArtificial Intelligence has the opportunity to impact how producers create music by giving reiterations of a track that follow a prompt given by the creator These prompts allow the AI to follow a certain style that the artist is trying to go for 46 AI has also been seen in musical analysis where it has been used for feature extraction pattern recognition and musical recommendations 47 Composition edit Artificial intelligence has had major impacts in the composition sector as it has influenced the ideas of composers producers and has the potential to make the industry more accessible to newcomers 48 With its development in music it has already been seen to be used in collaboration with producers Artists use these software to help generate ideas and bring out musical styles by prompting the AI to follow specific requirements that fit their needs 48 Software such as ChatGPT have been used by producers to do these tasks while other software such as Ozone11 have been used to automate time consuming and complex activities such as mastering 49 Future compositional impacts by the technology include style emulation and fusion and revision and refinement Development of these types of software can give ease of access to newcomers to the music industry 48 Copyright editFurther information Artificial intelligence and copyright In the United States the current legal framework tends to apply traditional copyright laws to AI despite its differences with the human creative process 50 However music outputs solely generated by AI are not granted copyright protection In the compendium of the U S Copyright Office Practices the Copyright Office has stated that it would not grant copyrights to works that lack human authorship and the Office will not register works produced by a machine or mere mechanical process that operates randomly or automatically without any creative input or intervention from a human author 51 In February 2022 the Copyright Review Board rejected an application to copyright AI generated artwork on the basis that it lacked the required human authorship necessary to sustain a claim in copyright 52 The situation in the European Union EU is similar to the US because its legal framework also emphasizes the role of human involvement in a copyright protected work 53 According to the European Union Intellectual Property Office and the recent jurisprudence of the Court of Justice of the European Union the originality criterion requires the work to be the author s own intellectual creation reflecting the personality of the author evidenced by the creative choices made during its production requires distinct level of human involvement 53 The reCreating Europe project funded by the European Union s Horizon 2020 research and innovation program delves into the challenges posed by AI generated contents including music suggesting legal certainty and balanced protection that encourages innovation while respecting copyright norms 53 The recognition of AIVA marks a significant departure from traditional views on authorship and copyrights in the realm of music composition allowing AI artists capable of releasing music and earning royalties This acceptance marks AIVA as a pioneering instance where an AI has been formally acknowledged within the music production 54 The recent advancements in artificial intelligence made by groups such as Stability AI OpenAI and Google has incurred an enormous sum of copyright claims leveled against generative technology including AI music Should these lawsuits succeed the machine learning models behind these technologies would have their datasets restricted to the public domain 55 Musical deepfakes editA more nascent development of AI in music is the application of audio deepfakes to cast the lyrics or musical style of a preexisting song to the voice or style of another artist This has raised many concerns regarding the legality of technology as well as the ethics of employing it particularly in the context of artistic identity 56 Furthermore it has also raised the question of to whom the authorship of these works is attributed As AI cannot hold authorship of its own current speculation suggests that there will be no clear answer until further rulings are made regarding machine learning technologies as a whole 57 Most recent preventative measures have started to be developed by Google and Universal Music group who have taken into royalties and credit attribution to allow producers to replicated the voices and styles of artists 58 Heart on My Sleeve edit In 2023 an artist known as ghostwriter977 created a musical deepfake called Heart on My Sleeve that copied the voices of Drake and The Weeknd by prompting an AI to create the track for them 59 The track was submitted for Grammy consideration for the best rap song and song of the year 60 It went viral and gained traction on TikTok and received a positive response from the audience leading to its official release on Apple Music Spotify and YouTube in April of 2023 61 Many believed the track was fully composed by an AI software but the producer claimed the songwriting production and voice were still done him 59 It would later be rescinded from any Grammy considerations due to it not following the guidelines necessary to be considered for a grammy award 61 The track would end up being removed from all music platforms by Universal Music Group 61 See also editAlgorithmic composition Automatic content recognition Computational models of musical creativity Generative artificial intelligence List of music software Music information retrieval OpenAI MuseNet and Jukebox music References edit D Herremans C H Chuan E Chew 2017 A Functional Taxonomy of Music Generation Systems ACM Computing Surveys 50 5 69 1 30 arXiv 1812 04186 doi 10 1145 3108242 S2CID 3483927 Dannenberg Roger Artificial Intelligence Machine Learning and Music Understanding PDF Semantic Scholar S2CID 17787070 Archived from the original PDF on August 23 2018 Retrieved August 23 2018 Erwin Panofsky Studies in Iconology Humanistic Themes in the Art of the Renaissance Oxford 1939 Dilly Heinrich 2020 Arnold Heinz Ludwig ed Panofsky Erwin Zum Problem der Beschreibung und Inhaltsdeutung von Werken der bildenden Kunst Kindlers Literatur Lexikon KLL in German Stuttgart J B Metzler pp 1 2 doi 10 1007 978 3 476 05728 0 16027 1 ISBN 978 3 476 05728 0 retrieved 2024 03 03 Miranda Eduardo Reck ed 2021 Handbook of Artificial Intelligence for Music PDF SpringerLink doi 10 1007 978 3 030 72116 9 ISBN 978 3 030 72115 2 a b Roads Curtis 1985 Research in music and artificial intelligence ACM Computing Surveys 17 2 163 190 doi 10 1145 4468 4469 Retrieved 2024 03 06 Zaripov Rudolf 1960 Ob algoritmicheskom opisanii processa sochineniya muzyki On algorithmic description of process of music composition Proceedings of the USSR Academy of Sciences 132 6 Katayose Haruhiro Inokuchi Seiji 1989 The Kansei Music System Computer Music Journal 13 4 72 77 doi 10 2307 3679555 ISSN 0148 9267 JSTOR 3679555 Johnson George 11 November 1997 Undiscovered Bach No a Computer Wrote It The New York Times Retrieved 29 April 2020 Dr Larson was hurt when the audience concluded that his piece a simple engaging form called a two part invention was written by the computer But he felt somewhat mollified when the listeners went on to decide that the invention composed by EMI pronounced Emmy was genuine Bach Mauricio Toro Myriam Desainte Catherine Camilo Rueda Formal semantics for interactive music scores a framework to design specify properties and execute interactive scenarios Journal of Mathematics and Music 8 1 Open Software System for Interactive Applications Retrieved 23 January 2018 Computer Music Group 2 cs cmu edu Retrieved on 2010 12 22 ChucK gt Strongly timed On the fly Audio Programming Language Chuck cs princeton edu Retrieved on 2010 12 22 Foundations of On the fly Learning in the ChucK Programming Language Driver Dustin 1999 03 26 Pro Profiles Stanford Laptop Orchestra SLOrk pg 1 Apple Retrieved on 2010 12 22 a b From Jingles to Pop Hits A I Is Music to Some Ears The New York Times 22 January 2017 Retrieved 2023 01 03 a b Need Music For A Video Jukedeck s AI Composer Makes Cheap Custom Soundtracks techcrunch com 7 December 2015 Retrieved 2023 01 03 What Will Happen When Machines Write Songs Just as Well as Your Favorite Musician motherjones com Retrieved 2023 01 03 Cookson Robert 7 December 2015 Jukedeck s computer composes music at touch of a button Financial Times Retrieved 2023 01 03 Jukedeck the software that writes music by itself note by note Wired UK Retrieved 2023 01 03 Robot rock how AI singstars use machine learning to write harmonies standard co uk March 2018 Retrieved 2023 01 03 TIKTOK OWNER BYTEDANCE BUYS AI MUSIC COMPANY JUKEDECK musicbusinessworldwide com 23 July 2019 Retrieved 2023 01 03 As TikTok s Music Licensing Reportedly Expires Owner ByteDance Purchases AI Music Creation Startup JukeDeck digitalmusicnews com 23 July 2019 Retrieved 2023 01 03 An AI generated music app is now part of the TikTok group sea mashable com 24 July 2019 Retrieved 2023 01 03 D Herremans E Chew 2016 MorpheuS Automatic music generation with recurrent pattern constraints and tension profiles IEEE Transactions on Affective Computing PP 1 arXiv 1812 04832 doi 10 1109 TAFFC 2017 2737984 S2CID 54475410 1 AIVA 2016 2 AI generated Rock Music the Making Of 3 Love Sick Composed with Artificial Intelligence Official Video with Lyrics Taryn Southern 4 Algo Rhythms the future of album collaboration 5 Welcome to Magenta Douglas Eck Published June 1 2016 Engel Jesse Resnick Cinjon Roberts Adam Dieleman Sander Eck Douglas Simonyan Karen Norouzi Mohammad 2017 Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders arXiv 1704 01279 a href Template Cite journal html title Template Cite journal cite journal a Cite journal requires journal help Open NSynth Super Google Creative Lab 2023 02 13 retrieved 2023 02 14 Cover Story Grimes is ready to play the villain Crack Magazine Retrieved 2023 02 14 What Machine Learning Taught the Band YACHT About Themselves Los Angeleno 2019 09 18 Retrieved 2023 02 14 6 Magenta Studio 7 MusicLM on Github Authored by Andrea Agostinelli Timo I Denk Zalan Borsos Jesse Engel Mauro Verzetti Antoine Caillon Qingqing Huang Aren Jansen Adam Roberts Marco Tagliasacchi Matt Sharifi Neil Zeghidour Christian Frank Published January 26 2023 8 Understanding What Makes MusicLM Unique Published January 27 2023 a b c Coldewey Devin December 15 2022 Try Riffusion an AI model that composes music by visualizing it a b Nasi Michele December 15 2022 Riffusion creare tracce audio con l intelligenza artificiale IlSoftware it Essayez Riffusion un modele d IA qui compose de la musique en la visualisant December 15 2022 a b 文章に沿った楽曲を自動生成してくれるAI Riffusion 登場 画像生成AI Stable Diffusion ベースで誰でも自由に利用可能 GIGAZINE a b Llano Eutropio December 15 2022 El generador de imagenes AI tambien puede producir musica con resultados de otro mundo Mubert launches Text to Music interface a completely new way to generate music from a single text prompt December 21 2022 MusicLM Generating Music From Text January 26 2023 5 Reasons Google s MusicLM AI Text to Music App is Different January 27 2023 Miranda Eduardo Reck ed 2021 Shibboleth Authentication Request doi 10 1007 978 3 030 72116 9 ISBN 978 3 030 72115 2 Retrieved 2024 04 03 a href Template Cite book html title Template Cite book cite book a website ignored help Zhang Yifei December 2023 Utilizing Computational Music Analysis and AI for Enhanced Music Composition Exploring Pre and Post Analysis Journal of Advanced Zoology 44 S 6 1377 1390 doi 10 17762 jaz v44is6 2470 S2CID 265936281 a b c Miranda Eduardo Reck ed 2021 Handbook of Artificial Intelligence for Music Foundations Advanced Approaches and Developments for Creativity Oxford handbooks 1st ed 2021 ed Cham Springer International Publishing ISBN 978 3 030 72115 2 Sunkel Cameron 2023 12 16 New Research Reveals Top AI Tools Utilized by Music Producers EDM com The Latest Electronic Dance Music News Reviews amp Artists Retrieved 2024 04 03 Art created by AI cannot be copyrighted says US officials what does this mean for music MusicTech Retrieved 2022 10 27 Can and should AI generated works be protected by copyright Hypebot 2022 02 28 Retrieved 2022 10 27 Re Second Request for Reconsideration for Refusal to Register A Recent Entrance to Paradise Correspondence ID 1 3ZPC6C3 SR 1 7100387071 PDF Report Copyright Review Board United States Copyright Office 2022 02 14 a b c Bulayenko Oleksandr Quintais Joao Pedro Gervais Daniel J Poort Joost February 28 2022 AI Music Outputs Challenges to the Copyright Legal Framework reCreating Europe Report Retrieved 2024 04 03 Ahuja Virendra June 11 2021 Artificial Intelligence and Copyright Issues and Challenges ILI Law Review Winter Issue 2020 Retrieved 2024 04 03 Samuelson Pamela 2023 07 14 Generative AI meets copyright Science 381 6654 158 161 doi 10 1126 science adi0656 ISSN 0036 8075 PMID 37440639 DeepDrake ft BTS GAN and TayloRVC An Exploratory Analysis of Musical Deepfakes and Hosting Platforms AI and Deepfake Voice Cloning Innovation Copyright and Artists Rights Google and Universal Music negotiate deal over AI deepfakes www ft com Retrieved 2024 04 03 a b Robinson Kristin 2023 10 11 Ghostwriter the Mastermind Behind the Viral Drake AI Song Speaks For the First Time Billboard Retrieved 2024 04 03 Drake The Weeknd deepfake song Heart on My Sleeve submitted to Grammys The FADER Retrieved 2024 04 03 a b c The AI deepfake of Drake and The Weeknd will not be eligible for a GRAMMY Mixmag Retrieved 2024 04 03 Further reading editUnderstanding Music with AI Perspectives on Music Cognition Archived 2021 01 10 at the Wayback Machine Edited by Mira Balaban Kemal Ebcioglu and Otto Laske AAAI Press Proceedings of a Workshop held as part of AI ED 93 World Conference on Artificial Intelligence in Education on Music Education An Artificial Intelligence Approach Tanguiane Tangian Andranick 1993 Artificial Perception and Music Recognition Lecture Notes in Artificial Intelligence Vol 746 Berlin Heidelberg Springer ISBN 978 3 540 57394 4 External links editThe Music Informatics Research Group Institut de Recherche et Coordination Acoustique Musique Interdisciplinary Centre for Research in Music Mixdevil Is AI Good Gor Music Producers OpenDream Retrieved from https en wikipedia org w index php title Music and artificial intelligence amp oldid 1218027031, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.