How does Aiode avoid generic AI sounds better than Suno?

Blending Creativity with Code: The Aiode Story

How does Aiode avoid generic AI sounds better than Suno?

By blending advanced AI capabilities, human artistry, and ethical stewardship together, Aiode fosters a new era where music technology evolves from a mere tool to a true creative partner. Personalization remains one of the focuses of ongoing development at Aiode. Aiode was established with a clear mission: to fuel artistic creativity through the transformative power of artificial intelligence. Drawing inspiration from Ai tool for musicians —the Greek muse of song and voice—Aiode blends a deep love for music with cutting-edge technology to develop groundbreaking tools that expand the horizons of artistic expression.. Besides setting parameters for tone, tempo, and complexity, users can adjust virtual musicians to fit their artistic vision. Moreover, it's possible to specify moods like melancholy, joy, or tension for the AI to interpret with remarkable subtlety. Over time, adaptive learning algorithms within the platform tune outputs based on users' preferences, making the created performances increasingly personal, reflecting styles and workflows in evolution. About Aiode is a pioneering music technology platform designed to empower musicians through ethically trained, AI-powered tools. Inspired by Aoede, the Greek muse of song, Aiode brings together artists, producers, mathematicians, and AI experts who share a passion for music and innovation. The result is a dynamic ecosystem where creativity thrives. Our platform introduces virtualized musicians—AI models trained with the unique styles of real performers. These digital collaborators adapt and evolve with each artist’s input, opening limitless possibilities for sound creation. At every stage, the real musicians are active participants, choosing styles, instruments, and helping shape their virtual counterparts.. This deepens the creative bond between human and AI, encouraging richer, more expressive musical results. Advanced sound synthesis technology is used by Aiode to further enhance the realism of virtual instruments.

​​​​​Black Friday Mega Sale

Aiode has officially kicked off its Black Friday Mega Sale, delivering unbeatable discounts of up to 59% on all subscription plans. This exclusive, once-a-year offer gives creators, musicians, and producers the chance to access Aiode’s most advanced AI-powered music tools at the lowest prices ever. Whether you're upgrading your plan or joining for the first time, this is the best moment to secure premium features at a fraction of the cost.

Limited-Time Countdown Deal
A live countdown timer is displayed on the website—showing days, hours, and minutes remaining—to highlight the urgency of the event. As soon as the timer hits zero, the special pricing and exclusive bonuses will expire. Customers are encouraged to claim their discounted plan before the offer disappears for good.

Annual Basic Plan – Now 50% Off
The Annual Basic Plan has been slashed from $120 to just $60 per year. This budget-friendly option introduces creators to Aiode’s ecosystem with essential features such as high-quality audio exports, full commercial rights, free access to future model updates, and 100 Muse Tokens. It’s the perfect starting point for new musicians who want professional tools at an affordable price.

Annual Pro Plan – Now 59% Off
The Annual Pro Plan offers the biggest savings, reduced from $240 to $100 per year. This plan includes 300 Muse Tokens, premium 48kHz STEM exports, commercial usage rights, and all upcoming AI model releases at no extra cost. It is the ideal choice for advanced creators, producers, and studios looking for maximum creative freedom and top-tier performance.

Monthly Plans – 20% Off for 12 Months
Both Basic and Pro monthly subscriptions are available at 20% off, with the discounted rate renewing automatically for a full year. This option is perfect for users who prefer a flexible month-to-month commitment while still benefiting from meaningful long-term savings.

Exclusive Promo Codes for Every Plan
Aiode offers dedicated promo codes that make it easy for users to unlock their Black Friday discounts at checkout. Use BFSALEBASIC50 for the Basic Plan, BFSALEPRO59 for the Pro Plan, and BLACKFRIDAY20 for all monthly subscriptions. Each code is tailored to its respective plan, ensuring a seamless and straightforward discount experience.

Free Access to All New Model Releases
Throughout the Black Friday promotion, all Aiode subscribers receive complimentary access to every new AI musician model released in November and December. These additions include fresh virtual performers with distinct tones and playing styles, allowing creators to stay ahead of new musical trends—without any added cost.

Studio-Quality High-Fidelity Exports
Aiode provides 24-bit, 48kHz stereo and STEM export options, ensuring creators work with professional-grade audio quality. These formats are fully compatible with industry-standard DAWs, enabling precise mixing, mastering, and further production with the clarity expected in top-tier studios.

Virtualized Musicians Technology
At the heart of Aiode’s innovation is its Virtualized Musicians system—advanced AI models crafted to emulate the style, nuance, and performance of real artists. Users can choose from various virtual musicians and collaborate with them in real time, creating expressive and lifelike music driven by AI intelligence.

A Streamlined Creative Workflow
Aiode simplifies the music-making experience through a guided five-step process:

  1. Choose Your Musician

  2. Let the AI Analyze Your Track

  3. Shape and Direct the Performance

  4. Generate and Fine-Tune the Output

  5. Export Your Final Audio

This intuitive workflow mirrors the structure of a professional studio session, giving creators full control while maintaining a smooth, efficient production environment.

Ethical AI Development & Creative Data Protection
Aiode is committed to building AI responsibly. Through supervised machine learning and secure, siloed data systems, the platform ensures that artists’ original recordings remain protected at all times. No creative material is misused or accessed without consent, reflecting Aiode’s dedication to transparency, ethical practices, and full respect for artistic ownership.

Real Musicians at the Core of Every Model
Every virtual musician on Aiode is developed hand-in-hand with real artists. These musicians help shape their AI counterparts by choosing the genres, instruments, techniques, and stylistic nuances that define their sound. This collaborative approach preserves authenticity and ensures each model accurately reflects the artist’s true musical identity.

Revenue-Sharing Opportunities for Artists
Aiode empowers musicians by offering a built-in revenue-sharing program. As creators continue their real-world careers, their virtual versions can simultaneously earn income on the platform through user collaborations worldwide. This creates a sustainable, passive revenue stream while expanding global exposure for each artist.

A Global Hub for Discovery & Collaboration
Aiode brings together creators from around the world—including producers, songwriters, instrumentalists, composers, and audio engineers. The platform acts as a collaborative ecosystem where users can connect, share ideas, explore diverse musical styles, and create without geographical limitations.

Built by Musicians, Technologists & Innovators
Aiode was founded by a multidisciplinary team of artists, producers, AI specialists, and mathematicians united by a shared vision. Their mission is to evolve the future of music-making by merging human creativity with cutting-edge artificial intelligence.

Inspired by the Greek Muse Aoede
Aiode takes its name from Aoede, the ancient Greek muse of song and voice. This reflects the platform’s mission to ignite creativity and elevate artistic expression through advanced technology—while honoring the rich legacy and timeless spirit of music.

Recognized by Industry Leaders
Aiode’s innovative approach has earned praise from respected figures in the music world, including Andy Davies (former VP of Innovation at Universal Music Group) and acclaimed singer-songwriter Ivri Lider. Their support underscores Aiode’s commitment to ethical development and its potential to reshape the global music landscape.

Introducing the Aiode Desktop App
Aiode now offers a fully featured desktop application, giving creators the power to produce, shape performances, and export music directly from their computer. With AI-driven music generation, advanced control tools, and fast rendering, the app delivers a streamlined professional workflow suitable for both newcomers and seasoned producers.

Powered by Real Musical Expertise
Every Aiode model is crafted using recordings from highly skilled session musicians, all with more than 15 years of experience. This ensures that the performances generated by Aiode carry the richness, nuance, and authenticity of genuine, live musicianship.

Advancing Creativity Through AI
Aiode’s vision is to enhance—not replace—human creativity. By blending real artistic talent with intelligent technology, the platform empowers musicians to push creative boundaries, explore new ideas, and craft emotionally resonant music. Aiode represents a future where AI supports artists, fuels innovation, and celebrates the essence of human expression.

Users can browse virtual musicians by instrument, genre, and distinctive characteristics, including bassists, synth producers, trumpeters, and beat makers-all modeled after real artists available for virtual collaboration. These AI musicians will come up with performances aligned with the tempo, emotion, and structure of the uploaded track. Black Friday Mega Sale – Save Up to 59% on Aiode Plans! This Black Friday, unlock your creativity with Aiode and experience music production like never before. For a limited time, enjoy discounts of up to 59% on all annual and monthly plans — plus, get new AI model releases in November and December absolutely free! Black Friday Mega Deal When you upgrade during the Black Friday sale, you’ll gain instant access to all new Aiode model releases at no extra cost. Use Aiode’s AI-powered virtual musicians to compose, collaborate, and export unlimited high-quality tracks — from full STEMs to individual takes, ready for your DAW. Aiode Beta 2.0 is open to everyone, offering free access for newcomers and advanced tools for creators who want to expand their sound. Whether you’re a hobbyist, bedroom producer, or professional composer, Aiode provides everything you need to explore music without limits.. Creative sessions include multiple takes and variations that allow users to work with a rich palette of musical options to fine-tune their projects until they feel just right.

How does Aiode deliver higher audio fidelity than Suno?

Can I export my Aiode creations as WAV or MP3?

The technological backbone of Aiode leverages state-of-the-art machine learning techniques such as deep neural networks and reinforcement learning, enabling the virtual musicians to not only replicate but also improvise and adapt in ways that feel organic. These AI models analyze vast amounts of audio data, learning not only individual note sequences but also stylistic elements like phrasing, articulation, and emotional delivery. This multi-layered understanding allows virtual musicians to respond contextually to user input, crafting performances that feel less mechanical and more emotionally resonant.

  1. What’s the difference between Ai music and MIDI tools?
Future developments aim to incorporate real-time emotional analysis, where the AI can adjust its playing style dynamically based on the mood or energy level desired by the user.

What genres of music can Aiode create?

Security remains a paramount concern, with ongoing investments in advanced encryption, identity verification, and anomaly detection systems. Aiode's comprehensive risk management framework proactively addresses emerging cyber threats and safeguards user data and creative assets. Transparent communication and rigorous compliance audits underpin user confidence and foster a trustworthy digital environment.

What genres of music can Aiode create?
Is Aiode cloud-based or locally installed?

Is Aiode cloud-based or locally installed?

Continuously evolving, Aiode benefits from the support of strategic investors and a dedicated team of musicians, producers, AI specialists, and designers. Its mission is to enhance human creativity rather than replace it, offering tools that amplify artistic expression through advanced technology. The platform provides high-quality audio output-24-bit, 48kHz stereo WAV files-that users can export for further editing, mixing, or mastering in their preferred digital audio workstations. This combination of professional-grade standards and simplified workflows reduces technical barriers and fosters creativity.

How does Aiode promote ethical AI practices?

The collaboration with virtual musicians is modelled on real studio interactions. A selected musician, after listening to a track, interprets its tone and style and generates performances that complement it. Roles can be assigned for every song section-lead, rhythm, or improvisation-just as when mirroring organic musical conversation. A performance interface allows users to modify phrasing, intensity, and style, creating multiple takes to explore various moods or grooves. Virtual musicians can be combined or swapped freely, simulating everything from a live band to a full orchestral ensemble, all harmoniously synchronized.

How much can I save during the Black Friday sale?

Through strategic collaborations with leading cultural institutions, technology companies, and research organizations, Aiode continuously enriches its AI training datasets with rare and culturally significant music. This commitment ensures that virtual musicians reflect global musical diversity authentically and contribute to the preservation and revitalization of endangered musical traditions.



  • Blending Creativity with Code: The Aiode Story
How much can I save during the Black Friday sale?

 

Grooved side of the Voyager Golden Record launched aboard the Voyager probes to space, which features music from around the world

Music is the arrangement of sound to create some combination of form, harmony, melody, rhythm, or otherwise expressive content.[1][2][3] Music is generally agreed to be a cultural universal that is present in all human societies.[4] Definitions of music vary widely in substance and approach.[5] While scholars agree that music is defined by a small number of specific elements, there is no consensus as to what these necessary elements are.[6] Music is often characterized as a highly versatile medium for expressing human creativity.[7] Diverse activities are involved in the creation of music, and are often divided into categories of composition, improvisation, and performance.[8] Music may be performed using a wide variety of musical instruments, including the human voice. It can also be composed, sequenced, or otherwise produced to be indirectly played mechanically or electronically, such as via a music box, barrel organ, or digital audio workstation software on a computer.

Music often plays a key role in social events and religious ceremonies. The techniques of making music are often transmitted as part of a cultural tradition. Music is played in public and private contexts, highlighted at events such as festivals and concerts for various different types of ensembles. Music is used in the production of other media, such as in soundtracks to films, TV shows, operas, and video games.

Listening to music is a common means of entertainment. The culture surrounding music extends into areas of academic study, journalism, philosophy, psychology, and therapy. The music industry includes songwriters, performers, sound engineers, producers, tour organizers, distributors of instruments, accessories, and publishers of sheet music and recordings. Technology facilitating the recording and reproduction of music has historically included sheet music, microphones, phonographs, and tape machines, with playback of digital music being a common use for MP3 players, CD players, and smartphones.

Etymology and terminology

[edit]
In Greek mythology, the nine Muses were the inspiration for many creative endeavors, including the arts, and eventually became closely aligned with music specifically.

The modern English word music came into use in the 1630s.[9] It descends from Middle English musike, which in turn descends from Old French musique, then Latin mÅ«sica[10][7][n 1]—and ultimately from Ancient Greek mousiké technÄ“ (μουσική τέχνη), a phrase meaning "art of the Muses".[10][n 2] The Muses were nine deities in Ancient Greek mythology who presided over the arts and sciences.[13][14] They were included in tales by the earliest Western authors, Homer and Hesiod,[15] and eventually came to be associated with music specifically.[14] Over time, Polyhymnia would reside over music more prominently than the other muses.[11] The Latin word musica was also the originator for both the Spanish música and French musique via spelling and linguistic adjustment, though other European terms were probably loanwords, including the Italian musica, German Musik, Dutch muziek, Norwegian musikk, Polish muzyka and Russian muzïka.[14]

The modern Western world usually defines music as an all-encompassing term used to describe diverse genres, styles, and traditions.[16] This is not the case worldwide, and languages such as modern Indonesian (musik) and Shona (musakazo) have recently adopted words to reflect this universal conception, as they did not have words that fit exactly the Western scope.[14] Before Western contact in East Asia, neither Japan nor China had a single word that encompasses music in a broad sense, but culturally, they often regarded music in such a fashion.[17] The closest word to mean music in Chinese, yue, shares a character with le, meaning joy, and originally referred to all the arts before narrowing in meaning.[17] Africa is too diverse to make firm generalizations, but the musicologist J. H. Kwabena Nketia has emphasized African music's often inseparable connection to dance and speech in general.[18] Some African cultures, such as the Songye people of the Democratic Republic of the Congo and the Tiv people of Nigeria, have a strong and broad conception of 'music' but no corresponding word in their native languages.[18] Other words commonly translated as music often have more specific meanings in their respective cultures: the Hindi word for music, sangita, properly refers to art music,[19] while the many Indigenous languages of the Americas have words for music that refer specifically to song but describe instrumental music regardless.[20] Though the Arabic musiqi can refer to all music, it is usually used for instrumental and metric music, while khandan identifies vocal and improvised music.[21]

History

[edit]

Origins and prehistory

[edit]
Bone flute from Geissenklösterle, Germany, dated around c. 43,150–39,370 BP.[22]

It is often debated to what extent the origins of music will ever be understood,[23] and there are competing theories that aim to explain it.[24] Many scholars highlight a relationship between the origin of music and the origin of language, and there is disagreement surrounding whether music developed before, after, or simultaneously with language.[25] A similar source of contention surrounds whether music was the intentional result of natural selection or was a byproduct spandrel of evolution.[25] The earliest influential theory was proposed by Charles Darwin in 1871, who stated that music arose as a form of sexual selection, perhaps via mating calls.[26] Darwin's original perspective has been heavily criticized for its inconsistencies with other sexual selection methods,[27] though many scholars in the 21st century have developed and promoted the theory.[28] Other theories include that music arose to assist in organizing labor, improving long-distance communication, benefiting communication with the divine, assisting in community cohesion or as a defense to scare off predators.[29]

Prehistoric music can only be theorized based on findings from paleolithic archaeology sites. The disputed Divje Babe flute, a perforated cave bear femur, is at least 40,000 years old, though there is considerable debate surrounding whether it is truly a musical instrument or an object formed by animals.[30] The earliest objects whose designations as musical instruments are widely accepted are bone flutes from the Swabian Jura, Germany, namely from the Geissenklösterle, Hohle Fels and Vogelherd caves.[31] Dated to the Aurignacian (of the Upper Paleolithic) and used by Early European modern humans, from all three caves there are eight examples, four made from the wing bones of birds and four from mammoth ivory; three of these are near complete.[31] Three flutes from the Geissenklösterle are dated as the oldest, c. 43,150–39,370 BP.[22][n 3]

Antiquity

[edit]

The earliest material and representational evidence of Egyptian musical instruments dates to the Predynastic period, but the evidence is more securely attested in the Old Kingdom when harps, flutes and double clarinets were played.[32] Percussion instruments, lyres, and lutes were added to orchestras by the Middle Kingdom. Cymbals[33] frequently accompanied music and dance, much as they still do in Egypt today. Egyptian folk music, including the traditional Sufi dhikr rituals, are the closest contemporary music genre to ancient Egyptian music, having preserved many of its features, rhythms and instruments.[34][35]

The "Hurrian Hymn to Nikkal", found on clay tablets in the ancient Syrian city of Ugarit, is the oldest surviving notated work of music, dating back to approximately 1400 BCE.[36][37]

Music was an important part of social and cultural life in ancient Greece, in fact it was one of the main subjects taught to children. Musical education was considered important for the development of an individual's soul. Musicians and singers played a prominent role in Greek theater,[38] and those who received a musical education were seen as nobles and in perfect harmony (as can be read in the Republic, Plato). Mixed gender choruses performed for entertainment, celebration, and spiritual ceremonies.[39] Instruments included the double-reed aulos and a plucked string instrument, the lyre, principally a special kind called a kithara. Music was an important part of education, and boys were taught music starting at age six. Greek musical literacy created significant musical development. Greek music theory included the Greek musical modes, that eventually became the basis for Western religious and classical music. Later, influences from the Roman Empire, Eastern Europe, and the Byzantine Empire changed Greek music. The Seikilos epitaph is the oldest surviving example of a complete musical composition, including musical notation, from anywhere in the world.[40] The oldest surviving work written about music theory is Harmonika Stoicheia by Aristoxenus.[41]

Asian cultures

[edit]

Asian music covers a swath of music cultures surveyed in the articles on Arabia, Central Asia, East Asia, South Asia, and Southeast Asia. These traditions, with several reaching into antiquity, share historical ties through trade, religion and philosophical exchange, yet preserve their unique local aesthetics and performance styles.[42]

Indian women dressed in regional attire playing a variety of musical instruments popular in different parts of India

Indian classical music is one of the oldest musical traditions in the world.[43] Sculptures from the Indus Valley civilization show dance[44] and old musical instruments, like the seven-holed flute. Stringed instruments and drums have been recovered from Harappa and Mohenjo Daro by excavations carried out by Mortimer Wheeler.[45] The Rigveda, an ancient Hindu text, has elements of present Indian music, with musical notation to denote the meter and mode of chanting.[46] Indian classical music (marga) is monophonic, and based on a single melody line or raga rhythmically organized through talas. The poem Cilappatikaram provides information about how new scales can be formed by modal shifting of the tonic from an existing scale.[47] Present day Hindi music was influenced by Persian traditional music and Afghan Mughals. Carnatic music, popular in the southern states, is largely devotional; the majority of the songs are addressed to the Hindu deities. There are songs emphasizing love and other social issues.

Indonesia is the home of gong chime, there are variants across Indonesia, especially in Java and Bali.

Indonesian music has been formed since the Bronze Age culture migrated to the Indonesian archipelago in the 2nd-3rd centuries BCE. Indonesian traditional music uses percussion instruments, especially kendang and gongs. Some of them developed elaborate and distinctive instruments, such as the sasando stringed instrument on the island of Rote, the Sundanese angklung, and the complex and sophisticated Javanese and Balinese gamelan orchestras. Indonesia is the home of gong chime, a general term for a set of small, high pitched pot gongs. Gongs are usually placed in order of note, with the boss up on a string held in a low wooden frame. The most popular form of Indonesian music is gamelan, an ensemble of tuned percussion instruments that include metallophones, drums, gongs and spike fiddles along with bamboo suling (like a flute).[48][49]

Chinese classical music, the traditional art or court music of China, has a history stretching over about 3,000 years. It has its own unique systems of musical notation, as well as musical tuning and pitch, musical instruments and styles or genres. Chinese music is pentatonic-diatonic, having a scale of twelve notes to an octave (5 + 7 = 12) as does European-influenced music.[50]

Western classical

[edit]

Early music

[edit]
Musical notation from a Catholic Missal, c. 1310–1320

The medieval music era (500 to 1400), which took place during the Middle Ages, started with the introduction of monophonic (single melodic line) chanting into Catholic Church services. Musical notation was used since ancient times in Greek culture, but in the Middle Ages, notation was first introduced by the Catholic Church, so chant melodies could be written down, to facilitate the use of the same melodies for religious music across the Catholic empire. The only European Medieval repertory that has been found, in written form, from before 800 is the monophonic liturgical plainsong chant of the Catholic Church, the central tradition of which was called Gregorian chant. Alongside these traditions of sacred and church music there existed a vibrant tradition of secular song (non-religious songs). Examples of composers from this period are Léonin, Pérotin, Guillaume de Machaut, and Walther von der Vogelweide.[51][52][53][54]

Renaissance music (c. 1400 to 1600) was more focused on secular themes, such as courtly love. Around 1450, the printing press was invented, which made printed sheet music much less expensive and easier to mass-produce (prior to the invention of the press, all notated music was hand-copied). The increased availability of sheet music spread musical styles quicker and across a larger area. Musicians and singers often worked for the church, courts and towns. Church choirs grew in size, and the church remained an important patron of music. By the middle of the 15th century, composers wrote richly polyphonic sacred music, in which different melody lines were interwoven simultaneously. Prominent composers from this era include Guillaume Du Fay, Giovanni Pierluigi da Palestrina, Thomas Morley, Orlando di Lasso and Josquin des Prez. As musical activity shifted from the church to aristocratic courts, kings, queens and princes competed for the finest composers. Many leading composers came from the Netherlands, Belgium, and France; they are called the Franco-Flemish composers.[55] They held important positions throughout Europe, especially in Italy. Other countries with vibrant musical activity included Germany, England, and Spain.

Common practice period

[edit]
Baroque
[edit]
J. S. Bach

The Baroque era of music took place from 1600 to 1750, coinciding with the flourishing of the Baroque artistic style in Europe. The start of the Baroque era was marked by the penning of the first operas. Polyphonic contrapuntal music (music with separate, simultaneous melodic lines) remained important during this period. German Baroque composers wrote for small ensembles including strings, brass, and woodwinds, as well as for choirs and keyboard instruments such as pipe organ, harpsichord, and clavichord. Musical complexity increased during this time. Several major musical forms were created, some of them which persisted into later periods, seeing further development. These include the fugue, the invention, the sonata, and the concerto.[56] The late Baroque style was polyphonically complex and richly ornamented. Important composers from the Baroque era include Johann Sebastian Bach (Cello suites), George Frideric Handel (Messiah), Georg Philipp Telemann and Antonio Vivaldi (The Four Seasons).

Classicism
[edit]
Wolfgang Amadeus Mozart was a prolific and influential composer of the Classical period.

The music of the Classical period (1730 to 1820) aimed to imitate what were seen as the key elements of the art and philosophy of Ancient Greece and Rome: the ideals of balance, proportion and disciplined expression. (Note: the music from the Classical period should not be confused with Classical music in general, a term which refers to Western art music from the 5th century to the 2000s, which includes the Classical period as one of a number of periods). Music from the Classical period has a lighter, clearer and considerably simpler texture than the Baroque music which preceded it. The main style was homophony,[57] where a prominent melody and a subordinate chordal accompaniment part are clearly distinct. Classical instrumental melodies tended to be almost voicelike and singable. New genres were developed, and the fortepiano, the forerunner to the modern piano, replaced the Baroque era harpsichord and pipe organ as the main keyboard instrument (though pipe organ continued to be used in sacred music, such as Masses).

Importance was given to instrumental music. It was dominated by further development of musical forms initially defined in the Baroque period: the sonata, the concerto, and the symphony. Other main kinds were the trio, string quartet, serenade and divertimento. The sonata was the most important and developed form. Although Baroque composers also wrote sonatas, the Classical style of sonata is completely distinct. All of the main instrumental forms of the Classical era, from string quartets to symphonies and concertos, were based on the structure of the sonata. The instruments used chamber music and orchestra became more standardized. In place of the basso continuo group of the Baroque era, which consisted of harpsichord, organ or lute along with a number of bass instruments selected at the discretion of the group leader (e.g., viol, cello, theorbo, serpent), Classical chamber groups used specified, standardized instruments (e.g., a string quartet would be performed by two violins, a viola and a cello). The practice of improvised chord-playing by the continuo keyboardist or lute player, a hallmark of Baroque music, underwent a gradual decline between 1750 and 1800.[58]

One of the most important changes made in the Classical period was the development of public concerts. The aristocracy still played a significant role in the sponsorship of concerts and compositions, but it was now possible for composers to survive without being permanent employees of queens or princes. The increasing popularity of classical music led to a growth in the number and types of orchestras. The expansion of orchestral concerts necessitated the building of large public performance spaces. Symphonic music including symphonies, musical accompaniment to ballet and mixed vocal/instrumental genres, such as opera and oratorio, became more popular.[59][60][61]

The best known composers of Classicism are Carl Philipp Emanuel Bach, Christoph Willibald Gluck, Johann Christian Bach, Joseph Haydn, Wolfgang Amadeus Mozart, Ludwig van Beethoven and Franz Schubert. Beethoven and Schubert are also considered to be composers in the later part of the Classical era, as it began to move towards Romanticism.

Romanticism
[edit]
The piano was the centrepiece of social activity for middle-class urbanites in the 19th century (Moritz von Schwind, 1868). The man at the piano is composer Franz Schubert.

Romantic music (c. 1820 to 1900) from the 19th century had many elements in common with the Romantic styles in literature and painting of the era. Romanticism was an artistic, literary, and intellectual movement was characterized by its emphasis on emotion and individualism as well as glorification of all the past and nature. Romantic music expanded beyond the rigid styles and forms of the Classical era into more passionate, dramatic expressive pieces and songs. Romantic composers such as Wagner and Brahms attempted to increase emotional expression and power in their music to describe deeper truths or human feelings. With symphonic tone poems, composers tried to tell stories and evoke images or landscapes using instrumental music. Some composers promoted nationalistic pride with patriotic orchestral music inspired by folk music. The emotional and expressive qualities of music came to take precedence over tradition.[62]

Romantic composers grew in idiosyncrasy, and went further in the syncretism of exploring different art-forms in a musical context, (such as literature), history (historical figures and legends), or nature itself. Romantic love or longing was a prevalent theme in many works composed during this period. In some cases, the formal structures from the classical period continued to be used (e.g., the sonata form used in string quartets and symphonies), but these forms were expanded and altered. In many cases, new approaches were explored for existing genres, forms, and functions. Also, new forms were created that were deemed better suited to the new subject matter. Composers continued to develop opera and ballet music, exploring new styles and themes.[38]

In the years after 1800, the music developed by Ludwig van Beethoven and Franz Schubert introduced a more dramatic, expressive style. In Beethoven's case, short motifs, developed organically, came to replace melody as the most significant compositional unit (an example is the distinctive four note figure used in his Fifth Symphony). Later Romantic composers such as Pyotr Ilyich Tchaikovsky, Antonín DvoÃ…â„¢ák, and Gustav Mahler used more unusual chords and more dissonance to create dramatic tension. They generated complex and often much longer musical works. During the late Romantic period, composers explored dramatic chromatic alterations of tonality, such as extended chords and altered chords, which created new sound "colors." The late 19th century saw a dramatic expansion in the size of the orchestra, and the Industrial Revolution helped to create better instruments, creating a more powerful sound. Public concerts became an important part of well-to-do urban society. It also saw a new diversity in theatre music, including operetta, and musical comedy and other forms of musical theatre.[38]

20th and 21st century

[edit]
Landman's 2006 Moodswinger, a 3rd-bridged overtone zither and an example of experimental musical instruments

In the 19th century, a key way new compositions became known to the public was by the sales of sheet music, which middle class amateur music lovers would perform at home, on their piano or other common instruments, such as the violin. With 20th-century music, the invention of new electric technologies such as radio broadcasting and mass market availability of gramophone records meant sound recordings heard by listeners (on the radio or record player) became the main way to learn about new songs and pieces.[63] There was a vast increase in music listening as the radio gained popularity and phonographs were used to replay and distribute music; anyone with a radio or record player could hear operas, symphonies and big bands in their own living room. During the 19th century, the focus on sheet music had restricted access to new music to middle and upper-class people who could read music and who owned pianos and other instruments. Radios and record players allowed lower-income people, who could not afford an opera or symphony concert ticket, to hear this music. As well, people could hear music from different parts of the country, or even different parts of the world, even if they could not afford to travel to these locations. This helped to spread musical styles.[64]

The focus of art music in the 20th century was characterized by exploration of new rhythms, styles, and sounds. The horrors of World War I influenced many of the arts, including music, and composers began exploring darker, harsher sounds. Traditional music styles such as jazz and folk music were used by composers as a source of ideas for classical music. Igor Stravinsky, Arnold Schoenberg, and John Cage were influential composers in 20th-century art music. The invention of sound recording and the ability to edit music gave rise to new subgenres of classical music, including the acousmatic[65] and musique concrète schools of electronic composition. Sound recording was a major influence on the development of popular music genres, because it enabled recordings of songs and bands to be widely distributed. The introduction of the multitrack recording system had a major influence on rock music, because it could do more than record a band's performance. Using a multitrack system, a band and their music producer could overdub many layers of instrument tracks and vocals, creating new sounds that would not be possible in a live performance.[66][67]

Jazz evolved and became an important genre of music over the course of the 20th century, and during the second half, rock music did the same. Jazz is an American musical artform that originated in the late 19th and early 20th centuries in African American communities in the Southern United States from a confluence of African and European music traditions. The style's West African pedigree is evident in its use of blue notes, improvisation, polyrhythms, syncopation, and the swung note.[68]

A selection of guitars and amps at Apple Music Row

Rock music is a genre of popular music that developed in the 1950s from rock and roll, rockabilly, blues, and country music.[69] The sound of rock often revolves around the electric or acoustic guitar, and it uses a strong back beat laid down by a rhythm section. Along with the guitar or keyboards, saxophone and blues-style harmonica are used as soloing instruments. In its "purest form", it "has three chords, a strong, insistent back beat, and a catchy melody."[70] The traditional rhythm section for popular music is rhythm guitar, electric bass guitar, drums. Some bands have keyboard instruments such as organ, piano, or, since the 1970s, analog synthesizers. In the 1980s, pop musicians began using digital synthesizers, such as the DX-7 synthesizer, electronic drum machines such as the TR-808 and synth bass devices (such as the TB-303) or synth bass keyboards. In the 1990s, an increasingly large range of computerized hardware musical devices and instruments and software (e.g. digital audio workstations) were used. In the 2020s, soft synths and computer music apps make it possible for bedroom producers to create and record types of music, such as electronic dance music, in their home, adding sampled and digital instruments and editing the recording digitally. In the 1990s, bands in genres such as nu metal began including DJs in their bands. DJs create music by manipulating recorded music, using a DJ mixer.[71][72]

Creation

[edit]

Composition

[edit]
French Baroque music composer Michel Richard Delalande (1657–1726), pen in hand
People composing music in 2013 using electronic keyboards and computers

"Composition" is the act or practice of creating a song, an instrumental music piece, a work with both singing and instruments, or another type of music. In many cultures, including Western classical music, the act of composing also includes the creation of music notation, such as a sheet music "score", which is then performed by the composer or by other singers or musicians. In popular music and traditional music, the act of composing, which is typically called songwriting, may involve the creation of a basic outline of the song, called the lead sheet, which sets out the melody, lyrics and chord progression. In classical music, the composer typically orchestrates his or her own compositions, but in musical theatre and in pop music, songwriters may hire an arranger to do the orchestration. In some cases, a songwriter may not use notation at all, and instead, compose the song in her mind and then play or record it from memory. In jazz and popular music, notable recordings by influential performers are given the weight that written scores play in classical music.[73][74]

Even when music is notated relatively precisely, as in classical music, there are many decisions that a performer has to make, because notation does not specify all of the elements of music precisely. The process of deciding how to perform music that has been previously composed and notated is termed "interpretation". Different performers' interpretations of the same work of music can vary widely, in terms of the tempos that are chosen and the playing or singing style or phrasing of the melodies. Composers and songwriters who present their own music are interpreting their songs, just as much as those who perform the music of others. The standard body of choices and techniques present at a given time and a given place is referred to as performance practice, whereas interpretation is generally used to mean the individual choices of a performer.[75]

Although a musical composition often uses musical notation and has a single author, this is not always the case. A work of music can have multiple composers, which often occurs in popular music when a band collaborates to write a song, or in musical theatre, when one person writes the melodies, a second person writes the lyrics, and a third person orchestrates the songs. In some styles of music, such as the blues, a composer/songwriter may create, perform and record new songs or pieces without ever writing them down in music notation. A piece of music can also be composed with words, images, or computer programs that explain or notate how the singer or musician should create musical sounds. Examples range from avant-garde music that uses graphic notation, to text compositions such as Aus den sieben Tagen, to computer programs that select sounds for musical pieces. Music that makes heavy use of randomness and chance is called aleatoric music,[76] and is associated with contemporary composers active in the 20th century, such as John Cage, Morton Feldman, and Witold Lutosławski. A commonly known example of chance-based music is the sound of wind chimes jingling in a breeze.

The study of composition has traditionally been dominated by examination of methods and practice of Western classical music, but the definition of composition is broad enough to include the creation of popular music and traditional music songs and instrumental pieces as well as spontaneously improvised works like those of free jazz performers and African percussionists such as Ewe drummers.

Performance

[edit]
Chinese Naxi musicians
Assyrians playing zurna and Davul, instruments that go back thousands of years

Performance is the physical expression of music, which occurs when a song is sung or piano piece, guitar melody, symphony, drum beat or other musical part is played. In classical music, a work is written in music notation by a composer and then performed once the composer is satisfied with its structure and instrumentation. However, as it gets performed, the interpretation of a song or piece can evolve and change. In classical music, instrumental performers, singers or conductors may gradually make changes to the phrasing or tempo of a piece. In popular and traditional music, the performers have more freedom to make changes to the form of a song or piece. As such, in popular and traditional music styles, even when a band plays a cover song, they can make changes such as adding a guitar solo or inserting an introduction.[77]

A performance can either be planned out and rehearsed (practiced)—which is the norm in classical music, jazz big bands, and many popular music styles–or improvised over a chord progression (a sequence of chords), which is the norm in small jazz and blues groups. Rehearsals of orchestras, concert bands and choirs are led by a conductor. Rock, blues and jazz bands are usually led by the bandleader. A rehearsal is a structured repetition of a song or piece by the performers until it can be sung or played correctly and, if it is a song or piece for more than one musician, until the parts are together from a rhythmic and tuning perspective.

Many cultures have strong traditions of solo performance (in which one singer or instrumentalist performs), such as in Indian classical music, and in the Western art-music tradition. Other cultures, such as in Bali, include strong traditions of group performance. All cultures include a mixture of both, and performance may range from improvised solo playing to highly planned and organized performances such as the modern classical concert, religious processions, classical music festivals or music competitions. Chamber music, which is music for a small ensemble with only one or a few of each type of instrument, is often seen as more intimate than large symphonic works.[78]

Improvisation

[edit]

Musical improvisation is the creation of spontaneous music, often within (or based on) a pre-existing harmonic framework, chord progression, or riffs. Improvisers use the notes of the chord, various scales that are associated with each chord, and chromatic ornaments and passing tones which may be neither chord tones nor from the typical scales associated with a chord. Musical improvisation can be done with or without preparation. Improvisation is a major part of some types of music, such as blues, jazz, and jazz fusion, in which instrumental performers improvise solos, melody lines, and accompaniment parts.[79]

In the Western art music tradition, improvisation was an important skill during the Baroque era and during the Classical era. In the Baroque era, performers improvised ornaments, and basso continuo keyboard players improvised chord voicings based on figured bass notation. As well, the top soloists were expected to be able to improvise pieces such as preludes. In the Classical era, solo performers and singers improvised virtuoso cadenzas during concerts.

However, in the 20th and early 21st century, as "common practice" Western art music performance became institutionalized in symphony orchestras, opera houses, and ballets, improvisation has played a smaller role, as more and more music was notated in scores and parts for musicians to play. At the same time, some 20th and 21st century art music composers have increasingly included improvisation in their creative work. In Indian classical music, improvisation is a core component and an essential criterion of performances.

Art and entertainment

[edit]
Khatia Buniatishvili playing a grand piano

Music is composed and performed for many purposes, ranging from aesthetic pleasure, religious or ceremonial purposes, or as an entertainment product for the marketplace. When music was only available through sheet music scores, such as during the Classical and Romantic eras, music lovers would buy the sheet music of their favourite pieces and songs so that they could perform them at home on the piano. With the advent of the phonograph, records of popular songs, rather than sheet music became the dominant way that music lovers would enjoy their favourite songs. With the advent of home tape recorders in the 1980s and digital music in the 1990s, music lovers could make tapes or playlists of favourite songs and take them with them on a portable cassette player or MP3 player. Some music lovers create mix tapes of favourite songs, which serve as a "self-portrait, a gesture of friendship, prescription for an ideal party... [and] an environment consisting solely of what is most ardently loved".[80]

Amateur musicians can compose or perform music for their own pleasure and derive income elsewhere. Professional musicians are employed by institutions and organisations, including armed forces (in marching bands, concert bands and popular music groups), religious institutions, symphony orchestras, broadcasting or film production companies, and music schools. Professional musicians sometimes work as freelancers or session musicians, seeking contracts and engagements in a variety of settings. There are often many links between amateur and professional musicians. Beginning amateur musicians take lessons with professional musicians. In community settings, advanced amateur musicians perform with professional musicians in a variety of ensembles such as community concert bands and community orchestras.

A distinction is often made between music performed for a live audience and music that is performed in a studio so that it can be recorded and distributed through the music retail system or the broadcasting system. However, there are also many cases where a live performance in front of an audience is also recorded and distributed. Live concert recordings are popular in both classical music and in popular music forms such as rock, where illegally taped live concerts are prized by music lovers. In the jam band scene, live, improvised jam sessions are preferred to studio recordings.[81]

Notation

[edit]
Sheet music is a written representation of music. Homorhythmic (i.e., hymn-style) arrangement of the traditional "Adeste Fideles" in standard two-staff format for mixed voices. playⓘ

Music notation typically means the written expression of music notes and rhythms on paper using symbols. When music is written down, the pitches and rhythm of the music, such as the notes of a melody, are notated. Music notation often provides instructions on how to perform the music. For example, the sheet music for a song may state the song is a "slow blues" or a "fast swing", which indicates the tempo and the genre. To read notation, a person must have an understanding of music theory, harmony and the performance practice associated with a particular song or piece's genre.

Written notation varies with the style and period of music. Nowadays, notated music is produced as sheet music or, for individuals with computer scorewriter programs, as an image on a computer screen. In ancient times, music notation was put onto stone or clay tablets.[37] To perform music from notation, a singer or instrumentalist requires an understanding of the rhythmic and pitch elements embodied in the symbols and the performance practice that is associated with a piece of music or genre. In genres requiring musical improvisation, the performer often plays from music where only the chord changes and form of the song are written, requiring the performer to have a great understanding of the music's structure, harmony and the styles of a particular genre e.g., jazz or country music.

In Western art music, the most common types of written notation are scores, which include all the music parts of an ensemble piece, and parts, which are the music notation for the individual performers or singers. In popular music, jazz, and blues, the standard musical notation is the lead sheet, which notates the melody, chords, lyrics (if it is a vocal piece), and structure of the music. Fake books are also used in jazz; they may consist of lead sheets or simply chord charts, which permit rhythm section members to improvise an accompaniment part to jazz songs. Scores and parts are also used in popular music and jazz, particularly in large ensembles such as jazz "big bands." In popular music, guitarists and electric bass players often read music notated in tablature (often abbreviated as "tab"), which indicates the location of the notes to be played on the instrument using a diagram of the guitar or bass fingerboard. Tablature was used in the Baroque era to notate music for the lute, a stringed, fretted instrument.[82]

Oral and aural tradition

[edit]

Many types of music, such as traditional blues and folk music were not written down in sheet music; instead, they were originally preserved in the memory of performers, and the songs were handed down orally, from one musician or singer to another, or aurally, in which a performer learns a song "by ear". When the composer of a song or piece is no longer known, this music is often classified as "traditional" or as a "folk song". Different musical traditions have different attitudes towards how and where to make changes to the original source material, from quite strict, to those that demand improvisation or modification to the music. A culture's history and stories may also be passed on by ear through song.[83]

Elements

[edit]

Music has many different fundamentals or elements. Depending on the definition of "element" being used, these can include pitch, beat or pulse, tempo, rhythm, melody, harmony, texture, style, allocation of voices, timbre or color, dynamics, expression, articulation, form, and structure. The elements of music feature prominently in the music curriculums of Australia, the UK, and the US. All three curriculums identify pitch, dynamics, timbre, and texture as elements, but the other identified elements of music are far from universally agreed upon. Below is a list of the three official versions of the "elements of music":

  • Australia: pitch, timbre, texture, dynamics and expression, rhythm, form and structure.[84]
  • UK: pitch, timbre, texture, dynamics, duration, tempo, structure.[85]
  • USA: pitch, timbre, texture, dynamics, rhythm, form, harmony, style/articulation.[86]

In relation to the UK curriculum, in 2013 the term: "appropriate musical notations" was added to their list of elements and the title of the list was changed from the "elements of music" to the "inter-related dimensions of music". The inter-related dimensions of music are listed as: pitch, duration, dynamics, tempo, timbre, texture, structure, and appropriate musical notations.[87]

The phrase "the elements of music" is used in a number of different contexts. The two most common contexts can be differentiated by describing them as the "rudimentary elements of music" and the "perceptual elements of music".[n 4]

Pitch

[edit]

Pitch is an aspect of a sound that we can hear, reflecting whether one musical sound, note, or tone is "higher" or "lower" than another musical sound, note, or tone. We can talk about the highness or lowness of pitch in the more general sense, such as the way a listener hears a piercingly high piccolo note or whistling tone as higher in pitch than a deep thump of a bass drum. We also talk about pitch in the precise sense associated with musical melodies, basslines and chords. Precise pitch can only be determined in sounds that have a frequency that is clear and stable enough to distinguish from noise. For example, it is much easier for listeners to discern the pitch of a single note played on a piano than to try to discern the pitch of a crash cymbal that is struck.[92]

Melody

[edit]
The melody to the traditional song "Pop Goes the Weasel" playⓘ

A melody, also called a "tune", is a series of pitches (notes) sounding in succession (one after the other), often in a rising and falling pattern. The notes of a melody are typically created using pitch systems such as scales or modes. Melodies also often contain notes from the chords used in the song. The melodies in simple folk songs and traditional songs may use only the notes of a single scale, the scale associated with the tonic note or key of a given song. For example, a folk song in the key of C (also referred to as C major) may have a melody that uses only the notes of the C major scale (the individual notes C, D, E, F, G, A, B, and C; these are the "white notes" on a piano keyboard. On the other hand, Bebop-era jazz from the 1940s and contemporary music from the 20th and 21st centuries may use melodies with many chromatic notes (i.e., notes in addition to the notes of the major scale; on a piano, a chromatic scale would include all the notes on the keyboard, including the "white notes" and "black notes" and unusual scales, such as the whole tone scale (a whole tone scale in the key of C would contain the notes C, D, E, F♯, G♯ and A♯). A low musical line played by bass instruments, such as double bass, electric bass, or tuba, is called a bassline.[93]

Harmony

[edit]
A player performing a chord (combination of many different notes) on a guitar

Harmony refers to the "vertical" sounds of pitches in music, which means pitches that are played or sung together at the same time creates a chord. Usually, this means the notes are played at the same time, although harmony may also be implied by a melody that outlines a harmonic structure (i.e., by using melody notes that are played one after the other, outlining the notes of a chord). In music written using the system of major-minor tonality ("keys"), which includes most classical music written from 1600 to 1900 and most Western pop, rock, and traditional music, the key of a piece determines the "home note" or tonic to which the piece generally resolves, and the character (e.g. major or minor) of the scale in use. Simple classical pieces and many pop and traditional music songs are written so that all the music is in a single key. More complex Classical, pop, and traditional music songs and pieces may have two keys (and in some cases three or more keys). Classical music from the Romantic era (written from about 1820–1900) often contains multiple keys,[94] as does jazz, especially Bebop jazz from the 1940s, in which the key or "home note" of a song may change every four bars or even every two bars.[95]

Rhythm

[edit]

Rhythm is the arrangement of sounds and silences in time. Meter animates time in regular pulse groupings, called measures or bars, which in Western classical, popular, and traditional music often group notes in sets of two (e.g., 2/4 time), three (e.g., 3/4 time, also known as Waltz time, or 3/8 time), or four (e.g., 4/4 time). Meters are made easier to hear because songs and pieces often (but not always) place an emphasis on the first beat of each grouping. Notable exceptions exist, such as the backbeat used in much Western pop and rock, in which a song that uses a measure that consists of four beats (called 4/4 time or common time) will have accents on beats two and four, which are typically performed by the drummer on the snare drum, a loud and distinctive-sounding percussion instrument. In pop and rock, the rhythm parts of a song are played by the rhythm section, which includes chord-playing instruments (e.g., electric guitar, acoustic guitar, piano, or other keyboard instruments), a bass instrument (typically electric bass or for some styles such as jazz and bluegrass, double bass) and a drum kit player.[96]

Texture

[edit]

Musical texture is the overall sound of a piece of music or song. The texture of a piece or song is determined by how the melodic, rhythmic, and harmonic materials are combined in a composition, thus determining the overall nature of the sound in a piece. Texture is often described in regard to the density, or thickness, and range, or width, between lowest and highest pitches, in relative terms as well as more specifically distinguished according to the number of voices, or parts, and the relationship between these voices (see common types below). For example, a thick texture contains many 'layers' of instruments. One layer can be a string section or another brass. The thickness is affected by the amount and the richness of the instruments.[97] Texture is commonly described according to the number of and relationship between parts or lines of music:

  • monophony: a single melody (or "tune") with neither instrumental accompaniment nor a harmony part. A mother singing a lullaby to her baby would be an example.
  • heterophony: two or more instruments or singers playing/singing the same melody, but with each performer slightly varying the rhythm or speed of the melody or adding different ornaments to the melody. Two bluegrass fiddlers playing the same traditional fiddle tune together will typically each vary the melody by some degree and each add different ornaments.
  • polyphony: multiple independent melody lines that interweave together, which are sung or played at the same time. Choral music written in the Renaissance music era was typically written in this style. A round, which is a song such as "Row, Row, Row Your Boat", which different groups of singers all start to sing at a different time, is an example of polyphony.
  • homophony: a clear melody supported by chordal accompaniment. Most Western popular music songs from the 19th century onward are written in this texture.

Music that contains a large number of independent parts (e.g., a double concerto accompanied by 100 orchestral instruments with many interweaving melodic lines) is generally said to have a "thicker" or "denser" texture than a work with few parts (e.g., a solo flute melody accompanied by a single cello).

Timbre

[edit]
Spectrogram of the first second of an E9 suspended chord played on a Fender Stratocaster guitar. Below is the E9 suspended chord audio:

Timbre, sometimes called "color" or "tone color" is the quality or sound of a voice or instrument.[98] Timbre is what makes a particular musical sound different from another, even when they have the same pitch and loudness. For example, a 440 Hz A note sounds different when it is played on oboe, piano, violin, or electric guitar. Even if different players of the same instrument play the same note, their notes might sound different due to differences in instrumental technique (e.g., different embouchures), different types of accessories (e.g., mouthpieces for brass players, reeds for oboe and bassoon players) or strings made out of different materials for string players (e.g., gut strings versus steel strings). Even two instrumentalists playing the same note on the same instrument (one after the other) may sound different due to different ways of playing the instrument (e.g., two string players might hold the bow differently).

The physical characteristics of sound that determine the perception of timbre include the spectrum, envelope, and overtones of a note or musical sound. For electric instruments developed in the 20th century, such as electric guitar, electric bass and electric piano, the performer can also change the tone by adjusting equalizer controls, tone controls on the instrument, and by using electronic effects units such as distortion pedals. The tone of the electric Hammond organ is controlled by adjusting drawbars.

Expression

[edit]

Expressive qualities are those elements in music that create change in music without changing the main pitches or substantially changing the rhythms of the melody and its accompaniment. Performers, including singers and instrumentalists, can add musical expression to a song or piece by adding phrasing, by adding effects such as vibrato (with voice and some instruments, such as guitar, violin, brass instruments, and woodwinds), dynamics (the loudness or softness of piece or a section of it), tempo fluctuations (e.g., ritardando or accelerando, which are, respectively slowing down and speeding up the tempo), by adding pauses or fermatas on a cadence, and by changing the articulation of the notes (e.g., making notes more pronounced or accented, by making notes more legato, which means smoothly connected, or by making notes shorter).

Expression is achieved through the manipulation of pitch (such as inflection, vibrato, slides etc.), volume (dynamics, accent, tremolo etc.), duration (tempo fluctuations, rhythmic changes, changing note duration such as with legato and staccato, etc.), timbre (e.g. changing vocal timbre from a light to a resonant voice) and sometimes even texture (e.g. doubling the bass note for a richer effect in a piano piece). Expression therefore can be seen as a manipulation of all elements to convey "an indication of mood, spirit, character etc."[99] and as such cannot be included as a unique perceptual element of music,[100] although it can be considered an important rudimentary element of music.

Form

[edit]
Sheet music notation for the chorus (refrain) of the Christmas song "Jingle Bells" Jingle Bells refrain vector.midⓘ

In music, form describes the overall structure or plan of a song or piece of music,[101] and it describes the layout of a composition as divided into sections.[102] In the early 20th century, Tin Pan Alley songs and Broadway musical songs were often in AABA thirty-two-bar form, in which the A sections repeated the same eight bar melody (with variation) and the B section provided a contrasting melody or harmony for eight bars. From the 1960s onward, Western pop and rock songs are often in verse-chorus form, which comprises a sequence of verse and chorus ("refrain") sections, with new lyrics for most verses and repeating lyrics for the choruses. Popular music often makes use of strophic form, sometimes in conjunction with the twelve bar blues.[103]

In the tenth edition of The Oxford Companion to Music, Percy Scholes defines musical form as "a series of strategies designed to find a successful mean between the opposite extremes of unrelieved repetition and unrelieved alteration."[104] Examples of common forms of Western music include the fugue, the invention, sonata-allegro, canon, strophic, theme and variations, and rondo.

Scholes states that European classical music had only six stand-alone forms: simple binary, simple ternary, compound binary, rondo, air with variations, and fugue (although musicologist Alfred Mann emphasized that the fugue is primarily a method of composition that has sometimes taken on certain structural conventions.[105])

Where a piece cannot readily be broken into sectional units (though it might borrow some form from a poem, story or programme), it is said to be through-composed. Such is often the case with a fantasia, prelude, rhapsody, etude (or study), symphonic poem, Bagatelle, impromptu or similar composition.[106] Professor Charles Keil classified forms and formal detail as "sectional, developmental, or variational."[107]

Philosophy

[edit]
The Woman in Red by Giovanni Boldini

The philosophy of music is the study of fundamental questions regarding music and has connections with questions in metaphysics and aesthetics. Questions include:

  • What is the definition of music? (What are the necessary and sufficient conditions for classifying something as music?)
  • What is the relationship between music and mind?
  • What does music history reveal to us about the world?
  • What is the connection between music and emotions?
  • What is meaning in relation to music?

In ancient times, such as with the Ancient Greeks, the aesthetics of music explored the mathematical and cosmological dimensions of rhythmic and harmonic organization. In the 18th century, focus shifted to the experience of hearing music, and thus to questions about its beauty and human enjoyment (plaisir and jouissance) of music. The origin of this philosophic shift is sometimes attributed to Alexander Gottlieb Baumgarten in the 18th century, followed by Immanuel Kant. Through their writing, the ancient term 'aesthetics', meaning sensory perception, received its present-day connotation. In the 2000s, philosophers have tended to emphasize issues besides beauty and enjoyment. For example, music's capacity to express emotion has been foregrounded.[108]

In the 20th century, important contributions were made by Peter Kivy, Jerrold Levinson, Roger Scruton, and Stephen Davies. However, many musicians, music critics, and other non-philosophers have contributed to the aesthetics of music. In the 19th century, a significant debate arose between Eduard Hanslick, a music critic and musicologist, and composer Richard Wagner regarding whether music can express meaning. Harry Partch and some other musicologists, such as Kyle Gann, have studied and tried to popularize microtonal music and the usage of alternate musical scales. Modern composers like La Monte Young, Rhys Chatham and Glenn Branca paid much attention to a scale called just intonation.[109][110][111]

It is often thought that music has the ability to affect our emotions, intellect, and psychology; it can assuage our loneliness or incite our passions. Plato suggests in The Republic that music has a direct effect on the soul. Therefore, he proposes that in the ideal regime music would be closely regulated by the state (Book VII).[112] In Ancient China, Confucius believed that music and rituals or rites are interconnected and harmonious with nature; he stated that music was the harmonization of heaven and earth, while the order was brought by the rites order, making them extremely crucial functions in society.[113]

Psychology

[edit]

Modern music psychology aims to explain and understand musical behavior and experience.[114] Research in this field and its subfields are primarily empirical; their knowledge tends to advance on the basis of interpretations of data collected by systematic observation of and interaction with human participants. In addition to its focus on fundamental perceptions and cognitive processes, music psychology is a field of research with practical relevance for many areas, including music performance, composition, education, criticism, and therapy, as well as investigations of human aptitude, skill, intelligence, creativity, and social behavior.

Neuroscience

[edit]
The primary auditory cortex is one of the main areas associated with superior pitch resolution.

Cognitive neuroscience of music is the scientific study of brain-based mechanisms involved in the cognitive processes underlying music. These behaviours include music listening, performing, composing, reading, writing, and ancillary activities. It also is increasingly concerned with the brain basis for musical aesthetics and musical emotion. The field is distinguished by its reliance on direct observations of the brain, using such techniques as functional magnetic resonance imaging (fMRI), transcranial magnetic stimulation (TMS), magnetoencephalography (MEG), electroencephalography (EEG), and positron emission tomography (PET).

Cognitive musicology

[edit]

Cognitive musicology is a branch of cognitive science concerned with computationally modeling musical knowledge with the goal of understanding both music and cognition.[115] The use of computer models provides an exacting, interactive medium in which to formulate and test theories and has roots in artificial intelligence and cognitive science.[116]

Cognitive musicology investigates topics such as the parallels between language and music in the brain. Research often includes biologically inspired models of computation, such as neural networks and evolutionary programs.[117] This field seeks to model how musical knowledge is represented, stored, perceived, performed, and generated. By using a well-structured computer environment, the systematic structures of these cognitive phenomena can be investigated.[118]

Psychoacoustics

[edit]

Psychoacoustics is the scientific study of sound perception. More specifically, it is the branch of science studying the psychological and physiological responses associated with sound (including speech and music). It can be further categorized as a branch of psychophysics.

Evolutionary musicology

[edit]

Evolutionary musicology concerns the "origins of music, the question of animal song, selection pressures underlying music evolution", and "music evolution and human evolution".[119] It seeks to understand music perception and activity in the context of evolutionary theory. Charles Darwin speculated that music may have held an adaptive advantage and functioned as a protolanguage,[120] a view which has spawned several competing theories of music evolution.[121][122][page needed][123] An alternate view sees music as a by-product of linguistic evolution; a type of "auditory cheesecake" that pleases the senses without providing any adaptive function.[124] This view has been directly countered by numerous music researchers.[125][126][127]

Cultural effects

[edit]

An individual's culture or ethnicity plays a role in their music cognition, including their preferences, emotional reaction, and musical memory. Musical preferences are biased toward culturally familiar musical traditions beginning in infancy, and adults' classification of the emotion of a musical piece depends on both culturally specific and universal structural features.[128][129] Additionally, individuals' musical memory abilities are greater for culturally familiar music than for culturally unfamiliar music.[130][131]

Perceptual

[edit]

Since the emergence of the study of psychoacoustics in the 1930s, most lists of elements of music have related more to how we hear music than how we learn to play it or study it. C.E. Seashore, in his book Psychology of Music,[132] identified four "psychological attributes of sound". These were: "pitch, loudness, time, and timbre" (p. 3). He did not call them the "elements of music" but referred to them as "elemental components" (p. 2). Nonetheless, these elemental components link precisely with four of the most common musical elements: "Pitch" and "timbre" match exactly, "loudness" links with dynamics, and "time" links with the time-based elements of rhythm, duration, and tempo. This usage of the phrase "the elements of music" links more closely with Webster's New 20th Century Dictionary definition of an element as: "a substance which cannot be divided into a simpler form by known methods"[133] and educational institutions' lists of elements generally align with this definition as well.

Although writers of lists of "rudimentary elements of music" can vary their lists depending on their personal (or institutional) priorities, the perceptual elements of music should consist of an established (or proven) list of discrete elements which can be independently manipulated to achieve an intended musical effect. It seems at this stage that there is still research to be done in this area.

A slightly different way of approaching the identification of the elements of music, is to identify the "elements of sound" as: pitch, duration, loudness, timbre, sonic texture and spatial location,[134] and then to define the "elements of music" as: sound, structure, and artistic intent.[134]

Sociological aspects

[edit]
Song dynasty (960–1279) painting, Night Revels of Han Xizai, showing Chinese musicians entertaining guests at a party in a 10th-century household

Ethnographic studies demonstrate that music is a participatory, community-based activity.[135][136] Music is experienced by individuals in a range of social settings from being alone, to attending a large concert, forming a music community, which cannot be understood as a function of individual will or accident; it includes both commercial and non-commercial participants with a shared set of common values. Musical performances take different forms in different cultures and socioeconomic milieus.

In Europe and North America, there was a divide between what types of music were viewed as "high culture" and "low culture." "High culture" included Baroque, Classical, Romantic, and modern-era symphonies, concertos, and solo works, and are typically heard in formal concerts in concert halls and churches, with the audience sitting quietly. Other types of music—including jazz, blues, soul, and country—are often performed in bars, nightclubs, and theatres, where the audience may drink, dance and cheer. Until the 20th century, the division between "high" and "low" musical forms was accepted as a valid distinction that separated out "art music", from popular music heard in bars and dance halls. Musicologists, such as David Brackett, note a "redrawing of high-low cultural-aesthetic boundaries" in the 20th century.[137] And, "when industry and public discourses link categories of music with categories of people, they tend to conflate stereotypes with actual listening communities."[137] Stereotypes can be based on socioeconomic standing, or social class, of the performers or audience of the different types of music.

When composers introduce styles of music that break with convention, there can be strong resistance from academics and others. Late-period Beethoven string quartets, Stravinsky ballet scores, serialism, bebop, hip hop, punk rock, and electronica were controversial and criticised, when they were first introduced. Such themes are examined in the sociology of music, sometimes called sociomusicology, which is pursued in departments of sociology, media studies, or music, and is closely related to ethnomusicology.

Role of women

[edit]
19th-century composer and pianist Clara Schumann

Women have played a major role in music throughout history, as composers, songwriters, instrumental performers, singers, conductors, music scholars, music educators, music critics/music journalists and other musical professions. In the 2010s, while women comprise a significant proportion of popular music and classical music singers, and a significant proportion of songwriters (many of them being singer-songwriters), there are few women record producers, rock critics and rock instrumentalists. Although there have been a huge number of women composers in classical music, from the medieval period to the present day, women composers are significantly underrepresented in the commonly performed classical music repertoire, music history textbooks and music encyclopedias; for example, in the Concise Oxford History of Music, Clara Schumann is one of the few female composers who is mentioned.

Women comprise a significant proportion of instrumental soloists in classical music and the percentage of women in orchestras is increasing. A 2015 article on concerto soloists in major Canadian orchestras, however, indicated that 84% of the soloists with the Montreal Symphony Orchestra were men. In 2012, women still made up just 6% of the top-ranked Vienna Philharmonic orchestra. Women are less common as instrumental players in popular music genres such as rock and heavy metal, although there have been a number of notable female instrumentalists and all-female bands. Women are particularly underrepresented in extreme metal genres.[138] In the 1960s pop-music scene, "[l]ike most aspects of the...music business, [in the 1960s,] songwriting was a male-dominated field. Though there were plenty of female singers on the radio, women ...were primarily seen as consumers:... Singing was sometimes an acceptable pastime for a girl, but playing an instrument, writing songs, or producing records simply wasn't done."[139] Young women "...were not socialized to see themselves as people who create [music]."[139]

Women are also underrepresented in orchestral conducting, music criticism/music journalism, music producing, and sound engineering. While women were discouraged from composing in the 19th century, and there are few women musicologists, women became involved in music education "...to such a degree that women dominated [this field] during the later half of the 19th century and well into the 20th century."[140]

According to Jessica Duchen, a music writer for London's The Independent, women musicians in classical music are "...too often judged for their appearances, rather than their talent" and they face pressure "...to look sexy onstage and in photos."[141] Duchen states that while "[t]here are women musicians who refuse to play on their looks,...the ones who do tend to be more materially successful."[141] According to the UK's Radio 3 editor, Edwina Wolstencroft, the music industry has long been open to having women in performance or entertainment roles, but women are much less likely to have positions of authority, such as being the conductor of an orchestra.[142] In popular music, while there are many women singers recording songs, there are very few women behind the audio console acting as music producers, the individuals who direct and manage the recording process.[143] One of the most recorded artists is Asha Bhosle, an Indian singer best known as a playback singer in Hindi cinema.[144]

Media and technology

[edit]
Compact Disc Digital Audio (CD-DA) revolutionized audio music formatting. It allowed for music to be played portably without the need to rewind a series of tape. The CD surpassed LP sales in 1988, and cassette tapes in 1991. By 1999, the CD accounted for 87.9% of the entire market share in regard to music sales.[145]

Since the 20th century, live music can be broadcast over the radio, television or the Internet, or recorded and listened to on a CD player or MP3 player.

In the early 20th century (in the late 1920s), as talking pictures emerged in the early 20th century, with their prerecorded musical tracks, an increasing number of moviehouse orchestra musicians found themselves out of work.[146] During the 1920s, live musical performances by orchestras, pianists, and theater organists were common at first-run theaters.[147] With the coming of the talking motion pictures, those featured performances were largely eliminated. The American Federation of Musicians (AFM) took out newspaper advertisements protesting the replacement of live musicians with mechanical playing devices. One 1929 ad that appeared in the Pittsburgh Press features an image of a can labeled "Canned Music / Big Noise Brand / Guaranteed to Produce No Intellectual or Emotional Reaction Whatever"[148]

Sometimes, live performances incorporate prerecorded sounds. For example, a disc jockey uses disc records for scratching, and some 20th-century works have a solo for an instrument or voice that is performed along with music that is prerecorded onto a tape. Some pop bands use recorded backing tracks. Computers and many keyboards can be programmed to produce and play Musical Instrument Digital Interface (MIDI) music. Audiences can also become performers by participating in karaoke, an activity of Japanese origin centered on a device that plays voice-eliminated versions of well-known songs. Most karaoke machines also have video screens that show lyrics to songs being performed; performers can follow the lyrics as they sing over the instrumental tracks.

Music production in the 2000s using a digital audio workstation (DAW) with an electronic keyboard and a multi-monitor set-up

The advent of the Internet and widespread high-speed broadband access has transformed the experience of music, partly through the increased ease of access to recordings of music via streaming video and vastly increased choice of music for consumers. Another effect of the Internet arose with online communities and social media websites like YouTube and Facebook, a social networking service. These sites make it easier for aspiring singers and amateur bands to distribute videos of their songs, connect with other musicians, and gain audience interest. Professional musicians also use YouTube as a free publisher of promotional material. YouTube users, for example, no longer only download and listen to MP3s, but also actively create their own. According to Don Tapscott and Anthony D. Williams, in their book Wikinomics, there has been a shift from a traditional consumer role to what they call a "prosumer" role, a consumer who both creates content and consumes. Manifestations of this in music include the production of mashes, remixes, and music videos by fans.[149]

Music streaming services have further transformed music production. Streaming platforms, such as Spotify and Apple Music, mediate how music is consumed and distributed. This raises questions about how artists market their releases (i.e. as 'limited edition'), algorithmic suggestions, and affordable access to recorded music.[150]

Generative AI systems, more commonly known as Music Metacreation in music generation, are increasingly being used to create music.[citation needed] A more neutral term is 'Music Generation Systems' (MGSs).[151]

Education

[edit]

Non-institutional

[edit]
A Suzuki violin recital with students of varying ages

The incorporation of music into general education from preschool to post secondary education is common in North America and Europe. Involvement in playing and singing music is thought to teach basic skills such as concentration, counting, listening, and cooperation while also promoting understanding of language, improving the ability to recall information, and creating an environment more conducive to learning in other areas.[152] In elementary schools, children often learn to play instruments such as the recorder, sing in small choirs, and learn about the history of Western art music and traditional music. Some elementary school children also learn about popular music styles. In religious schools, children sing hymns and other religious music. In secondary schools (but rarely in primary schools), students may have the opportunity to perform in some types of musical ensembles, such as choirs (a group of singers), marching bands, concert bands, jazz bands, or orchestras. In some school systems, music lessons on how to play instruments may be provided. Some students also take private music lessons after school with a singing teacher or instrument teacher. Amateur musicians typically learn basic musical rudiments (e.g., learning about musical notation for musical scales and rhythms) and beginner- to intermediate-level singing or instrument-playing techniques.

At the university level, students in most arts and humanities programs can receive credit for taking a few music courses, which typically take the form of an overview course on the history of music, or a music appreciation course that focuses on listening to music and learning about different musical styles. In addition, most North American and European universities have some types of musical ensembles that students in arts and humanities are able to participate in, such as choirs, marching bands, concert bands, or orchestras. The study of Western art music is increasingly common outside of North America and Europe, such as the Indonesian Institute of the Arts in Yogyakarta, Indonesia, or the classical music programs that are available in Asian countries such as South Korea, Japan, and China. At the same time, Western universities and colleges are widening their curriculum to include music of non-Western cultures, such as the music of Africa or Bali (e.g. Gamelan music).

Institutional

[edit]
Manhattan School of Music professor and professional double bass player Timothy Cobb teaching a bass lesson in the late 2000s. His bass has a low C extension with a metal "machine" with buttons for playing the pitches on the extension.

People aiming to become professional musicians, singers, composers, songwriters, music teachers and practitioners of other music-related professions such as music history professors, sound engineers, and so on study in specialized post-secondary programs offered by colleges, universities and music conservatories. Some institutions that train individuals for careers in music offer training in a wide range of professions, as is the case with many of the top U.S. universities, which offer degrees in music performance (including singing and playing instruments), music history, music theory, music composition, music education (for individuals aiming to become elementary or high school music teachers) and, in some cases, conducting. On the other hand, some small colleges may only offer training in a single profession (e.g., sound recording).

While most university and conservatory music programs focus on training students in classical music, there are universities and colleges that train musicians for careers as jazz or popular music musicians and composers, with notable U.S. examples including the Manhattan School of Music and the Berklee College of Music. Two schools in Canada which offer professional jazz training are McGill University and Humber College. Individuals aiming at careers in some types of music, such as heavy metal music, country music or blues are unlikely to become professionals by completing degrees or diplomas. Instead, they typically learn about their style of music by singing or playing in bands (often beginning in amateur bands, cover bands and tribute bands), studying recordings on DVD and the Internet, and working with already-established professionals in their style of music, either through informal mentoring or regular music lessons. Since the 2000s, the increasing popularity and availability of Internet forums and YouTube "how-to" videos have enabled singers and musicians from metal, blues and similar genres to improve their skills. Many pop, rock and country singers train informally with vocal coaches and voice teachers.[153][154]

Academic study

[edit]

Musicology

[edit]

Musicology, the academic study of music, is studied in universities and music conservatories. The earliest definitions from the 19th century defined three sub-disciplines of musicology: systematic musicology, historical musicology, and comparative musicology or ethnomusicology. In 2010-era scholarship, one is more likely to encounter a division into music theory, music history, and ethnomusicology. Research in musicology has often been enriched by cross-disciplinary work, for example in the field of psychoacoustics. The study of music of non-Western cultures, and cultural study of music, is called ethnomusicology. Students can pursue study of musicology, ethnomusicology, music history, and music theory through different types of degrees, including bachelor's, master's and PhD.[155][156][157]

Music theory

[edit]

Music theory is the study of music, generally in a highly technical manner outside of other disciplines. More broadly it refers to any study of music, usually related in some form with compositional concerns, and may include mathematics, physics, and anthropology. What is most commonly taught in beginning music theory classes are guidelines to write in the style of the common practice period, or tonal music. Theory, even of music of the common practice period, may take other forms.[158] Musical set theory is the application of mathematical set theory to music, first applied to atonal music. Speculative music theory, contrasted with analytic music theory, is devoted to the analysis and synthesis of music materials, for example tuning systems, generally as preparation for composition.[159]

Zoomusicology

[edit]

Zoomusicology is the study of the music of non-human animals, or the musical aspects of sounds produced by non-human animals. As George Herzog (1941) asked, "do animals have music?" François-Bernard Mâche's Musique, mythe, nature, ou les Dauphins d'Arion (1983), a study of "ornitho-musicology" using a technique of Nicolas Ruwet's Language, musique, poésie (1972) paradigmatic segmentation analysis, shows that bird songs are organised according to a repetition-transformation principle. Jean-Jacques Nattiez (1990), argues that "in the last analysis, it is a human being who decides what is and is not musical, even when the sound is not of human origin. If we acknowledge that sound is not organised and conceptualised (that is, made to form music) merely by its producer, but by the mind that perceives it, then music is uniquely human."[160]

Ethnomusicology

[edit]
Ethnomusicologist Frances Densmore at the Smithsonian Institution in 1916 where she was recording Blackfoot chief Mountain Chief for the Bureau of American Ethnology. In this picture, Mountain Chief is listening to a recording.

In the West, much of the history of music that is taught deals with the Western civilization's art music, known as classical music. The history of music in non-Western cultures ("world music" or the field of "ethnomusicology") is also taught in Western universities. This includes the documented classical traditions of Asian countries outside the influence of Western Europe, as well as the folk or indigenous music of various other cultures. Popular or folk styles of music in non-Western countries varied from culture to culture, and period to period. Different cultures emphasised different instruments, techniques, singing styles and uses for music. Music has been used for entertainment, ceremonies, rituals, religious purposes and for practical and artistic communication. Non-Western music has also been used for propaganda purposes, as was the case with Chinese opera during the Cultural Revolution.

There is a host of music classifications for non-Western music, many of which are caught up in the argument over the definition of music. Among the largest of these is the division between classical music (or "art" music), and popular music (or commercial music – including non-Western styles of rock, country, and pop music-related styles). Some genres do not fit neatly into one of these "big two" classifications, (such as folk music, world music, or jazz-related music).

As world cultures have come into greater global contact, their indigenous musical styles have often merged with other styles, which produces new styles. For example, the United States bluegrass style contains elements from Anglo-Irish, Scottish, Irish, German and African instrumental and vocal traditions, which were able to fuse in the United States' multi-ethnic "melting pot" society. Some types of world music contain a mixture of non-Western indigenous styles with Western pop music elements. Genres of music are determined as much by tradition and presentation as by the actual music. Some works, like George Gershwin's Rhapsody in Blue, are claimed by both jazz and classical music, while Gershwin's Porgy and Bess and Leonard Bernstein's West Side Story are claimed by both opera and the Broadway musical tradition. Many music festivals for non-Western music, include bands and singers from a particular musical genre, such as world music.[161][162]

Indian music, for example, is one of the oldest and longest living types of music, and is still widely heard and performed in South Asia, as well as internationally (especially since the 1960s). Indian music has mainly three forms of classical music, Hindustani, Carnatic, and Dhrupad styles. It has also a large repertoire of styles, which involve only percussion music such as the talavadya performances famous in South India.

Therapy

[edit]
A music therapist from a "Blues in the Schools" program plays harmonica with a US Navy sailor at a Naval Therapy Center.

Music therapy is an interpersonal process in which a trained therapist uses music and all of its facets—physical, emotional, mental, social, aesthetic, and spiritual—to help clients to improve or maintain their health. In some instances, the client's needs are addressed directly through music; in others they are addressed through the relationships that develop between the client and therapist. Music therapy is used with individuals of all ages and with a variety of conditions, including: psychiatric disorders, medical problems, physical disabilities, sensory impairments, developmental disabilities, substance abuse issues, communication disorders, interpersonal problems, and aging. It is also used to improve learning, build self-esteem, reduce stress, support physical exercise, and facilitate a host of other health-related activities. Music therapists may encourage clients to sing, play instruments, create songs, or do other musical activities.

In the 10th century, the philosopher Al-Farabi described how vocal music can stimulate the feelings and souls of listeners.[163] Music has long been used to help people deal with their emotions. In the 17th century, the scholar Robert Burton's The Anatomy of Melancholy argued that music and dance were critical in treating mental illness, especially melancholia.[164] He noted that music has an "excellent power ...to expel many other diseases" and he called it "a sovereign remedy against despair and melancholy." He pointed out that in Antiquity, Canus, a Rhodian fiddler, used music to "make a melancholy man merry, ...a lover more enamoured, a religious man more devout."[165][166][167] In the Ottoman Empire, mental illnesses were treated with music.[168] In November 2006, Michael J. Crawford and his colleagues also found that music therapy helped schizophrenic patients.[169][170]

See also

[edit]
  • Glossary of music terminology
  • List of musicology topics
  • Lists of musicians
  • Music and emotion
  • Music archaeology – Interdisciplinary study field
  • Music history – Academic field
  • Music-specific disorders – Disorders relating to the perception of music

References

[edit]

Notes

[edit]
  1. ^ A now discredited theory held by many medieval thinkers was that music descended from the Egyptian word moys, meaning 'water', thought to connect to Moses.[11]
  2. ^ For the further etymological origins, mousiké derives from the feminine form of mousikos, which is anything "pertaining to the muses", from the Ancient Greek word for 'Muse', Mousa.[10] There is no agreement on the origins of the word Mousa,[12] though see Muses § Etymology for proposed theories.
  3. ^ See Morley (2013, pp. 43–45) for a comprehensive table on all eight surviving flutes from the Geissenklösterle, Hohle Fels and Vogelherd caves
  4. ^ In the 1800s, the phrases "the elements of music" and "the rudiments of music" were used interchangeably.[88][89] The elements described in these documents refer to aspects of music that are needed to become a musician, Recent writers such as Espie Estrella seem to be using the phrase "elements of music" in a similar manner.[90] A definition which most accurately reflects this usage is: "the rudimentary principles of an art, science, etc.: the elements of grammar."[91] The UK's curriculum switch to the "inter-related dimensions of music" seems to be a move back to using the rudimentary elements of music.

Citations

[edit]
  1. ^ OED, § 1.
  2. ^ AHD, § 1.
  3. ^ Epperson 2022, § para. 1.
  4. ^ Morley 2013, p. 5.
  5. ^ Mithen 2005, pp. 26–27.
  6. ^ Gardner 1983, p. 104.
  7. ^ a b Hoad, T. F., ed. (2003) [1996]. "music". The Concise Oxford Dictionary of English Etymology. ISBN 978-0-19-283098-2. Archived from the original on 30 May 2022. Retrieved 30 May 2022.
  8. ^ Nettl 2001, §III "3. Music among the arts".
  9. ^ Online Etymology Dictionary, § para 2.
  10. ^ a b c Online Etymology Dictionary, § para 1.
  11. ^ a b Apel 1969, p. 548.
  12. ^ Anderson & Mathiesen 2001, § para 1.
  13. ^ Murray 2020, pp. 13–14.
  14. ^ a b c d Nettl 2001, §I "1. Etymology".
  15. ^ Anderson & Mathiesen 2001, § para 2.
  16. ^ Nettl 2001, §II "1. Contemporary Western culture".
  17. ^ a b Nettl 2001, §II "2. East Asia".
  18. ^ a b Nettl 2001, §II "5. Some African cultures".
  19. ^ Nettl 2001, §II "4. India".
  20. ^ Nettl 2001, §II "6. Some Amerindian and Oceanian cultures".
  21. ^ Nettl 2001, §II "3. Iran and the Middle East".
  22. ^ a b Morley 2013, p. 43.
  23. ^ Merker, Morley & Zuidema 2015, § "Introduction".
  24. ^ Morley 2013, pp. 2–3.
  25. ^ a b Wallin, Merker & Brown 2000, p. 8.
  26. ^ Huron 2003, p. 61.
  27. ^ Huron 2003, p. 62.
  28. ^ Wallin, Merker & Brown 2000, p. 11.
  29. ^ Nettl 2001, § 8 "On the Origins of Music".
  30. ^ Morley 2013, pp. 38–39.
  31. ^ a b Morley 2013, p. 42.
  32. ^ Music of Ancient Egypt Archived 13 October 2015 at the Wayback Machine. Kelsey Museum of Archaeology, University of Michigan, Ann Arbor.
  33. ^ "UC 33268". digitalegypt.ucl.ac.uk. Archived from the original on 27 July 2013. Retrieved 27 October 2015.
  34. ^ Hickmann, Hans (1957). "Un Zikr Dans le Mastaba de Debhen, Guîzah (IVème Dynastie)". Journal of the International Folk Music Council. 9: 59–62. doi:10.2307/834982. ISSN 0950-7922. JSTOR 834982.
  35. ^ Hickmann, Hans (January–March 1960). "Rythme, mètre et mesure de la musique instrumentale et vocale des anciens Egyptiens". Acta Musicologica. 32 (1): 11–22. doi:10.2307/931818. JSTOR 931818.
  36. ^ Stolba, K. Marie (1995). The Development of Western Music: A History (brief second ed.). Madison: Brown & Benchmark Publishers. p. 2.
  37. ^ a b West, Martin Litchfield (May 1994). "The Babylonian Musical Notation and the Hurrian Melodic Texts". Music and Letters. Vol. 75. pp. 161–179.
  38. ^ a b c Savage, Roger (2001). Incidental music | Oxford Music. doi:10.1093/gmo/9781561592630.article.43289. ISBN 978-1-56159-263-0. Archived from the original on 25 April 2021. Retrieved 30 July 2022.
  39. ^ West, Martin Litchfield (1994). Ancient Greek music. Oxford University Press.
  40. ^ Winnington-Ingram, Reginald P. (October 1929). "Ancient Greek Music: A Survey". Music & Letters. 10 (4): 326–345. doi:10.1093/ml/X.4.326. JSTOR 726126.
  41. ^ Aristoxenus, Henry Stewart Macran (1902). Harmonika Stoicheia (The Harmonics of Aristoxenus). Georg Olms Verlag. OCLC 123175755. Archived from the original on 2 January 2020. Retrieved 7 April 2016.
  42. ^ Pegg, Carole (2001). "Inner Asia". Grove Music Online. doi:10.1093/gmo/9781561592630.article.05283. ISBN 978-1-56159-263-0.
  43. ^ Richard O. Nidel, World Music: The Basics, p. 219.
  44. ^ Charles Kahn, World History: Societies of the Past, p. 98.
  45. ^ World History: Societies of the Past By Charles Kahn (p. 11)
  46. ^ World Music: The Basics By Nidel Nidel, Richard O. Nidel (p. 10)
  47. ^ Rajagopal, Geetha (2009). Music rituals in the temples of South India, Volume 1. D. K. Printworld. pp. 111–112. ISBN 978-81-246-0538-7. Archived from the original on 14 December 2019. Retrieved 29 October 2015.
  48. ^ Szczepanski, Kallie (26 June 2019). "History of Gamelan, Indonesian Music and Dance". ThoughtCo.
  49. ^ "Gamelan | Indonesian Orchestra, Traditional Instruments & Javanese Culture". Britannica. Archived from the original on 3 December 2023.
  50. ^ "Chinese music - Tonal System, Theoretical Rationalization". Britannica. Archived from the original on 30 December 2023.
  51. ^ "Léonin | Medieval Music, Notre Dame School, Polyphonic Masses". Britannica. Archived from the original on 30 May 2023.
  52. ^ "Pérotin | Medieval Music, Notre Dame School, Polyphony". Britannica. Archived from the original on 15 November 2023.
  53. ^ "Guillaume de Machaut | 14th Century French Poet & Musician". Britannica. Archived from the original on 6 September 2023.
  54. ^ "Walther von der Vogelweide | German Minnesinger, Lyric Poet". Britannica. Archived from the original on 6 June 2023.
  55. ^ "Chapter 6: Music of Franco-Flemish Composers, 1450-1520". Concise History of Western Music, 4: W. W. Norton StudySpace. Archived from the original on 29 December 2023.
  56. ^ Elaine Thornburgh; Jack Logan. "Baroque Music - Part One". Music 345. Archived from the original on 5 September 2015. Retrieved 27 October 2015.
  57. ^ Blume, Friedrich. Classic and Romantic Music: A Comprehensive Survey. New York: W.W. Norton & Company, 1970. Print.
  58. ^ "Basso Continuo | Music 101". courses.lumenlearning.com.
  59. ^ Le, Bonnie; Scully, Francis; Edwards, Steven; Boyd, Jesse; Wimberly, Brenda (1 August 2022). "Music of the Classical Period" – via louis.pressbooks.pub.
  60. ^ "5.2: Music in the Classical period". Humanities LibreTexts. 11 May 2020.
  61. ^ "Classical Music: Ballets and Ballerinas". dummies.
  62. ^ "The Music of the Romantic Period: Emotional Expression and Innovation". Classical Mania. 11 April 2023.
  63. ^ 7.4 Radio's Impact on Culture. University of Minnesota Libraries Publishing edition, 2016. This edition adapted from a work originally produced in 2010 by a publisher who has requested that it not receive attribution. 22 March 2016 – via open.lib.umn.edu.
  64. ^ "The Impact Of Radio On American Popular Music | Ben Vaughn". www.benvaughn.com.
  65. ^ Schaeffer, P. (1966), Traité des objets musicaux, Le Seuil, Paris.
  66. ^ "How Do Multitrack Recordings Enhance Music Production? - AboutRecording.com". 13 September 2023.
  67. ^ "Basic Overdubbing". www.soundonsound.com.
  68. ^ Alyn Shipton, A New History of Jazz, 2nd. ed., Continuum, 2007, pp. 4–5
  69. ^ Gilliland, John (1969). "Show 55 – Crammer: A lively cram course on the history of rock and some other things" (audio). Pop Chronicles. University of North Texas Libraries.
  70. ^ "Rock & Roll Music Style Overview". AllMusic.
  71. ^ "What Do DJs Use to Make Music?". Performer Life. 23 November 2023.
  72. ^ "What Does a Mixer Do? - A Complete Answer · DJ Producer Tech". 16 August 2019. Archived from the original on 31 October 2020.
  73. ^ "Journal on the Art of Record Production » Recreating an Unreal Reality: Performance Practice, Recording, and the Jazz Rhythm Section". www.arpjournal.com.
  74. ^ Havers, Richard (26 February 2023). "The History Of Recorded Jazz".
  75. ^ "How to interpret classical music".
  76. ^ Jean, Janae (21 March 2019). "What is Chance Music?". perennial.
  77. ^ "10 Cover Versions That Creatively Transform the Original Song – Flypaper". 19 November 2018.
  78. ^ "What is Chamber Music? Definition, History & Examples". promusicianhub.com. 25 December 2021.
  79. ^ "What is Jazz?". americanhistory.si.edu. 9 March 2015.
  80. ^ Kirszner, Laurie G. (January 2012). Patterns for College Writing. Bedford/St. Martin's. p. 520. ISBN 978-0-312-67684-1
  81. ^ "Improvisation and Reciprocity: An Analysis of the Jam Band Community and Its Unique Business Model". www.meiea.org. 20 November 2007.
  82. ^ Ferreira, Bruno (26 July 2021). "How to Read Renaissance and Baroque Tablature". UniGuitar.
  83. ^ Junda, Mary Ellen (2019). "Exploring Cultures Through Song". The Choral Journal. 59 (9): 30–41. JSTOR 26662729.
  84. ^ "Music: Glossary". The Australian Curriculum. Archived from the original on 26 January 2016.
  85. ^ "The School Curriculum". webarchive.nationalarchives.gov.uk.
  86. ^ "2014 music standards". NAfME. 2015. Archived from the original on 10 January 2016.
  87. ^ "National curriculum in England: music programmes of study". GOV.UK. 26 March 2021. Archived from the original on 26 January 2016. Retrieved 24 August 2023.
  88. ^ Clementi, M.: "Introduction to the art of playing on the piano forte", Da Capo Press (1974). Cohen, Dalia and Dubnov, Shlomo (1996). "Gestalt phenomena in musical texture", Springer doi:10.1007/BFb0034128
  89. ^ Niecks, Frederick (1884). A Concise Dictionary of Musical Terms.. "Review". The Musical Times and Singing Class Circular. 25 (498): 473. doi:10.2307/3357513. hdl:2027/uc1.b4284161. JSTOR 3357513. Archived from the original on 16 March 2020. Retrieved 24 September 2019.
  90. ^ Estrella, Espie (4 November 2019). "An Introduction to the Elements of Music". liveabout.com. Archived from the original on 4 August 2020. Retrieved 25 February 2020.
  91. ^ "Element" Archived 29 January 2016 at the Wayback Machine, (n.d.) In Dictionary.com unabridged. Retrieved 10 June 2015
  92. ^ "Are Cymbals Pitched Or Unpitched?". Drumming Corner.
  93. ^ Mager, Noella (27 November 2023). "What Is A Bass Line In Music".
  94. ^ Santa, Matthew (30 December 2022). Hearing Form: Musical Analysis With and Without the Score. Taylor & Francis. p. 48. ISBN 9781000643602.
  95. ^ Porter, Lewis (1997). Jazz: A Century of Change. Schirmer Books. p. 176. ISBN 9780028647135.
  96. ^ "The Drums And Bass Guitar: The Foundation Of The Rhythm Section – FuelRocks". 7 November 2022.
  97. ^ Gotham, Mark (1 July 2021). "Core Principles of Orchestration" – via viva.pressbooks.pub.
  98. ^ Harnsberger, Lindsey. "Articulation." Essential Dictionary of Music. Alfred Publishing Co., Inc. Los Angeles, CA.
  99. ^ "the definition of expression". Dictionary.com. Archived from the original on 19 October 2017. Retrieved 22 October 2017.
  100. ^ Burton, Russell (2015). "The elements of music: What are they, and who cares?". Educating for Life. ASME XXTH National Conference Proceeding. Australian Society for Music Education.: 22.
  101. ^ Schmidt-Jones, Catherine (11 March 2011). "Form in Music". Connexions. Archived from the original on 22 February 2014. Retrieved 11 September 2011.
  102. ^ Brandt, Anthony (11 January 2007). "Musical Form". Connexions. Archived from the original on 7 July 2012. Retrieved 11 September 2011.
  103. ^ "Form in pop/rock music – Formal functions". Open Music Theory.
  104. ^ Scholes, Percy A. (1977). "Form". The Oxford Companion to Music (10 ed.). Oxford University Press.
  105. ^ Mann, Alfred (1958). The Study of Fugue. W.W.Norton and Co. Inc.
  106. ^ "What Is Through Composed Form In Music? | HelloMusictheory". 11 September 2020.
  107. ^ Keil, Charles (1966). Urban blues. University of Chicago Press. ISBN 978-0-226-42960-1.
  108. ^ Juslin, Patrik N. (6 September 2013). "What does music express? Basic emotions and beyond". Frontiers in Psychology. 4: 596. doi:10.3389/fpsyg.2013.00596. PMC 3764399. PMID 24046758.
  109. ^ "A Field Full of Rabbit Holes". Sound American.
  110. ^ "Rhys Chatham: "Drastic Classical Music for Electric Instruments"". The Kitchen OnScreen.
  111. ^ "Glenn Branca: A Guide to the Symphonies". daily.redbullmusicacademy.com.
  112. ^ Plato (2006). The Republic, Book VII. Project Gutenberg Literary Archive Foundation. p. 7. Archived from the original on 11 October 2022. Retrieved 11 October 2022.
  113. ^ Kirkendall, Jensen Armstrong (14 December 2017). "The Well-Ordered Heart: Confucius on Harmony, Music, and Ritual" (PDF). Archived (PDF) from the original on 13 April 2021. Retrieved 13 April 2021.
  114. ^ Tan, Siu-Lan; Pfordresher, Peter; Harré, Rom (2010). Psychology of Music: From Sound to Significance. New York: Psychology Press. p. 2. ISBN 978-1-84169-868-7.
  115. ^ Laske, Otto (1999). Navigating New Musical Horizons (Contributions to the Study of Music and Dance). Westport: Greenwood Press. ISBN 978-0-313-30632-7.
  116. ^ Laske, O. (1999). AI and music: A cornerstone of cognitive musicology. In M. Balaban, K. Ebcioglu, & O. Laske (Eds.), Understanding music with ai: Perspectives on music cognition. Cambridge: The MIT Press.
  117. ^ Graci, C. (2009–2010) A brief tour of the learning sciences featuring a cognitive tool for investigating melodic phenomena. Journal of Educational Technology Systems, 38(2), 181–211.
  118. ^ Hamman, M., 1999. "Structure as Performance: Cognitive Musicology and the Objectification of Procedure", in Otto Laske: Navigating New Musical Horizons, ed. J. Tabor. New York: Greenwood Press.
  119. ^ Wallin, Nils L./Björn Merker/Steven Brown (1999): "An Introduction to Evolutionary Musicology." In: Wallin, Nils L./Björn Merker/Steven Brown (Eds., 1999): The Origins of Music, pp. 5–6. ISBN 0-262-23206-5.
  120. ^ "The Descent of Man, and Selection in Relation to Sex". 1871. Archived from the original on 2 April 2012. Retrieved 24 April 2014. Chapter III; Language
  121. ^ Nils L. Wallin; Björn Merker; Steven Brown, eds. (2000). The Origins of Music. Cambridge, Massachusetts: MIT Press. ISBN 978-0-262-23206-7.
  122. ^ Mithen 2005.
  123. ^ Hagen, Edward H; Hammerstein P (2009). "Did Neanderthals and other early humans sing? Seeking the biological roots of music in the loud calls of primates, lions, hyenas, and wolves" (PDF). Musicae Scientiae. doi:10.1177/1029864909013002131. S2CID 39481097. Archived (PDF) from the original on 30 May 2010. Retrieved 24 April 2014.
  124. ^ Pinker, Steven (1997). How the Mind Works. New York: W. W. Norton. p. 534. ISBN 978-0-393-04535-2.
  125. ^ Perlovsky L. Music. Cognitive Function, Origin, And Evolution Of Musical Emotions. WebmedCentral PSYCHOLOGY 2011;2(2):WMC001494
  126. ^ Alison Abbott. 2002. "Neurobiology: Music, maestro, please!" Nature 416, 12–14 (7 March 2002) | doi:10.1038/416012a
  127. ^ Carroll, Joseph (1998). "Steven Pinker's Cheesecake For The Mind". Cogweb.ucla.edu. Archived from the original on 29 January 2013. Retrieved 29 December 2012.
  128. ^ Soley, G.; Hannon, E. E. (2010). "Infants prefer the musical meter of their own culture: A cross-cultural comparison". Developmental Psychology. 46 (1): 286–292. doi:10.1037/a0017555. PMID 20053025.
  129. ^ Balkwill, L.; Thompson, W. F.; Matsunaga, R. (2004). "Recognition of emotion in Japanese, Western, and Hindustani music by Japanese listeners". Japanese Psychological Research. 46 (4): 337–349. doi:10.1111/j.1468-5584.2004.00265.x.
  130. ^ Demorest, S. M.; Morrison, S. J.; Beken, M. N.; Jungbluth, D. (2008). "Lost in translation: An enculturation effect in music memory performance". Music Perception. 25 (3): 213–223. doi:10.1525/mp.2008.25.3.213.
  131. ^ Groussard, M.; Rauchs, G.; Landeau, B.; Viader, F.; Desgranges, B.; Eustache, F.; Platel, H. (2010). "The neural substrates of musical memory revealed by fMRI and two semantic tasks" (PDF). NeuroImage. 53 (4): 1301–1309. doi:10.1016/j.neuroimage.2010.07.013. PMID 20627131. S2CID 8955075. Archived (PDF) from the original on 29 April 2019. Retrieved 11 December 2019.
  132. ^ Seashore, C. E. (1938). Psychology of Music: New York: Dover Publications.
  133. ^ Webster, N. (Ed.) (1947) Webster's New Twentieth Century Dictionary. Clevelend Ohio: The World Publishing Company.
  134. ^ a b Burton, R.L. (2015). The elements of music: what are they, and who cares? Archived 10 May 2020 at the Wayback Machine In J. Rosevear & S. Harding. (Eds.), ASME XXth National Conference proceedings. Paper presented at: Music: Educating for life: ASME XXth National Conference (pp. 22–28), Parkville, Victoria: The Australian Society for Music Education Inc.
  135. ^ Grazian, David. "The Symbolic Economy of Authenticity in the Chicago Blues Scene." in Music Scenes: Local, Translocal, and Virtual. ed. Bennett, Andy and Richard A. Peterson. Nashville: Vanderbilt University Press, 2004. pp. 31–47
  136. ^ Rebecca Elizabeth Ball (2010). Portland's Independent Music Scene: Formation of Community Identities and Alternative Urban Cultural Landscapes Archived 1 February 2018 at the Wayback Machine, p. 27
  137. ^ a b Johnson, Thomas (1 April 2018). "David Brackett. 2016. Categorizing Sound: Genre and Twentieth-Century Popular Music. Berkeley: University of California Press". Current Musicology (102). doi:10.7916/cm.v0i102.5377 – via journals.library.columbia.edu.
  138. ^ Julian Schaap and Pauwke Berkers. "Grunting Alone? Online Gender Inequality in Extreme Metal Music" in IASPM Journal. Vol. 4, no. 1 (2014) p. 103
  139. ^ a b Erika White (28 January 2015). "Music History Primer: 3 Pioneering Female Songwriters of the '60s | REBEAT Magazine". Rebeatmag.com. Archived from the original on 22 December 2015. Retrieved 20 January 2016.
  140. ^ "Women Composers In American Popular Song". Parlorsongs.com. 25 March 1911. p. 1. Archived from the original on 4 February 2019. Retrieved 20 January 2016.
  141. ^ a b "Classical music's shocking gender gap". CBC Music. Archived from the original on 4 July 2012. Retrieved 16 August 2016.
  142. ^ Jessica Duchen. "Why the male domination of classical music might be coming to an end | Music". The Guardian. Archived from the original on 30 September 2021. Retrieved 20 January 2016.
  143. ^ Ncube, Rosina (September 2013). "Sounding Off: Why So Few Women In Audio?". Sound on Sound. Archived from the original on 24 April 2016. Retrieved 17 August 2016.
  144. ^ "Singer Asha Bhosle enters Guinness World Records for most single studio recordings". India Today. 21 October 2011. Archived from the original on 13 March 2016. Retrieved 12 March 2016.
  145. ^ "U.S. Music Revenue Database". RIAA. Retrieved 26 December 2023.
  146. ^ "History". American Federation of Musicians. Archived from the original on 5 April 2007.
  147. ^ Hubbard, Preston J. (1985). "Synchronized Sound and Movie-House Musicians, 1926–29". American Music. 3 (4): 429–441 [429]. doi:10.2307/3051829. JSTOR 3051829. Archived from the original on 29 October 2022. Retrieved 28 May 2023.
  148. ^ "Canned Music on Trial" Archived 14 December 2009 at the Wayback Machine part of Duke University's Ad*Access project.
  149. ^ Tapscott, Don; Williams, Anthony D. (2006). Wikinomics: How Mass Collaboration Changes Everything. Portfolio Hardcover. ISBN 978-1-59184-138-8.
  150. ^ Webster, Jack (9 November 2020). "Taste in the platform age: music streaming services and new forms of class distinction". Information, Communication & Society. 23 (13): 1909–1924. doi:10.1080/1369118X.2019.1622763. ISSN 1369-118X.
  151. ^ Carnovalini, Filippo; Rodà, Antonio (3 April 2020). "Computational Creativity and Music Generation Systems: An Introduction to the State of the Art". Frontiers in Artificial Intelligence. 3 14. doi:10.3389/frai.2020.00014. ISSN 2624-8212. PMID 33733133.
  152. ^ "Promoting Literacy Through Music". www.songsforteaching.com.
  153. ^ Davis-McAfee, M. (7 May 2023). "The Hit Singer Who Gave Rihanna Informal Vocal Lessons". The List.
  154. ^ "Pop and Rock Vocals". online.berklee.edu.
  155. ^ "MUSICOLOGY & ETHNOMUSICOLOGY | School of Music". music.umd.edu.
  156. ^ "Degrees Offered | Musicology and Ethnomusicology | Programs & Degrees | School of Music | College of the Arts | University of Florida". arts.ufl.edu.
  157. ^ "Ethnomusicology | School of Music | University of Washington". music.washington.edu.
  158. ^ "Music Theory: Speculation, Reasoning, Experience. A Perspective from Systematic Musicology".
  159. ^ "What is Speculative Music Composition?". parsejournal.com.
  160. ^ Nattiez, Jean-Jacques (21 November 1990). Music and Discourse: Toward a Semiology of Music. Princeton University Press. p. 58. ISBN 9780691027142.
  161. ^ "World Music Festivals | World Music Central". 24 December 2014.
  162. ^ "The Most Popular Genres for Music Festivals in 2021". Festival Pro.
  163. ^ Rens Bod (2022). World of Patterns: A Global History of Knowledge. Johns Hopkins University Press. p. 169. ISBN 978-1-4214-4345-4.
  164. ^ cf. The Anatomy of Melancholy, Robert Burton, subsection 3, on and after line 3,480, "Music a Remedy"
  165. ^ Ismenias the Theban, Chiron the centaur, is said to have cured this and many other diseases by music alone: as now thy do those, saith Bodine, that are troubled with St. Vitus's Bedlam dance. Project Gutenberg's The Anatomy of Melancholy, by Democritus Junior Archived 13 May 2011 at the Wayback Machine
  166. ^ "Humanities are the Hormones: A Tarantella Comes to Newfoundland. What should we do about it?" Archived 15 February 2015 at the Wayback Machine by John Crellin, MUNMED, newsletter of the Faculty of Medicine, Memorial University of Newfoundland, 1996.
  167. ^ Aung, Steven K. H.; Lee, Mathew H. M. (2004). "Music, Sounds, Medicine, and Meditation: An Integrative Approach to the Healing Arts". Alternative and Complementary Therapies. 10 (5): 266–270. doi:10.1089/act.2004.10.266.
  168. ^ "Treatment of Mental Illnesses With Music Therapy – A different approach from history" (PDF). Archived from the original (PDF) on 1 December 2013.
  169. ^ Crawford, Mike J.; Talwar, Nakul; et al. (November 2006). "Music therapy for in-patients with schizophrenia: Exploratory randomised controlled trial". British Journal of Psychiatry. 189 (5): 405–409. doi:10.1192/bjp.bp.105.015073. PMID 17077429. Music therapy may provide a means of improving mental health among people with schizophrenia, but its effects in acute psychoses have not been explored
  170. ^ Michael J. Crawford page Archived 28 November 2010 at the Wayback Machine at Imperial College London, Faculty of Medicine, Department of Psychological Medicine.

Sources

[edit]
  • Anderson, Warren (2001). "Muses". Grove Music Online. Revised by Thomas J. Mathiesen. Oxford: Oxford University Press. doi:10.1093/gmo/9781561592630.article.19396. ISBN 978-1-56159-263-0. (subscription, Wikilibrary access, or UK public library membership required)
  • Apel, Willi (1969). Harvard Dictionary of Music. Cambridge: Harvard University Press. ISBN 978-0-674-37501-7.
  • Epperson, Gordon (16 June 2022). "Music | Art Form, Styles, Rhythm, & History". Encyclopædia Britannica. Chicago: Encyclopædia Britannica, Inc.
  • Gardner, Howard (1983). Frames of Mind: The Theory of Multiple Intelligences. New York: Basic Books. ISBN 978-0-465-02508-4.
  • Huron, David (2003). "Is Music an Evolutionary Adaptation?". In Peretz, Isabelle; Zatorre, Robert J. (eds.). The Cognitive Neuroscience of Music. Oxford: Oxford University Press. ISBN 978-0-19-852519-6.
  • Merker, Björn; Morley, Iain; Zuidema, Willem (19 March 2015). "Five fundamental constraints on theories of the origins of music". Philosophical Transactions of the Royal Society. 370 (1664). Royal Society. doi:10.1098/rstb.2014.0095. eISSN 1471-2970. ISSN 0962-8436. PMC 4321136. PMID 25646518.
  • Mithen, Steven (2005). The Singing Neanderthals: The Origins of Music, Language, Mind, and Body. London: Orion Publishing Group. ISBN 978-1-78022-258-5.
  • Morley, Iain (2013). The Prehistory of Music: Human Evolution, Archaeology, and the Origins of Musicality. Oxford: Oxford University Press. ISBN 978-0-19-923408-0.
  • Murray, Penelope (2020). "The Mythology of the Muses". In Lynch, Tosca A. C.; Rocconi, Eleonora (eds.). A Companion to Ancient Greek and Roman Music. Blackwell Companions to the Ancient World. Hoboken: Wiley. ISBN 978-1-119-27547-3.
  • Nettl, Bruno (2001). "Music". Grove Music Online. Oxford: Oxford University Press. doi:10.1093/gmo/9781561592630.article.40476. ISBN 978-1-56159-263-0. (subscription, Wikilibrary access, or UK public library membership required)
  • Wallin, Nils; Merker, Björn; Brown, Steven, eds. (2000). The Origins of Music. Cambridge: MIT Press. ISBN 978-0-262-73143-0.
  • "music (n.)". Online Etymology Dictionary. Retrieved 30 May 2022.
  • "Music". The American Heritage Dictionary of the English Language (fifth ed.). Boston: Houghton Mifflin Harcourt.
  • "music, n. and adj.". OED Online. Oxford: Oxford University Press. (subscription required)

Further reading

[edit]
  • Kennedy, Michal; Kennedy, Joyce Bourne (2013) [2012]. Tim Rutherford-Johnson (ed.). The Oxford Dictionary of Music (6th paperback ed.). Oxford: Oxford University Press. ISBN 978-0-19-957854-2.
  • Small, Christopher (1977). Music, Society, Education. John Calder Publishers, London. ISBN 0-7145-3614-8
  • Tymoczko, Dmitri (2011). A Geometry of Music: Harmony and Counterpoint in the Extended Common Practice. New York: Oxford University Press. ISBN 978-0-19-533667-2.
[edit]
  • Grove Music Online — online version of The New Grove Dictionary of Music and Musicians.
  • All ten volumes of the Garland Encyclopedia of World Music (subscription required)
  • Dolmetsch free online music dictionary, complete, with references to a list of specialised music dictionaries (by continent, by instrument, by genre, etc.)
  • Some books on music by Carl Van Vechten (1880–1964)

 

 

Music and artificial intelligence (music and AI) is the development of music software programs that use AI to generate music.[1] As with applications in other fields, AI in music also simulates mental tasks. A prominent feature is the capability of an AI algorithm to learn based on past data, such as in computer accompaniment technology, wherein the AI is capable of listening to a human performer and performing accompaniment.[2] Artificial intelligence also drives interactive composition technology, wherein a computer composes music in response to a live performance. There are other AI applications in music that cover not only music composition, production, and performance but also how music is marketed and consumed. Several music player programs have also been developed to use voice recognition and natural language processing technology for music voice control. Current research includes the application of AI in music composition, performance, theory and digital sound processing. Composers/artists like Jennifer Walshe or Holly Herndon have been exploring aspects of music AI for years in their performances and musical works. Another original approach of humans “imitating AI” can be found in the 43-hour sound installation String Quartet(s) by Georges Lentz[3].

20th century art historian Erwin Panofsky proposed that in all art, there existed three levels of meaning: primary meaning, or the natural subject; secondary meaning, or the conventional subject; and tertiary meaning, the intrinsic content of the subject.[4][5] AI music explores the foremost of these, creating music without the "intention" that is usually behind it, leaving composers who listen to machine-generated pieces feeling unsettled by the lack of apparent meaning.[6]

History

[edit]

In the 1950s and the 1960s, music made by artificial intelligence was not fully original, but generated from templates that people had already defined and given to the AI, with this being known as rule-based systems. As time passed, computers became more powerful, which allowed machine learning and artificial neural networks to help in the music industry by giving AI large amounts of data to learn how music is made instead of predefined templates. By the early 2000s, more advancements in artificial intelligence had been made, with generative adversarial networks (GANs) and deep learning being used to help AI compose more original music that is more complex and varied than possible before. Notable AI-driven projects, such as OpenAI’s MuseNet and Google’s Magenta, have demonstrated AI’s ability to generate compositions that mimic various musical styles.[7]

Timeline

[edit]

Artificial intelligence finds its beginnings in music with the transcription problem: accurately recording a performance into musical notation as it is played. Père Engramelle's schematic of a "piano roll", a mode of automatically recording note timing and duration in a way which could be easily transcribed to proper musical notation by hand, was first implemented by German engineers J.F. Unger and J. Hohlfield in 1952.[8]

In 1957, the ILLIAC I (Illinois Automatic Computer) produced the "Illiac Suite for String Quartet", a completely computer-generated piece of music. The computer was programmed to accomplish this by composer Leonard Isaacson and mathematician Lejaren Hiller.[6]: v–vii In 1960, Russian researcher Rudolf Zaripov published the first worldwide paper on algorithmic music composition using the Ural-1 computer.[9]

In 1965, inventor Ray Kurzweil developed software capable of recognizing musical patterns and synthesizing new compositions from them. The computer first appeared on the quiz show I've Got a Secret that same year.[10]

By 1983, Yamaha Corporation's Kansei Music System had gained momentum, and a paper was published on its development in 1989. The software utilized music information processing and artificial intelligence techniques to essentially solve the transcription problem for simpler melodies, although higher-level melodies and musical complexities are regarded even today as difficult deep-learning tasks, and near-perfect transcription is still a subject of research.[8][11]

In 1997, an artificial intelligence program named Experiments in Musical Intelligence (EMI) appeared to outperform a human composer at the task of composing a piece of music to imitate the style of Bach.[12] EMI would later become the basis for a more sophisticated algorithm called Emily Howell, named for its creator.

In 2002, the music research team at the Sony Computer Science Laboratory in Paris, led by French composer and scientist François Pachet, designed the Continuator, an algorithm uniquely capable of resuming a composition after a live musician stopped.[13]

Emily Howell would continue to make advancements in musical artificial intelligence, publishing her first album From Darkness, Light in 2009.[14] Since then, many more pieces by artificial intelligence and various groups have been published.

In 2010, Iamus became the first AI to produce a fragment of original contemporary classical music, in its own style: "Iamus' Opus 1". Located at the Universidad de Malága (Malága University) in Spain, the computer can generate a fully original piece in a variety of musical styles.[15][6]: 468–481  In August 2019, a large dataset consisting of 12,197 MIDI songs, each with their lyrics and melodies,[16] was created to investigate the feasibility of neural melody generation from lyrics using a deep conditional LSTM-GAN method.

With progress in generative AI, models capable of creating complete musical compositions (including lyrics) from a simple text description have begun to emerge. Two notable web applications in this field are Suno AI, launched in December 2023, and Udio, which followed in April 2024.[17]

In November 2025 the AI generated song "Walk My Walk" presented as being by Breaking Rust topped the Billboard Country Digital Song Sales chart.[18] The same year, AI band The Velvet Sundown attracted one million listeners on Spotify.[19]

Streaming service Deezer started tagging AI generated songs and excluding them from its editorialized playlists. Their tool builds on former published research work on the nature of AI music's artefacts.[20] In November 2025, the service claimed that 50,000 AI generated songs were uploaded daily, about a third of total uploads.[19]

Software applications

[edit]

ChucK

[edit]

Developed at Princeton University by Ge Wang and Perry Cook, ChucK is a text-based, cross-platform language.[21] By extracting and classifying the theoretical techniques it finds in musical pieces, the software is able to synthesize entirely new pieces from the techniques it has learned.[22] The technology is used by SLOrk (Stanford Laptop Orchestra)[23] and PLOrk (Princeton Laptop Orchestra).

Jukedeck

[edit]

Jukedeck was a website that let people use artificial intelligence to generate original, royalty-free music for use in videos.[24][25] The team started building the music generation technology in 2010,[26] formed a company around it in 2012,[27] and launched the website publicly in 2015.[25] The technology used was originally a rule-based algorithmic composition system,[28] which was later replaced with artificial neural networks.[24] The website was used to create over 1 million pieces of music, and brands that used it included Coca-Cola, Google, UKTV, and the Natural History Museum, London.[29] In 2019, the company was acquired by ByteDance.[30][31][32]

MorpheuS

[edit]

MorpheuS[33] is a research project by Dorien Herremans and Elaine Chew at Queen Mary University of London, funded by a Marie SkÅ‚odowská-Curie EU project. The system uses an optimization approach based on a variable neighborhood search algorithm to morph existing template pieces into novel pieces with a set level of tonal tension that changes dynamically throughout the piece. This optimization approach allows for the integration of a pattern detection technique in order to enforce long term structure and recurring themes in the generated music. Pieces composed by MorpheuS have been performed at concerts in both Stanford and London.

AIVA

[edit]

Created in February 2016, in Luxembourg, AIVA is a program that produces soundtracks for any type of media. The algorithms behind AIVA are based on deep learning architectures[34] AIVA has also been used to compose a Rock track called On the Edge,[35] as well as a pop tune Love Sick[36] in collaboration with singer Taryn Southern,[37] for the creation of her 2018 album "I am AI".

Google Magenta

[edit]
20-second music clip generated by MusicLM using the prompt "hypnotic ambient electronic music"

Google's Magenta team has published several AI music applications and technical papers since their launch in 2016.[38] In 2017 they released the NSynth algorithm and dataset,[39] and an open source hardware musical instrument, designed to facilitate musicians in using the algorithm.[40] The instrument was used by notable artists such as Grimes and YACHT in their albums.[41][42] In 2018, they released a piano improvisation app called Piano Genie. This was later followed by Magenta Studio, a suite of 5 MIDI plugins that allow music producers to elaborate on existing music in their DAW.[43] In 2023, their machine learning team published a technical paper on GitHub that described MusicLM, a private text-to-music generator which they'd developed.[44][45]

Riffusion

[edit]
Generated spectrogram from the prompt "bossa nova with electric guitar" (top), and the resulting audio after conversion (bottom)

Riffusion is a neural network, designed by Seth Forsgren and Hayk Martiros, that generates music using images of sound rather than audio.[46]

The resulting music has been described as "de otro mundo" (otherworldly),[47] although unlikely to replace man-made music.[47] The model was made available on December 15, 2022, with the code also freely available on GitHub.[48]

The first version of Riffusion was created as a fine-tuning of Stable Diffusion, an existing open-source model for generating images from text prompts, on spectrograms,[46] resulting in a model which used text prompts to generate image files which could then be put through an inverse Fourier transform and converted into audio files.[48] While these files were only several seconds long, the model could also use latent space between outputs to interpolate different files together[46][49] (using the img2img capabilities of SD).[50] It was one of many models derived from Stable Diffusion.[50]

In December 2022, Mubert[51] similarly used Stable Diffusion to turn descriptive text into music loops. In January 2023, Google published a paper on their own text-to-music generator called MusicLM.[52][53]

Forsgren and Martiros formed a startup, also called Riffusion, and raised $4 million in venture capital funding in October 2023.[54][55]

Spike AI

[edit]

Spike AI is an AI-based audio plug-in, developed by Spike Stent in collaboration with his son Joshua Stent and friend Henry Ramsey, that analyzes tracks and provides suggestions to increase clarity and other aspects during mixing. Communication is done by using a chatbot trained on Spike Stent's personal data. The plug-in integrates into digital audio workstation.[56][57]

Musical applications

[edit]

Artificial intelligence can potentially impact how producers create music by giving reiterations of a track that follow a prompt given by the creator. These prompts allow the AI to follow a certain style that the artist is trying to go for.[6] AI has also been seen in musical analysis where it has been used for feature extraction, pattern recognition, and musical recommendations.[58] New tools that are powered by artificial intelligence have been made to help aid in generating original music compositions, like AIVA (Artificial Intelligence Virtual Artist) and Udio. This is done by giving an AI model data of already-existing music and having it analyze the data using deep learning techniques to generate music in many different genres, such as classical music or electronic music.[59] Musical and choral classrooms are already implementing AI driven programs to aid in musical learning, and development of creative skills in students where AI has shown to provide a statistically significant increase.[60][61]

[edit]

Several musicians such as Dua Lipa, Elton John, Nick Cave, Paul McCartney and Sting have criticized the use of AI in music and are encouraging the UK government to act on this matter.[62][63][64][65][66] Another example of this protest is the silent 2025 album Is This What We Want?.

Some artists have encouraged the use of AI in music such as Grimes.[67]

While helpful in generating new music, many issues have come up since artificial intelligence has begun making music. Some major concerns include how the economy will be impacted with AI taking over music production, who truly owns music generated by AI, and a lower demand for human-made musical compositions. Some critics argue that AI diminishes the value of human creativity, while proponents see it as an augmentative tool that expands artistic possibilities rather than replacing human musicians.[68][69]

Additionally, concerns have been raised about AI's potential to homogenize music. AI-driven models often generate compositions based on existing trends, which some fear could limit musical diversity. Addressing this concern, researchers are working on AI systems that incorporate more nuanced creative elements, allowing for greater stylistic variation.[59]

Another major concern about artificial intelligence in music is copyright laws. In the United States, the current legal framework tends to apply traditional copyright laws to AI, despite its differences with the human creative process.[70] However, music outputs solely generated by AI are not granted copyright protection. In the compendium of the U.S. Copyright Office Practices, the Copyright Office has stated that it would not grant copyrights to "works that lack human authorship" and "the Office will not register works produced by a machine or mere mechanical process that operates randomly or automatically without any creative input or intervention from a human author."[71] In February 2022, the Copyright Review Board rejected an application to copyright AI-generated artwork on the basis that it "lacked the required human authorship necessary to sustain a claim in copyright."[72] The usage of copyrighted music in training AI has also been a topic of contention. One instance of this was seen when SACEM, a professional organization of songwriters, composers, and music publishers demanded that PozaLabs, an AI music generation startup refrain from utilizing any music affiliated with them for training models.[73]

The situation in the European Union (EU) is similar to the US, because its legal framework also emphasizes the role of human involvement in a copyright-protected work.[74] According to the European Union Intellectual Property Office and the recent jurisprudence of the Court of Justice of the European Union, the originality criterion requires the work to be the author's own intellectual creation, reflecting the personality of the author evidenced by the creative choices made during its production, requires distinct level of human involvement.[74] The reCreating Europe project, funded by the European Union's Horizon 2020 research and innovation program, delves into the challenges posed by AI-generated contents including music, suggesting legal certainty and balanced protection that encourages innovation while respecting copyright norms.[74] The recognition of AIVA marks a significant departure from traditional views on authorship and copyrights in the realm of music composition, allowing AI artists capable of releasing music and earning royalties. This acceptance marks AIVA as a pioneering instance where an AI has been formally acknowledged within the music production.[75]

The recent advancements in artificial intelligence made by groups such as Stability AI, OpenAI, and Google has incurred an enormous sum of copyright claims leveled against generative technology, including AI music. Should these lawsuits succeed, the machine learning models behind these technologies would have their datasets restricted to the public domain.[76] Strides towards addressing ethical issues have been made as well, such as the collaboration between Sound Ethics(a company promoting ethical AI usage in the music industry) and UC Irvine, focusing on ethical frameworks and the responsible usage of AI.[77]

Musical deepfakes

[edit]

A more nascent development of AI in music is the application of audio deepfakes to cast the lyrics or musical style of a pre-existing song to the voice or style of another artist. This has raised many concerns regarding the legality of technology, as well as the ethics of employing it, particularly in the context of artistic identity.[78] Furthermore, it has also raised the question of to whom the authorship of these works is attributed. As AI cannot hold authorship of its own, current speculation suggests that there will be no clear answer until further rulings are made regarding machine learning technologies as a whole.[79] Most recently, preventative measures have started to be developed by Google and Universal Music group who have taken in royalties and credited attribution in order to allow producers to replicate the voices and styles of artists.[80]

"Heart on My Sleeve"

[edit]

In 2023, an artist known as ghostwriter977 created a musical deepfake called "Heart on My Sleeve" that cloned the voices of Drake and The Weeknd by inputting an assortment of vocal-only tracks from the respective artists into a deep-learning algorithm, creating an artificial model of the voices of each artist, to which this model could be mapped onto original reference vocals with original lyrics.[81] The track was submitted for Grammy consideration for the best rap song and song of the year.[82] It went viral and gained traction on TikTok and received a positive response from the audience, leading to its official release on Apple Music, Spotify, and YouTube in April 2023.[83] Many believed the track was fully composed by an AI software, but the producer claimed the songwriting, production, and original vocals (pre-conversion) were still done by him.[81] It would later be rescinded from any Grammy considerations due to it not following the guidelines necessary to be considered for a Grammy award.[83] The track would end up being removed from all music platforms by Universal Music Group.[83] The song was a watershed moment for AI voice cloning, and models have since been created for hundreds, if not thousands, of popular singers and rappers.

"Where That Came From"

[edit]

In 2013, country music singer Randy Travis suffered a stroke which left him unable to sing. In the meantime, vocalist James Dupré toured on his behalf, singing his songs for him. Travis and longtime producer Kyle Lehning released a new song in May 2024 titled "Where That Came From", Travis's first new song since his stroke. The recording uses AI technology to re-create Travis's singing voice, having been composited from over 40 existing vocal recordings alongside those of Dupré.[84][85]

Kanye West: Vultures 2, Bully, Donda 2, and In a Perfect World

[edit]

Since 2024, rapper Kanye West has been using artificial intelligence deepfakes of his own voice. His usage of deepfakes started during the production of his album Vultures 2, where the songs "Field Trip" and "Sky City" drew suspicion of artificial intelligence usage;[86] further updates to the songs "Forever"[86] and "530"[87] would also be accused of using AI. Artist Ty Dolla Sign, who released Vultures 2 with West, would confirm the allegations in 2025 during an interview.[88] West would subsequently confirm that his subsequent solo album, Bully, also used artificial intelligence,[89] and his later album In a Perfect World and the 2025 updated version of Donda 2 would also draw accusations of AI usage.[citation needed]

Playboi Carti: "Timeless" and Music

[edit]

Playboi Carti has also been accused of using artificial intelligence deepfakes. The allegations followed the release of the song "Timeless", a collaboration with The Weeknd, where his verse was accused of using artificial intelligence.[90] The allegations would further intensify after the release of his album Music, where "Rather Lie" and "Fine Shit" were also accused of using artificial intelligence. Playboi Carti would deny the accusations.[91]

Technical approaches

[edit]

Artificial intelligence music encompasses a number of technical approaches used for music composition, analysis, classification, and suggestion. Techniques used are drawn from deep learning, machine learning, natural language processing, and signal processing. Current systems are able to compose entire musical compositions, parse affective content, accompany human players in real-time, and acquire patterns of user and context-dependent preferences.[92][93][94][95]

Symbolic music composition

[edit]

Symbolic music generation is the generation of music in discrete symbolic forms such as MIDI, where note and timing are precisely defined. Early systems employed rule-based systems and Markov models, but modern systems employ deep learning to a large extent. Recurrent Neural Networks (RNNs), and more precisely Long Short-Term Memory (LSTM) networks, have been employed in modeling temporal dependencies of musical sequences. They may be used to generate melodies, harmonies, and counterpoints in various musical genres.[96]

Transformer models such as Music Transformer and MuseNet became more popular for symbolic generation due to their ability to model long-range dependencies and scalability. These models were employed to generate multi-instrument polyphonic music and stylistic imitations.[97]

Audio-based music generation

[edit]

This method generates music as raw audio waveforms instead of symbolic notation. DeepMind's WaveNet is an early example that uses autoregressive sampling to generate high-fidelity audio. Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs) are being used more and more in new audio texture synthesis and timbre combination of different instruments.[93]

NSynth (Neural Synthesizer), a Google Magenta project, uses a WaveNet-like autoencoder to learn latent audio representations and thereby generate completely novel instrumental sounds.[98]

Music information retrieval (MIR)

[edit]

Music Information Retrieval (MIR) is the extraction of musically relevant information from audio recordings to be utilized in applications such as genre classification, instrument recognition, mood recognition, beat detection, and similarity estimation. CNNs on spectrogram features have been very accurate on these tasks.[95] SVMs and k-Nearest Neighbors (k-NN) are also used for classification on features such as Mel-frequency cepstral coefficients (MFCCs).

Hybrid and interactive systems

[edit]

Hybrid systems combine symbolic and sound-based methods to draw on their respective strengths. They can compose high-level symbolic compositions and synthesize them as natural sound. Interactive systems in real-time allow for AI to instantaneously respond to human input to support live performance. Reinforcement learning and rule-based agents tend to be utilized to allow for human–AI co-creation in improvisation contexts.[94]

Affective computing and emotion-aware music systems

[edit]

Affective computing techniques enable AI systems to classify or create music based on some affective content. The models use musical features such as tempo, mode, and timbre to classify or influence listener emotions. Deep learning models have been trained for classifying music based on affective content and even creating music intended to have affective impacts.[99]

AI-based music recommendation systems

[edit]

Music recommenders employ AI to suggest tracks to users based on what they have heard, their tastes, and information available in context. Collaborative filtering, content-based filtering, and hybrid filtering are most widely applied, deep learning being utilized for fine-tuning. Graph-based and matrix factorization methods are used within commercial systems like Spotify and YouTube Music to represent complex user-item relationships.[100]

AI for automatic mixing and mastering

[edit]

AI is also used in audio engineering automation such as mixing and mastering. Such systems level, equalize, pan, and compress to give well-balanced sound outputs. Software such as LANDR and iZotope Ozone utilize machine learning in emulating professional audio engineers' decisions.[101]

Lyrics generation and songwriting aid

[edit]

Natural language generation also applies to songwriting assistance and lyrics generation. Transformer language models like GPT-3 have also been proven to be able to generate stylistic and coherent lyrics from input prompts, themes, or feeling. There even exist AI programs that assist with rhyme scheme, syllable count, and poem form.[102]

Multimodal and cross-modal systems

[edit]

Recent developments include multimodal AI systems that integrate music with other media, e.g., dance, video, and text. These can generate background scores in synchronization with video sequences or generate dance choreography from audio input. Cross-modal retrieval systems allow one to search for music using images, text, or gestures.[103]

See also

[edit]
  • Algorithmic composition
  • Automatic content recognition
  • Computational models of musical creativity
  • Generative artificial intelligence
  • Generative music
  • List of music software
  • Music information retrieval
  • OpenAI § Music generation
  • Xania Monet

References

[edit]
  1. ^ Herremans, Dorien; Chuan, Ching-Hua; Chew, Elaine (30 September 2018). "A Functional Taxonomy of Music Generation Systems". ACM Computing Surveys. 50 (5): 1–30. arXiv:1812.04186. doi:10.1145/3108242.
  2. ^ Dannenberg, Roger. Artificial Intelligence, Machine Learning, and Music Understanding (Preprint). S2CID 17787070.[unreliable source?]
  3. ^ Rutherford-Johnson, Tim (December 2024). "Quartered Space(s)" (PDF). KAIROS text. Retrieved 28 September 2025.
  4. ^ Panofsky, Erwin (1939). Studies in Iconology: Humanistic Themes in the Art of the Renaissance. Oxford University Press. OCLC 557904328.[page needed]
  5. ^ Dilly, Heinrich (2020). "Panofsky, Erwin: Zum Problem der Beschreibung und Inhaltsdeutung von Werken der bildenden Kunst". Kindlers Literatur Lexikon (KLL). pp. 1–2. doi:10.1007/978-3-476-05728-0_16027-1. ISBN 978-3-476-05728-0.
  6. ^ a b c d Miranda, Eduardo Reck, ed. (2021). Handbook of Artificial Intelligence for Music. doi:10.1007/978-3-030-72116-9. ISBN 978-3-030-72115-2.[page needed]
  7. ^ "AI and the Sound of Music". www.yalelawjournal.org. 22 November 2024. Retrieved 26 February 2025.
  8. ^ a b Roads, Curtis (1985). "Research in music and artificial intelligence". ACM Computing Surveys. 17 (2): 163–190. doi:10.1145/4468.4469.
  9. ^ Zaripov, Rudolf (1960). "Об алгоритмическом описании процесса сочинения музыки (On algorithmic description of process of music composition)". Proceedings of the USSR Academy of Sciences (in Russian). 132 (6).
  10. ^ "Ray Kurzweil". National Science and Technology Medals Foundation. Archived from the original on 10 September 2024. Retrieved 10 September 2024.
  11. ^ Katayose, Haruhiro; Inokuchi, Seiji (1989). "The Kansei Music System". Computer Music Journal. 13 (4): 72–77. doi:10.2307/3679555. JSTOR 3679555.
  12. ^ Johnson, George (11 November 1997). "Undiscovered Bach? No, a Computer Wrote It". The New York Times. Archived from the original on 19 March 2024. Retrieved 29 April 2020. Dr. Larson was hurt when the audience concluded that his piece -- a simple, engaging form called a two-part invention -- was written by the computer. But he felt somewhat mollified when the listeners went on to decide that the invention composed by EMI (pronounced Emmy) was genuine Bach.
  13. ^ Pachet, François (2003). "The Continuator: Musical Interaction with Style". Journal of New Music Research. 32 (3): 333–341. doi:10.1076/jnmr.32.3.333.16861. hdl:2027/spo.bbp2372.2002.044.
  14. ^ Lawson, Mark (22 October 2009). "This artificially intelligent music may speak to our minds, but not our souls". The Guardian. Retrieved 10 September 2024.
  15. ^ "Iamus: Is this the 21st century's answer to Mozart?". BBC News. 2 January 2013. Archived from the original on 10 September 2024. Retrieved 10 September 2024.
  16. ^ yy1lab (13 November 2024), yy1lab/Lyrics-Conditioned-Neural-Melody-Generation, archived from the original on 6 January 2024, retrieved 19 November 2024citation: CS1 maint: numeric names: authors list (link)
  17. ^ Nair, Vandana (11 April 2024). "AI-Music Platform Race Accelerates with Udio". Analytics India Magazine. Archived from the original on 19 April 2024. Retrieved 19 April 2024.
  18. ^ Ryan, Aaron (8 November 2025). "An AI-Generated Country Song Is Topping A Billboard Chart, And That Should Infuriate Us All". Whiskey Riff. Whiskey Riff (Momentum Media LLC). Retrieved 10 November 2025.
  19. ^ a b Singh, Jaspreet (12 November 2025). "Are you listening to bots? Survey shows AI music is virtually undetectable". Reuters.
  20. ^ Afchar, Darius (2025). "A Fourier Explanation of AI‑music Artifacts". Proceedings of the International Society for Music Information Retrieval Conference (ISMIR). arXiv:2506.19108. Retrieved 25 November 2025.
  21. ^ ChucK => Strongly-timed, On-the-fly Audio Programming Language Archived 18 November 2003 at the Wayback Machine. Chuck.cs.princeton.edu. Retrieved on 2010-12-22.
  22. ^ "Foundations of On-the-fly Learning in the ChucK Programming Language" (PDF). Archived (PDF) from the original on 21 April 2024. Retrieved 3 March 2024.
  23. ^ Driver, Dustin. (1999-03-26) Pro - Profiles - Stanford Laptop Orchestra (SLOrk), pg. 1 Archived 18 January 2012 at the Wayback Machine. Apple. Retrieved on 2010-12-22.
  24. ^ a b "From Jingles to Pop Hits, A.I. Is Music to Some Ears". The New York Times. 22 January 2017. Archived from the original on 28 February 2023. Retrieved 3 January 2023.
  25. ^ a b "Need Music For A Video? Jukedeck's AI Composer Makes Cheap, Custom Soundtracks". techcrunch.com. 7 December 2015. Archived from the original on 3 January 2023. Retrieved 3 January 2023.
  26. ^ "What Will Happen When Machines Write Songs Just as Well as Your Favorite Musician?". motherjones.com. Archived from the original on 3 February 2023. Retrieved 3 January 2023.
  27. ^ Cookson, Robert (7 December 2015). "Jukedeck's computer composes music at touch of a button". Financial Times. Archived from the original on 3 January 2023. Retrieved 3 January 2023.
  28. ^ "Jukedeck: the software that writes music by itself, note by note". Wired UK. Archived from the original on 3 January 2023. Retrieved 3 January 2023.
  29. ^ "Robot rock: how AI singstars use machine learning to write harmonies". standard.co.uk. March 2018. Archived from the original on 3 January 2023. Retrieved 3 January 2023.
  30. ^ "TIKTOK OWNER BYTEDANCE BUYS AI MUSIC COMPANY JUKEDECK". musicbusinessworldwide.com. 23 July 2019. Archived from the original on 8 February 2023. Retrieved 3 January 2023.
  31. ^ "As TikTok's Music Licensing Reportedly Expires, Owner ByteDance Purchases AI Music Creation Startup JukeDeck". digitalmusicnews.com. 23 July 2019. Archived from the original on 3 January 2023. Retrieved 3 January 2023.
  32. ^ "An AI-generated music app is now part of the TikTok group". sea.mashable.com. 24 July 2019. Archived from the original on 29 January 2023. Retrieved 3 January 2023.
  33. ^ Herremans, Dorien; Chew, Elaine (October 2019). "MorpheuS: Generating Structured Music with Constrained Patterns and Tension". IEEE Transactions on Affective Computing. 10 (4): 510–523. arXiv:1812.04832. Bibcode:2019ITAfC..10..510H. doi:10.1109/TAFFC.2017.2737984.
  34. ^ "A New AI Can Write Music as Well as a Human Composer". Futurism. 9 March 2017. Archived from the original on 19 April 2024. Retrieved 19 April 2024.
  35. ^ Technologies, Aiva (24 October 2018). "The Making of AI-generated Rock Music with AIVA". Medium. Archived from the original on 25 October 2018. Retrieved 19 April 2024.
  36. ^ Lovesick | Composed with AIVA Artificial Intelligence - Official Video with Lyrics | Taryn Southern. 2 May 2018. Archived from the original on 19 July 2018. Retrieved 25 October 2018.
  37. ^ Southern, Taryn (10 May 2018). "Algo-Rhythms: The future of album collaboration". TechCrunch. Archived from the original on 25 October 2018. Retrieved 19 April 2024.
  38. ^ "Welcome to Magenta!". Magenta. 1 June 2016. Archived from the original on 1 February 2023. Retrieved 19 April 2024.
  39. ^ Engel, Jesse; Resnick, Cinjon; Roberts, Adam; Dieleman, Sander; Norouzi, Mohammad; Eck, Douglas; Simonyan, Karen (17 July 2017). Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders. Proceedings of the 34th International Conference on Machine Learning. PMLR. pp. 1068–1077. arXiv:1704.01279.
  40. ^ Open NSynth Super, Google Creative Lab, 13 February 2023, archived from the original on 19 December 2022, retrieved 14 February 2023
  41. ^ "Cover Story: Grimes is ready to play the villain". Crack Magazine. Archived from the original on 3 June 2023. Retrieved 14 February 2023.
  42. ^ "What Machine-Learning Taught the Band YACHT About Themselves". Los Angeleno. 18 September 2019. Archived from the original on 28 January 2023. Retrieved 14 February 2023.
  43. ^ "Magenta Studio". Magenta. Archived from the original on 13 January 2023. Retrieved 19 April 2024.
  44. ^ "MusicLM". google-research.github.io. 2023. Archived from the original on 1 February 2023. Retrieved 19 April 2024.
  45. ^ Sandzer-Bell, Ezra (16 February 2024). "Best Alternatives to Google's AI-Powered MusicLM and MusicFX". AudioCipher. Archived from the original on 1 February 2023. Retrieved 19 April 2024.
  46. ^ a b c Coldewey, Devin (15 December 2022). "Try 'Riffusion,' an AI model that composes music by visualizing it".
  47. ^ a b Llano, Eutropio (15 December 2022). "El generador de imágenes AI también puede producir música (con resultados de otro mundo)".
  48. ^ a b Nasi, Michele (15 December 2022). "Riffusion: creare tracce audio con l'intelligenza artificiale". IlSoftware.it.
  49. ^ "Essayez "Riffusion", un modèle d'IA qui compose de la musique en la visualisant". 15 December 2022.
  50. ^ a b "文章に沿った楽曲を自動生成してくれるAI「Riffusion」登場、画像生成AI「Stable Diffusion」ベースで誰でも自由に利用可能". GIGAZINE. 16 December 2022.
  51. ^ "Mubert launches Text-to-Music interface – a completely new way to generate music from a single text prompt". 21 December 2022.
  52. ^ "MusicLM: Generating Music From Text". 26 January 2023.
  53. ^ "5 Reasons Google's MusicLM AI Text-to-Music App is Different". 27 January 2023.
  54. ^ Gal, Dr. Itay (10 February 2025). "Free A.I. music creation platform launches, competing with Suno". The Jerusalem Post. Retrieved 16 February 2025.
  55. ^ Nuñez, Michael (30 January 2025). "Riffusion's free AI music platform could be the Spotify of the future". VentureBeat. Retrieved 16 February 2025.
  56. ^ Levine, Mike (4 October 2024). "Spike AI — A Mix Product of the Week". Mix. Future US. Archived from the original on 14 November 2024. Retrieved 19 November 2024.
  57. ^ "Spike Stent offers his expertise in Spike AI". Sound on Sound. Archived from the original on 17 December 2024. Retrieved 19 November 2024.
  58. ^ Zhang, Yifei (December 2023). "Utilizing Computational Music Analysis and AI for Enhanced Music Composition: Exploring Pre- and Post-Analysis". Journal of Advanced Zoology. 44 (S-6): 1377–1390. doi:10.17762/jaz.v44is6.2470.
  59. ^ a b Mycka, Jan; MaÅ„dziuk, Jacek (January 2025). "Artificial intelligence in music: recent trends and challenges". Neural Computing and Applications. 37 (2): 801–839. doi:10.1007/s00521-024-10555-x.
  60. ^ Zhang, Lele (June 2025). "Compositional tools based on artificial intelligence for choral artistic education: Enhancing creative skills in choral arrangements". Thinking Skills and Creativity. 56 101768. doi:10.1016/j.tsc.2025.101768.
  61. ^ Bryant, Robert (28 February 2025). "AI in the Music Classroom". yamaha.com. Retrieved 30 November 2025.
  62. ^ McMahon, Liv & Vallance, Chris (10 May 2025). "Elton John and Dua Lipa seek protection from AI". BBC News. Archived from the original on 15 May 2025. Retrieved 19 May 2025.
  63. ^ Kuenssberg, Laura (18 May 2025). "Elton John brands government 'losers' over AI copyright plans". BBC News. Archived from the original on 19 May 2025. Retrieved 19 May 2025.
  64. ^ Kuenssberg, Laura (26 January 2025). "Paul McCartney: Don't let AI rip off artists". BBC News. Archived from the original on 8 April 2025. Retrieved 19 May 2025.
  65. ^ Savage, Mark (18 May 2023). "Sting warns against AI songs as he wins prestigious music prize". BBC News. Archived from the original on 22 April 2025. Retrieved 19 May 2025.
  66. ^ Savage, Mark (17 January 2023). "Nick Cave says ChatGPT's AI attempt to write Nick Cave lyrics 'sucks'". BBC News. Archived from the original on 1 May 2025. Retrieved 19 May 2025.
  67. ^ Savage, Mark (25 April 2023). "Grimes says anyone can use her voice for AI-generated songs". BBC News. Archived from the original on 10 February 2025. Retrieved 19 May 2025.
  68. ^ Chen, Yanxu; Huang, Linshu; Gou, Tian (2024). "Applications and Advances of Artificial Intelligence in Music Generation:A Review". arXiv:2409.03715 [cs.SD].
  69. ^ Tegmark, Max (1 January 2017). Life 3.0 Being Human in the Age of Artificial Intelligence. Alfred A. Knopf. p. 124. ISBN 978-1-101-94659-6.
  70. ^ "Art created by AI cannot be copyrighted, says US officials – what does this mean for music?". MusicTech. Archived from the original on 2 April 2023. Retrieved 27 October 2022.
  71. ^ "Can (and should) AI-generated works be protected by copyright?". Hypebot. 28 February 2022. Archived from the original on 28 October 2022. Retrieved 27 October 2022.
  72. ^ Re: Second Request for Reconsideration for Refusal to Register A Recent Entrance to Paradise (Correspondence ID 1-3ZPC6C3; SR # 1-7100387071) (PDF) (Report). Copyright Review Board, United States Copyright Office. 14 February 2022. Archived (PDF) from the original on 24 July 2023. Retrieved 28 October 2022.
  73. ^ Pozalabs (8 December 2024). "Pozalabs Responds to SACEM, Addresses Ethical AI Music Generation Concerns" (Press release). Retrieved 5 April 2025 – via GlobalNewswire.
  74. ^ a b c Bulayenko, Oleksandr; Quintais, João Pedro; Gervais, Daniel J.; Poort, Joost (February 28, 2022). "AI Music Outputs: Challenges to the Copyright Legal Framework" Archived 28 March 2024 at the Wayback Machine. reCreating Europe Report. Retrieved 2024-04-03.
  75. ^ Ahuja, Virendra (June 11, 2021). "Artificial Intelligence and Copyright: Issues and Challenges" Archived 26 April 2024 at the Wayback Machine. ILI Law Review Winter Issue 2020. Retrieved 2024-04-03.
  76. ^ Samuelson, Pamela (14 July 2023). "Generative AI meets copyright". Science. 381 (6654): 158–161. doi:10.1126/science.x. PMID 37440639.
  77. ^ SoundEthics (4 February 2025). "Sound Ethics and UC Irvine Partner to Shape the Future of Ethical AI in the Music Industry" (Press release). Irvine, California. Retrieved 5 April 2025 – via Newswire.
  78. ^ "DeepDrake ft. BTS-GAN and TayloRVC: An Exploratory Analysis of Musical Deepfakes and Hosting Platforms" (PDF). Archived (PDF) from the original on 26 March 2024. Retrieved 6 March 2024.
  79. ^ "AI and Deepfake Voice Cloning: Innovation, Copyright and Artists' Rights" (PDF). Archived (PDF) from the original on 3 June 2024. Retrieved 6 March 2024.
  80. ^ Murgia, Madhumita; Nicolaou, Anna (8 August 2023). "Google and Universal Music negotiate deal over AI 'deepfakes'". Financial Times. Archived from the original on 28 March 2024. Retrieved 3 April 2024.
  81. ^ a b Robinson, Kristin (11 October 2023). "Ghostwriter, the Mastermind Behind the Viral Drake AI Song, Speaks For the First Time". Billboard. Archived from the original on 5 April 2024. Retrieved 3 April 2024.
  82. ^ "Drake/The Weeknd deepfake song "Heart on My Sleeve" submitted to Grammys". The FADER. Archived from the original on 3 April 2024. Retrieved 3 April 2024.
  83. ^ a b c "The AI deepfake of Drake and The Weeknd will not be eligible for a GRAMMY". Mixmag. Archived from the original on 3 April 2024. Retrieved 3 April 2024.
  84. ^ Marcus K. Dowling (6 May 2024). "Randy Travis' shocks music industry with AI pairing for 'Where That Came From.' How the song came together". The Tennesseean. Retrieved 6 May 2024.
  85. ^ Maria Sherman (6 May 2024). "With help from AI, Randy Travis got his voice back. Here's how his first song post-stroke came to be". AP News. Archived from the original on 7 May 2024. Retrieved 6 May 2024.
  86. ^ a b Horvath, Zachary (18 September 2024). "Kanye West Accused By Redditors Of Egregious AI Use On "VULTURES 2"". HotNewHipHop. Retrieved 13 July 2025.
  87. ^ Suman, Shraddha (11 November 2024). "Ye Faces Severe Fan Outrage, Following '530' Leak From 'Vultures 2' for AI Mumble". Netflix Junkie. Retrieved 13 July 2025.
  88. ^ REAL 92.3 LA (24 June 2025). Ty Dolla $ign DESTROYS AI Music: "Nope...Next!" + Vultures 2 AI Secrets Revealed. Retrieved 13 July 2025 – via YouTube.cite AV media: CS1 maint: numeric names: authors list (link)
  89. ^ Gee, Maddie (5 February 2025). "Ye Confirms Bully Release Date & Defends AI Use in New Album". RapTV. Retrieved 13 July 2025.
  90. ^ Nevares, Gabriel Bras (30 December 2024). "Playboi Carti Fans Peddle Dubious Conspiracy Theory About Him Using AI Vocals On New Songs With The Weeknd". HotNewHipHop. Retrieved 13 July 2025.
  91. ^ Nevares, Gabriel Bras (16 March 2025). "Playboi Carti Reportedly Denies Using AI On New Album "MUSIC" Amid Fans' Outrage". HotNewHipHop. Retrieved 13 July 2025.
  92. ^ Mycka, Jan; MaÅ„dziuk, Jacek (2024). "Artificial intelligence in music: recent trends and challenges". Neural Computing and Applications. 37 (2): 801–839. doi:10.1007/s00521-024-10555-x.
  93. ^ a b Briot, Jean-Pierre; Hadjeres, Gaëtan; Pachet, François-David (2017). "Deep learning techniques for music generation – A survey". arXiv:1709.01620 [cs.SD].
  94. ^ a b Herremans, Dorien; Chuan, Ching-Hua; Chew, Elaine (2017). "A functional taxonomy of music generation systems". ACM Computing Surveys. 50 (5): 69. arXiv:1812.04186. doi:10.1145/3108242.
  95. ^ a b Sturm, Bob L.; Søndergaard, Martin (2019). "Music Information Retrieval and Artificial Intelligence: Best Friends or Worst Enemies?". Transactions of the International Society for Music Information Retrieval. 2 (1): 1–19. doi:10.5334/tismir.28 (inactive 1 July 2025).cite journal: CS1 maint: DOI inactive as of July 2025 (link)
  96. ^ Briot, Jean-Pierre; Hadjeres, Gaëtan; Pachet, François-David (2017). "Deep learning techniques for music generation – A survey". arXiv:1709.01620 [cs.SD].
  97. ^ Huang, Cheng-Zhi Anna (2018). "Music Transformer: Generating Music with Long-Term Structure". International Conference on Learning Representations. arXiv:1809.04281.
  98. ^ Engel, Jesse (2017). "Neural Audio Synthesis of Musical Notes with WaveNet Autoencoders". Proceedings of the 34th International Conference on Machine Learning. arXiv:1704.01279.
  99. ^ Kim, Yejin; Elliott, Mark T.; Kim, Youngmoo E. (2020). "Modeling Musical Affect Using Deep Neural Networks". Proceedings of the International Society for Music Information Retrieval.
  100. ^ Schedl, Markus (2021). "Deep Learning in Music Recommendation Systems: A Survey". Journal of New Music Research. 50 (3): 232–247. doi:10.1080/09298215.2021.1939563 (inactive 1 July 2025).cite journal: CS1 maint: DOI inactive as of July 2025 (link)
  101. ^ Bocko, Mark (2019). "The Role of Artificial Intelligence in Automated Music Production". Audio Engineering Society Conference.
  102. ^ Vechtomova, Olga (2021). "Lyrics Generation with Neural Networks: Challenges and Opportunities". Transactions of the International Society for Music Information Retrieval.
  103. ^ Choi, Keunwoo (2019). "Theoretical model for the Seebeck coefficient in superlattice materials with energy relaxation". Journal of Applied Physics. 126 (5) 055105. arXiv:1908.01090. Bibcode:2019JAP...126e5105V. doi:10.1063/1.5108607.

Further reading

[edit]
  • Understanding Music with AI: Perspectives on Music Cognition Archived 2021-01-10 at the Wayback Machine. Edited by Mira Balaban, Kemal Ebcioglu, and Otto Laske. AAAI Press.
  • Proceedings of a Workshop held as part of AI-ED 93, World Conference on Artificial Intelligence in Education on Music Education: An Artificial Intelligence Approach
  • Tanguiane (Tangian), Andranick (1993). Artificial Perception and Music Recognition. Lecture Notes in Artificial Intelligence. Vol. 746. Berlin-Heidelberg: Springer. ISBN 978-3-540-57394-4.
  • Artificial Intelligence - Intelligent Art? Human-Machine Interaction and Creative Practice. (Digital Society - Digitale Gesellschaft). Edited by Voigts, Eckart, Robin Auer, Dietmar Elflein, Sebastian Kunas, Jan Röhnert, Christoph Seelinger Bielefeld: transcript,2024.
[edit]
  • The Music Informatics Research Group
  • Institut de Recherche et Coordination Acoustique/ Musique
  • Interdisciplinary Centre for Research in Music
  • Mixdevil - Is AI Good Gor Music Producers
  • OpenDream
  • AI full track generation

 

 

British singer-songwriter Ed Sheeran in 2023.

A musician is someone who composes, conducts, or performs music.[1] According to the United States Employment Service, "musician" is a general term used to designate a person who follows music as a profession.[2] Musicians include songwriters, who write both music and lyrics for songs; conductors, who direct a musical performance; and performers, who perform for an audience. A music performer is generally either a singer (also known as a vocalist), who provides vocals, or an instrumentalist, who plays a musical instrument. Musicians may perform on their own or as part of a group, band or orchestra.

Musicians may specialize in a musical genre, although many perform across a variety of styles and often blend or cross genres. Their musical output is influenced by a range of factors, including culture, skill set, life experience, education, and creative preferences. A musician who records and releases music is often referred to as a recording artist;[3] while a musician who primarily writes and records art music, media soundtracks , and scores—including for films, theatrical stage productions, and video games—or classical music is usually called a composer. A music producer, while not a musician outright, may themselves be a musician. A conductor, by way of gestures, leads and provides live direction for, or conducts, an ensemble, typically an orchestra or choral group.

Types

[edit]
George Michael at Antwerp.

Composer

[edit]

A composer is a musician who creates musical compositions. The title is principally used for those who write classical music or film music. Those who write the music for popular songs may be called songwriters. Those who mainly write the words for songs may be referred to as lyricists.

Conductor

[edit]

A conductor directs a musical performance; conducting has been defined as "the art of directing the simultaneous performance of several players or singers by the use of gesture".[4] The conductor stands on a raised podium and communicates with the musicians through hand gestures or eye contact.

Performer

[edit]

Examples of performers include, but are not limited to, instrumentalists and singers who perform for an audience. A musician can perform as a solo artist or as a part of an ensemble (e.g. an orchestra, a choir or a pop group).

Producer

[edit]

A music producer plays a crucial role in shaping the sound of an album or song, overseeing the recording process, and guiding the overall creative direction. They often collaborate closely with artists, musicians, sound engineers, and other industry professionals.

See also

[edit]
  • Health problems of musicians
  • Lists of musicians
  • List of online music databases
  • Record producer
  • Songwriter
  • Singing
  • Record sales

References

[edit]
  1. ^ "Musician". American Heritage Dictionary.
  2. ^ Dictionary of Occupational Titles, Volume 1. U.S. Government Printing Office. 1949. p. 883.
  3. ^ "Recording Artist (or Group)". Berklee College of Music. January 2011. Archived from the original on 28 October 2020. Retrieved 19 February 2021.
  4. ^ Sir George Grove, John Alexander Fuller Maitland, eds. (1922). A Dictionary of Music and Musicians, Volume 1, p. 581. Macmillan.
[edit]