Tech

Google AI tool uses written descriptions to create music

Share

A paper describing the results of a music-making artificial intelligence (AI) tool was published this week by Google researchers.

The AI music tool MusicLM is not the first to be released. Google, on the other hand, uses a limited set of descriptive words to illustrate musical creativity in the examples it provides.

AI demonstrates how human-like behavior has been taught to complex computer systems.

Tools like ChatGPT can quickly produce written documents that are comparable to human efforts. To operate intricate machine-learning models, ChatGPT and similar systems necessitate powerful computers. Late last year saw the launch of ChatGPT by OpenAI, a company based in San Francisco.

These systems, including AI voice generators, are programmed using a vast amount of data to learn and replicate various types of content. Written content, design elements, art, or music are all examples of computer-generated content.

ChatGPT has recently garnered a lot of attention for its capacity to generate intricate writings and other content from a straightforward natural language description.

MusicLM from Google

The MusicLM system is explained by Google engineers as follows:

A user’s first step is to think of a few words that best describe the kind of music they want the tool to make.

A user could, for instance, enter the following succinct phrase into the system: “a continuous calming violin backed by a soft guitar sound.” The descriptions that are entered may include various musical genres, instruments, or other sounds that already exist.

MusicLM produced a number of distinct music examples that were made available online. Some of the music that was made was based on just one or two words, like “jazz,” “rock,” or “techno.” Other examples were generated by the system from more in-depth descriptions that included entire sentences.

For example, Google researchers provide MusicLM with the following instructions: “The main soundtrack of an arcade game. It is fast-paced and upbeat, with a catchy electric guitar riff. The music is repetitive and easy to remember, but with unexpected sounds…”

In the final recording, the music seems to stay very close to what was described. According to the team, the system can attempt to produce a better description with more detail.

The machine-learning platforms used by ChatGPT are analogous to how the MusicLM model operates. Because they are trained on huge amounts of data, these tools can produce human-like results. The systems are fed a wide variety of materials to enable them to acquire complex skills for creating realistic works.

According to the team, the system can also create examples based on a person’s own singing, humming, whistling, or instrument playing, in addition to creating new music from written descriptions.

The tool “produces high-quality music…over several minutes, while being faithful to the text conditioning signal,” according to the researchers.

The MusicLM models have not yet been made available to the general public by the Google team. This is in contrast to ChatGPT, which was made accessible online in November for users to try out.

However, MusicCaps, a “high-quality dataset” composed of over 5,500 music-writing pairs prepared by professional musicians, was announced by Google. This action was taken by the researchers to aid in the creation of additional AI music generators.

According to the MusicLM researchers, they are confident that they have developed a novel instrument that will enable anyone to quickly and easily produce music selections of high quality. However, the team stated that it also recognizes some machine learning-related risks.

“biases present in the training data” was one of the most significant issues the researchers identified. Including too much of one side and not enough of the other could be considered bias. “About appropriateness for music generation for cultures underrepresented in the training data,” the researchers stated, “raises a question.”

The team stated that it intends to continue to study any system results that could be regarded as cultural appropriation. Through additional development and testing, the objective would be to reduce biases.

In addition, the researchers stated that they intend to continue improving the system to include better voice and music quality, text conditioning, and the generation of lyrics.

Raeesa Sayyad

Recent Posts

How to Reduce the Interest Rate on Your Credit Card

There is a real cost to carrying a balance on a credit card, even though… Read More

17 hours ago

Best Free Reverse Phone Lookup Services USA: Top Five Picks

Even with technological advancements, people are still getting scammed online and through phone calls. The… Read More

20 hours ago

Micah Raskin: Interested in Becoming A Professional Poker Player? Here Are Some Tips to Get You Started

Are you interested in becoming a professional poker player? If so, you're in the right… Read More

22 hours ago

Beyond Neon Lights: A Traveler’s Guide to Las Vegas

Known as the "Entertainment Capital of the World," Las Vegas is well-known for its opulent… Read More

22 hours ago

Extend the Life of High-End Watches in New York City with First Class Timepieces’ Trade-In Program

NEW YORK, NY - In the bustling heart of New York City, First Class Timepieces… Read More

1 day ago

How Can You Protect Your Disability Claim From the Investigation Tools Used by Disability Insurance Companies

You may have counted on your long term disability insurance claim to be treated like… Read More

1 day ago