Notebookcheck Logo

Nvidia's Fugatto 1 can synthesize audio to create new sounds

Nvidia's Fugatto 1 can synthesize audio to create new sounds (Image Source: Fugatto Github)
Nvidia's Fugatto 1 can synthesize audio to create new sounds (Image Source: Fugatto Github)
Nvidia has unveiled a new generative AI model that can synthesize audio to create unique sounds. The company says the model will be marketed to the entertainment industry for music, films, and video games.

Nvidia has shown off a new generative AI model that can synthesize audio through simple text instructions and contextual audio inputs to create unique sounds. Nvidia envisions Fugatto 1 "as a tool for creatives, empowering them to quickly bring their sonic fantasies and unheard sounds to life—an instrument for imagination, not a replacement for creativity."

In its research paper, the team says Large Language Models (LLMs) trained on text can learn to infer instructions from inputs, but LLMs trained purely on audio cannot do that. Audio does not have data that shows how it was created. 

Nvidia's Fugatto 1 uses a specialized dataset that pulls from a wide gamut of sounds and a method for understanding and controlling instructions called ComposeableART. This allows the model to create an emergent dataset that can help the model combine different sounds, even ones it wasn't trained to handle.

Nvidia has shown off a few examples of the model in action on Fugatto's Github page, such as the ability to synthesize the sound of a dog barking in time with electronic dance music, a typewriter that whispers every letter typed, and even a saxophone that meows or barks.

So far, Nvidia has no plans to release the model publicly.

Read all 1 comments / answer
static version load dynamic
Loading Comments
Comment on this article
Please share our article, every link counts!
Mail Logo
> Expert Reviews and News on Laptops, Smartphones and Tech Innovations > News > News Archive > Newsarchive 2024 11 > Nvidia's Fugatto 1 can synthesize audio to create new sounds
Rohith Bhaskar, 2024-11-26 (Update: 2024-11-26)

OSZAR »