bnew

Veteran
Joined
Nov 1, 2015
Messages
63,087
Reputation
9,641
Daps
172,661

Using AI to Decode Animal Communication with Aza Raskin​




76,397 views Aug 1, 2023
From crows to dolphins, gelada monkeys to primrose flowers - Aza Raskin, co-founder of Earth Species Project, shares how the latest advances in AI help us to better understand and learn from other species. In this talk, learn how our ability to communicate with other species could transform the way humans relate to the rest of nature. This talk was recorded at Summit At Sea in May 2023.
 

MushroomX

Packers Stockholder
Supporter
Joined
Aug 17, 2013
Messages
28,073
Reputation
9,389
Daps
118,190
Reppin
Wisconsin
:russ: Feel free to be Aquaman first my breh. If you can convince me that you can get a Great White to invest in Crypto, then maybe you have something.
 

mastermind

Rest In Power Kobe
Joined
May 1, 2012
Messages
64,034
Reputation
6,422
Daps
170,562
I remember thinking about this shyt, should be pretty straightforward. Induce certain behaviors then use the AI to map out all external vocalizations, body movements etc. shyt like this makes me excited for the possibilities.
You should listen to that episode because their are ethical issues that need to be considered too
 

GnauzBookOfRhymes

Superstar
Joined
May 7, 2012
Messages
12,596
Reputation
2,862
Daps
48,082
Reppin
NULL
I think LLMs are tailor made for this. Hopefully in a couple of years, we can find ways to "understand" animals better. AI powered cat and dog translators would sell like hot cakes I feel :russ:

A lot of pet owners gonna be depressed when they find out how much their dogs/cats hate their guts but don’t do anything because their instinct for survival tells them not to hurt the person who feeds/houses them :mjlol:
 

GnauzBookOfRhymes

Superstar
Joined
May 7, 2012
Messages
12,596
Reputation
2,862
Daps
48,082
Reppin
NULL
You should listen to that episode because their are ethical issues that need to be considered too

I dunno. I’m definitely one of those “try your hardest to suppress your ethical/moral misgivings” when it comes to animals being delicious or providing certain types of entertainment/education. I buy the free range eggs even though they’re more expensive but free range chicken is double the price and half the size lol.

Except for bullfighting and/or animal combat. That’s just barbaric.
 

mastermind

Rest In Power Kobe
Joined
May 1, 2012
Messages
64,034
Reputation
6,422
Daps
170,562
I dunno. I’m definitely one of those “try your hardest to suppress your ethical/moral misgivings” when it comes to animals being delicious or providing certain types of entertainment/education. I buy the free range eggs even though they’re more expensive but free range chicken is double the price and half the size lol.

Except for bullfighting and/or animal combat. That’s just barbaric.
I think you should listen to the part about ethics.
 

GnauzBookOfRhymes

Superstar
Joined
May 7, 2012
Messages
12,596
Reputation
2,862
Daps
48,082
Reppin
NULL
I think you should listen to the part about ethics.

Just did and thanks. I was so focused on the idea of translating their language I didn't even really think about potential consequences of humans trying to communicate back. I've no scientific basis for this but something tells me that animals will not be so easily fooled.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,087
Reputation
9,641
Daps
172,661


Scientists Have Reported a Breakthrough In Understanding Whale Language​

Researchers have identified new elements of whale vocalizations that they propose are analogous to human speech, including vowels and pitch.

By Jordan Pearson
December 7, 2023, 10:19am


1701962301590-your-paragraph-text.png

IMAGE:
REINHARD DIRSCHERL VIA GETTY IMAGES


Researchers have identified previously unknown elements of whale vocalizations that may be analogous to human speech, a new study reports.

Sperm whales are giants of the deep, with healthy adults having no known predators. Scientists studying their vocalizations have already picked out key elements of their communication, namely clicks, sequences of which are called codas. Now, researchers led by Gašper Beuš from the University of California, Berkeley report the discovery that the acoustic properties of these clicks—for example, pitch—are “on many levels analogous to human vowels and diphthongs,” which is when one vowel sound morphs into another such as in the word “coin.” The researchers even identify two unique “coda vowels” that are “actively exchanged” in conversation between whales, which they term the a-vowel and i-vowel.

The researchers explain in their paper, published as a preprint online this week, that the first clue that so-called spectral properties could be meaningful for whale speech was provided by AI. Beuš previously developed a deep learning model for human language called fiwGAN which “was trained to imitate sperm whale codas and embed information into these vocalizations.” Not only did the AI predict elements of whale vocalizations already thought to be meaningful, such as clicks, but it also singled out acoustic properties.

To follow up on the AI’s tip, the researchers analyzed a dataset of 3948 sperm whale codas recorded with hydrophones placed directly on whales between 2014 and 2018. They only analyzed one channel from the hydrophones to control for underwater effects and whale movement, and removed click timing from their visualization to better isolate patterns in the acoustic properties themselves.



These visualizations vindicated the AI’s prediction: The whales reliably exchanged codas with one or two formants—frequency peaks in the sound wave—below the 10kHz range. The researchers termed these codas “vowels,” with single-formant codas being a-vowels and two-formant codas being i-vowels. “This is by analogy to human vowels which differ in their formant frequencies,” the authors wrote. They also identified upward and downward frequency “trajectories” in these codas, which they considered analogous to diphthongs in human language.

Considering that these coda vowel patterns were very distinct and not intermixed, plus the existence of diphthongs, the researchers argue that whales are controlling the frequency of their vocalizations.

“Under our proposed view, whale clicks are equivalent to the pulses of vocal folds in human speech production,” the authors wrote. “In other words, we treat clicks as the source and the sperm whales’ resonant body (the nasal complex, including the spermaceti organ) as the filter that modulates resonant frequencies.”

The analogies to human speech are readily apparent. The authors note, for example, that vocal tone in Mandarin can change the meaning of otherwise identical syllables.

“If our findings are correct, it means that the communication of sperm whales is much more complex and can carry more information than previously thought,” the researchers concluded.
 

GnauzBookOfRhymes

Superstar
Joined
May 7, 2012
Messages
12,596
Reputation
2,862
Daps
48,082
Reppin
NULL
This is dope usage of AI.

AI is going to open up perspectives/dimensions in nature that we have never experienced or have only theorized.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,087
Reputation
9,641
Daps
172,661

The sperm whale 'phonetic alphabet' revealed by AI​

1 day ago

By Katherine Latham and Anna Bressanin,

Share

p0j9d4hj.jpg.webp

Amanda Cotton/Project CETI
Sperm whale communication may have similarities to human language (Credit: Amanda Cotton/Project CETI)

Researchers studying sperm whale communication say they've uncovered sophisticated structures similar to those found in human language.

In the inky depths of the midnight zone, an ocean giant bears the scars of the giant squid she stalks. She searches the darkness, her echolocation pulsing through the water column. Then she buzzes – a burst of rapid clicks – just before she goes in for the kill.

But exactly how sperm whales catch squid, like many other areas of their lives, remains a mystery. "They're slow swimmers," says Kirsten Young, a marine scientist at the University of Exeter. Squid, on the other hand, are fast. "How can [sperm whales] catch squid if they can only move at 3 knots [5.5 km/h or 3.5mph]? Are the squid moving really slowly? Or are the whales stunning them with their vocalisations? What happens down there? Nobody really knows," she says.

Sperm whales are not easy to study. They spend much of their lives foraging or hunting at depths beyond the reach of sunlight. They are capable of diving over 3km (10,000ft) and can hold their breath for two hours.

p0j9d575.jpg.webp

Amanda Cotton/Project CETI
Sperm whales are in constant communication with one another, even when foraging alone at depth (Credit: Amanda Cotton/Project CETI)

"At 1000m (3300ft) deep, many of the group will be facing the same way, flanking each other – but across an area of several kilometres," says Young. "During this time they're talking, clicking the whole time." After about an hour, she says, the group rises to the surface in synchrony. "They'll then have their rest phase. They might be at the surface for 15 to 20 minutes. Then they'll dive again," she says.

At the end of a day of foraging, says Young, the sperm whales come together at the surface and rub against each other, chatting while they socialise. "As researchers, we don't see a lot of their behaviour because they don't spend that much time at the surface," she says. "There's masses we don't know about them, because we are just seeing a tiny little snapshot of their lives during that 15 minutes at the surface."

It was around 47 million years ago that land-roaming cetaceans began to gravitate back towards the ocean – that's 47 million years of evolution in an environment alien to our own. How can we hope to easily understand creatures that have adapted to live and communicate under such different evolutionary pressures to ourselves?

"It's easier to translate the parts where our world and their world overlap – like eating, nursing or sleeping," says David Gruber, lead and founder of the Cetacean Translation Initiative (Ceti) and professor of biology at the City University of New York. "As mammals, we share these basics with others. But I think it's going to get really interesting when we try to understand the areas of their world where there's no intersection with our own," he says.

p0j9d582.jpg.webp

Project CETI
The Dominica Sperm Whale Project has been listening to sperm whales for almost 20 years (Credit: Project CETI)

Now, from elephants to dogs, modern technology is helping researchers to sift through enormous datasets, and uncover previously unknown diversity and complexity in animal communication. And Ceti's researchers say they, too, have used AI to decode a "sperm whale phonetic alphabet".

In 2005, Shane Gero, biology lead for Ceti, founded The Dominica Sperm Whale Project to study the social and vocal behaviour of around 400 sperm whales that live in the Eastern Caribbean. Almost 20 years – and thousands of hours of observation – later, the researchers have discovered intricacies in whale vocalisations never before observed, revealing structures within sperm whale communication akin to human language.

We're at base camp. This is a new place for humans to be – David Gruber

Sperm whales live in multi-level, matrilineal societies – groups of daughters, mothers and grandmothers – while the males roam the oceans, visiting the groups to breed. They are known for their complex social behaviour and group decision-making, which requires sophisticated communication. For example, they are able to adapt their behaviour as a group when protecting themselves from predators like orcas or humans.

Sperm whales communicate with each other using rhythmic sequences of clicks, called codas. It was previously thought that sperm whales had just 21 coda types. However, after studying almost 9,000 recordings, the Ceti researchers identified 156 distinct codas. They also noticed the basic building blocks of these codas which they describe as a "sperm whale phonetic alphabet" – much like phonemes, the units of sound in human language which combine to form words. (Watch the video below to hear some of the variety in sperm whale vocalisations the AI identified.)

p0j9t32m.jpg

2:25
The secret coda of whales (Video by Anna Bressanin and Katherine Latham)


Pratyusha Sharma, a PhD student at MIT and lead author of the study, describes the "fine-grain changes" in vocalisations the AI identified. Each coda consists of between three and 40 rapid-fire clicks. The sperm whales were found to vary the overall speed, or the "tempo", of the codas, as well as to speed up and slow down during the delivery of a coda, in other words, making it "rubato". Sometimes they added an extra click at the end of a coda, akin, says Sharma, to "ornamentation" in music. These subtle variations, she says, suggest sperm whale vocalisations could carry a much richer amount of information than previously thought.

"Some of these features are contextual," says Sharma. "In human language, for example, I can say 'what' or 'whaaaat!?'. It's the same word, but to understand the meaning you have to listen to the whole sound," she says.

The researchers also found the sperm whale "phonemes" could be used in a combinatorial fashion, allowing the whales to construct a vast repertoire of distinct vocalisations. The existence of a combinatorial coding system, write the report authors, is a prerequisite for " duality of patterning" – a linguistic phenomenon thought to be unique to human language – in which meaningless elements combine to form meaningful words.

p0j9d5fh.jpg.webp

Project CETI
In 2023, drone footage captured the sights and sounds of a sperm whale calf's birth. Now researchers are analysing the whales' vocalisations from the event (Credit: Project CETI)

However, Sharma emphasises, this is not something they have any evidence of as yet. "What we show in sperm whales is that the codas themselves are formed by combining from this basic set of features. Then the codas get sequenced together to form coda sequences." Much like humans combine phonemes to create words, and then words to create sentences.

So, what does all this tell us about sperm whales' intelligence? Or their ability to reason, or store and share information?

"Well, it doesn't tell us anything yet," says Gruber. "Before we can get to those amazing questions, we need to build a fundamental understanding of how [sperm whales communicate] and what's meaningful to them. We see them living very complicated lives, the coordination and sophistication in their behaviours. We're at base camp. This is a new place for humans to be – just give us a few years. Artificial intelligence is allowing us to see deeper into whale communication than we've ever seen before."

But not everyone is convinced, with experts warning of an anthropocentric focus on language which risks forcing us to view things from one perspective.

More like this:

The scientists learning to speak whale

Scientists built this listening network to detect nuclear bomb tests. It found blue whales instead

The unknown giants of the deep oceans

Young, though, describes the research as an "incremental step" towards understanding these giants of the deep. "We're starting to put the pieces of the puzzle together," she says. And perhaps if we could listen and really understand something like how important sperm whales' grandmothers are to them – something that resonates with humans, she says, we could drive change in human behaviour in order to protect them.

Categorised as " vulnerable" by the International Union for Conservation of Nature (IUCN), sperm whales are still recovering from commercial hunting by humans in the 19th and 20th Centuries. And, although such whaling has been banned for decades, sperm whales face new threats such as climate change, ocean noise pollution and ship strikes.

However, Young adds, we're still a long way off from understanding what sperm whales might be saying to each other. "We really have no idea. But the better we can understand these amazing animals, the more we'll know about how we can protect them."

--
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,087
Reputation
9,641
Daps
172,661



Google’s New AI Is Trying to Talk to Dolphins—Seriously​


A new AI model produced by computer scientists in collaboration with dolphin researchers could open the door to two-way animal communication.

By Isaac Schultz Published April 15, 2025 | Comments (25)

A bottlenose dolphin underwater.
A bottlenose dolphin underwater. Photo: טל שמע

In a collaboration that sounds straight out of sci-fi but is very much grounded in decades of ocean science, Google has teamed up with marine biologists and AI researchers to build a large language model designed not to chat with humans, but with dolphins.

The model is DolphinGemma, a cutting-edge LLM trained to recognize, predict, and eventually generate dolphin vocalizations, in an effort to not only crack the code on how the cetaceans communicate with each other—but also how we might be able to communicate with them ourselves. Developed in partnership with the Wild Dolphin Project (WDP) and researchers at Georgia Tech, the model represents the latest milestone in a quest that’s been swimming along for more than 40 years.

A deep dive into a dolphin community​


Since 1985, WDP has run the world’s longest underwater study of dolphins. The project investigates a group of wild Atlantic spotted dolphins (S. frontalis) in the Bahamas. Over the decades, the team has non-invasively collected underwater audio and video data that is associated with individual dolphins in the pod, detailing aspects of the animals’ relationships and life histories.

The project has yielded an extraordinary dataset—one packed with 41 years of sound-behavior pairings like courtship buzzes, aggressive squawks used in cetacean altercations, and “signature whistles” that act as dolphin name tags.

This trove of labeled vocalizations gave Google researchers what they needed to train an AI model designed to do for dolphin sounds what ChatGPT does for words. Thus, DolphinGemma was born: a roughly 400-million parameter model built on the same research that powers Google’s Gemini models.

DolphinGemma is audio-in, audio-out—the model “listens” to dolphin vocalizations and predicts what sound comes next—essentially learning the structure of dolphin communication.

AI and animal communication​


Artificial intelligence models are changing the rate at which experts can decipher animal communication. Everything under the Sun—from dog barks and bird whistles—is easily fed into large language models which then can use pattern recognition and any relevant contexts to sift through the noise and posit what the animals are “saying.”

Last year, researchers at the University of Michigan, Mexico’s National Institute of Astrophysics, and the Optics and Electronics Institute used an AI speech model to identify dog emotions, gender, and identity from a dataset of barks.

Cetaceans, a group that includes dolphins and whales, are an especially good target for AI-powered interpretation because of their lifestyles and the way they communicate. For one, whales and dolphins are sophisticated, social creatures, which means that their communication is packed with nuance. But the clicks and shrill whistles the animals use to communicate are also easy to record and feed into a model that can unpack the “grammar” of the animals’ sounds. Last May, for example, the nonprofit Project CETI used software tools and machine learning on a library of 8,000 sperm whale codas, and found patterns of rhythm and tempo that enabled the researchers to create the whales’ phonetic alphabet.

Talking to dolphins with a smartphone​


The DolphinGemma model can generate new, dolphin-like sounds in the correct acoustic patterns, potentially helping humans engage in real-time, simplified back-and-forths with dolphins. This two-way communication relies on what a Google blog referred to as Cetacean Hearing Augmentation Telemetry, or CHAT—an underwater computer that generates dolphin sounds the system associates with objects the dolphins like and regularly interact with, including seagrass and researchers’ scarves.

“By demonstrating the system between humans, researchers hope the naturally curious dolphins will learn to mimic the whistles to request these items,” the Google Keyword blog stated. “Eventually, as more of the dolphins’ natural sounds are understood, they can also be added to the system.”

CHAT is installed on modified smartphones, and the researchers’ idea is to use it to create a basic shared vocabulary between dolphins and humans. If a dolphin mimics a synthetic whistle associated with a toy, a researcher can respond by handing it over—kind of like dolphin charades, with the novel tech acting as the intermediary.

Future iterations of CHAT will pack in more processing power and smarter algorithms, enabling faster responses and clearer interactions between the dolphins and their humanoid counterparts. Of course, that’s easily said for controlled environments—but raises some serious ethical considerations about how to interface with dolphins in the wild should the communication methods become more sophisticated.

A summer of dolphin science​


Google plans to release DolphinGemma as an open model this summer, allowing researchers studying other species, including bottlenose or spinner dolphins, to apply it more broadly. DolphinGemma could be a significant step toward scientists better understanding one of the ocean’s most familiar mammalian faces.

We’re not quite ready for a dolphin TED Talk, but the possibility of two-way communication is a tantalizing indicator of what AI models could make possible.
 
Top