ÄûÃʵ¼º½

News

Your voice gives away valuable personal information, so how do you keep that data safe?

With speech technologies becoming increasingly common, researchers want to make sure we don’t give away more information than we mean to.
A person speaking into a smartwatch with a silver mesh band, displaying a waveform on the screen.
When someone talks, a lot of information about their health, cultural background, education level and so on is embedded in the speech signal, Professor Tom Bäckström says. Photo: Adobe Stock.

You can probably quickly tell from a friend’s tone of voice whether they’re feeling happy or sad, energetic or exhausted. Computers can already do a similar analysis, and soon they’ll be able to extract a lot more information. It’s something we should all be concerned about, according to Associate Professor in Speech and Language Technology, Tom Bäckström. Personal information encoded in your voice could lead to increased insurance premiums or to advertising that exploits your emotional state. Private information could also be used for harassment, stalking or even extortion. 

‘When someone talks, a lot of information about their health, cultural background, education level and so on is embedded in the speech signal. That information gets transmitted with the speech, even though people don’t realise it,’ says Bäckström, an engineering researcher at Aalto University. For example, even subtle patterns of intonation or word choice can be a giveaway as to your political preferences, while clues in breathing or voice quality may correlate with certain health conditions. 

One important risk is that medical information inferred from voice recordings could affect insurance prices or be used to market medication. Yet Bäckström also highlights the potential for indirect harm. ‘The fear of monitoring or the loss of dignity if people feel like they’re constantly monitored—that’s already psychologically damaging,’ he says. For example, employers might extract personal information from voice recordings which could be used against employees or to screen candidates, or exes might use such tools for stalking or harassment.

A person in a light blue shirt and grey blazer stands with one hand on their hip. A modern artwork is visible in the background.
Professor Tom Bäckström. Photo: Aalto University / Mikko Raskinen.

While Bäckström says that the technology to get access to all that information is ‘not quite there yet’, researchers are working to develop protective measures before the problem becomes too big. 

So how can engineers such as Bäckström tackle these problems?

Protecting against abuses means ensuring that only the information that’s strictly necessary is transmitted and that this information is securely delivered to the intended recipient. One approach is to separate out the private information and only transmit the information needed to provide a service. Speech can also be processed locally on a phone or computer rather than sent to the cloud, and acoustic technologies can be used to make sure that sounds are only recorded from (or audible in) a specific place.

These are relatively new challenges, driven by rapid technological changes and the growth of large data collections. In 2019, Bäckström and a few others established an . The team just published a tool that can address one of the field’s fundamental questions: how much information is there in a recording of speech?

‘To ensure privacy, you decide that only a certain amount of information is allowed to leak, and then you build a tool which guarantees that,’ he explains. ‘But with speech, we don’t really know how much information there is. It’s really hard to build tools when you don’t know what you’re protecting, so the first thing is to measure that information.’

The paper offers a metric which can be used to tell how precisely a speaker’s identity can be narrowed down based on the features of a recording, such as the pitch of their speech or its linguistic content. Existing metrics provide measurements in terms of recognition risk, giving an estimate of whether the speaker in a recording can be matched with a specific feature—for example, the likelihood of being able to tell if the speaker has Parkinson’s disease. Bäckström says those approaches are more difficult to understand and generalize. The new metric is the first to capture how much information is contained in an audio clip.

Better science means better tools

Bäckström sees the research as a step towards informing people about the privacy of different speech technologies. ‘I dream of being able to say that, for example, if you give a recording to whatever service, then at a cost of 10 euros, that company will be able to narrow your identity down to, let’s say, a thousand people. That’s something people understand, so it could be reflected in the user interface. Then we can start to discuss things in concrete terms,’ he says.

Having useful metrics isn’t just needed for communicating with the public. It’s also important for designing and evaluating tools to protect privacy. In a paper just published in the Proceedings of IEEE, Bäckström’s team provided the first comprehensive overview of different threats and possible protection strategies, as well as highlighting paths for further research. The paper also covers privacy risks to people who aren’t using speech services, for example, when data from your voice might be captured as background noise in a recording.

The study highlights that preserving privacy isn’t just a technical issue but also a question of the user’s psychology and perceptions, as well as user interface design. 

‘The interface should have ways of communicating how private an interaction is,’ says Bäckström. It should also communicate the system’s competence or confidence to help prevent accidental information leaks or incorrect actions. ‘Communicating those things in the appropriate way helps build long-term trust in a service,’ he adds. 

For Bäckström, addressing privacy concerns doesn’t have to be burdensome but can actually mean improving a product or service. For example, stripping out private information from speech would mean less data is transmitted, bringing down network traffic and reducing costs. 

‘We often see privacy and utility as somehow contradictory forces, but many privacy technologies have utility benefits as well,’ he concludes.

More information:

Link to the paper:

  • Updated:
  • Published:
Share
URL copied!

Read more news

Three people having a discussion at a table with laptops. Text: Visiting Professorships at TU Graz, October 1, 2026 - January 31, 2027.
Cooperation, Research & Art, Studies, University Published:

Apply Now: Unite! Visiting Professorships at TU Graz

TU Graz, Austria, invites experienced postdoctoral researchers to apply for two fully funded visiting professorships. The deadline for expressions of interest is 20 February 2026, and the positions will begin on 1 October 2026.

A modern lobby with a large brown sectional sofa, colourful artwork, and a staircase. A '50' logo is on the back wall.
Press releases Published:

Hanaholmen’s 50th anniversary exhibition lives on online – making the history of Finnish–Swedish cooperation accessible worldwide

MeMo Institute at Aalto University has produced a virtual 3D version of the anniversary exhibition of Hanaholmen.
Research & Art Published:

Soil Laboratory Exhibition – Exploring the Dialogue Between Human and the Earth in Utsjoki

Soil Laboratory explores the relationship between humans and the earth as a living landscape through ceramic practices in Utsjoki.
Three people walking in winter next to a sign that says 'Aalto University' with snow-covered trees and buildings in the background.
Research & Art Published:

The Finnish Cultural Foundation awarded grants for science and art

A total of 15 individuals or groups from Aalto University received grants