HeAR
Stay organized with collections
Save and categorize content based on your preferences.
Health Acoustics Representations (HeAR) is a machine learning (ML) model that
produces embeddings based on health acoustic data. The embeddings can be used to
efficiently build AI models for health acoustic-related tasks (for example,
identifying disease status from cough sounds, or measuring lung function using
exhalation sounds made during spirometry), requiring less data and less compute
than having to fully train a model without the embeddings or the pretrained
model.
HeAR has been trained on 300+ million two-second audio clips comprising of five
types of non-speech health acoustic events in two-second audio clips: coughing,
breathing, throat clearing, laughing, and speaking.
For details about how to use the model and how it was trained, see the HeAR model card.
Common Use Cases
The following sections present some common use cases for the model. You're free
to pursue any use case, as long as it adheres to the
Health AI Developer Foundations terms of use.
Data-efficient classification or regression
HeAR Foundation can be used for data-efficient classification or regression
tasks, including:
- Classifying respiratory conditions like COVID-19, tuberculosis, and COPD
based on cough and breath sounds
- Identifying different types of health acoustic events, such as coughs,
wheezes, and snores
- Classifying the severity of respiratory diseases based on acoustic features
- Measuring quantities assumed to be somewhat proportional to acoustic
intensity, for example urine flow or spirometry
With a small amount of labeled data, you can train a model on top of HeAR
embeddings. Furthermore, the embedding from each acoustic sample only needs to
be generated once and can be used as an input for a variety of different
classifiers, with very little additional compute.
For an example of how to use the model to train classifiers, see the
HeAR linear classifier notebook
in Colab.
Next Steps
Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.
Last updated 2025-03-18 UTC.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-03-18 UTC."],[],[],null,["# HeAR\n\nHealth Acoustics Representations (HeAR) is a machine learning (ML) model that\nproduces embeddings based on health acoustic data. The embeddings can be used to\nefficiently build AI models for health acoustic-related tasks (for example,\nidentifying disease status from cough sounds, or measuring lung function using\nexhalation sounds made during spirometry), requiring less data and less compute\nthan having to fully train a model without the embeddings or the pretrained\nmodel.\n\nHeAR has been trained on 300+ million two-second audio clips comprising of five\ntypes of non-speech health acoustic events in two-second audio clips: coughing,\nbreathing, throat clearing, laughing, and speaking.\n\nFor details about how to use the model and how it was trained, see the [HeAR model card](/health-ai-developer-foundations/hear/model-card).\n\nCommon Use Cases\n----------------\n\nThe following sections present some common use cases for the model. You're free\nto pursue any use case, as long as it adheres to the\n[Health AI Developer Foundations terms of use](/health-ai-developer-foundations/terms).\n\n### Data-efficient classification or regression\n\nHeAR Foundation can be used for data-efficient classification or regression\ntasks, including:\n\n- Classifying respiratory conditions like COVID-19, tuberculosis, and COPD based on cough and breath sounds\n- Identifying different types of health acoustic events, such as coughs, wheezes, and snores\n- Classifying the severity of respiratory diseases based on acoustic features\n- Measuring quantities assumed to be somewhat proportional to acoustic intensity, for example urine flow or spirometry\n\nWith a small amount of labeled data, you can train a model on top of HeAR\nembeddings. Furthermore, the embedding from each acoustic sample only needs to\nbe generated once and can be used as an input for a variety of different\nclassifiers, with very little additional compute.\n\nFor an example of how to use the model to train classifiers, see the\n[HeAR linear classifier notebook](https://github.com/google-health/hear/blob/master/notebooks/train_data_efficient_classifier.ipynb)\nin Colab.\n\nNext Steps\n----------\n\n- [Get started using the model](/health-ai-developer-foundations/hear/get-started)"]]