Category: Africa & AI

  • My Notes on Exploring Google’s Health Foundation Models

    (Note: This post reflects my personal opinions and may not reflect those of my employer)

    Article content
    Example of the HeAR encoder that generates a machine learning representation (known as “embeddings”)

    This image is a spectrogram representing my name, “Abdoulaye,” generated from my voice audio by HeAR (Health Acoustic Representations). HeAR is one of the recently released Health AI foundation models by Google. I’ve been captivated by these foundation models lately, spending time digging into them, playing with the demos and notebooks, reading ML papers about the models, and also learning more about embeddings in general and their usefulness in low-resource environments. All of this started after playing with a couple of the notebooks.

    Embeddings are numerical representations of data. AI models learn to create these compact summaries (vectors) from various inputs like images, sounds, or text, capturing essential features. These information-rich numerical representations are useful because they can serve as a foundation for developing new, specialized AI models, potentially reducing the amount of task-specific data and development time required. This efficiency is especially crucial in settings where large, labeled medical datasets may be scarce.

    • If you would like to read further into what Embeddings are, Vicki Boykis’ essay is such a great free resource; this essay is also ideal to learn or dive into machine learning. I know many of my previous colleagues from the telco and engineering world will love this: https://vickiboykis.com/what_are_embeddings/
    • For a technical perspective on their evolution, check out the word2vec paper: https://arxiv.org/abs/1301.3781

    The HeAR model, which processed my voice audio, is trained on over 300 million audio clips (e.g., coughs, breathing, speech). Its application can extend to identifying acoustic biomarkers for conditions like TB or COVID-19. It utilizes a Vision Transformer (ViT) to analyze spectrograms. Below, you can see an example of sneezing being detected within an audio file, and later, throat clearing detected at the end.

    Health event detector demo

    Article content
    Health event detector demo

    This release also includes other open-weight foundation models, each designed to generate high-quality embeddings:

    Derm Foundation (Skin Images) This model processes dermatology images to produce embeddings, aiming to make AI development for skin image analysis more efficient by reducing data and compute needs. It facilitates the development of tools for various tasks, such as classifying clinical conditions or assessing image quality.

    Explore the Derm Foundation model site for more information and to download the model use this link.

    CXR Foundation (Chest X-rays) The CXR Foundation model produces embeddings from chest X-ray images, which can then be used to train models for various chest X-ray related tasks. The models were trained on very large X-ray datasets. What got my attention, some models within the collection, like ELIXR-C, use an approach inspired by CLIP (contrastive language-image pre-training) to link images with text descriptions, enabling powerful zero-shot classification. This means the model might classify an X-ray for a condition it wasn’t specifically trained on, simply by understanding a text description of that condition which i find fascinating. The embeddings generated can also be used to train models that can detect diseases like tuberculosis without a large amount of data; for instance, “models trained on the embeddings derived from just 45 tuberculosis-positive images were able to achieve diagnostic performance non-inferior to radiologists.” This data efficiency is particularly valuable in regions with limited access to large, labeled datasets. Read the paper for more details.

    Retrieve images by text queries

    Article content
    Retrieve images by text queries demo

    Path Foundation (Pathology Slides) Google’s Path Foundation model is trained on large-scale digital pathology datasets to produce embeddings from these complex microscopy images. Its primary purpose is to enable more efficient development of AI tools for pathology image analysis. This approach supports tasks like identifying tumor tissue or searching for similar image regions, using significantly less data and compute. See the impressive Path Foundation demos on HuggingFace.

    Article content
    Path foundation demos

    Outlier Tissue Detector Demo

    These models are provided as Open Weight with the goal of enabling developers and researchers to download and adapt them, fostering the creation of localized AI tools. In my opinion, this is particularly exciting for regions like Africa, where such tools could help address unique health challenges and bridge gaps in access to specialist diagnostic capabilities.

    For full acknowledgment of contributions from various institutions, including partners like the Center for Infectious Disease Research in Zambia, please refer to the detailed in the paper.

    For those interested in the architectural and training methodologies, here are some of the pivotal papers and concepts relevant to these foundation models:

    #AIforHealth #FoundationModels #GlobalHealth #AIinAfrica #ResponsibleAI #MedTech #Innovation #GoogleResearch #Embeddings #MachineLearning #DeepLearning

  • Google Translate Expands to 110 New Languages, Including 31 from Africa

    Google Translate has taken a significant step towards greater inclusivity by adding 110 new languages to their service, including 31 from Africa. This expansion means that millions of people who previously lacked access to translation services now have a tool to communicate and connect with a wider world.

    This achievement is the result of a concerted multi years effort by Isaac Caswell, the Google Translate Research team, and numerous community collaborators. They faced unique challenges, as building high-quality translations for languages with limited digital resources is complex. To address this, they developed an approach that relies on “monolingual” data – text in a single language – instead of solely relying on translated text. This method, called “zero-shot learning,” allows for the creation of translations for languages not explicitly trained on, though it’s important to remember that this is still a developing technology.

    Monolingual

    What does this mean?

    • More Accessibility: People speaking these newly included languages now have a tool to access information, communicate, and break down language barriers.
    • Language Preservation: This expansion helps preserve and promote less commonly used languages, which is crucial for maintaining linguistic diversity.
    • New Opportunities: The inclusion of languages like Punjabi and Romani opens doors for those communities, aiding in digital navigation and accessing information.

    While this is a significant step, it’s important to note that Google Translate, while utilizing powerful AI technology, is not a replacement for the expertise of professional translators, however the app is useful for millions of people. This new approach, powered by the Palm 2 language model will require ongoing refinement and feedback to improve accuracy.

    The languages added are significant:

    A few notables ones:

    • Punjabi (Shahmukhi): This language, written in the Shahmukhi script, is spoken by millions in Pakistan and India. Its inclusion expands communication and access to information for a large community.
    • Romani: Romani, the language of romani communities with presence in many european countries, has historically been underrepresented in technology. Its inclusion in Google Translate is a step towards recognizing and supporting this community.
    • N’ko: Created in the 1940s, N’ko uses a unique script to unify Manding languages in West Africa. This addition supports literacy and cultural preservation efforts.
    • Tamazight (with Tifinagh script): Tamazight is spoken by millions of Berber people in North Africa. Its inclusion acknowledges their cultural diversity and language heritage.

    This expansion is a positive step towards a more inclusive digital world, but it’s important to be aware of its limitations. While AI is an exciting tool, it requires continual development and feedback to refine its capabilities. The addition of these new languages is a testament to the potential of technology to foster communication and understanding, but it’s crucial to remember that it’s a journey, not a destination.

    Languages by Region

    APAC

    • Southern Asia
      • Bhutan: Dzongkha
      • India: Awadhi, Bodo, Khasi, Kokborok, Marwadi, Santali, Tulu
      • Nepal: Nepalbhasa (Newari)
      • Pakistan: Baluchi, Punjabi (Shahmukhi)
    • Eastern Asia
      • China: Cantonese, Tibetan
      • Hong Kong: Cantonese
      • Tibet: Tibetan
    • Southeast Asia
      • East Timor: Tetum
      • Indonesia: Acehnese, Balinese, Batak Karo, Batak Simalungun, Batak Toba, Betawi, Iban, Madurese, Makassar, Minang
      • Malaysia: Malay (Jawi)
      • Myanmar: Hakha Chin, Jingpo, Shan
      • Philippines: Bikol, Hiligaynon, Kapampangan, Pangasinan, Waray
    • Melanesia
      • Fiji: Fijian
      • Papua New Guinea: Tok Pisin
    • Micronesia
      • Guam: Chamorro
      • Micronesia: Chuukese
      • Marshall Islands: Marshallese
    • Central Asia
      • Mongolia: Buryat
    • Polynesia
      • Tahiti: Tahitian
      • Tonga Islands: Tongan

    EMEA

    • Western Asia
      • Afghanistan: Dari
    • Northern Africa
      • Algeria: Tamazight
      • Morocco: Tamazight, Tamazight (Tifinagh)
      • Sudan: Acholi, Dinka, Luo, Nuer
    • Eastern Europe
      • Austria: Romani
      • Bosnia and Herzegovina: Romani
      • Denmark: Romani
      • Finland: Romani
      • Germany: Romani
      • Hungary: Romani
      • Kosovo: Romani
      • Montenegro: Romani
      • North Macedonia: Romani
      • Poland: Romani, Silesian
      • Romania: Romani
      • Russia: Avar, Bashkir, Buryat, Chechen, Chuvash, Crimean Tatar, Komi, Meadow Mari, Ossetian, Tuvan, Udmurt, Yakut
      • Serbia: Romani
      • Slovakia: Romani
      • Sweden: Romani
      • Ukraine: Crimean Tatar
    • Western Africa
      • Benin: Fon
      • Burkina Faso: Dyula
      • Côte d’Ivoire: Baoulé, Dyula
      • Gabon: Fon
      • Gambia: Wolof
      • Ghana: Dyula, Fon, Ga
      • Guinea: N’Ko, Susu
      • Guinea-Bissau: N’Ko, Susu
      • Mali: Dyula, N’Ko
      • Mauritania: Wolof
      • Nigeria: Fon, Tiv
      • Senegal: Wolof
      • Sierra Leone: Susu
      • Togo: Fon
    • Southern Africa
      • Botswana: Tswana
      • Eswatini: Swati
      • Lesotho: Swati
      • South Africa: Ndebele (South), Swati, Tswana, Venda
    • Eastern Africa
      • Burundi: Rundi
      • Ethiopia: Afar, Luo, Nuer
      • Kenya: Luo
      • Malawi: Tumbuka
      • Mauritius: Mauritian Creole
      • Mozambique: Ndau, Swati, Venda
      • Rwanda: Kiga
      • Seychelles: Seychellois Creole
      • South Sudan: Acholi, Dinka, Luo, Nuer
      • Tanzania: Bemba, Luo, Tumbuka
      • Uganda: Acholi, Alur, Kiga, Luo
      • Zambia: Bemba, Dombe, Tumbuka
      • Zimbabwe: Dombe, Ndau, Venda
    • Middle Africa
      • Central African Republic: Sango
      • Chad: Sango
      • Congo: Kituba, Kikongo
      • DRC: Alur, Bemba, Kituba, Kikongo, Luba, Sango
    • Northern Europe
      • Faroe Islands: Faroese
      • Isle of Man: Manx
      • Latvia: Latgalian
      • Norway: Sami (North), Romani
    • Western Europe
      • France: Breton, Occitan
      • Netherlands: Limburgish, Papiamento
    • Southern Europe
      • Italy: Friulian, Ligurian, Lombard, Sicilian, Venetian
      • Portugal: Portuguese (Portugal)

    LATAM

    Jamaica: Jamaican Patois

    Central America

    Guatemala: Q’eqchi’, Mam

    Mexico: Nahuatl (Eastern Huaste), Q’eqchi’, Mam, Yucatec Maya, Zapotec

    Belize: Q’eqchi’

    South America

    Brazil: Hunsrik

    Caribbean

    Caribbean Netherlands: Papiamento