The Case for Micro Foundation Models to Support Robust Edge Intelligence

Tomoyoshi Kimura, Ashitabh Misra, Yizhuo Chen, Denizhan Kara, Jinyang Li, Tianshi Wang, Ruijie Wang, Joydeep Bhattacharyya, Jae Kim, Prashant Shenoy, Mani Srivastava, Maggie Wigness, Tarek Abdelzaher

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

This paper advocates the concept of micro foundation models (µFMs), recently introduced by the authors to describe a category of self-supervised pre-training solutions that we argue are necessary to support robust intelligent inference tasks in Internet of Things (IoT) applications. The work is motivated by the fact that collecting sufficient amounts of labeled data in IoT applications to train AI/ML tasks is challenging due to the difficulties in labeling such data after the fact. In the absence of sufficient labeled data, supervised training solutions become brittle and prone to overfitting. Self-supervised training obviates the collection of labeled data, allowing pre-training with the more readily available unlabeled data instead. Specifically, the µFMs discussed in this paper use self-supervised pre-training to develop an encoder that maps input data into a semantically-organized latent representation in a manner agnostic to the downstream inference task. Our preliminary work shows that this (unsupervised) encoder can be moderately sized, yet produce a latent representation that simultaneously supports the fine-tuning of multiple downstream inference tasks, each at a minimal labeling cost. We demonstrate the efficacy of this pre-training/fine-tuning pipeline using a vibration-based µFM as a running case study. The study shows that the fine-tuning of inference tasks on top of the aforementioned encoder-produced latent representation needs orders of magnitude fewer labels than supervised training solutions, and that the resulting tasks are significantly more robust to environmental changes and easier to adapt to domain shifts compared to their supervised counterparts. Furthermore, we show that inference algorithms based on our example µFM can be executed in real time on a Raspberry Pi device, making the approach viable for the IoT space. We conclude that µFMs are a preferred (and likely necessary) route to support robust intelligent sensing on IoT devices in subareas where labeled data collection is challenging. The paper is a call for the research community to invest in µFM research for IoT applications.

Original languageEnglish (US)
Title of host publicationProceedings - 2024 IEEE 6th International Conference on Cognitive Machine Intelligence, CogMI 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages23-31
Number of pages9
ISBN (Electronic)9798350386721
DOIs
StatePublished - 2024
Event6th IEEE International Conference on Cognitive Machine Intelligence, CogMI 2024 - Washington, United States
Duration: Oct 28 2024Oct 30 2024

Publication series

NameProceedings - 2024 IEEE 6th International Conference on Cognitive Machine Intelligence, CogMI 2024

Conference

Conference6th IEEE International Conference on Cognitive Machine Intelligence, CogMI 2024
Country/TerritoryUnited States
CityWashington
Period10/28/2410/30/24

Keywords

  • Foundation Models
  • Internet of Things
  • Self-Supervised Learning

ASJC Scopus subject areas

  • Artificial Intelligence
  • Computer Science Applications
  • Computer Vision and Pattern Recognition
  • Information Systems and Management
  • Cognitive Neuroscience

Fingerprint

Dive into the research topics of 'The Case for Micro Foundation Models to Support Robust Edge Intelligence'. Together they form a unique fingerprint.

Cite this