TinyLLM

A Framework for Training and Deploying Language Models at the Edge

🚀 Introducing TinyLLM: Empowering AI at the Edge 🌟

TinyLLM is a lightweight framework designed to train, fine-tune, and deploy smaller language models (30 - 124M) at edge devices for supporting embedded sensing/ IoT platforms.

TinyLLM: You provide data, and TinyLLM generates a custom-trained tiny language model tailored for your embedded sensing application. TinyLLM also simplifies their deployment on resource-constrained platform!

Features of TinyLLM

  • 🧩 Pre-training flexibility: Allows pre-training, enabling the addition of custom data to improve accuracy.
  • đź“š Dataset preparation made easy: Flexibility to prepare pre-training datasets by merging multiple datasets seamlessly.
  • 🔄 Fine-tuning support: Fine-tune various models like LLaMA, Phi, and more to suit diverse use cases.
  • 📦 Optimized for small-scale LLMs, making AI accessible even on devices with limited resources.
  • 🔓 Fully open-source and Apache-v2 licensed for maximum flexibility.

đź’ˇ Featured Application: Embedded Sensing

TinyLLM excels in embedded sensing, enabling almost real-time activity recognition and other sensor-based AI tasks with minimal latency. In one of our benchmarks, TinyLLM achieved state-of-the-art accuracy while running on a microcontroller-class device—also matching the performance of larger off-the-shelf LLMs like Meta’s Llama 3 and Microsoft’s Phi 3. Further, we have created custom foundational models for applications like hand gesture tracking support, localisation, and breathing rate detection.

👉 Check out the framework, experiment with the code, or dive deeper into the technical details:
🌍 Website: TinyLLM.org 🔗 GitHub Repository: github.com/weiserlab/TinyLLM
đź“„ ArXiv Paper: (Kandala et al., 2024)

#TinyLLM #EdgeAI #EmbeddedSensing #MachineLearning #OpenSource #AI

From the left, the images show the way the sensor data is collected for human gestures and robot localisation. The third image shows the accuracy of TinyLLM models compared to off-the-shelf LLMs.

References

2024

  1. TinyLLM: A Framework for Training and Deploying Language Models at the Edge Computers
    Savitha Viswanadh Kandala, Pramuka Medaranga, and Ambuj Varshney
    2024