Hello! đ Iâm a PhD student in Computer Science at the National University of Singapore, advised by Prof. Ambuj Varshney (https://www.comp.nus.edu.sg/cs/people/ambuj/). I explore how next-gen embedded tags are evolving toward architectures that emphasize local processing and peer-to-peer coordination for energy-sustainable operation at scale. My work also examines how smaller foundational models enable practical on-device AI in highly constrained IoT environments đđ˛.
Released TinyLLM! A Framework for Training and Deploying Language Models at the Edge
Sep 09, 2024
Poster accepted at ACM MobiCom 2024. Title: âYour Data, Your Model: A Framework for Training and Deploying Foundational Language Models on Embedded Devices.â
Sep 07, 2024
Workshop paper accepted at ACM S3 Workshop (held in conjunction with MobiCom 2024).
@misc{tinyllm,title={TinyLLM: A Framework for Training and Deploying Language Models at the Edge Computers},author={Kandala, Savitha Viswanadh and Medaranga, Pramuka and Varshney, Ambuj},year={2024},eprint={2412.15304},archiveprefix={arXiv},primaryclass={cs.LG},url={https://arxiv.org/abs/2412.15304},}
A Framework for Training and Deploying Foundational Language Models for Embedded Sensing
Savitha Viswanadh Kandala, and Ambuj Varshney
In Proceedings of the 30th Annual International Conference on Mobile Computing and Networking, Washington D.C., DC, USA, 2024
Large language models have attracted a significant recent interest, with an overwhelming efforts and emphasis on scaling their parameter size to support general-purpose and emergent capabilities. However, memory and processing requirements for model inference also scales proportionally with the modelâs parameter size. This makes it challenging for state-of-the-art, larger models to perform inference locally on edge and mobile devices, even though such models could benefit numerous tasks on these systems. Consequently, these devices are often restricted to accessing larger models through network calls. This approach introduces challenges related to increased inference latency due to network delays and provider capacity. It also raises concerns about sharing private information with third-party vendors. To address these issues, we are developing a system called OTTER. This system tackles the particular problem by enabling the training of smaller yet highly capable foundational language models. As a result of the reduced parameter size, these models can run locally even on constrained edge devices, such as mobile phones and wearables, and are able to provide low-latency responses compared with their larger remotely hosted counterparts. We present our ongoing work describing the framework as applied to training a smaller foundational model for embedded sensing application for tracking a personâs breathing with high accuracy comparable to models orders of magnitude larger in size. Our results demonstrate that carefully pre-trained and fine-tuned smaller sized models outperform much larger counterparts for some tasks, while inferring locally on the constrained edge and mobile devices.
@inproceedings{s3,author={Kandala, Savitha Viswanadh and Varshney, Ambuj},title={A Framework for Training and Deploying Foundational Language Models for Embedded Sensing},year={2024},isbn={9798400704895},publisher={Association for Computing Machinery},address={New York, NY, USA},url={https://doi.org/10.1145/3636534.3695901},doi={10.1145/3636534.3695901},booktitle={Proceedings of the 30th Annual International Conference on Mobile Computing and Networking},pages={2236â2238},numpages={3},location={Washington D.C., DC, USA},series={ACM MobiCom '24},}
Engineering End-to-End Remote Labs Using IoT-Based Retrofitting
Savitha Viswanadh Kandala, Akshit Gureja, Nagesh Walchatwar, and 8 more authors
@article{journal_rtl,author={Viswanadh Kandala, Savitha and Gureja, Akshit and Walchatwar, Nagesh and Agrawal, Rishabh and Sinha, Shiven and Chaudhari, Sachin and Vaidhyanathan, Karthik and Choppella, Venkatesh and Bhimalapuram, Prabhakar and Kandath, Harikumar and Hussain, Aftab},journal={IEEE Access},title={Engineering End-to-End Remote Labs Using IoT-Based Retrofitting},year={2025},volume={13},number={},pages={1106-1132},doi={10.1109/ACCESS.2024.3523066},}