Revolutionary Algorithm Shrinks Large Language Models for Local Use
Researchers from **Princeton and Stanford** have unveiled a groundbreaking algorithm called **CALDERA (Calibration Aware Low precision DEcomposition with low Rank Adaptation)**. This technique effectively compresses the data within large language models (LLMs), such as those used in applications like translation and customer service. The method focuses on reducing redundancies and precision within the LLM layers, allowing for **local storage and processing on consumer devices** such as smartphones and laptops. One of the key benefits of this innovation is the potential increase in **privacy and cost-efficiency**. By compressing LLMs, users no longer need to send data to centralized servers, addressing the high-energy and expensive problem traditionally associated with LLM use. This advancement creates the potential for **on-device AI** that requires less computational power and bandwidth. **CALDERA combines low-precision and low-rank techniques** to achieve greater compression than would be possible with either method alone. Low-precision reduces the bit count used by computers to store information, while low-rank minimizes data redundancy in the weight matrices, which are crucial components of LLMs. Testing their approach on Meta AI's open-source models Llama 2 and Llama 3, the research team reported a performance improvement of up to 5% in certain tasks, offering a more energy-efficient model iteration while maintaining accuracy. While not intended for scenarios requiring maximum precision, compressed LLMs are ideal for tasks that require less detail, offering **personalization and enhanced privacy** by allowing model fine-tuning on individual devices. Nevertheless, researchers caution that local LLM processing could heavily drain device memory and may need further enhancement. Presented in December at the Conference on Neural Information Processing Systems (NeurIPS), this work is a collaborative effort bolstered by **US science and defense agencies**. It represents a significant step toward bringing LLM capabilities to personal devices and democratizing AI technology.