Enables collaboration between small, on-device LLMs and larger cloud-based models for cost-efficient language processing.
Minions is a communication protocol that empowers small, on-device language models to work in concert with powerful cloud-based LLMs. By primarily utilizing local models for context reading, it drastically reduces cloud costs while maintaining, or even improving, output quality. This approach is particularly beneficial for tasks requiring extensive contextual understanding without the expense of sending large amounts of data to the cloud.