Salesforce Argues for Small Models in Enterprise AI
- •Skyrocketing compute costs and power demands make massive models increasingly impractical for most enterprise applications.
- •Smaller, domain-specific models like XGen 7B often match or exceed the performance of larger general-purpose systems.
- •Orchestrating multiple specialized models into a single deployment enhances transparency, sustainability, and overall reliability.
While the AI industry often fixates on massive parameter counts, Salesforce AI Research suggests that for enterprise needs, bigger is not always better. The primary barrier is the 'cost to serve,' where every parameter in a model requires a mathematical calculation—specifically a floating-point operation—multiplied by every token of input. This creates a direct link between model size and the massive compute power and energy required for deployment.
Performance is not a one-dimensional metric where larger models always win. In specific business domains like technical support or knowledge retrieval, smaller models can be trained on high-quality, curated data to meet or exceed the capabilities of their larger counterparts. Salesforce’s XGen 7B serves as a prime example, leveraging specialized training strategies to outperform models with significantly more parameters. This approach allows companies to use data they already own while avoiding the ethical pitfalls of massive, unvetted datasets.
The future of efficient AI lies in orchestration, where multiple specialized small models are coordinated like a human team. One model might handle information retrieval while another manages user interaction, allowing each to be independently validated. This modular strategy significantly reduces carbon emissions while building 'trust'—a critical factor for applications requiring high levels of accountability.