inference optimization AI News & Updates
DeepSeek Introduces Sparse Attention Model Cutting Inference Costs by Half
DeepSeek released an experimental model V3.2-exp featuring "Sparse Attention" technology that uses a lightning indexer and fine-grained token selection to dramatically reduce inference costs for long-context operations. Preliminary testing shows API costs can be cut by approximately 50% in long-context scenarios, addressing the critical challenge of server costs in operating pre-trained AI models. The open-weight model is freely available on Hugging Face for independent verification and testing.
Skynet Chance (-0.03%): Lower inference costs make AI deployment more economically accessible and sustainable, potentially enabling better monitoring and alignment research through reduced resource barriers. However, it also enables broader deployment of powerful models, creating a minor mixed effect on control mechanisms.
Skynet Date (+0 days): Reduced inference costs enable more sustainable AI scaling and wider deployment, but this is primarily an efficiency gain rather than a capability breakthrough that would accelerate uncontrolled AI development. The modest deceleration reflects that economic sustainability may slow rushed deployment.
AGI Progress (+0.02%): The sparse attention breakthrough represents meaningful architectural progress in making transformer models more efficient at handling long-context operations, addressing a fundamental limitation in current AI systems. This optimization enables more practical deployment of advanced capabilities needed for AGI.
AGI Date (+0 days): Cutting inference costs by half significantly reduces economic barriers to scaling and deploying advanced AI systems, enabling more organizations to experiment with and advance long-context AI applications. This efficiency breakthrough accelerates the practical timeline for developing and deploying AGI-relevant capabilities.
Spanish Startup Raises $215M for AI Model Compression Technology Reducing LLM Size by 95%
Spanish startup Multiverse Computing raised €189 million ($215M) Series B funding for its CompactifAI technology, which uses quantum-computing inspired compression to reduce LLM sizes by up to 95% without performance loss. The company offers compressed versions of open-source models like Llama and Mistral that are 4x-12x faster and reduce inference costs by 50%-80%, enabling deployment on devices from PCs to Raspberry Pi. Founded by quantum physics professor Román Orús and former banking executive Enrique Lizaso Olmos, the company claims 160 patents and serves 100 customers globally.
Skynet Chance (-0.03%): Model compression technology makes AI more accessible and deployable on edge devices, but doesn't inherently increase control risks or alignment challenges. The focus on efficiency rather than capability enhancement provides marginal risk reduction through democratization.
Skynet Date (+0 days): While compression enables broader AI deployment, it focuses on efficiency rather than advancing core capabilities that would accelerate dangerous AI development. The technology may slightly slow the concentration of AI power by enabling wider access to compressed models.
AGI Progress (+0.02%): Significant compression advances (95% size reduction while maintaining performance) represent important progress in AI efficiency and deployment capabilities. This enables more widespread experimentation and deployment of capable models, contributing to overall AI ecosystem development.
AGI Date (+0 days): The dramatic cost reduction (50%-80% inference savings) and ability to run capable models on edge devices accelerates AI adoption and experimentation cycles. Broader access to efficient AI models likely speeds up overall progress toward more advanced systems.