Back to home

Articles tagged with "AI Efficiency, Sparse Attention, Deep Learning, Language Models, AI Optimization"

DeepSeek tests “sparse attention” to slash AI processing costs

DeepSeek tests “sparse attention” to slash AI processing costs

DeepSeek, a Chinese AI company facing export restrictions on advanced AI chips, has developed "DeepSeek Sparse Attention" (DSA) to enhance processing efficiency in its latest language model, DeepSeek-V3.2-Exp. This technique, similar to sparse transformers used by OpenAI and Google Research, aims to reduce computational costs. DeepSeek claims its implementation achieves "fine-grained sparse attention" and has cut API prices by 50%. The company's focus on optimizing performance with limited resources highlights the ongoing efforts to enhance AI models while managing processing costs.

Ars Technica

No more articles to load

We use cookies

We use cookies to ensure you get the best experience on our website. For more information on how we use cookies, please see our cookie policy.