DEV Community

# moe

Posts

👋 Sign in for the ability to sort posts by relevant, latest, or top.
Edge computing et AMD MI300X pour training et inférence : alternative à NVIDIA

Edge computing et AMD MI300X pour training et inférence : alternative à NVIDIA

7 min read
Mixture of Experts Implementation using Granite4: Harnessing Specialization with the Latest Granite Family Model

Mixture of Experts Implementation using Granite4: Harnessing Specialization with the Latest Granite Family Model

5
11 min read
Understanding Mixture of Experts (MoE)

Understanding Mixture of Experts (MoE)

5 min read
🚀 LLMs are getting huge. But do we need all that firepower all the time?

🚀 LLMs are getting huge. But do we need all that firepower all the time?

1
1 min read
A Slightly Technical Deep Dive into DeepSeek R1

A Slightly Technical Deep Dive into DeepSeek R1

3
3 min read
DBRX, Grok, Mixtral: Mixture-of-Experts is a trending architecture for LLMs

DBRX, Grok, Mixtral: Mixture-of-Experts is a trending architecture for LLMs

7 min read
loading...