Abstract AlgorithmsAbstract Algorithms

  • Home
  • All Posts
  • All Series
  • About

Category

mixture of experts

1 article

Sparse Mixture of Experts: How MoE LLMs Do More With Less Compute

TLDR: Mixture of Experts (MoE) replaces the single dense Feed-Forward Network (FFN) layer in each Transformer block with N independent expert FFNs plus a learned router. Only the top-K experts activate per token — so total parameters far exceed activ...

Apr 17, 2026•28 min read

Abstract Algorithms

Exploring the fascinating world of algorithms, data structures, and software engineering through clear explanations and practical examples.

Navigation

  • Home
  • All Posts
  • All Series
  • About

Series

  • LLM Engineering
  • Machine Learning Fundamentals
  • How It Works: Internals Explained
  • System Design Interview Prep

Popular Topics

  • Databases
  • distributed systems
  • transactions
  • Isolation Levels
  • llm
  • Machine Learning

Author

Abstract Algorithms

Abstract Algorithms

@abstractalgorithms

1 followers on Hashnode

© 2026 Abstract Algorithms. All rights reserved.

Powered by Hashnode