Abstract AlgorithmsAbstract Algorithms

  • Home
  • All Posts
  • All Series
  • About

Category

ai-safety

1 article in this category

RLHF in Practice: From Human Preferences to Better LLM Policies

TLDR: Reinforcement Learning from Human Feedback (RLHF) helps align language models with human preferences after pretraining and SFT. The typical pipeline is: collect preference comparisons, train a reward model, then optimize a policy (often with KL...

Mar 9, 2026•8 min read

Abstract Algorithms

Exploring the fascinating world of algorithms, data structures, and software engineering through clear explanations and practical examples.

Navigation

  • Home
  • All Posts
  • All Series
  • About

Popular Topics

  • llm
  • Machine Learning
  • System Design
  • nlp
  • Deep Learning
  • basics

Author

Abstract Algorithms

Abstract Algorithms

@abstractalgorithms

Ā© 2026 Abstract Algorithms. All rights reserved.

Powered by Hashnode