Abstract AlgorithmsAbstract Algorithms

  • Home
  • All Posts
  • All Series
  • About

Category

alignment

2 articles in this category

RLHF in Practice: From Human Preferences to Better LLM Policies

TLDR: Reinforcement Learning from Human Feedback (RLHF) helps align language models with human preferences after pretraining and SFT. The typical pipeline is: collect preference comparisons, train a reward model, then optimize a policy (often with KL...

Mar 9, 2026•8 min read

RLHF Explained: How We Teach AI to Be Nice

TLDR: A raw LLM is a super-smart parrot that read the entire internet — including its worst parts. RLHF (Reinforcement Learning from Human Feedback) is the training pipeline that transforms it from a pattern-matching engine into an assistant that is ...

Mar 9, 2026•5 min read

Abstract Algorithms

Exploring the fascinating world of algorithms, data structures, and software engineering through clear explanations and practical examples.

Navigation

  • Home
  • All Posts
  • All Series
  • About

Popular Topics

  • llm
  • Machine Learning
  • System Design
  • nlp
  • Deep Learning
  • basics

Author

Abstract Algorithms

Abstract Algorithms

@abstractalgorithms

© 2026 Abstract Algorithms. All rights reserved.

Powered by Hashnode