Abstract AlgorithmsAbstract Algorithms

  • Home
  • All Posts
  • All Series
  • About

Category

lora

3 articles across 2 sub-topics

Ai(2)

PEFT, LoRA, and QLoRA: A Practical Guide to Efficient LLM Fine-Tuning

TLDR: Full fine-tuning updates every model weight, which is expensive in memory, compute, and storage. PEFT methods update only a small trainable slice. LoRA learns low-rank adapters on top of frozen base weights. QLoRA pushes efficiency further by q...

Mar 9, 2026•13 min read
LoRA Explained: How to Fine-Tune LLMs on a Budget

LoRA Explained: How to Fine-Tune LLMs on a Budget

TLDR: Fine-tuning a 7B-parameter LLM updates billions of weights and requires expensive GPUs. LoRA (Low-Rank Adaptation) freezes the original weights and trains only tiny adapter matrices that are added on top. 90%+ memory reduction; zero inference l...

Mar 9, 2026•13 min read
Fine Tuning(1)
Fine-Tuning LLMs: The Complete Engineer's Guide to SFT, LoRA, and RLHF

Fine-Tuning LLMs: The Complete Engineer's Guide to SFT, LoRA, and RLHF

TLDR: A pretrained LLM is a generalist. Fine-tuning makes it a specialist. Supervised Fine-Tuning (SFT) teaches it your domain's language through labeled examples. LoRA does the same with 99% fewer trainable parameters. RLHF shapes its behavior using...

Apr 18, 2026•30 min read

Abstract Algorithms

Exploring the fascinating world of algorithms, data structures, and software engineering through clear explanations and practical examples.

Navigation

  • Home
  • All Posts
  • All Series
  • About

Series

  • LLM Engineering
  • Apache Spark Engineering
  • Python Programming

Popular Topics

  • #apache-spark14
  • intermediate12
  • Python6
  • performance6
  • Structured Streaming4
  • big data4

Author

Abstract Algorithms

Abstract Algorithms

@abstractalgorithms

© 2026 Abstract Algorithms. All rights reserved.

Powered by Hashnode