File I/O and Exception Handling in Python
Why try/except in Python is not just error-catching โ it's a flow control tool that makes production code robust
Abstract AlgorithmsAI-assisted content. This post may have been written or enhanced with AI tools. Please verify critical information independently.
๐ The Config File That Took Down a Friday Deployment
Picture this: it's 5 PM on a Friday. A developer pushes a new service to production. The deployment succeeds, but five minutes later the service is dead โ a cascade of AttributeError: 'NoneType' object has no attribute 'get' errors flooding the logs.
The root cause? A configuration file that wasn't present on the new server. The script that loaded it looked like this:
# This code works perfectly in development โ and silently kills production
import json
def load_config():
f = open("config.json")
data = json.load(f)
f.close()
return data
config = load_config()
db_host = config["database"]["host"] # AttributeError if config is None
Three things go wrong here. First, if config.json doesn't exist, open() raises FileNotFoundError and the program terminates with a raw traceback โ no graceful message, no fallback, no logging. Second, the file is opened but close() is only called if nothing goes wrong between open() and close() โ a raised exception skips it, leaving the file handle open. Third, the caller has no way to distinguish between "config file missing" and "config file has invalid JSON" โ both just crash.
This is what production code without error handling looks like. And every Python developer eventually ships a version of it.
LBYL vs EAFP: Two Different Philosophies
Developers coming from Java, C#, or C tend to write defensive code in a style called LBYL โ Look Before You Leap. The idea is to check for error conditions before attempting an operation:
# LBYL style โ check first, then act
import os
def load_config_lbyl():
if not os.path.exists("config.json"):
return {}
if not os.access("config.json", os.R_OK):
return {}
f = open("config.json")
data = json.load(f)
f.close()
return data
This works, but it has a fundamental flaw: the gap between the check and the operation. You check that the file exists on line 4, but between that check and the open() call on line 7, another process could delete the file. This race condition is called TOCTOU โ Time Of Check To Time Of Use. Your guard is already stale by the time you use it.
Python's idiomatic answer is EAFP โ Easier to Ask Forgiveness than Permission. Instead of guarding every operation, you attempt it and handle the exception if something goes wrong:
# EAFP style โ attempt first, handle exceptions
import json
def load_config_eafp():
try:
with open("config.json") as f:
return json.load(f)
except FileNotFoundError:
print("Config file not found โ using defaults.")
return {}
except json.JSONDecodeError as e:
print(f"Config file is malformed: {e}")
return {}
This version is shorter, avoids the TOCTOU race, handles exactly the exceptions it knows about, and gives a useful message for each failure mode. The with statement ensures the file is always closed โ even if json.load() raises an exception. This is the Python way.
The rest of this post builds your complete mental model for writing production-grade file I/O and exception handling in Python.
๐ Opening Files the Pythonic Way: open(), Modes, and the with Statement
The open() built-in is the gateway to all file operations in Python. It takes a file path and an optional mode string that controls how the file is opened.
File Opening Modes
| Mode | Name | Behaviour |
"r" | Read (text) | Opens for reading. Fails if file does not exist. Default mode. |
"w" | Write (text) | Creates or overwrites the file. Destroys existing content. |
"a" | Append (text) | Opens for writing at the end. Creates file if missing. |
"x" | Exclusive create | Creates file; raises FileExistsError if it already exists. |
"r+" | Read + write | Opens for both; file must exist. Does not truncate. |
"rb" | Read (binary) | Reads raw bytes โ use for images, PDFs, or any non-text data. |
"wb" | Write (binary) | Writes raw bytes; creates or overwrites the file. |
"ab" | Append (binary) | Appends raw bytes to end of file. |
The mode "r" is the default โ calling open("myfile.txt") is identical to open("myfile.txt", "r").
Reading a File: Three Methods
Once a file is open, you have three ways to read its contents:
# All three reading methods demonstrated
with open("sample.txt", "r", encoding="utf-8") as f:
full_text = f.read() # Reads entire file into one string
print(full_text)
with open("sample.txt", "r", encoding="utf-8") as f:
first_line = f.readline() # Reads exactly one line (including \n)
second_line = f.readline() # Reads the next line
print(first_line, second_line)
with open("sample.txt", "r", encoding="utf-8") as f:
all_lines = f.readlines() # Returns a list of all lines
for line in all_lines:
print(line.strip()) # .strip() removes the trailing \n
For large files, the most memory-efficient approach is to iterate the file object directly โ Python reads one line at a time without loading the entire file:
# Memory-efficient line-by-line reading โ ideal for large log files
with open("large_logfile.txt", "r", encoding="utf-8") as f:
for line in f: # File object is iterable
process(line.strip()) # Never loads whole file into memory
Writing to a File
# Writing and appending
with open("output.txt", "w", encoding="utf-8") as f:
f.write("First line\n")
f.write("Second line\n")
f.writelines(["Third line\n", "Fourth line\n"])
# Append to an existing file (does not overwrite)
with open("output.txt", "a", encoding="utf-8") as f:
f.write("This is appended\n")
Why with Is Always the Right Choice
The with statement uses Python's context manager protocol. When you enter a with block, Python calls __enter__() on the object. When the block exits โ for any reason, including an exception โ Python calls __exit__(), which closes the file handle automatically.
Without with, you must call f.close() yourself, and any exception between open() and close() will skip the cleanup, leaking an operating system file descriptor. Python processes have a hard limit on open file descriptors. Leak enough of them and your process will eventually fail with OSError: [Errno 24] Too many open files.
Always use with open(...) as f:. It is not optional for production code.
โ๏ธ Python's Exception System: Hierarchy, try/except/else/finally, and Advanced Patterns
Before you can handle exceptions intelligently, you need to understand where they come from and how they relate to each other.
The Exception Hierarchy You Actually Need to Know
Python exceptions form a class hierarchy. Every exception is an object that inherits from BaseException. The classes you'll interact with most are:
BaseException
โโโ SystemExit โ raised by sys.exit() โ do not catch unless cleaning up
โโโ KeyboardInterrupt โ Ctrl+C โ do not swallow silently
โโโ Exception โ Parent of all "normal" program errors
โโโ ValueError โ Bad argument value (e.g., int("abc"))
โโโ TypeError โ Wrong type (e.g., "2" + 2)
โโโ AttributeError โ Object has no attribute
โโโ IndexError โ List index out of range
โโโ KeyError โ Dict key not found
โโโ StopIteration โ Iterator exhausted โ used by for loops
โโโ RuntimeError โ Miscellaneous runtime errors
โโโ OSError โ OS-level errors (file, network, permissions)
โโโ FileNotFoundError โ File or directory does not exist
โโโ PermissionError โ Insufficient permissions
โโโ IsADirectoryError โ Tried to open a directory as a file
โโโ TimeoutError โ OS timeout
This hierarchy matters for your except clauses. Catching OSError will catch FileNotFoundError, PermissionError, and TimeoutError โ because they are all subclasses. Catching Exception will catch almost everything except SystemExit and KeyboardInterrupt. Catching BaseException catches those too โ almost never what you want.
Rule of thumb: Catch the most specific exception class you can name. Catching broad exceptions (except Exception:) hides bugs.
The Full try/except/else/finally Pattern
Python's exception block has four clauses:
import json
def parse_config(filepath):
try:
# The try block: code that might raise an exception
with open(filepath, "r", encoding="utf-8") as f:
data = json.load(f)
except FileNotFoundError:
# except: runs only if the named exception was raised
print(f"Config file '{filepath}' not found.")
return {}
except json.JSONDecodeError as e:
# You can catch multiple exception types, each in its own clause
print(f"Config file has invalid JSON at line {e.lineno}: {e.msg}")
return {}
except (PermissionError, IsADirectoryError) as e:
# Catch multiple exception types in one clause using a tuple
print(f"Cannot read config file: {e}")
return {}
else:
# else: runs ONLY if no exception was raised in the try block
# This is the "success path" โ a clean way to signal normal flow
print(f"Config loaded successfully: {len(data)} keys")
return data
finally:
# finally: runs ALWAYS โ whether an exception occurred or not
# Use it for cleanup that must happen regardless of outcome
print("parse_config() finished.")
The else clause is one of the least-used but most expressive parts of Python's exception system. It answers the question "what do I do when everything went right?" without mixing that logic into the try block itself. If you put the success-path code inside try, you risk accidentally catching exceptions that are raised by that code โ not by the operation you were guarding.
Re-raising Exceptions and Exception Chaining
Sometimes you want to catch an exception, log it or enrich it, and then re-raise it so the caller can still see it:
def load_user_data(user_id):
filepath = f"data/user_{user_id}.json"
try:
with open(filepath) as f:
return json.load(f)
except FileNotFoundError as e:
# Re-raise with additional context using "raise X from Y"
# This preserves the original exception as the __cause__ attribute
raise RuntimeError(f"User data for id={user_id} is missing") from e
The raise X from Y syntax creates an exception chain. When Python prints the traceback, it will show both the original FileNotFoundError and the new RuntimeError, connected with "The above exception was the direct cause of the following exception." This is invaluable when debugging โ you see both the root cause and the higher-level context.
If you just want to re-raise the same exception without modification, use a bare raise inside an except block:
except OSError:
log_error("File operation failed")
raise # re-raises the current exception unchanged
Suppressing Exceptions Intentionally with contextlib.suppress
For cases where you genuinely want to ignore a specific exception, contextlib.suppress is cleaner than a bare try/except/pass:
from contextlib import suppress
import os
# Delete a temp file if it exists; ignore the error if it doesn't
with suppress(FileNotFoundError):
os.remove("tmp_cache.json")
# Without suppress โ noisier
try:
os.remove("tmp_cache.json")
except FileNotFoundError:
pass # silence is the intent, but "pass" hides that intent
contextlib.suppress makes the intent explicit: "I am deliberately ignoring this exception class." Use it sparingly โ it is appropriate only when the exception truly represents a normal, expected condition (like trying to remove a file that might already be gone).
๐ How Python Finds the Right Handler: Exception Propagation Through the Call Stack
Understanding what happens after an exception is raised requires a mental model of the call stack and how Python searches for a matching handler.
The diagram below traces a FileNotFoundError raised inside open() as it travels up a three-function call stack. Python searches for a matching except clause in the current frame first, then in each caller frame going upward. The finally block in every frame that has one will execute during unwinding, regardless of whether a handler is found. If Python reaches the top of the call stack without finding a handler, it prints the full traceback and terminates the program.
`mermaid graph TD A[main calls load_config] --> B[load_config calls open] B --> C{File exists?} C -->|Yes| D[Return file handle] C -->|No| E[FileNotFoundError raised] E --> F{except clause in open frame?} F -->|No| G[Unwind to load_config frame] G --> H{except FileNotFoundError in load_config?} H -->|Yes| I[Handler executes - exception resolved] H -->|No| J[Unwind to main frame] J --> K{except clause in main?} K -->|Yes| L[Handler executes - exception resolved] K -->|No| M[Interpreter catches - full traceback printed] I --> N[finally blocks run on the way out] L --> N D --> O[else block runs - success path]
Notice that exception propagation moves *upward* through callers, not downward into callees. The `finally` blocks are always executed during this unwinding process โ this is the guarantee that makes file handle cleanup reliable. The `else` block runs only on the successful path, when `open()` returns normally and no exception was raised. This clean separation of success and failure paths is what makes Python's exception model so readable once you internalize it.
---
## ๐ File I/O in the Real World: CSV, JSON Configs, Log Files, and Binary Data
The open-read-close pattern extends to several common file formats Python developers encounter daily.
### Processing CSV Files with the `csv` Module
Never parse CSV by splitting on commas manually โ quoted fields with embedded commas will break it silently. Python's `csv` module handles all edge cases:
```python
import csv
def read_employees(filepath):
employees = []
try:
with open(filepath, "r", encoding="utf-8", newline="") as f:
# newline="" is required for csv.reader to handle line endings correctly
reader = csv.DictReader(f)
for row in reader:
# Each row is an OrderedDict keyed by the header row
employees.append({
"name": row["name"],
"department": row["department"],
"salary": float(row["salary"]),
})
except FileNotFoundError:
print(f"CSV file not found: {filepath}")
except KeyError as e:
print(f"Missing expected column: {e}")
except ValueError as e:
print(f"Could not parse salary as a number: {e}")
return employees
Loading a JSON Configuration File
JSON configuration loading is one of the most common file I/O tasks in Python services:
import json
from pathlib import Path
DEFAULT_CONFIG = {
"log_level": "INFO",
"database": {"host": "localhost", "port": 5432},
"max_connections": 10,
}
def load_config(config_path="config.json"):
path = Path(config_path)
try:
with path.open("r", encoding="utf-8") as f:
user_config = json.load(f)
# Merge user config on top of defaults โ user values win
return {**DEFAULT_CONFIG, **user_config}
except FileNotFoundError:
print(f"No config found at {config_path!r} โ using defaults.")
return DEFAULT_CONFIG.copy()
except json.JSONDecodeError as e:
print(f"Invalid JSON in config: {e}")
return DEFAULT_CONFIG.copy()
Note the use of pathlib.Path โ it is the modern, object-oriented alternative to os.path string manipulation and is the preferred approach in new Python 3 code.
Appending to a Log File
Append mode ("a") is perfect for log files โ it adds to the end without erasing existing content, and it creates the file if it does not yet exist:
from datetime import datetime
def append_log(message, logfile="app.log"):
timestamp = datetime.now().strftime("%Y-%m-%d %H:%M:%S")
entry = f"[{timestamp}] {message}\n"
try:
with open(logfile, "a", encoding="utf-8") as f:
f.write(entry)
except OSError as e:
# Fallback: print to stderr if we cannot write the log
import sys
print(f"Log write failed: {e}", file=sys.stderr)
Reading Binary Files (Images, PDFs)
Binary mode ("rb") returns raw bytes objects rather than decoded strings โ essential when handling any non-text data:
def read_file_header(filepath, num_bytes=8):
"""Read the first N bytes of a file to inspect its magic number."""
try:
with open(filepath, "rb") as f:
header = f.read(num_bytes)
return header
except FileNotFoundError:
return None
# Check if a file is a PNG (PNG magic bytes: 89 50 4E 47 0D 0A 1A 0A)
header = read_file_header("photo.png")
if header and header[:4] == b'\x89PNG':
print("File is a valid PNG image")
Binary mode is also the correct choice for reading any file whose encoding is unknown โ you get the raw bytes and decide what to do with them.
๐งช Three Worked Examples: From Fragile to Production-Ready
This section walks through three complete, runnable examples that apply everything covered above. Each one starts with context, shows the full implementation, then highlights what makes it robust.
Example 1: Robust Config Loader with Defaults and Validation
A production config loader needs to handle missing files, malformed JSON, and missing required keys โ all without crashing the calling service:
import json
from pathlib import Path
REQUIRED_KEYS = {"database_url", "secret_key", "debug"}
def load_application_config(path="config.json"):
"""
Load application config from a JSON file.
Returns a config dict on success.
Raises RuntimeError with a clear message on unrecoverable failure.
"""
config_path = Path(path)
try:
with config_path.open("r", encoding="utf-8") as f:
config = json.load(f)
except FileNotFoundError:
raise RuntimeError(
f"Required config file not found: {config_path.resolve()}\n"
"Create config.json before starting the service."
) from None
except json.JSONDecodeError as e:
raise RuntimeError(
f"Config file contains invalid JSON: {e}\n"
f"Fix line {e.lineno} in {config_path}."
) from e
missing = REQUIRED_KEYS - set(config.keys())
if missing:
raise RuntimeError(
f"Config is missing required keys: {sorted(missing)}"
)
return config
# Usage
try:
config = load_application_config()
print(f"Service starting with debug={config['debug']}")
except RuntimeError as e:
print(f"Startup failed: {e}")
raise SystemExit(1)
This example uses raise X from None to suppress the original traceback when the user-facing error message is already complete, and raise X from e to chain the original exception when the internals matter for debugging.
Example 2: Streaming Log Processor with Error Recovery
A streaming processor reads a potentially large log file line by line. Malformed lines should be counted and skipped, not crash the whole run:
import re
from datetime import datetime
LOG_PATTERN = re.compile(
r"\[(?P<timestamp>[^\]]+)\]\s+(?P<level>INFO|WARN|ERROR)\s+(?P<message>.+)"
)
def process_log_file(filepath, target_level="ERROR"):
"""
Stream-parse a log file and yield entries at or above target_level.
Skips malformed lines and reports a summary at the end.
"""
counts = {"parsed": 0, "skipped": 0, "matched": 0}
results = []
try:
with open(filepath, "r", encoding="utf-8", errors="replace") as f:
for line_number, line in enumerate(f, start=1):
line = line.strip()
if not line:
continue
match = LOG_PATTERN.match(line)
if not match:
counts["skipped"] += 1
continue
counts["parsed"] += 1
if match.group("level") == target_level:
counts["matched"] += 1
results.append({
"line": line_number,
"timestamp": match.group("timestamp"),
"message": match.group("message"),
})
except FileNotFoundError:
print(f"Log file not found: {filepath}")
return [], counts
except OSError as e:
print(f"Could not read log file: {e}")
return [], counts
else:
print(
f"Processed {counts['parsed']} lines, "
f"skipped {counts['skipped']} malformed, "
f"found {counts['matched']} {target_level} entries."
)
return results, counts
# Usage
errors, stats = process_log_file("service.log", target_level="ERROR")
for entry in errors:
print(f"Line {entry['line']}: {entry['message']}")
Note errors="replace" in the open() call โ this substitutes the Unicode replacement character (\ufffd) for any byte sequences that cannot be decoded, instead of raising UnicodeDecodeError. For log files from systems with mixed encodings, this is often the right default.
Example 3: CSV-to-JSON Converter with Atomic Writes
This example converts a CSV file to JSON and writes the result atomically โ meaning either the full file is written or nothing is written, avoiding partially written output files:
import csv
import json
import os
import tempfile
from pathlib import Path
def csv_to_json(csv_path, json_path, encoding="utf-8"):
"""
Convert a CSV file to a JSON array file.
Uses an atomic write: the output file is only replaced when
the new content is fully written, preventing partial writes.
"""
csv_path = Path(csv_path)
json_path = Path(json_path)
rows = []
# Step 1: Read the CSV
try:
with csv_path.open("r", encoding=encoding, newline="") as f:
reader = csv.DictReader(f)
rows = list(reader)
except FileNotFoundError:
raise FileNotFoundError(f"Input CSV not found: {csv_path}") from None
except csv.Error as e:
raise ValueError(f"Malformed CSV: {e}") from e
if not rows:
raise ValueError(f"CSV file is empty: {csv_path}")
# Step 2: Write JSON atomically via a temp file in the same directory
tmp_fd, tmp_path = tempfile.mkstemp(dir=json_path.parent, suffix=".tmp")
try:
with os.fdopen(tmp_fd, "w", encoding=encoding) as tmp_file:
json.dump(rows, tmp_file, indent=2, ensure_ascii=False)
# Atomic rename โ on POSIX this is guaranteed atomic
os.replace(tmp_path, json_path)
except OSError:
# Clean up the temp file if something went wrong
with suppress(OSError):
os.remove(tmp_path)
raise
print(f"Converted {len(rows)} rows from {csv_path} to {json_path}")
return len(rows)
from contextlib import suppress
# Usage
try:
count = csv_to_json("employees.csv", "employees.json")
print(f"Success: {count} records written")
except (FileNotFoundError, ValueError, OSError) as e:
print(f"Conversion failed: {e}")
The atomic write pattern โ write to a temp file, then os.replace() โ is important whenever you are updating a file that other processes might be reading. On POSIX systems, os.replace() is guaranteed to be atomic at the filesystem level.
๐ Lessons Learned: What Five Years of File I/O Bugs Teach You
Never skip with. Every time you write f = open(...) without a with block, you are betting that no exception will occur between that line and f.close(). You will eventually lose that bet.
EAFP is not "ignore errors" โ it's "handle errors at the right level." Catching FileNotFoundError and returning a default is not sloppy; it is appropriate when missing config is a recoverable situation. Catching bare Exception everywhere and returning None is sloppy.
Always specify encoding="utf-8" explicitly. Python's default encoding depends on the operating system โ it is cp1252 on many Windows systems, which will silently corrupt files with accented characters when you move code between platforms. Explicit is always better than implicit.
Use exception chaining (raise X from Y). When you catch a low-level exception and raise a higher-level one, preserve the original with from e. Future-you will be grateful when debugging a production incident at 2 AM.
else in try/except is underused. Put your success-path logic there, not at the end of the try block. It makes your intent clear: "this code only runs when nothing went wrong."
Atomic writes prevent partial files. If your process is killed mid-write, a temp-file-plus-rename pattern means readers either see the old complete file or the new complete file โ never a partial one.
Binary mode for binary files, text mode for text files. This seems obvious until you try to parse a PDF in text mode and get UnicodeDecodeError on the first byte.
๐ TLDR: File I/O and Exception Handling in One Paragraph
TLDR: Python favors EAFP over LBYL โ attempt the operation, then handle specific exceptions if it fails. Always use
with open(...) as f:to guarantee file handles are closed. Catch the most specific exception class available (FileNotFoundError, notException). Usetry/except/else/finallyโelsefor the success path,finallyfor cleanup that always runs. Chain exceptions withraise X from Yto preserve root-cause context. For production file writing, use an atomic temp-file-plus-rename pattern to avoid partial writes.
๐ Practice Quiz: Test Your File I/O and Exception Handling Knowledge
What is the default file mode when you call
open("data.txt")with no mode argument?- A)
"w"โ write mode - B)
"a"โ append mode - C)
"r"โ read mode - D)
"r+"โ read/write mode
Show answer
Correct Answer: C โ"r"(read mode, text).open(path)defaults to"r". The call fails withFileNotFoundErrorif the file does not exist. To create a file, you need"w","a", or"x"mode.- A)
Why does the
withstatement guarantee the file is closed even if an exception is raised inside the block?- A) Python automatically wraps all file operations in try/finally internally
- B) The context manager protocol calls
__exit__()when thewithblock exits for any reason, including exceptions - C)
withopens the file in a special read-only mode that auto-closes - D) The garbage collector immediately closes file handles when they go out of scope
Show answer
Correct Answer: B โ__exit__()is called unconditionally. Thewithstatement calls__enter__()on entry and__exit__()on exit. The__exit__()method of Python's file objects always closes the handle, regardless of whether the block exited normally or via an exception. The garbage collector (option D) is non-deterministic โ it may close handles eventually, but not immediately.
You want to delete a temp file if it exists, but silently do nothing if it is already gone. Which is the most Pythonic approach?
- A) Check
os.path.exists()before callingos.remove() - B) Wrap
os.remove()intry/except FileNotFoundError: pass - C) Use
contextlib.suppress(FileNotFoundError)aroundos.remove() - D) Use a bare
try/except:to suppress all exceptions
Show answer
Correct Answer: C โcontextlib.suppress(FileNotFoundError). Both B and C are functionally correct EAFP, butcontextlib.suppressis more expressive: it makes the intent explicit rather than usingpasswhich looks like an oversight. Option A is LBYL and suffers from the TOCTOU race condition โ the file could be deleted between theexists()check and theremove()call. Option D catches all exceptions, which hides unrelated bugs.- A) Check
In
try/except/else/finally, when does theelseclause execute?- A) When an exception is raised in the
tryblock - B) When the
exceptclause handles an exception - C) When the
tryblock completes without raising any exception - D) Always, regardless of whether an exception was raised
Show answer
Correct Answer: C โelseruns only when no exception was raised intry. This is the key distinction:elseis the "success path." Code inelseis outside the protection of thetryblock, meaning exceptions raised there will not be caught by the precedingexceptclauses โ they propagate normally.finally(option D) runs always;elseruns only on success.- A) When an exception is raised in the
What is the difference between
raise RuntimeError("msg") from eandraise RuntimeError("msg") from None?- A) There is no difference โ both re-raise the same exception
- B)
from echains the original exception as__cause__;from Nonesuppresses the original traceback entirely - C)
from Noneis invalid Python syntax - D)
from esuppresses the original exception;from Nonepreserves it
Show answer
Correct Answer: B.raise X from esetsX.__cause__ = eand Python prints "The above exception was the direct cause of the following exception" in the traceback โ invaluable for debugging.raise X from NonesetsX.__suppress_context__ = True, which completely hides the original exception from the traceback. Usefrom Noneonly when the original exception would confuse the user and your new message is fully self-sufficient.
Open-ended challenge: You are building a service that reads a YAML configuration file on startup. The file may not exist on the first run (and should be created with defaults), or it may contain invalid YAML, or it may be present but missing required keys. Design a
load_or_create_config(path)function that handles all three cases with appropriate error recovery, uses atomic writes for the initial file creation, and raises aRuntimeErroronly for truly unrecoverable states (such as a permission error). What exceptions will you catch, what will you log, and what will you return in each case?There is no single correct answer โ consider the trade-offs between fail-fast and graceful degradation, and how an operator monitoring the service logs would understand what happened.
๐ Related Posts in the Python Programming Series
- Python Basics: Variables, Types, and Control Flow โ Start here if you are new to Python. Covers the label model, truthiness, f-strings, and loop patterns that underpin everything in this post.
- Python OOP: Classes, Dataclasses, and Dunder Methods โ Learn how Python's
__enter__and__exit__dunder methods enable the context manager protocol used bywith open(...). - Pythonic Code: Idioms Every Developer Should Know โ Context managers, comprehensions, unpacking, and the rest of the idiomatic Python toolkit that makes production code readable.

Written by
Abstract Algorithms
@abstractalgorithms
More Posts
RAG vs Fine-Tuning: When to Use Each (and When to Combine Them)
TLDR: RAG gives LLMs access to current knowledge at inference time; fine-tuning changes how they reason and write. Use RAG when your data changes. Use fine-tuning when you need consistent style, tone, or domain reasoning. Use both for production assi...
Fine-Tuning LLMs with LoRA and QLoRA: A Practical Deep-Dive
TLDR: LoRA freezes the base model and trains two tiny matrices per layer โ 0.1 % of parameters, 70 % less GPU memory, near-identical quality. QLoRA adds 4-bit NF4 quantization of the frozen base, enabling 70B fine-tuning on 2ร A100 80 GB instead of 8...
Build vs Buy: Deploying Your Own LLM vs Using ChatGPT, Gemini, and Claude APIs
TLDR: Use the API until you hit $10K/month or a hard data privacy requirement. Then add a semantic cache. Then evaluate hybrid routing. Self-hosting full model serving is only cost-effective at > 50M tokens/day with a dedicated MLOps team. The build ...
Watermarking and Late Data Handling in Spark Structured Streaming
TLDR: A watermark tells Spark Structured Streaming: "I will accept events up to N minutes late, and then I am done waiting." Spark tracks the maximum event time seen per partition, takes the global minimum across all partitions, subtracts the thresho...
