⚡ Performance Engineering

Deep Technical Explorations in Systems Optimization

From bare-metal microcontroller registers to large-scale LLM inference pipelines. We obsess over every microsecond, every cache miss, every memory access pattern.

perf.sh
$ perf stat -e cycles,instructions,cache-misses ./inference
# Performance counter stats:
42,891,234,567 cycles
98,234,567,890 instructions # 2.29 IPC
1,234,567 cache-misses # 0.001%

✓ 3.2x throughput improvement

Recent Posts

View all
llm inference expert

Agent-Based Simulation: Using 10,000 AI Agents to Generate Synthetic Training Data

How to build a simulated marketplace where 10,000 LLM-powered agents shop, negotiate, write reviews, and ask questions — generating millions of realistic training examples. Agent architecture, environment design, persona systems, data extraction pipelines, quality scoring, and the full cost analysis for 1M synthetic examples.

· 30 min read
#synthetic-data#agent-simulation#environment +1
llm inference expert

Code Dataset Curation: Deduplication, License Filtering, and Quality Scoring for LLM Training

How to build a training dataset for code LLMs from raw GitHub data — license filtering with SPDX headers and permissive-only policies, exact and near-deduplication with MinHash at function-level granularity, quality scoring with lint, cyclomatic complexity, and documentation ratio, language distribution strategies, and PII removal. Based on The Stack v2 methodology.

· 35 min read
#code-data#deduplication#license-filtering +2
llm inference expert

Data Mixing: Optimal Proportions of Code, Math, Web, and Books for LLM Training

A deep dive into data mixing for LLM pretraining -- how to allocate 15T tokens across web text, code, math, and books. Covers empirical results from Llama 3, the 'code helps everything' finding, synthetic math data generation, curriculum learning schedules that shift proportions during training, and a complete implementation of a configurable multi-source data mixer with dynamic scheduling.

· 35 min read
#data-mixing#proportions#code +4

Stay Updated

Get notified when new deep-dive technical articles are published.

Subscribe via RSS