Practical guide to multimodal embedding and reranker models that extend traditional RAG pipelines to handle text, images, and other modalities in a shared embedding space. Covers model loading, encoding mixed-modality inputs, and computing cross-modal similarities with concrete code examples and performance considerations.
Comprehensive reference on coding agent architecture covering six main building blocks of agentic systems (tool use, context management, memory, prompt caching, etc.) and how they differ from raw LLMs and reasoning models. Explains why systems like Claude Code outperform standalone models through their surrounding harness design rather than model capability alone.
A comprehensive Chinese technical guide ("御舆") that deconstructs AI Agent architecture, specifically analyzing Claude Code's design patterns including conversation loops, tool permission pipelines, context compression, and the Agent Harness runtime framework. Provides a transferable mental model for building production-grade agent systems across different frameworks without relying on prompt engineering tutorials.
In-depth technical analysis of Claude Code's source architecture, covering the agent loop, context engineering, tool system, and production-grade error recovery strategies. Includes a companion project (Claude Code From Scratch) with ~4000 lines of TypeScript/Python and 11-chapter tutorial for building your own AI programming agent from scratch.
Comprehensive reference guide organizing 45+ LLM architectures with visual model cards and detailed explanations of attention variants (MHA, GQA, sliding window, etc.) used in modern models. Includes both a web gallery and printable poster, serving as a practical learning resource for understanding contemporary transformer architectures.
A comprehensive AI engineering curriculum spanning 260+ lessons across 20 phases (~290 hours) covering fundamentals from linear algebra to autonomous agent swarms in Python, TypeScript, Rust, and Julia. Each lesson produces reusable artifacts (prompts, skills, agents, MCP servers) that can be immediately integrated into AI coding workflows, with personalized learning paths based on existing ML/DL knowledge.
Comprehensive overview of inference-time scaling techniques for LLMs, covering methods like chain-of-thought prompting, self-consistency, best-of-N ranking, and rejection sampling with verifiers. The author shares practical experimentation results (achieving 15% to 52% accuracy improvement) and categorizes approaches from both academic literature and proprietary LLM implementations, making it directly applicable to deployed systems.
Comprehensive overview of alternative LLM architectures beyond standard transformers, including diffusion models, linear attention hybrids, state space models (SSMs), and specialized architectures like code world models. The article surveys emerging approaches aimed at improving efficiency and modeling performance, with comparisons to current SOTA transformer-based models like DeepSeek R1, Llama 4, and Qwen3.
Practical guide covering four main LLM evaluation methods: multiple-choice benchmarks, verifiers, leaderboards, and LLM judges, with code examples and analysis of their strengths/weaknesses. Essential reading for engineers comparing models, interpreting benchmarks, and measuring progress on their own projects.
Deep dive into Qwen3 architecture implementation from scratch in PyTorch, covering the open-weight model family's design choices and building blocks. Provides practical code examples and architectural patterns directly applicable to understanding modern LLM internals and building custom variations.