HuggingFace Blog · 3d ago · 8 · tutorial rag library inference

Practical guide to multimodal embedding and reranker models that extend traditional RAG pipelines to handle text, images, and other modalities in a shared embedding space. Covers model loading, encoding mixed-modality inputs, and computing cross-modal similarities with concrete code examples and performance considerations.

Ahead of AI · 8d ago · 8 · agent workflow tutorial

Comprehensive reference on coding agent architecture covering six main building blocks of agentic systems (tool use, context management, memory, prompt caching, etc.) and how they differ from raw LLMs and reasoning models. Explains why systems like Claude Code outperform standalone models through their surrounding harness design rather than model capability alone.

GitHub Trending AI · 11d ago · 7 · agent architecture tutorial workflow

A comprehensive Chinese technical guide ("御舆") that deconstructs AI Agent architecture, specifically analyzing Claude Code's design patterns including conversation loops, tool permission pipelines, context compression, and the Agent Harness runtime framework. Provides a transferable mental model for building production-grade agent systems across different frameworks without relying on prompt engineering tutorials.

GitHub Trending AI · 11d ago · 9 · agent architecture tutorial open source workflow

In-depth technical analysis of Claude Code's source architecture, covering the agent loop, context engineering, tool system, and production-grade error recovery strategies. Includes a companion project (Claude Code From Scratch) with ~4000 lines of TypeScript/Python and 11-chapter tutorial for building your own AI programming agent from scratch.

Ahead of AI · 21d ago · 8 · research tutorial open source

Comprehensive reference guide organizing 45+ LLM architectures with visual model cards and detailed explanations of attention variants (MHA, GQA, sliding window, etc.) used in modern models. Includes both a web gallery and printable poster, serving as a practical learning resource for understanding contemporary transformer architectures.

GitHub Trending AI · 24d ago · 7 · tutorial workflow agent open source

A comprehensive AI engineering curriculum spanning 260+ lessons across 20 phases (~290 hours) covering fundamentals from linear algebra to autonomous agent swarms in Python, TypeScript, Rust, and Julia. Each lesson produces reusable artifacts (prompts, skills, agents, MCP servers) that can be immediately integrated into AI coding workflows, with personalized learning paths based on existing ML/DL knowledge.

Ahead of AI · 78d ago · 8 · inference prompt engineering tutorial research

Comprehensive overview of inference-time scaling techniques for LLMs, covering methods like chain-of-thought prompting, self-consistency, best-of-N ranking, and rejection sampling with verifiers. The author shares practical experimentation results (achieving 15% to 52% accuracy improvement) and categorizes approaches from both academic literature and proprietary LLM implementations, making it directly applicable to deployed systems.

Ahead of AI · 158d ago · 7 · research benchmark tutorial

Comprehensive overview of alternative LLM architectures beyond standard transformers, including diffusion models, linear attention hybrids, state space models (SSMs), and specialized architectures like code world models. The article surveys emerging approaches aimed at improving efficiency and modeling performance, with comparisons to current SOTA transformer-based models like DeepSeek R1, Llama 4, and Qwen3.

Ahead of AI · 189d ago · 7 · benchmark tutorial workflow

Practical guide covering four main LLM evaluation methods: multiple-choice benchmarks, verifiers, leaderboards, and LLM judges, with code examples and analysis of their strengths/weaknesses. Essential reading for engineers comparing models, interpreting benchmarks, and measuring progress on their own projects.

Ahead of AI · 218d ago · 8 · tutorial open source research

Deep dive into Qwen3 architecture implementation from scratch in PyTorch, covering the open-weight model family's design choices and building blocks. Provides practical code examples and architectural patterns directly applicable to understanding modern LLM internals and building custom variations.