Search by topic, company, or concept and scan results quickly.
Why it matters: HQQ enables engineers to deploy massive LLMs on consumer-grade hardware with minimal setup. By removing the need for calibration data and drastically reducing quantization time, it simplifies the pipeline for optimizing and testing state-of-the-art models at scale.
Why it matters: This article details how Pinterest uses advanced ML and LLMs to understand complex user intent, moving beyond simple recommendations to goal-oriented assistance. It offers a practical blueprint for building robust, extensible recommendation systems from limited initial data.
Why it matters: This article details how Netflix scaled real-time recommendations for live events to millions of users, solving the "thundering herd" problem. It offers a robust, two-phase architectural pattern for high-concurrency, low-latency updates, crucial for distributed systems engineers.
Why it matters: DSF revolutionizes AI network scaling by overcoming traditional fabric limitations. Its disaggregated architecture, packet spraying, and advanced congestion control ensure high-performance, lossless connectivity for massive GPU clusters, crucial for the future of large-scale AI model training.
Why it matters: This article details how Netflix built a real-time distributed graph to unify disparate data from microservices, enabling complex relationship analysis and personalized experiences. It showcases a robust stream processing architecture for internet-scale data.
Why it matters: This article details Meta's innovations in LLM inference parallelism, offering critical strategies for engineers to achieve high throughput, low latency, and better resource efficiency when deploying large language models at scale. It provides practical solutions for optimizing performance.
Why it matters: This article introduces Sapling's innovative directory branching solution for monorepos, enabling scalable version management and merging without compromising performance or developer experience. It's crucial for engineers working with large codebases to maintain agility.
Why it matters: This article details how Meta is re-architecting its core network infrastructure to handle the massive data demands of AI, offering insights into large-scale network design for future-proof, high-capacity connectivity.
Why it matters: This article offers engineers actionable design principles to reduce IT hardware's environmental impact, fostering sustainability and cost savings through circularity and emissions reduction in data center infrastructure.
Why it matters: Postgres 18's new I/O methods offer performance gains, but their effectiveness depends heavily on storage architecture. Understanding the trade-offs between io_uring and worker processes helps engineers optimize database throughput and cost-efficiency for I/O-bound workloads.