Oct 2025AI Designing and Building Custom Reinforcement Learning Environments for Fine-tuning LLMsN. Bantilan
Oct 2025AI Enabling VLLM V1 on AMD GPUs With TritonThomas Parnell, IBM Research & Aleksandr Malyshev, AMD
Oct 2025AI Scaling KV Caches for LLMs: How LMCache + NIXL Handle Network and Storage...- J. Jiang & M. Khazraee
Oct 2025AI Serving PyTorch LLMs at Scale: Disaggregated Inference With Kubernetes and Llm-dM. Ayoub & C. Liu
Oct 2025AI Unlocking Performance: Harnessing LLMs To Streamline GPU Kernel Development in...Jiannan Wang
Oct 2025AI Verl: A Flexible and Efficient RL Framework for LLMsHongpeng Guo & Ziheng Jiang, ByteDance Seed
Oct 2025AI A Practical Field Guide to Optimizing Cost, Speed & Accuracy of LLMsNiels Bantilan, Union.ai
Oct 2025AI LLM Inference: A Comparative Guide to Modern Open-Source RuntimesAleksandr Shirokov, Wildberries
Oct 2025AI Vibe Coding Your First LLM End-to-End ApplicationGreg Loughnane & Chris Alexiuk, AI Makerspace
Oct 2025AI Vibe Coding Your First LLM End-to-End ApplicationGreg Loughnane & Chris Alexiuk, AI Makerspace
Jun 2025DevOps From AI agents to LLMs as judges: Reshaping observability in the era of generative AIDiana Todea
Jun 2025AI From LLM-as-a-Judge To Human-in-the-Loop: Rethinking Evaluat...Eric Pugh & Fernando Rejon Barrera
Jun 2025AI From Traces To Action: Auto-Instrumenting LLMs for Observability...Aditya Soni & Anshika Tiwari
Jun 2025AI Optimizing LLM Performance With Caching Strategies in OpenSearchUri Rosenberg & Sherin Chandy