Join us

ContentUpdates from DevOpsDayLA...
Link
@faun shared a link, 5 months ago
FAUN.dev()

GPT-5 Thinking in ChatGPT (aka Research Goblin) is shockingly good at search

GPT-5's“thinking” modeljust leveled up. It's not just answering queries—it’s doing full-on research. Picture deep, multi-step Bing searches mixed with tool use and reasoning chains. It reads PDFs. Analyzes them. Suggests what to do next. Then actually does it. All from your phone. What’s changing:L.. read more  

GPT-5 Thinking in ChatGPT (aka Research Goblin) is shockingly good at search
Link
@faun shared a link, 5 months ago
FAUN.dev()

From Zero to GPU: A Guide to Building and Scaling Production-Ready CUDA Kernels

Hugging Face just dropped Kernel Builder—a full-stack toolchain for building, versioning, and shippingcustom CUDA kernels as native PyTorch ops. Kernels arearchitecture-aware,semantically versioned, andpullable straight from the Hub. It tracks changes with lockfiles and bakes inDocker deploysout of.. read more  

Link
@faun shared a link, 5 months ago
FAUN.dev()

Best Practices for High Availability of LLM Based on AI Gateway

Alibaba Cloud’s AI Gateway just got sharper. It now handlesreal-time overload protectionandLLM fallback routingusing passive health checks, first packet timeouts, and traffic shaping. It proxies both BYO and cloud LLMs—think PAI-EAS, Tongyi Qianwen—and redirects load spikes or failures on the fly. F.. read more  

Best Practices for High Availability of LLM Based on AI Gateway
Link
@faun shared a link, 5 months ago
FAUN.dev()

Simplifying Large-Scale LLM Processing across Instacart with Maple

Instacart builtMaple, a backend brain for handling millions of LLM prompts—fast, cheap, and shared across teams. It’s not just another service. Maple runs onTemporal,PyArrow, andS3, strip-mines away provider-specific boilerplate, auto-batches prompts, retries failures, and slashes LLM costs by up t.. read more  

Simplifying Large-Scale LLM Processing across Instacart with Maple
Link
@faun shared a link, 5 months ago
FAUN.dev()

Hermes V3: Building Swiggy’s Conversational AI Analyst

Swiggy just gave its GenAI tool, Hermes, a serious glow-up. What started as a simple text-to-SQL bot is now acontext-aware AI analystthat lives inside Slack. The upgrade? Not just tweaks—an overhaul. Think: vector-based prompt retrieval, session-level memory, an Agent orchestration layer, and a SQL.. read more  

Hermes V3: Building Swiggy’s Conversational AI Analyst
Link
@faun shared a link, 5 months ago
FAUN.dev()

Why language models hallucinate

OpenAI sheds light on the persistence ofhallucinationsin language models due to evaluation methods favoring guessing over honesty, requiring a shift towards rewarding uncertainty acknowledgment. High model accuracy does not equate to the eradication of hallucinations, as some questions are inherentl.. read more  

Why language models hallucinate
Link
@faun shared a link, 5 months ago
FAUN.dev()

The Big LLM Architecture Comparison

Architectures since GPT-2 still ride transformers. They crank memory and performance withRoPE, swapGQAforMLA, sprinkle in sparseMoE, and roll sliding-window attention. Teams shiftRMSNorm. They tweak layer norms withQK-Norm, locking in training stability across modern models. Trend to watch:In 2025,.. read more  

The Big LLM Architecture Comparison
Story
@laura_garcia shared a post, 5 months ago
Software Developer, RELIANOID

RELIANOID Load Balancer Community Edition v7 on AWS using Terraform

🚀 New Guide Available! Learn how to quickly deploy RELIANOID Load Balancer Community Edition v7 on AWS using Terraform. Our step-by-step article shows you how to provision everything automatically — from VPCs and subnets to EC2 and key pairs — in just minutes. 👉 https://www.relianoid.com/resources/k..

Knowledge base Deploy RELIANOID Load Balancer Community Edition v7 with Terraform on AWS
Link
@faun shared a link, 5 months ago
FAUN.dev()

Measuring Developer Productivity with Amazon Q Developer and Jellyfish

Amazon Q Developer now plugs into Jellyfish. Teams get a clearer view of how AI fits into the real flow of work—prompt usage, code adoption, PR throughput. Not just surface stats. The setup pipes data from AWS S3 straight into Jellyfish’s analytics engine. It tags AI users, tracks velocity gains, an.. read more  

Measuring Developer Productivity with Amazon Q Developer and Jellyfish
Link
@faun shared a link, 5 months ago
FAUN.dev()

Sandboxed to Compromised: New Research Exposes Credential Exfiltration Paths in AWS Code Interpreters

Researchers poked holes insandboxed Bedrock AgentCore code interpreters—and found a way to leak execution role credentials through theMicroVM Metadata Service (MMDS). No outside network? Doesn’t matter. The exploit dodges basic string filters in requests and lets non-agentic code swipe AWS creds to .. read more  

DevOpsDayLA is Southern California's premier conference focused on the human side of technology delivery. For 15 years, we've brought together SoCal practitioners who understand that great software isn't just about tools, it's about people working together effectively.

The 2026 Lens: DevOps in an AI World: You're working in a world where AI is everywhere, embedded in your tools, requested by your management, and reshaping how teams deliver software. Whether you're embracing these changes or navigating the challenges they create, your experience matters.

We're looking for stories about how you and your organizations are adapting your mindsets, processes, and team dynamics in this rapidly evolving landscape. How did your team integrate AI into existing workflows? What changed about collaboration when AI entered the picture? Where do humans remain critical in automated processes? How do you build the next generation of engineers when entry-level work is automated? What are the real implementation challenges you're facing in Southern California's entertainment, gaming, aerospace, retail, healthcare, and manufacturing companies?

What makes DevOpsDayLA different: We focus on culture, collaboration, and cross-industry implementation stories rather than pure technical deep-dives. We're all about the people and processes that make technology successful in an AI world, not the technology itself.

We especially welcome new speakers, under-represented voices, and fresh perspectives from across SoCal's diverse tech landscape.

Attending

To attend DevOps Day LA purchase a  SCALE ticket via the SCALE website. Your ticket will provide access to both events.

Sponsor

DevOpsDayLA sponsors have the opportunity to offer job opportunities, demonstrations of the latest tools, and world-class service offerings to facilitate organizations’ DevOps journey. The conference is presented with the support of the Southern California Linux Expo (SCALE). It provides a relaxed, casual, and diverse environment where experienced professionals, evangelists, agile practitioners, developers, job seekers, business executives, educators, students, and curiosity seekers can connect and attend talks on current topics.

For additional information, please contact us at los-angeles@devopsdays.org.

Speak at DevOpsDay LA

DevOpsDay LA sessions are selected through our calling for presentations for our 2026 event. We have posted a list of desired topics and best practices for submissions.

Friday, March 6, 2026 - 10:00 to 17:00