Join us

ContentUpdates and recent posts about Grafana Tempo..
Link
@kala shared a link, 3 months, 3 weeks ago
FAUN.dev()

The 1 Billion Token Challenge: Finding the Perfect Pre-training Mix

Researchers squeezed GPT-2-class performance out of a model trained on just1 billion tokens- 10× less data - by dialing in a sharp dataset mix:50% finePDFs, 30% DCLM-baseline, 20% FineWeb-Edu. Static mixing beat curriculum strategies. No catastrophic forgetting. No overfitting. And it hit90%+of GPT-.. read more  

The 1 Billion Token Challenge: Finding the Perfect Pre-training Mix
Link
@kala shared a link, 3 months, 3 weeks ago
FAUN.dev()

Jensen Huang's Stark Warning: China's 1 Million AI Workers vs America's 20,000

Nvidia CEO Jensen Huang, in some leaked comments, didn’t mince words: U.S. export bans aren’t hobbling China’s AI game - they’re fueling it. He pointed to Huawei’s 910C chip edging close to H100 territory, a forecast putting China ahead in AI compute by 2027, and a fast-growing local chip industry n.. read more  

Jensen Huang's Stark Warning: China's 1 Million AI Workers vs America's 20,000
Link
@kala shared a link, 3 months, 3 weeks ago
FAUN.dev()

Context Management in Amp

Amp stretches the context window into something more useful. It pulls in system prompts, tool info, runtime metadata, even AGENTS.md files - fuel for agentic behavior. It gives devs serious control: edit messages, fork threads, drop in files with @mentions, hand off conversations, or link threads to.. read more  

Context Management in Amp
Link
@kala shared a link, 3 months, 3 weeks ago
FAUN.dev()

Google to release Nano Banana Pro next week

Google dropsGemini 3and the newNano Banana Pronext week. Big swing at image generation - now tied tight to Gemini 3 Pro. Early glimpses in Google Vids hint Nano Banana Pro is built for sharper visuals in creative tools. System shift:Google’s stacking its apps behind a single backbone: Gemini 3 Pro. .. read more  

Google to release Nano Banana Pro next week
Link
@kala shared a link, 3 months, 3 weeks ago
FAUN.dev()

Inside Cursor - Sixty days with the AI coding decacorn

Cursor is shaking up recruiting by treating the hiring process as more about the person than the job, resulting in a fast-growing team of exceptional individuals drawn in by the company's compelling mission and focus on challenging technical problems. Women in product and engineering roles are a kno.. read more  

Link
@kala shared a link, 3 months, 3 weeks ago
FAUN.dev()

Introducing structured output for Custom Model Import in Amazon Bedrock

Amazon Bedrock’s Custom Model Import just got structured output support. Now LLMs can lock their responses to your JSON schema - no prompt hacks, no cleanup after... read more  

Link
@kala shared a link, 3 months, 3 weeks ago
FAUN.dev()

LaTeX, LLMs and Boring Technology 

LLMs are tearing down LaTeX's old walls. Syntax hell, cryptic errors, clunky formatting - easier now. Whether baked into editors or running solo, these models smooth the pain. Why does it work so well? LaTeX has history. Mountains of examples. It's the perfect training set. That puts newer contender.. read more  

Link
@kala shared a link, 3 months, 3 weeks ago
FAUN.dev()

Building a Healthcare Robot from Simulation to Deployment with NVIDIA Isaac

NVIDIA just droppedIsaac for Healthcare v0.4, and it’s a big one. Headliner: the newSO-ARM starter workflow- a full-stack sim2real pipeline built for surgical robotics. It covers the whole loop: spin up synthetic and real-world data capture, train withGR00t N1.5, and deploy straight to 6-DOF hardwar.. read more  

Link
@kala shared a link, 3 months, 3 weeks ago
FAUN.dev()

The Fatal Math Error Killing Every AI Architecture - Including The New Ones

LLMs are fading as JEPA (Joint Embedding Predictive Architecture) emerges with joint, embedding, predictive architecture. JEPA is a step towards true intelligence by avoiding the flat, finite spreadsheet trap of Euclidean space and opting for a toroidal model... read more  

Link
@devopslinks shared a link, 3 months, 3 weeks ago
FAUN.dev()

Visibility at Scale: How Detects Sensitive Data Exposure

Segment gutted its old permissions table—bloated, slow, tangled in logic - and replaced it with a lean, service-based setup. The new stack runs onPostgres,Redis, and a sharply tunedGo API, cutting query times from 1400ms to under 100ms. Clean, fast, and centralized... read more  

Visibility at Scale: How Detects Sensitive Data Exposure
Grafana Tempo is a distributed tracing backend built for massive scale and low operational overhead. Unlike traditional tracing systems that depend on complex databases, Tempo uses object storage—such as S3, GCS, or Azure Blob Storage—to store trace data, making it highly cost-effective and resilient. Tempo is part of the Grafana observability stack and integrates natively with Grafana, Prometheus, and Loki, enabling unified visualization and correlation across metrics, logs, and traces.

Technically, Tempo supports ingestion from major tracing protocols including Jaeger, Zipkin, OpenCensus, and OpenTelemetry, ensuring easy interoperability. It features TraceQL, a domain-specific query language for traces inspired by PromQL and LogQL, allowing developers to perform targeted searches and complex trace-based analytics. The newer TraceQL Metrics capability even lets users derive metrics directly from trace data, bridging the gap between tracing and performance analysis.

Tempo’s Traces Drilldown UI further enhances usability by providing intuitive, queryless analysis of latency, errors, and performance bottlenecks. Combined with the tempo-cli and tempo-vulture tools, it delivers a full suite for trace collection, verification, and debugging.

Built in Go and following OpenTelemetry standards, Grafana Tempo is ideal for organizations seeking scalable, vendor-neutral distributed tracing to power observability at cloud scale.