NVIDIA’s AI Red Team nailed three security sinkholes in LLMs: reckless use of exec/eval, RAG pipelines that grab too much data, and markdown that doesn't get cleaned. These cracks open doors to remote code execution, sneaky prompt injection, and link-based data leaks.
The fix-it trend: App security’s leaning hard into sandboxed runtimes, tighter data perms, and markdown that can’t stab you.









