Maligned #14 - Agents, Video, and Practical Control
Monday. New week. Let’s get into it.
AI Agent Security Demands Fundamental Rethink
Perplexity’s white paper on AI agent security highlights what many of us building these systems already know: current security models are insufficient. Agents break traditional code-data separation, authority boundaries, and execution predictability, introducing novel attack surfaces like indirect prompt injection and confused-deputy scenarios. The emphasis on sandboxed execution and deterministic policy enforcement for high-consequence actions is spot on. This isn’t just about patching; it’s about re-architecting security from the ground up to account for emergent behaviours and complex tool interactions. Enterprises rolling out agents need to pay close attention to these architectural shifts, as the risks are far higher than with static models.
Multimodal Reasoning Benchmarks Expose Agentic AI Weakness
We keep hearing about multimodal models doing everything, but new benchmarks like MM-CondChain reveal a significant gap: deep compositional reasoning. Current MLLMs struggle with multi-layer conditional logic, especially when it’s visually grounded and requires tracking multiple objects or attributes. The “if X and Y, then Z, otherwise A” scenarios, common in GUI automation or medical imaging workflows, lead to sharp performance drops as complexity grows. This isn’t just an academic curiosity; it directly impacts the reliability of AI agents in complex, real-world tasks. We need more than superficial understanding; we need models that can genuinely reason through chains of conditions.
Video LLMs Begin Streaming “Thinking While Watching”
For truly interactive AI, processing video streams isn’t enough; models need to reason in real-time. Video Streaming Thinking (VST) proposes a neat solution: amortising LLM reasoning latency by activating “thinking while watching” over incoming video clips. This paradigm improves timely comprehension and coherent cognition without sacrificing responsiveness. The results on online benchmarks like StreamingBench are promising, showing substantial speed improvements while maintaining competitive performance on offline tasks. This is a practical step towards responsive video agents that can interact fluidly, whether it’s for remote assistance or surgical guidance. Building real-time understanding is critical for production systems.
OmniStream Offers a Unified Visual Backbone for Agents
The promise of a general-purpose vision model, capable of perception, reconstruction, and action across diverse inputs, moved a step closer with OmniStream. By integrating causal spatiotemporal attention and 3D positional embeddings, it supports efficient, frame-by-frame processing. Pre-training on 29 datasets for multiple tasks – static and temporal representation, geometric reconstruction, vision-language alignment – shows ambition. While it doesn’t aim for benchmark dominance, demonstrating consistent performance across varied tasks with a frozen backbone is meaningful. This could simplify agent architectures considerably, moving away from fragmented, specialised vision components towards a more integrated understanding of the visual world.
Precise Multi-Subject Video Generation Becomes Achievable
Generating complex video with specific subjects performing controlled motions has been a persistent challenge. DreamVideo-Omni presents a unified framework tackling this, allowing harmonious multi-subject customisation with “omni-motion” control. The two-stage training, integrating comprehensive control signals for appearances, global motion, local dynamics, and camera movements, is clever. Especially interesting is the use of group and role embeddings to disentangle motion signals for specific identities, resolving multi-subject ambiguity. Incorporating latent identity reward feedback learning also helps preserve identity, a common failure point. This moves video generation capabilities significantly closer to production-ready, highly controllable creative tools.
Elastic Latent Interfaces Improve Diffusion Transformer Efficiency
Diffusion Transformers (DiTs) are powerful, but their fixed computational cost, tied directly to image resolution, limits practical deployment and efficient scaling. ELIT, or Elastic Latent Interface Transformer, offers a smart solution by decoupling input image size from compute. It introduces a learnable, variable-length latent token sequence, allowing dynamic adjustment of compute based on constraints. By prioritising important regions and learning importance-ordered representations, ELIT delivers substantial gains in FID and FDD scores while reducing FLOPs. This is a sensible architectural improvement that offers fine-grained control over the latency-quality trade-off, making DiTs more practical for real-world applications where computational budgets vary.
See you next week.
Maligned - AI news by Mal