- Agent Pulse
- Posts
- #49: Forget LLMs? NVIDIA Says Small Models Will Power AI Agents
#49: Forget LLMs? NVIDIA Says Small Models Will Power AI Agents
AgentPulse: Your Weekly Dose of AI Agents News.

AArena | Submit Agent | Leaderboard | Landscape Map | Agencies | Advertise
Hey, it’s OP.
In today’s Agent Pulse:
📢 Small Models Will Power AI Agents
📢 A README for AI Agents
⚔️ Gemma is #1 on Agent Arena
✨ Featured Agents
📡 Agent Signals
🎓 Free courses
📚 Must-Read Papers
🗺️ Landscape Map
You read, we listen. Got feedback? Just hit reply - we’d love to hear from you.
Enjoying Agent Pulse?
Support independent AI insights for just $5/month. Upgrade to Premium →
Turn AI into Your Income Engine
Ready to transform artificial intelligence from a buzzword into your personal revenue generator
HubSpot’s groundbreaking guide "200+ AI-Powered Income Ideas" is your gateway to financial innovation in the digital age.
Inside you'll discover:
A curated collection of 200+ profitable opportunities spanning content creation, e-commerce, gaming, and emerging digital markets—each vetted for real-world potential
Step-by-step implementation guides designed for beginners, making AI accessible regardless of your technical background
Cutting-edge strategies aligned with current market trends, ensuring your ventures stay ahead of the curve
Download your guide today and unlock a future where artificial intelligence powers your success. Your next income stream is waiting.
📢 TOP Headlines
Why Small Language Models Could Power the Next Wave of AI Agents

NVIDIA just dropped a research paper with a bold claim: small language models (SLMs), not giant LLMs, are the real future of agentic AI. At first glance, it sounds counterintuitive - after all, LLMs like GPT-5 and Claude dominate headlines. But NVIDIA’s argument is refreshingly pragmatic.
Here’s the gist:
Most agents don’t need “infinite wisdom.” The majority of agent tasks - like calling APIs, formatting structured data, or generating boilerplate code - are narrow, repetitive, and rule-bound. You don’t need a 175B-parameter model to schedule a meeting or process a JSON output.
SLMs are getting seriously capable. Models like Microsoft’s Phi-3 (7B), NVIDIA’s own Nemotron-H (2–9B), and Salesforce’s xLAM-2 (8B) are matching or beating older 30B–70B models at reasoning, tool calling, and code generation. That means small is no longer “weak.”
Economics matter. Serving an SLM is 10–30× cheaper than an LLM. They run faster, burn less energy, and can even live on consumer GPUs (hello, on-device agents). For businesses, that means scalability without bankrupting on inference costs.
Modularity beats monoliths. Instead of one all-knowing LLM, NVIDIA suggests a “Lego-style” approach: combine a handful of small, specialized agents for specific jobs, and bring in a big model only when truly needed.
One striking takeaway: NVIDIA estimates that in open-source agent frameworks like MetaGPT or Cradle, 40–70% of LLM calls could already be swapped out for SLMs without losing performance.
Why this matters:
If you’re building agents, SLMs mean lower costs, faster iteration, and easier fine-tuning. You can adapt overnight instead of waiting weeks for an LLM fine-tune.
If you’re using agents, it points toward a future where AI tools are cheaper, more private (local execution), and more abundant - not bottlenecked by a handful of massive cloud APIs.
If you’re investing in the AI space, this shift could reshape infrastructure bets - from centralized LLM superclouds toward distributed, specialized SLM ecosystems.
Our take: NVIDIA isn’t saying LLMs are obsolete. They’ll still shine in complex, open-ended reasoning. But for the everyday plumbing of agent workflows, smaller, specialized models may quietly take over - and that’s a shift worth watching closely.
Agents.md — A README for AI Agents

OpenAI and partners just introduced Agents.md, a simple but powerful standard: a Markdown file that lives alongside your code to guide AI agents on how to interact with your project. Think of it as a README, but for ai agents.
Instead of bloating human-facing docs with build commands, test scripts, or linting rules, you put those details into AGENTS.md
. Now, tools like OpenAI Codex, Cursor, Google’s Gemini CLI, and Sourcegraph’s Amp can automatically parse it and “know” how to behave inside your repo.
Why this matters
Standardization over fragmentation
Until now, every AI coding agent had its own configs and quirks. Agents.md offers a single, shared format - so your project works across ecosystems instead of locking into one tool.Smarter automation, fewer mistakes
Agents can run builds, tests, and PR checks the way your team does them - because you’ve spelled it out once in agents.md. This reduces the risk of an agent “hallucinating” commands or ignoring team conventions.Modularity for complex codebases
You can drop multipleagents.md
files into different subfolders of a repo. That means a monorepo with frontend, backend, and infra can give tailored instructions to agents at each level.
The bigger picture
This isn’t just an OpenAI experiment. Factory.ai, Sourcegraph, Google, and OpenAI are all backing the spec. If adoption spreads, agents.md
could become a de facto interface layer between humans and AI agents, much like README.md
or package.json
became foundational in their eras.
It’s also part of a larger trend: moving from “generalist AI agents that figure it out” to “AI agents that reliably follow project-specific rules.” In other words, agents that behave like trained teammates, not freelancers guessing at your process.
What to watch
Adoption curve: If GitHub, GitLab, or major frameworks start auto-scaffolding
agents.md
, it could spread fast.Ecosystem lock-in vs. open standard: Right now it’s open, but if one vendor twists the spec, we could see fragmentation creep back in.
Beyond code repos: Imagine a
agents.md
-style file guiding business agents - customer support, sales ops, finance automation—on how to safely interact with company workflows.
Our take: agents.md
may look like a tiny addition, but it hints at a bigger future: AI agents won’t just “understand language,” they’ll understand your context - because you’ve documented it once in a standard way.
Typing is a thing of the past.
Typeless turns your raw, unfiltered voice into beautifully polished writing - in real time.
It works like magic, feels like cheating, and allows your thoughts to flow more freely than ever before.
Your voice is your strength. Typeless turns it into a superpower.
⚔️ AArena: All-in-One AI Workspace
Get Sh*t Done with AI.
Bring your task. Get best results.
TOP 5 AI (Aug 25):
Gemma 3 4b
Qwen3 32B
Grok 4
Llama 3.3 70B
ChatGPT-4o
✨ Featured Agents
TeammatesAI: Autonomous AI Teammates
TensorStax: Autonomous AI Agents for Data Engineering
TheLibrarian.io: WhatsApp AI Personal Assistant
Agentverse: Search and Discover AI Agents
📡 TOP Agent Signals
xAI Grok 2.5 is open-sourced now - Source
Reynolds introduces Rey, its cutting-edge AI agent for dealerships - Source
Cambridge Investment Research implements agentic AI - Source
Composio announces its Universal MCP Gateway - Source
Confluent announces Streaming Agents for Confluent - Source
Druva expands DruAI with intelligent agents - Source
LambdaTest launches private beta of Agent-to-Agent Testing - Source
AI agent transactions will trigger new payment disputes - Source
91% of organizations are already using AI agents - Source
Former Zillow exec targets $1.3T market
The wealthiest companies tend to target the biggest markets. For example, NVIDIA skyrocketed nearly 200% higher in the last year with the $214B AI market’s tailwind.
That’s why investors are so excited about Pacaso.
Created by a former Zillow exec, Pacaso brings co-ownership to a $1.3 trillion real estate market. And by handing keys to 2,000+ happy homeowners, they’ve made $110M+ in gross profit to date. They even reserved the Nasdaq ticker PCSO.
No wonder the same VCs behind Uber, Venmo, and eBay also invested in Pacaso. And for just $2.90/share, you can join them as an early-stage Pacaso investor today.
Paid advertisement for Pacaso’s Regulation A offering. Read the offering circular at invest.pacaso.com. Reserving a ticker symbol is not a guarantee that the company will go public. Listing on the NASDAQ is subject to approvals.
🎓 FREE AI Courses
New! LangGraph: Deep Research
Anthropic:
AI Fluence Course, designed for everyday users of AI.
Google: Introduction to LLM
BeeAI: Agent Communication Protocol
HuggingFace: Model Context Protocol (MCP)
Microsoft: Generative AI for Beginners
OpenAI: Advanced Prompt Engineering

📚 Must Read Papers
New! NVIDIA: Small Language Models are the Future of Agentic AI (Doc)
Google:
Microsoft: Governing Agents (Doc)
OpenAI: A Practical Guide to building Agents (Doc)
AWS: Agentic AI frameworks (Doc)
KPMG: AI Quarterly Pulse Survey: Q2 2025 (Doc)
Stanford University: Future of Work with AI Agents (Doc)
Thomson Reuters: Agentic AI 101 (Doc)
BCG: AI at Work (Doc)
ServiceNow: Enterprise AI Maturity Index 2025 (Doc)
IBM: Agentic AI in Financial Services (Doc)
Capgemini: Rise of Agentic AI (Doc)
Accenture: Technology vision 2025 (Doc)
Infosys: Agentic Enterprise AI Playbook (Doc)

🗺️ AI Agents Landscape Map
We’re mapping thousands of AI Agents. Find your next AI agent today — explore thousands of options on our interactive Landscape Map.
How'd we do? |
Reach 23,000+ Readers:
Newsletter is read by VCs, founders, engineers, managers and tech professionals.
Reply