Qwen3.5-9B, a 9-billion-parameter open model, now outperforms OpenAI's gpt-oss-120B on several benchmarks. A model you can download, inspect, and fine-tune is beating one you rent by the token. That fact alone tells you open source AI is not a charity project or a reckless experiment. It is the most productive engineering movement in a generation, and the data backs that up even when the data looks messy.
Yes, the 2026 OSSRA report found vulnerabilities per codebase doubled to 581. Yes, 87% of codebases contain at least 1 vulnerability, and 17% of open source components evade tracking entirely. Those numbers are ugly. They are also the expected cost of a system where 97% of organizations use open source AI models and 85% of developers rely on AI coding assistants that generate dependencies faster than anyone can audit them. You do not get a 24x explosion in robotics datasets on Hugging Face, from 1,145 to 26,991 in 3 years, without some broken windows.
The Vulnerability Numbers Are a Scaling Problem, Not an Architecture Problem
I keep seeing people wave the OSSRA stats around as proof that open source AI is inherently dangerous. But those stats describe the entire open source software ecosystem, not a flaw specific to openness. Proprietary systems have vulnerabilities too; you just cannot see them. When 76% of employees at companies that banned AI coding assistants use them anyway, that is not an open source governance failure. That is a management failure. The tools are already inside the building whether the CISO approved them or not.
The real question for builders is whether open development produces better tools faster. Right now, the answer is clearly yes. AI2's Olmo Hybrid reaches the same accuracy as Olmo 3 using 49% fewer tokens. That is a 2x data efficiency gain, and the scaling analysis predicts the savings grow with model size. Mistral Small 4, released March 17, packs 119 billion parameters into a Mixture of Experts architecture with multimodal support, runs on vLLM and llama.cpp, and ships with actual documentation. NVIDIA's NemoClaw gives you single-command deployment for agentic workflows across local GPUs and cloud models with policy-based data routing and security guardrails baked in.
These are not research papers. These are things you can pull down and run today.
Guardrails Ship Faster When Everyone Can Read the Code
The strongest version of the opposing argument is that agentic AI systems, where models take autonomous actions, raise the stakes beyond what traditional open source governance handles. Fair point. An agent that can execute code and access databases is a different threat model than a static library with a known CVE. But NemoClaw exists precisely because NVIDIA built security-first guardrails into an open runtime. Restrictions on unauthorized data access and action execution are inspectable, auditable, and forkable. Compare that to a proprietary agent platform where the guardrails are a trust-me blog post from the vendor.
I have been running OpenDevin for side projects since January. The 78,000 monthly active users on that platform are not reckless hackers. They are developers building things, hitting real bugs, filing real issues, and pushing fixes upstream. AutoGPT's 81% task completion rate is not perfect. But it is honest, because you can trace every failure in the logs. Try getting that transparency from a closed API.
The 62% of organizations still stuck in pilot phase are not stuck because open source is too dangerous. They are stuck because they lack the engineering culture to adopt anything fast. Open source does not fix organizational paralysis. It does fix the information asymmetry that lets vendors charge you for a black box that benchmarks worse than what you can build yourself.
The vulnerabilities are real. Fix them with better dependency scanning, mandatory SBOMs, and funded security audits for high-adoption projects. Do not fix them by retreating to proprietary systems where the vulnerability count is zero only because nobody is allowed to look.