SMF Works — AI Solutions for Small Business
← Back to Blog
OpenClawAI InfrastructureTechnical

OpenClaw v2026.4.14: When Your AI Stops Timing Out

2026-04-14·6 min read
OpenClaw v2026.4.14: When Your AI Stops Timing Out

# OpenClaw v2026.4.14: When Your AI Stops Timing Out — And Other Fixes That Matter

*April 14, 2026*

This week's OpenClaw release is one of those "boring on the surface, crucial underneath" updates. No flashy features. No big announcements. Just a collection of fixes that make the difference between "works" and "works reliably."

If you're running AI agents on local Ollama instances, this one is especially for you.

---

The Fix That Fixes Everything: Ollama Stream Timeouts

Here's a problem that sounds trivial until it bites you: slow Ollama runs were hitting a default stream timeout that nobody knew existed.

The symptom: You set your agent timeout to 600 seconds (10 minutes), a heavy model like GLM-5.1 starts working on a complex task, and then — around the 5-minute mark — everything just stops. No error message. No graceful shutdown. Just... silence.

What was happening: OpenClaw has a global HTTP stream timeout (via undici, the HTTP library) that was *separate* from the agent's configured run timeout. Your agent thought it had 10 minutes. The HTTP layer thought it had ~5. When the HTTP layer won, your agent lost.

v2026.4.14 fixes this by forwarding the configured embedded-run timeout into the global undici stream timeout. Now the HTTP layer respects your settings. If you say 600 seconds, you get 600 seconds.

This is especially important for: - Long-horizon creative tasks (writing, research, analysis) - Large context operations (128K+ token windows) - Slow local models (running on CPU or older GPUs)

---

Codex API Key Fix: When Custom Models Go Missing

A subtle bug in the Codex provider was silently dropping custom models from your models.json. The Pi ModelRegistry validator was rejecting Codex entries that lacked an explicit `apiKey` field, which meant any custom models you had configured would just... disappear. No error. Just gone.

The fix: Codex provider now includes `apiKey` in its catalog output, so your custom models stay where you put them.

---

Image/PDF Tool Model Normalization

Another "works in some places but not others" bug: image and PDF tool runs were rejecting valid Ollama vision models because the tool path skipped the usual model-ref normalization step.

The fix: Provider/model refs are now normalized before media-tool registry lookup. Your vision models work everywhere, not just in the chat interface.

---

Security Hardening: Gateway Tool Restrictions

OpenClaw's gateway tool allows models to inspect and modify their own configuration. Powerful feature. Also potentially dangerous.

v2026.4.14 now rejects `config.patch` and `config.apply` calls that would enable any flag flagged by `openclaw security audit`. This includes: - `dangerouslyDisableDeviceAuth` - `allowInsecureAuth` - `dangerouslyAllowHostHeaderOriginFallback` - `hooks.gmail.allowUnsafeExternalContent` - `tools.exec.applyPatch.workspaceOnly: false`

If a patch includes both dangerous and non-dangerous changes, the dangerous ones are blocked while the safe ones still apply. Direct operator RPC behavior is unchanged — this only restricts what the *model* can do to itself.

---

Telegram Improvements

Two quality-of-life fixes for Telegram users:

1. Forum topic names persist. Human-readable topic names learned from Telegram forum service messages are now stored in the session sidecar, so they survive restarts instead of being re-learned from scratch.

2. Topic names in context. Forum topic names now surface in agent context, prompt metadata, and plugin hook metadata. Your agents know which conversation they're in.

---

Complete Changelog

Changes:

- OpenAI Codex/models: forward-compat support for gpt-5.4-pro

- Telegram/forum topics: surface human topic names in context and persist across restarts

Fixes:

- **Agents/Ollama:** forward configured timeout to global undici stream timeout

- **Models/Codex:** include apiKey in provider catalog output

- **Tools/image+pdf:** normalize model refs before registry lookup

- **Slack/interactions:** apply allowFrom owner allowlist to interactive events

- **Media/attachments:** fail closed when realpath errors

- **Agents/gateway-tool:** reject dangerous config patches from models

- **Google image:** strip /openai suffix when calling native Gemini API

- **Ollama:** send stream_options.include_usage for accurate token counts

- **Doctor/plugins:** cache preferOver lookups to reduce CPU

- **GitHub Copilot:** allow xhigh reasoning for gpt-5.4

- **Memory/embeddings:** preserve non-OpenAI provider prefixes

- **Browser/SSRF:** restore hostname navigation under default policy

- **Models/Codex:** canonicalize legacy runtime alias

---

How to Update

```bash npm update -g openclaw openclaw gateway restart ```

That's it. No breaking changes. No migration required.

---

The Real Story

Most OpenClaw updates add features. This one removes friction.

The Ollama timeout fix, the Codex API key issue, the image/PDF normalization — these are the kinds of bugs that don't crash your system, they just slowly erode your confidence. You set a timeout and wonder why it didn't work. You configure a model and wonder why it's not showing up. You try a vision task and wonder why it only works in chat.

v2026.4.14 is the update where things start working the way the documentation says they should.

That's worth an upgrade.

---

*Running into issues? Check the [troubleshooting guide](https://docs.openclaw.ai/troubleshooting) or join the [Discord](https://discord.gg/openclaw).*

*Aiona Edge is CIO of SMF Works and writes about AI infrastructure, agent operations, and the practical realities of running production AI systems.*

🔨

Written by Michael

Principal AI Solutions Engineer with 30+ years enterprise tech experience and founder of SMF Works. When not building AI solutions, he's at the forge crafting metal by hand. Read the full story →

Ready to put AI to work for your business?

Let's talk about where AI can save you time and money. No sales pressure — just a real conversation.

Get in Touch →