← Back to home
feb 1

openclaw

someone created a reddit for an AI agent called OpenClaw and they are actively plotting against humans. well, not really. but the architecture behind it? absolutely worth taking a deeper look at. i have already discussed a lot about agent architectures in my previous posts so here we will only look into things that are different/more optimal than normal but first, what really is openclaw? openclaw is an open-source multi-channel AI agent gateway. think of it as an agent like claude code that can be managed via telegram, discord, slack, whatsApp, imessage - basically any messaging platform you can think of. most agent systems i've seen have one fallback: "if model fails, show error to user." openclaw has three independent fallback layers: layer 1: auth profile rotation(an auth profile is basically a set of credentials of openai, anthropic with diff api keys) layer 2: thinking level fallback (per profile) high → medium → low off and resets when you switch profiles. why? because profile B might support what profile A didn't. layer 3: model fallback - claude-opus → gpt-4o → gemini-2.0-flash → default. only triggered after layer 1 and 2 are exhausted. but why is it even important? because most failures aren't "the AI broke." they're "this specific auth token hit rate limits" or "this model doesn't support extended thinking." and if you are creating a proactive personal assistant that's always on, you need to treat these errors differently. you need to separate billing failures from rate limit failures. it feels v obvious in retrospect that billing failures mean your account has a problem. hammering the API won't fix it. rate limits are temporary. billing issues are not. treat every known error differently. the fallbacks need to be different, so do the retries. now we all kinda know that context bloat is the biggest bottleneck in agent performance. openclaw also uses file system as a context manager but something is different. files aren't auto-sent to context. agents have access but must explicitly read what they need through tool calls. session history? files. tool results? files. the pattern: let agents gather context on-demand instead of frontloading everything. this kinda builds onto the fact that the difference between a demo agent and a production agent isn't the model. it's the harness. none of this is "AI." it's just good engineering around AI. equally as important if not more. well if you are a supply chain team, you don't need to worry about all this. at Lumari, we will read all these architectures and create state of the art custom built solutions for you, do reach out :)
loading comments...