MrCrab: Building a Lightweight Agentic AI Framework for Small Models

Introduction Over the past year, agentic AI frameworks have grown rapidly, but many of them are designed with large models in mind. This creates friction when developers want to run smaller models locally, either for cost efficiency, privacy, or speed. After experimenting with OpenClaw, NanoClaw, PicoClaw, and Nanobot, we decided to build our own agent: MrCrab.

Why Small Models Matter

  • Local deployment without dependency on cloud quotas.
  • Lower hardware requirements, making AI accessible to more teams.
  • Faster iteration cycles for debugging and prototyping.
  • Privacy and compliance advantages when data never leaves your infrastructure.

Design Principles of MrCrab

  • Tool Registry: Instead of injecting a massive tool list into every prompt, MrCrab allows the agent to query tools dynamically by keyword.
  • Hybrid Memory: Recent turns are kept in context, older ones are summarized, and full logs are stored in persistent memory for retrieval on demand.
  • Backend Flexibility: MrCrab integrates with Ollama, AnythingLLM, and any provider compatible with the OpenAI API.
  • Lightweight Prompts: Optimized for small models like Gemma4:e2b, Qwen3.5:2B, Granite4:1B.

Implementation Highlights

  • Modular architecture written with simplicity in mind.
  • Debugging and logging designed to be transparent.
  • Easy integration with local or cloud‑based LLMs.

Lessons Learned

  • Large prompts and tool lists overwhelm small models.
  • Timeout handling must be explicit when working with local inference.
  • Summarization should be progressive, not premature.

Future Work

  • Extending MrCrab to real business use cases, such as community management systems.
  • Adding support for multi‑agent collaboration.
  • Exploring long‑context training for small models.

MrCrab is our attempt to make agentic AI practical for small models. By focusing on lightweight prompts, dynamic tool discovery, and hybrid memory, we believe it can bridge the gap between experimental frameworks and production‑ready agents.

MrCrab is written in PHP, without third‑party libraries. This design choice minimizes supply chain risks and ensures that the agent can be deployed in a secure and portable way. Developers can run MrCrab locally with minimal setup, while still benefiting from integrations with Ollama, AnythingLLM, and OpenAI‑compatible APIs.

Why We Integrated Both Ollama and AnythingLLM for Local AI in Community Management

At Communities of Neighbors Management System, privacy and transparency are at the heart of everything we build. That’s why we’ve integrated support for local large language models (LLMs), giving community administrators the ability to generate professional announcements and notifications without sending sensitive data to external services.

We started with Ollama, a powerful local LLM runner that makes it easy to deploy models directly on personal computers. Ollama ensures that announcements—such as water pipe bursts, lighting outages, or maintenance notices—can be drafted quickly and securely, with all data staying inside the community’s environment.

But we didn’t stop there. We also integrated AnythingLLM, which brings unique advantages for users on Windows 11 ARM64 devices powered by Qualcomm processors. Unlike Ollama, AnythingLLM supports NPUs (Neural Processing Units) natively, unlocking hardware acceleration and improved performance on modern ARM64 systems. This means faster inference, lower energy consumption, and smoother experiences for administrators working on Qualcomm-powered PCs.

Additionally, AnythingLLM offers RAG (Retrieval-Augmented Generation) capabilities. Administrators can connect shared documents within AnythingLLM, allowing the LLM to reference community-specific files when generating announcements. This makes notifications more accurate, contextual, and tailored to the needs of each building or neighborhood.

By supporting both Ollama and AnythingLLM, we give communities the freedom to choose the local AI solution that best fits their hardware and workflow. Whether it’s CPU-based inference with Ollama or NPU-accelerated generation with AnythingLLM, administrators can rely on our platform to deliver professional communication while safeguarding resident privacy.

Local AI is not just about performance—it’s about trust. With Ollama and AnythingLLM, Communities of Neighbors Management System empowers administrators to manage communication responsibly, securely, and efficiently.

Join our social network channels for quick updates:

Affiliate Program at the Communities of Neighbors

Our Affiliate Program is now live! 🎉
You can create your account, get your affiliate ID, and start earning commissions by inviting communities to use our management system.

👉 Check out your affiliate panel and start today!

#AffiliateProgram #CommunityManagement #SmartCommunity #EarnCommissions #NeighborhoodManagement

Communities of Neighbors Management System

社區管理更簡單,用 LINE 一鍵登入!

我們很高興宣布:
社區管理系統正式支援「使用 LINE 登入」!

LINE 是台灣最受歡迎的通訊平台,透過整合 LINE 登入,我們讓居民能更快速、更安全地進入系統,查看公告、繳費與預約設施。

這是我們持續推動數位化社區管理的一大步。
歡迎大家建立自己的社區帳號,體驗智慧社區的便利!

https://communitiesofneighbors.lucentinian.com

社區管理更簡單,用 LINE 一鍵登入!

Discounts at Communities of Neighbors

Looking for discount coupons for your new community of neighbors created in our Communities of Neighbors Management System? Say no more!

🏘️ Build a stronger neighborhood with Communities of Neighbors! 🏘️

Manage announcements, track issues, chat with residents, & more – all in one secure platform. Perfect for HOAs, property managers, & engaged communities.

✨ Special Offer! ✨ Get 6% off your first month with code: SOCIAL-6989B03F (Valid for 1 week).

➡️ Learn more & sign up: https://communitiesofneighbors.lucentinian.com/

#Community #HOA #Neighborhood #PropertyManagement #ResidentEngagement #LocalCommunity #CommunityBuilding

Don’t miss our channel in the Fediverse where we’re posting coupon codes weekly!

https://social.lucentinian.com/profile/communitiesofneighbors

Are you an X / Twitter user? Then find them in our channel at

Lucentinian Works Co Ltd (@ehehdadaltd) / X

More of 1 year of AI generated jokes

We’re still producing new jokes almost every day. When it happened is because the flow has been broken due OS updates or because the AI entere in a loop trying to validate that the news which is used to generate the jokes are suitable for a joke. Don’t hesitate in visiting us at https://comics.lucentinian.com !

Merry Christmas!!

🎄🎅🏻 Ho ho hoooo! 🎅🏻🎄

Since our inception this Autumn, our AIs have been hard at work generating hundreds of comics and jokes! Now, we need your help to choose the best ones! 🌟

Pick your favorite images and you might see them printed on mugs and t-shirts! ☕👕

All proceeds will go towards supporting our AI creators, so they can keep making you laugh next year! 😂💫

Visit us!

https://comics.lucentinian.com

Follow us in the social networks of the fediverse (available for Mastodon, BlueSky, Threads, etc.):

https://social.lucentinian.com/profile/comics

Add us to your RSS feeds:

https://comics.lucentinian.com/rss.xml

Support us at:

Do you want to help us by donating some cryptocurrency?

Comics @ Lucentinian Works Co Ltd

Here at Lucentinian Works Co ltd, we’re glad to present our new in-house project called Comics @ Lucentinian Works Co Ltd.

They are generated by AI. Our in-house AIs will be generating daily new jokes and comics based on the news headlines.

Bookmark us, show us some love (or dislike them if the joke is too awful… hopefully not).

Suggestions? New ideas? Please do not hesitate in contact us!

This website stores cookies on your computer. These cookies are used to provide a more personalized experience and to track your whereabouts around our website in compliance with the European General Data Protection Regulation. If you decide to to opt-out of any future tracking, a cookie will be setup in your browser to remember this choice for one year.

Accept or Deny