Welcome to our engineering blog – a collection of technical deep dives into the challenges we’re solving while building orra.
Here we share the architectural decisions, implementation details, and hard-won lessons from building a production-grade orchestration system for LLM applications. From semantic caching to execution planning, from vector operations to state management, we document our journey through the technical complexities that arise when building reliable AI systems.
No fluff, just engineering insights for builders who appreciate the details.