
Suppose you’re a sports enthusiast or, better still, a scout looking for a fast and interactive way to obtain reliable information about players, games, schedules, etc. You need insights into player performance, longitudinal trends, and statistics. This information will help you build rosters or choose which teams to support as a fan, so it must be accurate, factual, and up-to-date. Would you trust an LLM chatbot with this task?
Providing a dependable, domain-specific chatbot experience with inherently probabilistic LLMs under the hood is a challenge faced by one of our clients, a sports analytics company. We have successfully helped them bring their product from inception to production, and now it’s consistently serving over 50,000 prompts a month while processing more than 6 billion tokens.
To ensure output accuracy, we also developed an AI-driven engine that powers the chatbot’s analytical capabilities.
Large language models excel at generating human-like text that looks plausible even when incorrect. An LLM-based service that provides data-driven analytics for high-stakes decision-makers like sports scouts must mitigate generative AI’s tendency to “hallucinate” with solutions that optimize for accuracy and reliability.
To that end, Janea Systems’ engineers set out to build a fast, secure, configurable, and trustworthy AI engine with deep analytical abilities to undergird our client’s chatbot.
Industry: Enterprise Software, AI/ML Big Data, SaaS, Cloud Computing
Technologies: Python, JavaScript
Solutions: AI/ML Big Data
Modern, state-of-the-art LLMs like Claude or Gemini boast impressive accuracy metrics, oscillating between 80 and 90%, depending on the benchmark. Nevertheless, even error rates as low as 10% can compund across multi-step, agentic workflows used in domain-specific chatbots such as our client’s.
All in all, our team encountered the following challenges in delivering the system to full specification:
Bringing the project to successful fruition required a concerted engineering effort, encompassing the following solutions:
Going forward, we will further personalize the user experience through features like extended chat memory, and expand the array of purpose-built agents to support a wider range of tasks.
Are you exploring a similar product? Let us share what we learned about analytics-first, LLM-powered systems that stay fast, factual, and secure at scale. Our team can help you design and deliver multi-agent workflows with robust evaluation, private-cloud implementation, and regulatory compliance.
Ready to discuss your software engineering needs with our team of experts?