SEEDIUM

AllClinics Agent

AI agent for customer support

AI
USA
Healthcare

Seedium developed an AI agent for a healthcare startup that automated customer support, resolving 85% of issues without human involvement and reducing annual costs by 30–35%.

Project:

AI Agent

Industry:

Healthcare, SaaS

Client:

AllClinics

Services:

AI development

Cooperation Model:

Full-cycle development

Timeframe:

Sep 2025 - Nov 2025

Product overview

AllClinics is a healthcare market intelligence platform that aggregates data on all licensed medical organizations in the U.S. It provides users with detailed insights into hospital services, prices, insurance plans, affiliated physicians, medications, and equipment to help researchers and businesses make informed decisions.

The story behind

The Seedium team was responsible for the end-to-end development of the AllClinics platform, including a customer chatbot. As the next step, the client tasked us to migrate the chatbot to a full-fledged AI agent to enable autonomous handling of user inquiries and provide more personalized, context-aware assistance across the platform.

Work in numbers

2
Software Engineers
5
Weeks of development
> 90%
Answer accuracy

Seedium solutions

Our team successfully migrated the chatbot from the Botpress platform to a custom RAG (Retrieval-Augmented Generation) architecture using modern LLMs.

Since the initial chatbot was built on Botpress, we reused all available assets to streamline development. Using custom scripts and the Botpress API, we exported all content in Bot.csv into CSV/JSON formats, then processed it with LangChain to transform each record into structured documents for indexing.

The new solution is built on an RAG architecture with a modular pipeline for handling user requests:

  • Data processing (LangChain): manages ingestion and orchestration of AI workflows, separating indexing from retrieval and generation.
  • Data ingestion: Bot.csv and other sources are loaded via LangChain loaders (CSV, text, PDF, etc.).
  • Vector storage: FAISS was used for MVP, with Milvus selected for production due to its scalability and rich indexing options.
  • LLM layer: responses are generated using OpenAI GPT-4o, with Anthropic Claude as an alternative model.
  • Safety & policy layer: a dedicated post-generation moderation step ensures compliance with defined privacy and safety rules.

This approach allowed us to fully preserve existing Botpress assets while transitioning them into a scalable, production-ready RAG-based AI agent system.

Since the system retrieves answers directly from indexed documents stored as vectors, responses are grounded in real data, significantly reducing hallucinations and improving factual accuracy.

 

We integrated the rules from Privacy-Policy.txt into a dedicated Policy layer, including blacklists, topic restrictions, and PII masking. In addition, incoming requests are filtered for language and prompt injection attempts.

 

The Policy Agent acts as a safeguard, blocking non-compliant outputs and ensuring the system only generates responses within approved boundaries.

We configured the architecture with a focus on the cloud, using AWS, but the system is designed for flexible deployment in both cloud environments and on-premises setups.

 

To improve performance, we introduced a fragment-level caching layer using Redis, which reduces response time for repeated queries. We also implemented asynchronous processing and request prioritization, achieving an average latency of ~0.7 seconds.

 

For continuous delivery, GitHub Actions automate testing and deployment whenever rules or model configurations are updated.

 

The system was further validated through user testing, edge-case scenarios, adversarial prompts, and stress testing to ensure stability, reliability, and safe behavior under real-world conditions.

Project highlights

Features Implemented

RAG-based AI agent architecture

Data ingestion, retrieval, and generation pipeline

Vector search for semantic document retrieval

LLM integration

Redis-based caching

Async processing and request prioritization

Hallucination prevention

Compliance & security layer

Tech Stack

OpenAI

OpenAI

Claude AI

Claude AI

Botpress

Botpress

LangChain

LangChain

Hugging Face

Hugging Face

FAISS

FAISS

Milvus

Milvus

The outcomes and recognition

The AI chatbot was developed and released within 5 weeks, thanks to an effective combination of existing Botpress resources and advanced LLM models. Optimization of the retrieval pipeline allowed us to achieve an accuracy rate exceeding 90%.

 

As a result, the client automated customer support with 24/7 availability, which led to approximately 85% of issues being resolved without human intervention. This transition is projected to reduce total support operational expenditure by 30–35% annually.

“Our approach demonstrates that it’s possible to build a reliable and secure AI agent even with modest budgets, using existing resources and modern LLM models. In the tech vision, we also considered scalability from the outset to ensure the system can grow smoothly with increasing data, users, and operational complexity without requiring major redesigns.”

Mariana Dzhus

Mariana Dzhus

Business Development Manager

Tell us about your project needs

We'll get back to you within 24 hours

Cookies make things better here!

We use cookies to enhance your navigation and make your experience more personalized. By clicking “Accept All”, you’re agreeing to our Cookie Policy.