FinanceAI

How FinanceAI uses Kurai to power their newest projects.

Industry

FinTech & SaaS

Location

Jakarta, Indonesia

Employees

150

Identity Provider

FinanceAI

Workloads

RAG System, LLM Integration, Pinecone Vector Database, FastAPI

About

FinanceAI is a B2B fintech startup providing automated accounting and financial planning software for SMBs. Their platform handles millions of transactions monthly and serves 50,000+ businesses across North America and Europe.

Challenge

FinanceAI's customer support team was overwhelmed with 15,000+ tickets per month. Common questions about invoice processing, reconciliation, and tax regulations took 15+ minutes per ticket, leading to 48-hour response times and 65% customer dissatisfaction.

Solution

Kurai built a RAG-powered support assistant using GPT-5 and Pinecone. We indexed 50,000+ documentation pages, transaction histories, and regulatory documents. The assistant handles 80% of inquiries automatically with 94% accuracy, escalating only complex cases to humans.

Results

60% reduction in support tickets handled by humans...

Response time improved from 48 hours to 2 minutes...

Customer satisfaction increased from 35% to 89%...

$450K annual savings in support costs...

RAG-Powered Support Assistant: Transforming Customer Experience

The Problem

FinanceAI was scaling fast—but so were their support costs. With 15,000+ monthly tickets and a team of 20 support agents, they were drowning in repetitive questions. “How do I reconcile this invoice?” “What’s the tax deduction for this expense?” “Why did my transaction fail?”

CEO Jennifer Chen said: “We were adding 5 new customers for every new support hire. Our growth was unsustainable.”

The Solution

Kurai deployed a production RAG system in 6 weeks:

Architecture:

  • Embeddings: OpenAI text-embedding-3-small (512 dimensions)
  • Vector Database: Pinecone (5M vectors, 95% recall)
  • LLM: GPT-5 Turbo with function calling
  • Backend: FastAPI with WebSocket support
  • Frontend: React widget embedded in help center

Data Pipeline:

  1. Ingestion: Scraped documentation, knowledge base, and regulatory docs
  2. Chunking: Semantic chunking preserving context (avg. 500 tokens)
  3. Embedding: Batch processing with OpenAI API ($0.0001/1K tokens)
  4. Indexing: Upserted to Pinecone with metadata filters (category, date, jurisdiction)

Query Pipeline:

  1. User question → Embed query
  2. Similarity search (k=5) with metadata filter
  3. Context + question → GPT-5 Turbo
  4. Answer with citations → User

Guardrails:

  • Fallback to human if confidence < 80%
  • PII redaction before processing
  • Audit logging for compliance
  • Daily accuracy monitoring

The Results

Immediate Impact (Month 1):

  • 9,000 tickets automated (60%)
  • Avg. response time: 48 hours → 2 minutes
  • CSAT score: 35% → 89%

Cost Savings (Annual):

  • Support headcount: -$350K (3 fewer agents needed)
  • LLM API costs: +$45K
  • Net savings: $450K/year

Scalability:

  • System handles 100K queries/month
  • P95 latency: 850ms
  • 99.94% uptime (4.3 hours downtime/year)

Customer Feedback

“Instead of waiting 2 days for an answer, I get it in seconds with links to the exact regulations. It’s like having a CPA on call 24/7.” — Michael Torres, CFO at TechScale Inc.

What’s Next

Phase 2 roadmap includes:

  • Multi-language support (Spanish, French, German)
  • Proactive notifications for common issues
  • Integration with accounting software (QuickBooks, Xero)
  • Voice-based support using Whisper API

Technology Stack

  • LLM: OpenAI GPT-5 Turbo
  • Embeddings: OpenAI text-embedding-3-small
  • Vector DB: Pinecone (Production tier)
  • Backend: Python FastAPI + Celery
  • Frontend: React + TypeScript
  • Monitoring: Prometheus + Grafana
  • Logging: ELK Stack
  • Infrastructure: AWS (ECS, Fargate, RDS)

Key Metrics

MetricBeforeAfterImprovement
Monthly tickets15,0006,00060% reduction
Response time48 hours2 minutes99.3% faster
CSAT score35%89%+154%
Support cost/ticket$15$380% reduction
Resolution accuracyN/A94%New capability

Timeline

  • Week 1-2: Data ingestion and preprocessing
  • Week 3-4: RAG pipeline development and testing
  • Week 5: Beta testing with 500 users
  • Week 6: Full deployment to all 50K users

Lessons Learned

  1. Data quality matters: We spent 40% of effort cleaning documentation
  2. Citations build trust: Users were 3x more satisfied when sources were shown
  3. Human-in-the-loop essential: 20% escalation prevents bad experiences
  4. Monitor continuously: Accuracy dropped from 96% → 94% after tax law updates; auto-retraining fixed it

FinanceAI is now scaling to 100K customers without adding headcount. Their support team focuses on complex, high-value inquiries while the RAG assistant handles the rest.

Trusted by Industry Leaders

Empowering innovators, shaping the future

David Gutierrez

David Gutierrez

CTO at TechFlow AI

"The RAG system they built for us reduced our support tickets by 60%. Their expertise in LLM integration is unmatched."

Pierluigi Camomillo

Pierluigi Camomillo

VP Engineering at DataScale

"They migrated our monolith to microservices seamlessly. We saw a 40% cost reduction and significantly improved scalability."

Ella Svensson

Ella Svensson

Founder at MediSort Health

"Their ML-powered patient triage system transformed our operations. 70% faster triage with 94% accuracy—sim incredible results."

Alexa Rios

Alexa Rios

Chief Product Officer at ShopMax

"The recommendation engine they built increased our AOV by 32%. Highly recommended for any e-commerce business looking to leverage AI!"