RAG-Sys:
Revolutionizing Enterprise LLM Deployment

RagSys, Crossing Minds' innovative Few-Shot In-Context-Learning Engine, supercharges LLM performance without the pain of traditional fine-tuning.

Discover a cost-effective, scalable solution to enhance AI language models for any tasks.

Unleash the Full Potential of Large Language Models

Dramatic Reduction in Prompt Engineering

Our platform, built on a cutting-edge stack, delivers real-time, scalable performance.

Say goodbye to troubleshooting fine tuning pipelines.

Seamless Integration with Any LLM

Deploy RAG-Sys with any LLM, from Anthropic and OpenAI to open-source alternatives.

Seamlessly switch between models without losing optimizations, future-proofing your AI stack.

Enhanced Retrieval with RAG Embeddings

Our proprietary RAG embeddings ensure better understanding and rapid information retrieval, even with massive datasets.

This technology drives more contextually relevant and accurate LLM outputs.

Groundbreaking Accuracy Lift

RAG-Sys consistently outperforms traditional fine-tuning, with up to 55.8% improvement on key benchmarks like HellaSwag.

Significant enhancements in truthfulness, emotion detection, and commonsense reasoning across various LLMs.

Designed for Enterprise Scale and Flexibility

Size
Pattern
Color
Fabric Type
Style

Scalable Architecture & Rules

RAG-Sys is designed for enterprise-scale deployment, efficiently handling large datasets and complex retrieval tasks.

Our infrastructure scales seamlessly from proof-of-concept to full production, ensuring consistent performance as your AI needs grow.

Customizable Dataset Creation

Our intuitive dashboard enables rapid development of domain-specific knowledge bases.

Easily create and iterate on custom datasets, tailoring RAG-Sys to your unique business requirements without extensive data engineering.

Revolutionizing Task-Specific Performance

RAG-Sys achieves superior task-specific performance without resource-intensive fine-tuning.

Rapidly adapt LLMs to new tasks or domains, saving computational resources and accelerating deployment cycles.

Key Innovations

Adaptive Knowledge Repository

At the heart of RAG-Sys lies a dynamic, self-improving knowledge base:

  • Custom Example Database: Create and maintain a tailored database of examples specific to your use cases and domain expertise. This allows your ML team to build a proprietary knowledge base that continuously enhances your LLM's performance in your unique business context.
  • Model-Agnostic Design: Our adaptive few-shot database is engineered to be compatible across various LLM architectures. This flexibility allows you to switch between different LLM providers or versions without losing your accumulated knowledge and optimizations.

  • Continuous Learning: The system features an automated feedback loop that refines and expands its knowledge base in real-time, ensuring your AI capabilities evolve alongside your business needs.
RAG-Sys features a self-improving knowledge base with a custom example database, model-agnostic design, and continuous learning, enhancing LLM performance and adapting to unique business needs.

Advanced RAG

RAG-Sys transcends traditional RAG limitations:


  • Entropy-Maximizing Selection: Our proprietary algorithms ensure LLMs receive a diverse, information-rich input, improving response quality and reducing redundancy.

  • Quality-Weighted Retrieval: Multi-factor scoring system prioritizes high-quality, relevant information, significantly reducing hallucinations and improving factual accuracy.

  • Domain-Specific Customization: Flexible rule engine allows seamless integration of business logic and regulatory requirements into the retrieval process.
RAG-Sys overcomes traditional RAG limitations with entropy-maximizing selection, quality-weighted retrieval, and domain-specific customization, enhancing input quality, reducing redundancy, and integrating business logic.

Efficient Few-Shot Learning

Redefining few-shot learning for enterprise LLM deployment:


  • Optimal Example Selection: Leveraging advanced information theory, RAG-Sys identifies the most informative examples for in-context learning, dramatically improving task performance.

  • Accelerated Fine-Tuning: By optimizing the retrieval model instead of the entire LLM, RAG-Sys achieves fine-tuning speeds up to 1000x faster than traditional methods.

  • Transfer Learning Across Models: Retrieval engines trained on one LLM can be efficiently transferred to another, allowing you to leverage your optimizations across different models and providers.
RAG-Sys redefines few-shot learning for enterprise LLMs with optimal example selection, accelerated fine-tuning, and transfer learning across models, enhancing performance and efficiency.

ICLEB

Organaization

Model

ERR@10

nDCG@10

Crossing Minds

cm-ragsys-rlaif-mini-v1

0.860

0.701

Salesforce

SFR-Embedding-2_R

0.775

0.610

Cohere

rerank-english-v3.0

0.773

0.618

Snowflake

snowflake-artic-embed-m-v1.5

0.751

0.596

OpenAI

text-embedding-3-small

0.751

0.606

Nvidia

NV-Embed-v2

0.741

0.612

Model Distillation in FinTech

Context: FinTech Data

  • One of our client is a large FinTech Enterprise
  • Billions of credit card transactions
  • Unstructured use of abbreviations

Example

New York Home Hardware Distributors

NY HOME HARDWARE D

NEW YORK HHDW DISTR

Task: Entity Deduplication with LLM

  • Extract and clean the Merchant Name using a Strong LLM (Claude 3.5 Sonnet)
  • Distill the Strong Teacher LLM into a Cheap Student LLM while preserving accuracy (Claude 3 Haiku)
  • Generate 4k training data from Claude 3.5 Sonnet (Teacher)
  • Fine-tune the student using this 4k dataset

Method

Accuracy

Total Time (min)

Static Few Shots

0.65

5

Fine Tuning

0.88

660

RAGSys

0.91

6

eCommerce Product Catalog Enrichment

Context

  • Leading B2C Marketplace
  • Millions of end-users are creating hundreds of millions of products
  • Extract products tags to boost search and recommendation

Example

  • Use an LLM to extract Product Tags based on the product details
  • Leverage a Manually Curated Set of Tags for Train and Validation

LLM Tags completion

Stone Nail File, Nail Art, Manicure
Amazing Stone Nail File
"best nail file i have ever used".
its never wears outs, and the tapered chiseled end is
wonderful
Measures 4" long x 1/4" wide
you get 1
Pink or Green

Hidden Tags:

stone nail file, nail art, nail polish, nail tools, manicure, manipedi, pedicure, gift, nail health, nail file

Generated Tags:

stone nail file, nail art, manicure, nail tools, nail care, nail file, pink nail file, green nail file, nail grooming, nail accessories, nail health, pedicure

Method

Precision

Recall

Zero Shot

0.1196

0.1326

Static Few Shots

0.1295

0.1415

RAGSys

0.2286

0.2559

LLM Tags Completion: Results.  Average over 1k items in the test set. LLM: gpt-4o

Live Tuning for Entity Deduplication

Context: B2B Marketplace Catalog

  • One of our client is a leading B2B marketplace in their industry
  • 5000+ merchants, 1M+ items
  • All merchants are creating items manually, creating many duplicates
  • They need to consolidate the items catalog

Task: Entity Deduplication with LLM

Given two items titles, are they the same product?

Example 1

Product A:

Google Chromecast Ultra 4K Streaming Media Player

Product B:

Chromecast Ultra

Answer:

Yes

Example 2

Product A:

Dell Alienware M15

Product B:

Alienware M17

Answer:

No

Get an overview of Crossing Minds and its features.
Find out how to take personalized experiences to the next level.
A/B test and customize the smartest recommendations for your unique scenario.
CB Insights Awards Retail Tech 100 in 2022CB Insights Top AI 100 companies in 2022Martech Breakthrough Awards 2022
trusted by brands like

Request a demo

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.