Gen AI Fine Tuning

The Future of GenAI Fine-Tuning

Fine-tuning large language models (LLMs) for specific tasks has become a critical challenge.

Crossing Minds' RAG-Sys offers a groundbreaking approach that transforms how businesses leverage and customize GenAI technologies.

Your ultimate LLM fine tuning tool.

Unparalleled Adaptability

Instantly switch between diverse domains without retraining.

RAGSys allows effortless adaptation to legal, ecommerce, or technical fields by simply changing the knowledge base.

Maximize versatility while saving time and resources.

Real-Time Knowledge Integration

Stay cutting-edge with immediate knowledge updates.

RAGSys integrates new information in real-time, ensuring your LLM always leverages the latest data.

Deliver relevant, up-to-date responses in fast-paced environments.

Precision and Transparency Redefined

Achieve new levels of accuracy and accountability. RAGSys grounds responses in factual information, dramatically reducing hallucinations.

With clear source attribution, it offers unmatched transparency crucial for high-stakes industries.

Unmatched Efficiency and Scalability

Transform the economics of LLM deployment. RAGSys focuses on efficient retrieval rather than constant retraining, significantly reducing costs and environmental impact.

Scale seamlessly to handle growing data volumes without proportional resource increases.

Dynamic Knowledge Integration Without Retraining

Traditional fine-tuning requires extensive retraining to incorporate new knowledge, a process that's both time-consuming and computationally expensive. RAGSys changes the game by allowing real-time knowledge integration:

  • Instant Updates: New information can be added to the knowledge base and immediately utilized by the LLM, ensuring up-to-date responses without model retraining.

  • Flexible Knowledge Management: Easily add, remove, or modify domain-specific information without touching the underlying LLM architecture.

  • Reduced Computational Overhead: Eliminate the need for frequent large-scale model retraining, significantly reducing computational costs and environmental impact.

Precision and Consistency in Outputs

LLMs are prone to hallucinations and inconsistencies, especially when dealing with specialized or rapidly changing information. RAGSys addresses this head-on:

  • Fact-Grounded Responses: By retrieving and incorporating relevant information, RAGSys ensures LLM outputs are anchored in factual data.Consistency
  • Across Queries: The retrieval mechanism helps maintain consistent responses to similar queries, enhancing reliability in critical applications.
  • Transparent Source Attribution: RAGSys can provide references to the sources of information used in generating responses, adding a layer of explainability often missing in traditional LLM outputs.

Adaptive Domain Expertise

While general-purpose LLMs struggle with specialized domains, RAGSys enables rapid adaptation to specific industries or use cases:


Advanced capabilities:

  • Effortless Domain Switching: Swap out knowledge bases to instantly repurpose the same LLM for different domains, from legal to medical to technical support.

  • Granular Expertise Layers: Layer multiple knowledge bases to create nuanced, multi-disciplinary expertise tailored to specific organizational needs.

  • Continuous Learning: The system can learn from interactions and feedback, constantly refining its domain expertise without the need for model-wide updates.

Scalability and Efficiency Reimagined

RAGSys transforms the scalability and efficiency landscape of LLM deployments:
Resource Optimization:

  • Focus computational resources on retrieval and integration rather than massive model retraining, allowing for more efficient scaling.

  • Distributed Knowledge Architecture: Leverage distributed knowledge bases, enabling organizations to manage and update information across various departments or geographical locations seamlessly.

  • Adaptive Performance: The system can dynamically allocate resources based on query complexity and retrieval needs, ensuring optimal performance under varying loads.

ICLEB

Organaization

Model

ERR@10

nDCG@10

Crossing Minds

cm-ragsys-rlaif-mini-v1

0.860

0.701

Salesforce

SFR-Embedding-2_R

0.775

0.610

Cohere

rerank-english-v3.0

0.773

0.618

Snowflake

snowflake-artic-embed-m-v1.5

0.751

0.596

OpenAI

text-embedding-3-small

0.751

0.606

Nvidia

NV-Embed-v2

0.741

0.612

Model Distillation in FinTech

Context: FinTech Data

  • One of our client is a large FinTech Enterprise
  • Billions of credit card transactions
  • Unstructured use of abbreviations

Example

New York Home Hardware Distributors

NY HOME HARDWARE D

NEW YORK HHDW DISTR

Task: Entity Deduplication with LLM

  • Extract and clean the Merchant Name using a Strong LLM (Claude 3.5 Sonnet)
  • Distill the Strong Teacher LLM into a Cheap Student LLM while preserving accuracy (Claude 3 Haiku)
  • Generate 4k training data from Claude 3.5 Sonnet (Teacher)
  • Fine-tune the student using this 4k dataset

Method

Accuracy

Total Time (min)

Static Few Shots

0.65

5

Fine Tuning

0.88

660

RAGSys

0.91

6

eCommerce Product Catalog Enrichment

Context

  • Leading B2C Marketplace
  • Millions of end-users are creating hundreds of millions of products
  • Extract products tags to boost search and recommendation

Example

  • Use an LLM to extract Product Tags based on the product details
  • Leverage a Manually Curated Set of Tags for Train and Validation

LLM Tags completion

Stone Nail File, Nail Art, Manicure
Amazing Stone Nail File
"best nail file i have ever used".
its never wears outs, and the tapered chiseled end is
wonderful
Measures 4" long x 1/4" wide
you get 1
Pink or Green

Hidden Tags:

stone nail file, nail art, nail polish, nail tools, manicure, manipedi, pedicure, gift, nail health, nail file

Generated Tags:

stone nail file, nail art, manicure, nail tools, nail care, nail file, pink nail file, green nail file, nail grooming, nail accessories, nail health, pedicure

Method

Precision

Recall

Zero Shot

0.1196

0.1326

Static Few Shots

0.1295

0.1415

RAGSys

0.2286

0.2559

LLM Tags Completion: Results.  Average over 1k items in the test set. LLM: gpt-4o

Live Tuning for Entity Deduplication

Context: B2B Marketplace Catalog

  • One of our client is a leading B2B marketplace in their industry
  • 5000+ merchants, 1M+ items
  • All merchants are creating items manually, creating many duplicates
  • They need to consolidate the items catalog

Task: Entity Deduplication with LLM

Given two items titles, are they the same product?

Example 1

Product A:

Google Chromecast Ultra 4K Streaming Media Player

Product B:

Chromecast Ultra

Answer:

Yes

Example 2

Product A:

Dell Alienware M15

Product B:

Alienware M17

Answer:

No

Get an overview of Crossing Minds and its features.
Find out how to take personalized experiences to the next level.
A/B test and customize the smartest recommendations for your unique scenario.
CB Insights Awards Retail Tech 100 in 2022CB Insights Top AI 100 companies in 2022Martech Breakthrough Awards 2022
trusted by brands like

Request a demo

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.