Semantic SQL · Any database · Your SQL client

Ask your data
what it means.

MEANS, CLASSIFY, TOPICS, SUMMARIZE — 100+ AI operators that work like native SQL. Connect any database. Query from DBeaver, DataGrip, or Tableau. No Python. No notebooks. No new tools.

PostgreSQLMySQLSnowflakeBigQueryClickHouseS3
query.sql
SELECT name, region,
       SENTIMENT(feedback) AS mood
FROM   customer_reviews
WHERE  feedback IMPLIES
         'might cancel soon'
ORDER  BY mood;
nameregionmood
Sarah K.Northeast-0.82
James R.West-0.71
2 rows·41ms·cached

Down the rabbit hole

Start with a single semantic filter. By end of week, you're running verified AI aggregations across your entire warehouse.

Day 1Semantic filtering
filter.sql
SELECT *
FROM   support_tickets
WHERE  body MEANS
  'billing dispute';
→ 23 rows matching semantic intent
Day 3AI creates the dimensions
topics.sql
SELECT
  TOPICS(tweet, 4) AS topic,
  CLASSIFY(tweet,
    'political, not-political')
    AS political,
  COUNT(*) AS tweets,
  AVG(likes) AS avg_likes
FROM   twitter_archive
GROUP BY topic, political;
→ You didn't define the topics. The data organized itself.
Week 1Your own operators, verified
operator.sql
CREATE SEMANTIC OPERATOR
  compliance_risk(clause VARCHAR)
  RETURNS VARCHAR
  PROMPT 'Evaluate this contract clause
    for regulatory risk. Return: low,
    medium, or high with one sentence
    explaining why.
    Clause: {{ input.clause }}';

SELECT compliance_risk(clause):takes(3,
         'Pick the most conservative')
       AS risk_assessment
FROM   contracts;
→ Custom operator + ensemble verification. Built in SQL, runs in SQL.
wonderland

Your data never moves

DataRabbit speaks the PostgreSQL wire protocol. Connect from psql, DBeaver, DataGrip, Tableau, or anything that talks to Postgres. Queries run where your data lives.

01

Bring your databases

PostgreSQL, MySQL, Snowflake, BigQuery, ClickHouse, S3, Parquet. Data never leaves your infrastructure.

02

Write semantic SQL

Use your existing SQL client. Add operators like MEANS, CLASSIFY, or SUMMARIZE alongside standard SQL.

03

LLM once, SQL forever

The system fingerprints data shapes — not individual values. A million phone numbers might have 10 formats. 10 LLM calls generate SQL expressions. The expressions run on every future row. No LLM needed.

100+ built-in operators

Filtering, classification, summarization, parsing, validation, embeddings — all callable from SQL.

Intent & meaning

MEANSIMPLIESCONTRADICTSSIMILAR_TO

Classification

CLASSIFYSENTIMENTINTENTTOXICITY

Dimensional

TOPICSTHEMESCLUSTERCONSENSUSSUMMARIZE

Data quality

VALIDATEDEDUPECORRECTFILL

Parsing

PARSESMART_JSONPARSE_ADDRESSPARSE_NAME

Embeddings

EMBEDVECTOR_SEARCHEMBED_COLUMN

Build your own operators

Define custom AI functions in SQL DDL — no Python, no deploy pipeline. CREATE it, then SELECT with it.

operator.sql
CREATE SEMANTIC OPERATOR
  compliance_check(text VARCHAR)
  RETURNS VARCHAR
  PROMPT 'Check if this text violates
    our compliance policy.
    Return: pass, warn, or fail.
    Text: {{ input.text }}';

Then use it immediately:

SELECT clause, compliance_check(clause)
FROM   contracts
WHERE  compliance_check(clause)
       MEANS 'fail';

Built for production, not demos

LLM outputs are non-deterministic. DataRabbit has first-class primitives to make them reliable, auditable, and cost-efficient.

Takes

Run N model variations in parallel. An evaluator picks the best. No serial retry loops.

Shape compiler

Fingerprints data structures, generates SQL expressions, caches the code not the values. A million rows, ~10 LLM calls. Cost drops toward zero over time.

Query memory

Every query is embedded and summarized. Search past analyses by meaning. 'Did anyone look at churn by region?' surfaces the answer — and the SQL.

Few-shot training

Flag any operator result as 'correct.' Future calls automatically include your validated examples. No ML pipeline. No fine-tuning. Just click 'this was right.'

Your team already knows SQL.
Now it knows AI.

Free tier includes 5,000 credits. No credit card. Connect any database in 60 seconds. If you already have a SQL client open, you're ready.

Credits scale with AI usage. Cache hits are free. The more you use it, the cheaper it gets.

Works with PostgreSQL, MySQL, Snowflake, BigQuery, ClickHouse, S3, Parquet. Connects via pgwire — use psql, DBeaver, DataGrip, Tableau, or any Postgres-compatible client.