Blog Splash

Gestell has solved search-based reasoning

RAG as sickness and cure

March 31st, 2025

Why your LLM can’t work with your data

LLMs were made to be tools of knowledge and intelligence, not to search through messy databases. The current Retrieval Augmented Generation (‘RAG’) tooling that exists expects LLMs to be experts at search, and it consequently fails to achieve results. Fundamentally, RAG is an overwhelming task for an LLM, the attempt to find truth in data devolves as more context of ‘similar chunks’ are shoved into a transformer which degrades the quality of the attention ranking mechanism

How can truth be found in data without search? It can’t. Search is an unavoidable paradigm in dealing with the real-world and real business use-cases for LLMs. Information that hasn’t been trained into an LLM must somehow be brought into the purview of it; however, the static weights applied to the transformer architecture define a model’s capabilities. It is these weights that ‘weigh down’ the search mechanism. New unknown information suffers as an infinite regress develops where each additional token that isn’t the ground truth of a query further starves the transformer of attention to the truth

RAG is thus both a sickness and a cure to models. Database structuring begins to attempt to solve this issue but fails when it mistakes the contents of reasoning (context) with the actual process of reasoning. Vector stores that measure the distance between words and find similarities using nearest neighbor algorithms aren’t coterminous with actual reasoning. Even specialized ‘embeddings’ can only capture certain keywords. This is why vector stores fail to achieve actual search-based reasoning and we have proclamations of the ‘failure of RAG’

‘Agentic’ search has been presented as a solution to the RAG paradigm. Here, rather than vectors powering search, smaller agents must comb through data to ingest, summarize and present it to a ‘master’ model. This search process may sound appealing but devolves into essentially the same challenge of vector-based search, except instead of keyword embeddings powering the search algorithm, it is agents. RAG is unavoidable in the basic sense that retrieval must augment the generative knowledge of a model. The core challenge of RAG, however, still remains though, truth and reasoning are not the results of either keywords or ‘agents’ presenting similar conceptual chunks into a context window

Why you need an ETL

Truth is the synthesis of context and reasoning. Truth requires context - context rightly delivered and presented, concisely and apparently, to an attentive model. When attention is overwhelmed, synthesis is impossible. When context is missing, hallucination is the result. What is needed to enable search-based reasoning is not simply chunks of similar information or relevant concepts but a fulsome ETL that can Extract the truth a model needs, Transform it into concise, legible formats and Load it into the proper context conducive for reasoning

The gordian knot of the search-mechanism and the fundamental architecture of a transformer is circumvented by an ETL, as an ETL doesn’t allow a model to ‘search’ for data but rather ‘delivers’ the exact contents needed to the model. An ETL is a complement to a model, an assistant in its reasoning. The challenge with this ETL model for LLM reasoning is that most of the data LLMs need to reason over isn’t already in structured formats which would be accessible via traditional ETL architecture. Most of the data LLMs need is trapped within unstructured, unavailable formats

A new ETL is thus needed

What is Gestell?

Gestell is a new ETL for LLMs - a comprehensive data structuring system that transforms unstructured data into AI-ready databases. Gestell covers the entire process from Enframing to Disclosure, enabling accurate, scalable search-based reasoning for your language models

What does Gestell do differently?

Gestell solves the RAG problem by disempowering the model from doing search in favor of an ETL providing the exact information the model needs to make a decision. Our ‘search’ process then is not comparable to normal ‘retrieval’ processes which attempt to ‘find’ chunks of similar context within a database, but rather is a method of ‘delivering’ truth to a model that can then think about the data and come to a decision. What we call Disclosure (search-based reasoning) is in fact the limiting of the information delivered to a model, such that the exact context is given to a model so that it can reason effectively about it. Gestell prioritizes attention and context as a means of enabling reasoning

Graphic Describing Gestell's Process

Key features of Gestell:

1. Integrated ETL: Simply upload your data, in any format, across any discipline and let Gestell handle the rest from Enframing to Disclosure

2. Customizable: Using natural language, instruct how you want Gestell to structure your data

3. Scalable, immediately: Gestell is the only solution for search-based reasoning on the market proven to scale beyond 50k pages, and is ready to deploy immediately

4. Cost Efficient: Gestell is ~40% cheaper than solutions like Google’s Document AI (which only has pdf parsing). Gestell’s cheaper pricing is also entirely inclusive of every filetype and format (images and video included) while also serving to structure the data being processed

Gestell vs. the competition

We often hear from users that they have either relied on AI workspaces (for example: AWS or Google Vertex) or have built their products upon point solutions like traditional vector stores

AI workspaces

Mainstream AI workspaces exemplify the current RAG paradigm; they enable simple data upload and then naively throw everything into a vector store, expecting quality results from search. The challenge of the AI workspaces is that they show the appearance of competency at search-based reasoning through ease-of-use. These tools almost all enable no-code deployments and plentiful connectors to bring data into a platform; however, their architecture is fundamentally flawed resulting in failure at production-level deployment of search-based reasoning

Point solutions / build-it-yourself

Outside of AI workspaces, many people opt to build their own data processing, structuring and search systems. These systems are made up of individual tools or point solutions. Point solutions like vector stores or knowledge graphs have all attempted to enable deeper complexity or higher performance search through specializing a specific part of the ‘RAG’ stack. Companies that don’t rely on AI workspaces will often cobble together 5+ point solutions to build search-based reasoning. There are two primary challenges with point solutions: they weren’t made with other solutions in mind and they take a long time to deploy in tandem. Since they specialize one part of the stack they fail to converse eloquently with other parts of search architecture, resulting in over-emphasis. For example, some RAG tooling may excel at single-document retrieval but fail when you add additional data or different formats. Additionally, these tools all take months of engineering work to actually deploy, with no guarantee of scalable results. A demo might be able to be pieced together using these tools, but production-scale real-world use-cases are often unable to be shipped

Below is a comparison table of Gestell vs. the competition:

Gestell LogoGestellPinecone LogoPineconeChroma LogoChromaGoogle LogoGoogleAWS LogoAWSDatabricks LogoDatabricksSnowflake LogoSnowflakeCohere LogoCohere

Integrated ETL

End-to-end from Enframing to Disclosure

Ingestion - Parsing

Systematic breakdown and analysis of input data

Ingestion - Intelligent Chunking

LLM-enabled segmentation for search optimization

Ingestion - Multi-modal

Handles diverse data types simultaneously

Customizable, specialized vector store

Tailored database optimized for vector embeddings

Integrated Knowledge Graphs

Information network showing contextual relationships

1

Re-rankers

Refined search result prioritization system

Accurate Scalability

Maintains precision while handling increasing data volumes (50k+ pages)

Feature Extraction

Identifies and isolates key data characteristics

Table Structuring and Extraction

Organizes and retrieves tabular information effectively

RLHF enablement

Supports human feedback for model improvement

Full pipeline structuring optimization

End-to-end process refinement for maximum efficiency across the stack

Code / No-Code

Flexible implementation across technical expertise

BOTHCODEBOTHCODE2BOTHBOTHBOTHCODE

Model-Agnostic

Works with any AI framework or model

7 day a week support

White-Glove support - you will have the founder's phone number

Reach out, Gestell is ready to deploy today