Skip to content

Chunking strategy comparison

This table maps the chunker names you will see across different vendors and sources, shows how boundaries are chosen, and summarizes the most common tradeoffs.

ComplexityChunker strategyNames you'll seeHow boundaries are chosenRetrieval unitTypical knobsAdvantagesDisadvantages / common failure modesGood fit
0No chunking"No chunking", "whole document", "document-as-unit"Don't split; store each record or document as-isWhole doc or recordMostly retrieval knobs such as top-k and filtersSimplest; preserves full context; minimal ingest logicCoarse embeddings; lost-in-the-middle risk; higher latency or cost if you stuff whole docs into promptsFAQs, short tickets, short posts, already atomic content
1Fixed-size (tokens or chars)Weaviate fixed-size chunking, Pinecone fixed-size chunking, Chonkie TokenChunkerCut every N tokens or characters regardless of meaningChunkchunk_size, overlap, chunk_overlapFast, cheap, deterministic baseline; easy to A/B testBreaks sentences and ideas; can orphan facts at boundariesQuick baseline; messy text; speed-first ingestion
2Sliding windowSliding window, overlap chunking, windowed chunking, Chonkie Slumber ChunkerLike fixed-size, but step size is smaller than chunk size so chunks heavily overlapChunkoverlap ratio, step sizeReduces boundary loss without smarter semanticsRedundant index; higher embedding and storage cost; duplicate-heavy retrievalWhen boundary misses hurt but you still need a deterministic path
3Sentence-basedSentence splitting, paragraph splitting, Chonkie SentenceChunkerSplit on sentence boundaries and then pack to a size limitChunktokenizer, max tokens, overlap, min sentencesCleaner thought units; fewer mid-sentence breaksSentence length varies; a single sentence can still be too context-poorQA and summarization where sentence integrity matters
4Recursive delimiter chunkingRecursive chunking, RecursiveCharacterTextSplitter, Chonkie RecursiveChunkerTry paragraph breaks first, then sentence breaks, then words until the chunk fitsChunkseparator list, max size, min chunk sizeSolid default; respects common structure more than fixed-sizeStill heuristic; weird formatting can defeat it; still returns broken textArticles, blog posts, papers, reports
5Document-structure awareDocument-based chunking, header-based chunking, element packing, Unstructured smart chunkingParse the format and split on structure such as headers, tags, pages, or semantic elementsChunkmax_characters, new_after_n_chars, section rulesBetter topical coherence; less random tearing; robust across file types when partitioning worksDepends on extraction quality; bad title detection can create tiny or wrong chunksPDFs, HTML, Markdown, manuals, policies
6Table-awareTable chunking, specialized table chunkersIdentify tables and keep them isolated; split oversized tables separatelyTable chunksize limit, table serializationPreserves row and column structureLarge tables can dominate retrieval and prompt budgetFinancials, specs, legal tables
7Code-awareCode chunking, AST chunkersParse code syntax and split by logical code blocks such as functions, classes, or modulesCode chunklanguage detection, chunk size, docstring or import preservationPreserves runnable units; improves code retrievalCross-file dependencies remain hard; large functions still need splittingRepos, SDK docs, notebooks, code-heavy knowledge bases
8Semantic similarity chunkingSemantic chunking, context-aware chunking, Unstructured by_similarity, Chonkie SemanticChunkerEmbed local sentences or elements and place boundaries where similarity drops or merge only if similarity remains highChunkembedding model, similarity threshold, window size, max sizeBetter topical purity; fewer mixed-topic chunksMore ingest-time compute; brittle if topic drift is gradualDense legal, academic, or technical narrative text
9Neural boundary detectionNeural chunking, learned boundary detectionA learned model predicts good boundaries from semantic-coherence patternsChunkmodel choice, min chunk sizeLess hand-tuning; can outperform naive separator listsOpaque decisions; domain mismatch can create odd splitsMixed corpora where heuristics underperform
10LLM-based chunkingLLM-based chunking, proposition extractionAsk an LLM to propose boundaries or rewrite text into retrieval-friendly unitsChunkprompts, target size, cost and latency budgetHigh semantic quality on complex textSlow, expensive, nondeterministic; risk of summary driftHigh-value docs where quality beats cost
11Agentic chunkingAgentic chunking, strategy-selection chunkingAn agent inspects each document and chooses a strategy or mix, possibly adding metadataChunkmodel choice, allowed tools, budgetFlexible across wildly different document typesHighest complexity; inconsistent across runs without strong monitoringRegulatory and multi-format enterprise corpora
12Adaptive chunkingAdaptive chunkingKeep one approach but dynamically adjust size or overlap depending on content density or structureChunkdynamic size ranges, density rulesAvoids one-size-fits-all behavior within a documentHarder to predict index stats and reproduce exactlyLong docs with mixed dense and sparse sections
13Late chunkingLate chunking, embed-first-split-later chunkingEmbed the whole document with full context and derive chunk embeddings afterwardChunk with document-aware embeddingslong-context embedding model, split rules, pooling methodReduces context loss in embeddings; improves cross-reference understandingRequires long-context embeddings and more computeTechnical or legal docs with cross-references
14Hierarchical chunkingHierarchical chunking, parent-child chunking, multi-level chunkingCreate multiple layers of chunks and retrieve coarse-to-fineMulti-level chunkslevels, sizes per level, parent-child retrieval logicAnswers broad and specific questions more gracefullyMore indexing and retrieval logic; still breaks textTextbooks, handbooks, manuals, long contracts
15POMA chunksets plus cheatsheetsChunksets, hierarchical chunksets, breadcrumb context, cheatsheetsParse sentence-by-sentence, infer hierarchy, and group sentences into unbreakable root-to-leaf pathsChunkset, later compiled into a cheatsheetMinimal exposed knobs; relies on structure inferenceFacts arrive with contextual lineage attached; structured, token-efficient contextRequires an integrated ingestion and retrieval design; different mental model from standard chunk retrievalHigh-stakes grounding in policies, legal, compliance, and manuals

For the narrative framing around this table, go back to the RAG chunking guide.

Continue reading