RAG Chunker
Split text into retrieval-friendly chunks with token-aware sizing and overlap.
Defaults: Token-based sizing, 350 tokens per chunk, 50 token overlap, tokenizer gpt-tokenizer-gpt4. Max recommended input: 200k characters.
Pure TypeScript GPT-4 tokenizer (cl100k_base, no WASM required)