ChatGPT caps file uploads at 100MB as of 2026. Claude does too. Your Excel file with 500,000+ rows? Probably bigger than that – sales data, transaction logs, CRM exports all hit that wall.
The real problem isn’t file size.
Tokens – the chunks of text the model processes in one request. A 50MB CSV? Could be 20 million tokens when you include your instructions and the model’s response. ChatGPT Plus caps at 128K tokens (as of 2026). Claude Pro offers 200K. The file uploads. The AI reads the first 30% of your data, truncates the rest. You don’t notice until you check the output.
What AI Data Cleaning Does (and Where It Fails)
AI data cleaning works by pattern recognition. Upload messy data – duplicate customer names, inconsistent date formats, missing ZIP codes. The model identifies what should be uniform. Standardizes “New York” and “NY” into one format. Fills gaps by inferring from similar rows. Detects outliers manual review would miss.
The failure mode: numeric data gets misinterpreted. ZIP code 01002 (valid for Amherst, Massachusetts) looks like the number 1,002 to the model. It “corrects” it. Product SKU “00456” becomes “456.” Your clean data now has errors the messy version didn’t.
Why? LLMs tokenize text. Leading zeros look like formatting mistakes unless you tell the model they’re intentional. Community forums report this constantly. Tutorials? Never.
Think of it like autocorrect on your phone – it “fixes” names it doesn’t recognize. The model does the same with codes it interprets as numbers.
Pro tip: When uploading numeric data like ZIP codes, IDs, or SKUs, instruct the model: “Treat all values in column X as strings. Do not interpret them as numbers or remove leading zeros.” Add this to every data cleaning prompt that involves codes.
The Practical Workflow
Start with ChatGPT Plus ($20/month as of 2026) or Claude Pro (same price). Both let you upload files directly. Free tiers work but cap usage – Claude’s free plan allows roughly 15 messages per 5-hour window, fewer when servers are busy.
Prepare the file. Export your data as CSV (not Excel). Remove any sheets, charts, or macros. Keep only the data table. If the file exceeds 50MB, split it – more on that below.
Upload and describe. Drag the CSV into the chat. “This file contains customer records with inconsistent company names, missing email domains, and duplicate entries. Standardize company names, fill missing emails where possible, and flag duplicates without deleting them.”
Output format. “Return the cleaned data as a CSV file” or “Provide a downloadable table.” Without this? You’ll get a preview or text output instead of a usable file.
Spot-check. Download the output. Open it in Excel or a text editor. Check the first 20 rows, the last 20 rows, and 5-10 random rows in the middle. Look for: dropped columns, merged cells that shouldn’t be merged, numeric fields that lost leading zeros.
Iterate if needed. If the model made mistakes, upload the result back: “You removed leading zeros from the ZIP code column. Restore them as strings.” It’ll fix it in the next pass.
Files Over 100MB
Dataset too large? Split it into chunks. Export every 100,000 rows as a separate file. Clean each file with the same prompt – keeps the cleaning rules consistent. Merge the outputs manually (or use Python pandas if you code).
For truly massive files (10M+ rows), OpenRefine – free, open-source – handles scale better than ChatGPT. But steeper learning curve. You define clustering rules manually. AI is faster for small-to-medium datasets; traditional tools scale better.
ChatGPT vs Claude: What Actually Matters
Both clean data. Their strengths differ.
| Feature | ChatGPT Plus | Claude Pro |
|---|---|---|
| Price | $20/month (2026) | $20/month (2026) |
| File upload limit | 100MB | 100MB (estimated) |
| Context window | 128K tokens | 200K tokens |
| Best for | Quick fixes, small files, Structured Outputs | Large text context, long prompts, technical data |
| Usage limits | More generous message cap | Stricter limits (message cap per 5-hour window) |
ChatGPT’s edge: Structured Outputs. Introduced in August 2024, this feature forces the model to return data in a predefined JSON schema. You define the exact structure – column names, data types, required fields. The model guarantees compliance. 93% reliability on OpenAI’s benchmark. And the schema doesn’t count toward your token limit (a cost optimization buried in the docs).
Claude’s edge: larger context window. 200K tokens holds more data per request than ChatGPT’s 128K. Got a 75MB file that’s mostly text (logs, transcripts, unstructured notes)? Claude handles it in one pass. ChatGPT truncates.
Structured tabular data (CSV, spreadsheets)? Use ChatGPT with Structured Outputs. Large unstructured text? Claude.
3 Ways AI Data Cleaning Breaks
Hallucination in duplicates. You ask the model to merge duplicate records. It uses fuzzy matching – “John Smith, 123 Main St” and “John Smith, 456 Oak Ave” look similar. The model merges them. They’re two different people. Data loss. There’s no documented false positive rate for LLM-based deduplication. The fix: ask the model to flag duplicates, not delete them. Review flagged records manually.
Token limits hit mid-file. 70MB CSV with 300K rows. The model processes the first 150K rows, hits its context window, stops. Returns a “cleaned” file with half your data missing. Error message is vague (“output truncated” or “response too long”). Split large files into chunks before uploading. Don’t wait for the model to truncate.
Cost for repeated iterations. Each upload: input tokens. Each output: output tokens. Iterate 5 times on a 50MB file? You’re burning through tokens fast. ChatGPT Plus has a message cap (not unlimited). Claude Pro has stricter limits. Plan your prompt carefully on the first try. Don’t experiment iteratively unless you’re on an enterprise plan.
When AI Isn’t the Right Tool
AI works for pattern-based errors – typos, inconsistent formatting, obvious duplicates. Doesn’t work for:
Domain-specific rules. Your company defines “duplicate” as “same email + same phone + transaction within 30 days.” The model won’t know that. You need SQL or Python with explicit logic.
Regulatory compliance. Financial data, healthcare records, GDPR-sensitive information shouldn’t go into ChatGPT or Claude unless you’re using enterprise versions with data processing agreements. The free/Plus/Pro tiers train on your data by default (you can opt out – check your settings).
Real-time pipelines. Ingest data continuously (streaming logs, API feeds)? Batch-uploading to ChatGPT every hour isn’t sustainable. Use dedicated tools like OpenRefine or Alteryx (starts at $250/user/month as of 2026).
A 2025 study on AI-assisted medical data cleaning found 6x improvement in throughput and 46% accuracy boost – but only when combined with human oversight. AI flagged issues. Humans made final calls. Fully automated AI cleaning, no review? Introduced errors in 8-15% of cases depending on data type.
What does that mean for your data? Don’t trust AI to clean anything you can’t afford to lose. Always spot-check.
What About GPT-4o’s Structured Outputs Feature?
OpenAI introduced Structured Outputs in August 2024. This feature guarantees JSON schema adherence – 93% reliability on OpenAI’s benchmark. When you clean data with ChatGPT, you can define a schema: column names, data types, required fields. The model returns data that matches your schema exactly.
The catch? It’s an API feature. ChatGPT Plus users can approximate it by specifying output format in prompts (“return as CSV with these exact columns”), but the guarantee isn’t there. Developers using the API get the full benefit – schema tokens don’t count toward your input limit, a cost optimization nobody explains upfront.
For non-coders: Think of Structured Outputs as a template. You tell the model: “I need a table with columns A, B, C. Column A must be a string, B must be a number, C must be a date.” The model won’t deviate. No more “the model added an extra column I didn’t ask for” surprises.
Frequently Asked Questions
Can I use free ChatGPT for data cleaning?
Yes, but no file uploads. Copy-paste data into the chat – caps at a few thousand rows. Anything larger needs ChatGPT Plus ($20/month as of 2026).
What happens if my CSV file has special characters or non-English text?
GPT-4 and Claude handle UTF-8 encoding well. Accented characters (é, ñ, ü) and non-Latin scripts (Chinese, Arabic) usually work. Risk: the model might “correct” names it thinks are typos. Vietnamese name “Nguyễn” could become “Nguyen” if the model assumes the diacritic is an error. Your prompt should say: “Preserve all diacritics and special characters exactly as written.” Turns out that one line prevents 90% of these issues.
How do I know if the AI deleted rows or just cleaned them?
Count rows. In Excel: Ctrl+Down (Windows) or Cmd+Down (Mac) jumps to the last row. Note the row number. Upload to ChatGPT. Download the result. Check the row count again. Lower? AI dropped rows. Ask it: “Why did the output have fewer rows than the input?” It’ll explain (or admit the error).
Your Next Step
Export one messy file from your CRM, database, or spreadsheet. Keep it small – 1,000 to 10,000 rows. Upload it to ChatGPT Plus with this prompt:
“This file contains [describe the problem: duplicate names, inconsistent formats, missing values]. Clean it by [describe the fix: standardizing, filling gaps, flagging duplicates]. Treat all ID columns as strings. Return a downloadable CSV.”
Download the result. Compare it to the original. You’ll see where AI saves time and where it introduces risk. That 10-minute test tells you more than any tutorial.
FAQ
Does AI data cleaning work offline?
No. ChatGPT and Claude are cloud services. Need offline cleaning? Use OpenRefine (free, runs locally) or Python with pandas. Both require more setup but don’t send your data to external servers.
Can AI clean data in Excel without exporting to CSV?
Not directly. ChatGPT and Claude can’t read .xlsx files natively – you’ll get a parsing error or incomplete results. Always export as CSV first. Multiple sheets? Export each one separately.
Is AI data cleaning accurate enough for production use?
Low-stakes data (marketing lists, internal reports)? Yes with spot-checking. High-stakes data (financial transactions, medical records, legal documents)? No – AI should flag issues for human review, not make final corrections. A 2024 study found that AI-powered data cleaning surpasses traditional methods in efficiency and accuracy but still requires validation. Never deploy AI-cleaned data to production without manual verification of at least a sample. The 8-15% error rate from fully automated cleaning isn’t worth the risk.