Hi there! I’m dealing with a frustrating performance issue in my n8n automation workflow and hoping someone here can point me in the right direction.
My Setup:
I’ve got a Google Spreadsheet containing 135,000+ business records. My n8n workflow processes job postings and needs to verify if each company name appears in this master list.
Current Challenge:
The Google Sheets integration completely fails when handling this data volume. When I try Get Rows:
- Sometimes throws
Maximum call stack size exceedederror - Other times just hangs indefinitely without returning results
Solutions I’ve Attempted:
- Filtered queries on the “Company Name” column - still crashes from data size
- Exported data to JSON format using a Python script locally
- File import attempts in n8n:
File operationsnode - JSON parsing issues with binary dataHTTP Requestfrom GitHub raw file - works but extremely slow parsing, can’t pin due to 12MB+ size
- Manual data entry via Set node - browser crashes from memory overload
- Code node with workflow cache (
this.getWorkflowStaticData) - doesn’t persist between executions - Batch processing ideas - still blocked by initial data loading problems
What I Need:
A reliable method to:
- Quickly verify company existence in this large dataset
- Avoid re-processing all 135k records on each workflow execution
- Stay within n8n memory constraints
Any suggestions for caching strategies, external databases, or alternative file hosting approaches? How do others handle large reference datasets in n8n?
Thanks for any help!