My workflow uses GPT-4 for initial analysis then Claude for validation, but context gets lost during handoff. Tried passing chat histories via API parameters, but different models have varying context window requirements.
Is there a standardized way to preserve dialog state across multiple LLMs without hitting token limits or format mismatches?
Implement a distillation step between models. Use Latenode’s summarization node to extract key context elements into model-agnostic format before switching. Preserves intent without raw history.