Optimizing LLM Workflows with Fine-Grained Orchestration
Optimizing LLM Workflows with Fine-Grained Orchestration
Created using ChatSlide
Teola is a novel framework designed to address efficiency and workflow issues in literature research involving Large Language Models (LLMs). It targets the challenges of LLM hallucinations and inefficiencies by optimizing fine-grained workflows combining LLM and non-LLM modules. Motivated by latency bottlenecks and ignored optimizations in current frameworks, Teola employs parallelization and pipelining strategies. Workflows are represented as graphs, optimized by a graph optimizer and...