## **Feature Description** <!-- I've profiled the current MTP runtime implementation in my testing environment. Key findings: - litellm import time: 2.36-2.51 seconds startup latency (measured across multiple runs) - Memory overhead: Around 143MB spike during import - Comparative performance: 5.2x slower than OpenAI client (2.51s vs 0.48s) - Root cause: Immediate imports in `llm_connector.py:23` and `types.py:19` Based on my findings, lazy loading appears to be a promising approach: - Replace immediate `import litellm` with function-level imports - Defer initialization until first actual LLM call - Optional: Background preloading via non-blocking thread --> ## **Examples** ```python # Current (problematic): import litellm # - Around 2s delay at module load def call_model(...): response = litellm.completion(...) # Already loaded # Proposed (lazy loading): def get_litellm(): import litellm # - Only imported when actually needed return litellm def call_model(...): litellm = get_litellm() # - Import happens here response = litellm.completion(...) -->