You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When training with num_threads > 1, operations are consistently throttled by rate limits, even when configured with high API limits.
Current Behavior
Training attempts with multiple threads (num_threads > 1) consistently hit rate limits
This occurs even when API limits are set high
Currently no automatic fallback mechanism exists when rate limits are hit
Proposed Solution
Implementation of an automatic fallback system that would:
Accept multiple LLM configurations
Automatically switch to alternate configurations when rate limits are encountered
Related Issues
While there are existing issues discussing litellm router configurations, this proposal suggests directly implementing fallback functionality within the core system.
The text was updated successfully, but these errors were encountered:
Rate Limiting Issues with Multi-threaded Training
Issue
When training with
num_threads > 1
, operations are consistently throttled by rate limits, even when configured with high API limits.Current Behavior
num_threads > 1
) consistently hit rate limitsProposed Solution
Implementation of an automatic fallback system that would:
Related Issues
While there are existing issues discussing litellm router configurations, this proposal suggests directly implementing fallback functionality within the core system.
The text was updated successfully, but these errors were encountered: