Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Rate Limiting #2048

Open
samschiff19 opened this issue Jan 16, 2025 · 0 comments
Open

Rate Limiting #2048

samschiff19 opened this issue Jan 16, 2025 · 0 comments

Comments

@samschiff19
Copy link

Rate Limiting Issues with Multi-threaded Training

Issue

When training with num_threads > 1, operations are consistently throttled by rate limits, even when configured with high API limits.

Current Behavior

  • Training attempts with multiple threads (num_threads > 1) consistently hit rate limits
  • This occurs even when API limits are set high
  • Currently no automatic fallback mechanism exists when rate limits are hit

Proposed Solution

Implementation of an automatic fallback system that would:

  1. Accept multiple LLM configurations
  2. Automatically switch to alternate configurations when rate limits are encountered

Related Issues

While there are existing issues discussing litellm router configurations, this proposal suggests directly implementing fallback functionality within the core system.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant