You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
LlamaServerlessAzureRestEndpointModel which is used to run 405B models sets ignore_eos: str = "false" by default. This is passed to the api as a string and as a consequence it's not set correctly. This causes the model to continue post EOS and generate random tokens till max_token limit.
Fix: Need to set ignore_eos: bool = False. I have tested this fix for Calendar Planning.
We will need to test other bool flags like skip_special_tokens and use_beam_search as well. Similar str flags are there in Mistral model class too.
The text was updated successfully, but these errors were encountered:
eureka-ml-insights/eureka_ml_insights/models/models.py
Line 279 in 1713e79
LlamaServerlessAzureRestEndpointModel which is used to run 405B models sets ignore_eos: str = "false" by default. This is passed to the api as a string and as a consequence it's not set correctly. This causes the model to continue post EOS and generate random tokens till max_token limit.
Fix: Need to set ignore_eos: bool = False. I have tested this fix for Calendar Planning.
We will need to test other bool flags like skip_special_tokens and use_beam_search as well. Similar str flags are there in Mistral model class too.
The text was updated successfully, but these errors were encountered: