Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Benchmarking o1-preview-2024-09-12 #135

Open
haesleinhuepf opened this issue Sep 12, 2024 · 3 comments
Open

Benchmarking o1-preview-2024-09-12 #135

haesleinhuepf opened this issue Sep 12, 2024 · 3 comments

Comments

@haesleinhuepf
Copy link
Owner

If anyone knows anyone who is tier 5 in openai (@royerloic maybe ?), they could benchmark the new o1 model. I am just tier3 and have to wait...

https://x.com/OpenAI/status/1834278218888872042

https://platform.openai.com/docs/models/o1

@jkh1
Copy link
Collaborator

jkh1 commented Sep 13, 2024

It seems that this model has hidden tokens that you still get billed for (see note in the docs here: https://platform.openai.com/docs/guides/reasoning/how-reasoning-works) which may explain why it's for tier 5 users 😃
This could become an expensive experiment.

@haesleinhuepf
Copy link
Owner Author

Yeah I know. I think claude does something similar. That's why I'm curious if it's really so much better than the models we tested so far.

@haesleinhuepf
Copy link
Owner Author

haesleinhuepf commented Sep 24, 2024

Ok, I have access now. Just FYI: My first 12 prompts cost $6.68 for o1-preview, hence running the entire benchmark would cost about $300. (updated)

@haesleinhuepf haesleinhuepf mentioned this issue Sep 25, 2024
9 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants