-
Notifications
You must be signed in to change notification settings - Fork 323
Add recent fireworks models #566
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
| input = 0.56 | ||
| output = 1.68 | ||
| cache_read = 0.28 | ||
|
|
||
| [limit] | ||
| context = 160_000 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
isn't this an interleaved thinking model?
| knowledge = "2025-04" | ||
| open_weights = true | ||
|
|
||
| [cost] | ||
| input = 0.60 | ||
| output = 2.20 | ||
| cache_read = 0.30 | ||
|
|
||
| [limit] | ||
| context = 198_000 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
isn't this interleaved thinking model?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It is, but I dont know if it requires provider support. Does it "just work"? If that is the case kimi-k2-thinking is also missing interleaved thinking
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Also I based the definitions on the previous version of the models which don't document it either. should I just add it?
|
Yeah it looks like for fireworks it is Only models I think that'd be:
I think that's all of them for them |
16_384 seems to be the ui limit
|
I added interleaved thinking. |
Adding new models availbe through Fireworks.
Pricing data taken from their model library.
Usure about the max output tokens, reused the context limit.