Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add more models to default pack [Ongoing] #29

Open
scosman opened this issue Nov 12, 2024 · 8 comments
Open

Add more models to default pack [Ongoing] #29

scosman opened this issue Nov 12, 2024 · 8 comments
Labels
enhancement New feature or request

Comments

@scosman
Copy link
Collaborator

scosman commented Nov 12, 2024

We like to have a well-tested default pack of models (each model-provider pair is tested, including testing that structured output works reliably).

What models should we add next? 👍 a model comment for support, and I'll prioritize by this list.

See the separate issue tracking end-user option to add models.

@scosman
Copy link
Collaborator Author

scosman commented Nov 12, 2024

✅ Request: Qwen 2.5

@scosman scosman added the enhancement New feature or request label Nov 12, 2024
@scosman
Copy link
Collaborator Author

scosman commented Dec 10, 2024

✅ Llama 3.3 70b

@scosman scosman changed the title Add more models to default pack Add more models to default pack [Ongoing] Dec 10, 2024
@scosman
Copy link
Collaborator Author

scosman commented Dec 30, 2024

✅ Deepseek V3

@leonardmq
Copy link
Contributor

Adding Google as a provider would be a nice one as Gemini 2.0 is coming out to prod in the next few weeks

@scosman
Copy link
Collaborator Author

scosman commented Jan 6, 2025

@leonardmq I can add 2.0 when it's out. Is there a reason to want direct Google/Vertex access instead of Gemini via OpenRouter?

@leonardmq
Copy link
Contributor

@leonardmq I can add 2.0 when it's out. Is there a reason to want direct Google/Vertex access instead of Gemini via OpenRouter?

My assumption was that having Google/Vertex as a provider would enable adding support for fine-tuning for models from Google/Vertex, similar to how Kiln currently supports fine-tuning for OpenAI models. From what I have seen, it seems OpenRouter does not support submitting fine-tuning jobs, though I may have missed some functionality there.

My primary use case right now involves creating synthetic datasets and fine-tuning models for specific tasks to then identify the best-performing and most cost-effective one - and Gemini 2.0 Flash is likely to be quite relevant there so would be great if we could fine-tune as easily as we can OpenAI models

@scosman
Copy link
Collaborator Author

scosman commented Jan 6, 2025

Ah, fine tuning support should be its own issue (feel free to create it). Fine tuning is a lot more work than just a model addition which is a few lines of config. You're correct OpenRouter is only for inference. I need to look at the Vertex docs; I know they have some API driven fine-tuning, but need to look at their serverless hosting options. In the meantime, just export a JSONL and use their UI?

@scosman
Copy link
Collaborator Author

scosman commented Jan 21, 2025

Deepseek R1 - I’ll add this when I’m back

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants