-
Notifications
You must be signed in to change notification settings - Fork 16
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add a base path for ollama #25
Comments
Howdy @oppenheimer- I totally get the request here. The hold up is that there is some functionality in Caret that can only be done with models that support functional calling. Unfortunately I don't think the api here provides that information. I think the easiest solution would be to allow you to import the models and then have a UI for marking if they support function calling or not. No current ETA on when I'll be able to get around to that. Going to leave this open for now. |
Maybe there's a way to implement a 'preflight' check with Ollama when users save options or select a model, which would verify compatibility. However, I recommend a simpler approach: let users learn through trial and error which models work for their setup. This avoids unnecessary complexity in the implementation. Then give simple advice for working models (Phi 4, Llama 3.2/3.3 etc.) |
I'm not particularly keen on users having to trial and error which models work for which functionality in Caret. I think an import flow threads the needle the best. Caret can import most of the data from ollama or other providers an then the user just has to adjust a couple settings for each they want to import. I might take a look at this over the next few weeks. But this would also be a great first issue if anyone from the community wants to take a crack at it |
maybe similar to smart second brain or other plugins that use ollama.
the route "/api/tags" delivers all the models to populate the list.
Originally posted by @oppenheimer- in #5 (comment)
The full API docs are here:
Ollama API docs
"name": "llama3.2"
}'
will reveal the required information
if im not mistaken, the context_length can be acquired with the model family name.
The text was updated successfully, but these errors were encountered: