-
Notifications
You must be signed in to change notification settings - Fork 6.5k
-
Notifications
You must be signed in to change notification settings - Fork 6.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Advanced Paste > Paste with AI] Custom Model / Endpoint Selection #32960
Comments
It would be nice to have local models too. |
@minzdrav This would be enabled by my proposed change - Ollama provides partial support for the OpenAI API schema, so you'd be able to point the plugin at your local model |
In particular, supporting an Azure OpenAI endpoint would be great first implementation. It would be even better if the Azure implementation supported Managed Identities so we don't end up with the unmanageable mess of API key distribution and rotation. |
supporting Groq would be nice too |
IMPORTANT |
bump... |
bump |
Has anyone started working on this item? |
To my knowledge, no The basics should be pretty easy to implement though! All you'd need to do to allow for a different api-compatible host and model is add two text fields to the settings page (model, URL) and link them in exactly the same way that the chatgpt token field is currently linked into the app (as far as I know, they are just additional inputs into the same function in the associated library) Obviously making it "Microsoft-quality" will require more work on documentation and integration - see the points @htcfreek has raised in this thread for examples of these I'd be happy to take a look, but I won't be able to for at least a week so you may be better placed than me. |
@nathancartlidge , @tjtanaa
Are you referring to my comment regarding the Group Policies above. |
Yeah, that's what I was referring to! It's a great addition, but also the kind of thing I'd completely overlook when building this sort of feature :) I hadn't seen that thread before, thanks for bringing it up - from a cursory reading it does look like their work could currently be independent from this, as it seems to be exclusively non-ai features - however, I agree that it could make sense to combine them for the sake of reduced development overheads. |
Thank you very much for the suggestions @nathancartlidge . I have a prototype version which leads me to think there are some changes that I am thinking of making. I would be great if I could get some inputs. I am planning to target local LLM Usecase on PC without dedicated GPU. (In most cases, there is only enough resources to host one model at a time).
Other feature improvements would be adding some common usecases as quick access on menu, such as
Moreover, I also saw that there is a branch I am new to |
@nathancartlidge , @tjtanaa @tjtanaa |
Sure. That's a better approach. Let's get the inputs from the core team first. Does it totally disable paste-with-ai feature or it will restrict which llm service endpoints that a user can use? |
Currently it disable totally and is based on the name expected to disable Online AI feature. But I can imagine that we add two or three new policies:
I can help later with implementing this. |
@htcfreek @nathancartlidge Just to let you know as shown in the prototype video in #33109, model selection would be an option. I don't know if custom endpoints were specifically shown, but I believe it was part of the implementation plan. |
Tuning in here! I started taking a swing at it and have it pretty much working well with Ollama. I used Phi-3 mini and the results are great on my Nvidia 4090. Happy to share my results if interested. |
Is the development in your fork? |
Once this is done, we can begin merging the idea of custom presets discussed in #33109 with the different AI models, as well as the offline features that are directly baked in. |
I tried adding my Google Gemini API key to the AI paste feature but it does not work. I might just try the OpenAI key for 5 bucks and see how this feature works. Really enjoying PowerToys thus far, thank you! |
This feature has not been added yet. |
As clipboard may contain important data, such as account and password. Local model should be the default option. Hope to see the feature. |
Is there any update on this feature? I’m really looking forward to it! 😊 |
Wouldn't it be easier to add support for OpenRouter? That should cover most of the popular AI models, and make things easier on the integration too, yes? |
Using OpenRouter requires you to use their credit platform I believe. This would not be useful to people who wish to use their own API keys. |
I'd like to also advocate for Ollama support. |
Description of the new feature / enhancement
It should be possible to configure the model used (currently fixed as
gpt3.5-turbo
) and endpoint (currently fixed as OpenAI's) to arbitrary valuesScenario when this would be used?
Sending requests to an alternative AI endpoint (eg a local model, internal company hosted models, alternative ai providers), or ensuring higher-quality conversions (eg by pointing requests at gpt-4o)
Supporting information
Microsoft's documentation appears to suggest that the underlying library used for AI completions supports other libraries, it just needs to be provided with an endpoint.
The currently used model is a hardcoded string in this repository
The text was updated successfully, but these errors were encountered: