Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature Request]: Autogen online? #3546

Open
summersonnn opened this issue Sep 19, 2024 · 0 comments
Open

[Feature Request]: Autogen online? #3546

summersonnn opened this issue Sep 19, 2024 · 0 comments
Labels
enhancement New feature or request

Comments

@summersonnn
Copy link

summersonnn commented Sep 19, 2024

Hello,

I’ve been using vLLM to successfully run my local language model. Currently, I’m using AutoGen to connect to my vLLM server, define tools, run them, and everything works smoothly in offline mode, including calling the appropriate functions.

When using the standard vLLM, serving was straightforward with the vllm serve command. However, with AutoGen, since it’s just a basic script, I can’t serve it the same way as before. I now have to implement an additional layer using FastAPI or something similar on top of the vLLM server. This is a complex task and involves a lot of effort. Handling load balancing will also be a pain.

In short, is there a way to use AutoGen in an online setting? Perhaps a smart method to serve it, or a way to connect it back to the vLLM endpoint and continue interacting as before?

I’m curious to know what the standard approach is.

Many thanks.

Describe the solution you'd like

Maybe an autogen server?

@summersonnn summersonnn added the enhancement New feature or request label Sep 19, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant