Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LLM Optimization with DirectML reply only displays "O"s #1282

Open
yichunx1 opened this issue Aug 6, 2024 · 3 comments
Open

LLM Optimization with DirectML reply only displays "O"s #1282

yichunx1 opened this issue Aug 6, 2024 · 3 comments
Labels
DirectML DirectML

Comments

@yichunx1
Copy link

yichunx1 commented Aug 6, 2024

Describe the bug
I followed all the steps from LLM Optimization with DirectML. I was able to find the ONNX model and was able to start gradio UI. But no matter what I entered in the chat box, the reply is always a series of "O"s, as shown in the following screenshot.
bug_screenshot

To Reproduce
I followed this for setup:
https://github.com/microsoft/Olive/blob/main/examples/README.md#important
I also pip installed pillow because it's not in the requirement.txt
Then I followed this for ONNX conversion and run chat app.
https://github.com/microsoft/Olive/tree/main/examples/directml/llm
I also tried the gradio 4.29.0 but it seems not compatible.

Expected behavior
The reply should be text instead of "O"s.

Olive config
Add Olive configurations here.

Olive logs
Add logs here.

Other information

  • OS: Windows 11 Pro
  • Olive version: 0.7.0
  • ONNXRuntime package and version: 1.16.1
  • gradio: 3.42.0

Additional context
Add any other context about the problem here.

@jambayk
Copy link
Contributor

jambayk commented Aug 7, 2024

@PatriceVignola do you have any insights on this?

@yichunx1
Copy link
Author

yichunx1 commented Aug 7, 2024

I just tried different models.
When I use phi3 mini 128k, the answer is a few line of "////////"s.
I also tried Mistral 7b, it shows error, saying that the model cannot be found (but I can see the optimized model is in the folder with others).
Fortunately, when I go for gemma 7b, the output is normal.
Does anyone knows why?

@PatriceVignola
Copy link
Contributor

@yichunx1 Which GPU are you using? And which onnxruntime-directml version are you using?

@devang-ml devang-ml added the DirectML DirectML label Aug 8, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
DirectML DirectML
Projects
None yet
Development

No branches or pull requests

4 participants