Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

optimizing whisper without audio_decoder #1236

Open
Aiurus opened this issue Jul 15, 2024 · 3 comments
Open

optimizing whisper without audio_decoder #1236

Aiurus opened this issue Jul 15, 2024 · 3 comments

Comments

@Aiurus
Copy link

Aiurus commented Jul 15, 2024

Describe the bug
I tried to optimize whisper-tiny.en model without audio_decoder, but error occurred.

To Reproduce

  1. python3 prepare_whisper_configs.py --model_name openai/whisper-tiny.en --no_audio_decoder
  2. olive run --config whisper_cpu_int8.json --setup
  3. olive run --config whisper_cpu_int8.json 2> /dev/null
    After running 3rd command, the model should be generated in 'model' folder but error occured.

Expected behavior
When I tried with audio_decoder, the code works well.

Olive logs
[olive_evaluator.py:236:generate_metric_user_config_with_model_io] Model input shapes are not static. Cannot use inferred input shapes for creating dummy data. This will cause an error when creating dummy data for tuning.

Other information

  • OS: Debian
@jambayk
Copy link
Contributor

jambayk commented Jul 15, 2024

Can you try again by removing the "evaluator": "common_evaluator" from the template. There might be an issue with the evaluator but it is not required.

If it still fails, please share the full log from the run.

@Aiurus Aiurus closed this as completed Jul 15, 2024
@Aiurus Aiurus reopened this Jul 16, 2024
@Aiurus

This comment was marked as resolved.

@jambayk
Copy link
Contributor

jambayk commented Jul 16, 2024

We don't provide an option to remove this mode. It was added by onnxruntime-extensions in this PR microsoft/onnxruntime-extensions#681

please install the previous version of onnxruntime-extensions 0.10.1 and rerun the workflow. you can add "clean_run_cache" : true at the same level as

"type": "AppendPrePostProcessingOps",
to only rerun this pass.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants