Compile the model online in EP #1111
wujiangGitHub
started this conversation in
Support for Targets (OS / EPs / Hardware)
Replies: 3 comments 3 replies
-
Is this an onnxruntime specific issue? Or is there a solution with onnxruntime that genai is not making accessible? |
Beta Was this translation helpful? Give feedback.
3 replies
-
Update: ORT GenAI now supports OpenVINO EP, please use the latest commits from https://github.com/microsoft/onnxruntime main branch |
Beta Was this translation helpful? Give feedback.
0 replies
-
Hello, how is the performance of the large language generation model?
…------------------ 原始邮件 ------------------
发件人: "microsoft/onnxruntime-genai" ***@***.***>;
发送时间: 2025年7月5日(星期六) 下午3:54
***@***.***>;
***@***.******@***.***>;
主题: Re: [microsoft/onnxruntime-genai] Compile the model online in EP (Discussion #1111)
Update: ORT GenAI now supports OpenVINO EP, please use the latest commits from https://github.com/microsoft/onnxruntime main branch
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: ***@***.***>
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello, some of the backends of onnxruntime require online compilation of models, such as openvino. Some CPUs with poor online compilation of large models may take more than 30 minutes. Is there any solution for this? Or are there any requirements for supporting such backends?
Beta Was this translation helpful? Give feedback.
All reactions