Skip to content
This repository was archived by the owner on Oct 25, 2024. It is now read-only.

update ipex api #1650

Draft
wants to merge 2 commits into
base: main
Choose a base branch
from
Draft

update ipex api #1650

wants to merge 2 commits into from

Conversation

LJ-underdog
Copy link
Contributor

Type of Change

feature or bug fix or documentation or others
API changed or not
yes

Description

update ipex api
detail description
image

Signed-off-by: intellinjun <jun.lin@intel.com>
Copy link

github-actions bot commented Jul 3, 2024

⛈️ Required checks status: Has failure 🔴

Warning
If you do not have the access to re-run the CI-Summary bot, please contact VincyZhang for help. If you push a new commit, all of the workflow will be re-triggered.

Groups summary

🟢 Format Scan Tests workflow
Check ID Status Error details
format-scan (pylint) success
format-scan (bandit) success
format-scan (cloc) success
format-scan (cpplint) success

These checks are required after the changes to intel_extension_for_transformers/neural_chat/models/model_utils.py.

🔴 NeuralChat Unit Test
Check ID Status Error details
neuralchat-unit-test-baseline cancelled 🚫
neuralchat-unit-test-PR-test failure download
Generate-NeuralChat-Report skipped

These checks are required after the changes to intel_extension_for_transformers/neural_chat/models/model_utils.py.

🟡 Chat Bot Test workflow
Check ID Status Error details
call-inference-llama-2-7b-chat-hf / inference test no_status
call-inference-mpt-7b-chat / inference test no_status

These checks are required after the changes to intel_extension_for_transformers/neural_chat/models/model_utils.py.


Thank you for your contribution! 💜

Note
This comment is automatically generated and will be updates every 180 seconds within the next 6 hours. If you have any other questions, contact VincyZhang or XuehaoSun for help.

@a32543254 a32543254 requested a review from changwangss July 3, 2024 02:15
@a32543254
Copy link
Contributor

@changwangss could you help review this pr?

model.eval(),
dtype=torch_dtype,
inplace=True,
level="O1",
auto_kernel_selection=True,
)
except AssertionError:
model = intel_ipex.optimize(
Copy link
Contributor

@changwangss changwangss Jul 3, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Could you also change it (Line 849)? as I know, ipex.llm.optimize is recommanded to run bf16 LLMs inference by IPEX.

Copy link
Contributor

@a32543254 a32543254 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Signed-off-by: intellinjun <jun.lin@intel.com>
@a32543254 a32543254 marked this pull request as draft July 5, 2024 03:15
@a32543254
Copy link
Contributor

convert to draft since the intel_ipex.llm.optimize will make the model cannot run.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
3 participants