-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
failed to create the serving #17
Comments
Please check your IPEX version. The trace shows that the IPEX version is mismatch with Pytorch. PyTorch version 2.2.0 |
the issue is solved by install oneccl. |
Hi, thanks for bringing this up and for the suggestion! We'll look into adding the packages you listed as well as whatever other ones are missing |
It seems like you're facing an issue with setting up a neural chat server. To resolve this, you need to ensure that you have the required packages installed and properly configured. Here are some steps to follow:
|
I tried to create the serving on my system, but failed with the below error:
(emon_analyzer) [root@SPR-1 emon_data_analyzer]# neuralchat_server start --config_file ./config/neuralchat.yaml
2024-03-19 11:38:57,005 - numexpr.utils - INFO - Note: detected 224 virtual cores but NumExpr set to maximum of 64, check "NUMEXPR_MAX_THREADS" environment variable.
2024-03-19 11:38:57,005 - numexpr.utils - INFO - Note: NumExpr detected 224 cores but "NUMEXPR_MAX_THREADS" not set, so enforcing safe limit of 8.
2024-03-19 11:38:57,005 - numexpr.utils - INFO - NumExpr defaulting to 8 threads.
2024-03-19 11:38:57,348 - datasets - INFO - PyTorch version 2.2.0+cpu available.
[2024-03-19 11:38:57,430] [ ERROR] - Failed to start server.
[2024-03-19 11:38:57,430] [ ERROR] - partially initialized module 'intel_extension_for_pytorch' has no attribute '_C' (most likely due to a circular import)
yaml config file:
host: 0.0.0.0
port: 8000
model_name_or_path: "Intel/neural-chat-7b-v3-1"
#model_name_or_path: "/home/zluo2/TableLlama-model"
#tokenizer_name_or_path: ""
#peft_model_path: "./models/emon_llama"
device: "cpu"
asr:
enable: false
args:
# support cpu, hpu, xpu, cuda
device: "cpu"
# support openai/whisper series
model_name_or_path: "openai/whisper-small"
# only can be set to true when the device is set to "cpu"
bf16: false
tts:
enable: false
args:
device: "cpu"
voice: "default"
stream_mode: false
output_audio_path: "./output_audio.wav"
asr_chinese:
enable: false
tts_chinese:
enable: false
args:
device: "cpu"
spk_id: 0
stream_mode: false
output_audio_path: "./output_audio.wav"
retrieval:
enable: true
args:
input_path: "./rag_data/emon-sample"
#vector_database: "Qdrant"
#retrieval_type: "bm25"
safety_checker:
enable: false
ner:
enable: false
args:
spacy_model: "en_core_web_lg"
tasks_list: ['textchat', 'retrieval']
The text was updated successfully, but these errors were encountered: