intel-analytics / ipex-llm

Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, Axolotl, etc.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

intel gpu and ollama error

dayskk opened this issue · comments

微信图片_20240605135354

The following error occurred after running for a period of time, please refer to the attachment.
Currently, no reproduction method has been found.
GPU accelerates Ollama to run Qwen1.5 7b, and it appears while running