You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
(ChatGLM) PS D:\Code\Python\ChatGLM-6B\ChatGLM-webui> python .\webui.py --model-path ..\THUDM\chatglm-6b-int4-qe\
Explicitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision.
Explicitly passing a `revision` is encouraged when loading a configuration with custom code to ensure no malicious code has been contributed in a newer revision.
Explicitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision.
No compiled kernel found.
Compiling kernels : C:\Users\lxy52\.cache\huggingface\modules\transformers_modules\quantization_kernels_parallel.c
Compiling gcc -O3 -fPIC -pthread -fopenmp -std=c99 C:\Users\lxy52\.cache\huggingface\modules\transformers_modules\quantization_kernels_parallel.c -shared -o C:\Users\lxy52\.cache\huggingface\modules\transformers_modules\quantization_kernels_parallel.so
Kernels compiled : C:\Users\lxy52\.cache\huggingface\modules\transformers_modules\quantization_kernels_parallel.so
Cannot load cpu kernel, don't use quantized model on cpu.Using quantization cacheApplying quantization to glm layersGPU memory: 8.59 GBNo compiled kernel found.Compiling kernels : C:\Users\lxy52\.cache\huggingface\modules\transformers_modules\quantization_kernels_parallel.cCompiling gcc -O3 -fPIC -pthread -fopenmp -std=c99 C:\Users\lxy52\.cache\huggingface\modules\transformers_modules\quantization_kernels_parallel.c -shared -o C:\Users\lxy52\.cache\huggingface\modules\transformers_modules\quantization_kernels_parallel.soKernels compiled : C:\Users\lxy52\.cache\huggingface\modules\transformers_modules\quantization_kernels_parallel.soTraceback (most recent call last): File "D:\Code\Python\ChatGLM-6B\ChatGLM-webui\webui.py", line 52, in <module> init() File "D:\Code\Python\ChatGLM-6B\ChatGLM-webui\webui.py", line 24, in init load_model() File "D:\Code\Python\ChatGLM-6B\ChatGLM-webui\modules\model.py", line 61, in load_model prepare_model() File "D:\Code\Python\ChatGLM-6B\ChatGLM-webui\modules\model.py", line 42, in prepare_model model = model.half().quantize(4).cuda() File "C:\Users\lxy52/.cache\huggingface\modules\transformers_modules\modeling_chatglm.py", line 1281, in quantize load_cpu_kernel(**kwargs) File "C:\Users\lxy52/.cache\huggingface\modules\transformers_modules\quantization.py", line 390, in load_cpu_kernel cpu_kernels = CPUKernel(**kwargs) File "C:\Users\lxy52/.cache\huggingface\modules\transformers_modules\quantization.py", line 157, in __init__ kernels = ctypes.cdll.LoadLibrary(kernel_file) File "D:\Application\Miniconda3\envs\ChatGLM\lib\ctypes\__init__.py", line 452, in LoadLibrary return self._dlltype(name) self._handle = _dlopen(self._name, mode)rnels_parallel.so' (or one of its dependencies). Try using the full path with constructor syntax.
The text was updated successfully, but these errors were encountered:
目录如下:
在
/d/Code/Python/ChatGLM-6B/ChatGLM-webui
下运行python .\webui.py --model-path ..\THUDM\chatglm-6b-int4-qe\
会报如下错误,加载chatglm-6b-int4也会报错, 是目录不能用回退的方式加载么?还是什么原因,上周某个版本就可以的,更新了后就不行了The text was updated successfully, but these errors were encountered: