jankais3r / LLaMA_MPS

Run LLaMA (and Stanford-Alpaca) inference on Apple Silicon GPUs.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

op Height/Width dimensions must be less than 16384

itrcz opened this issue · comments

MacOS 14.0
MacBook Pro M1 Max

autocomplete and instruction-response give same result:

$python3 chat.py --ckpt_dir models/7B-alpaca --tokenizer_path models/tokenizer.model --max_batch_size 8 --max_seq_len 256
Seed: 30112
Loading checkpoint
Loaded in 94.05 seconds
Running the fine-tuned 'alpaca' model in an instruction-response mode.
Instruction: hello
loc("mps_transpose"("(mpsFileLoc): /AppleInternal/Library/BuildRoots/d8ee83b8-11b4-11ee-a66d-46d450270006/Library/Caches/com.apple.xbs/Sources/MetalPerformanceShadersGraph/mpsgraph/MetalPerformanceShadersGraph/Core/Files/MPSGraphUtilities.mm":206:0)): error: 'anec.transpose' op Height/Width dimensions must be less than 16384
Response: hello, hello

seems like it not works as expected. I have waited couple minutes and I got just "hello, hello"