Mozilla-Ocho / llamafile

Distribute and run LLMs with a single file.

Home Page:https://llamafile.ai

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Signal SIGILL (Illegal Instructions)

DjagbleyEmmanuel opened this issue · comments

commented

llama_new_context_with_model: n_ctx = 2048
llama_new_context_with_model: n_batch = 2048
llama_new_context_with_model: n_ubatch = 512
llama_new_context_with_model: freq_base = 10000.0
llama_new_context_with_model: freq_scale = 1
llama_kv_cache_init: CPU KV buffer size = 384.00 MiB
llama_new_context_with_model: KV self size = 384.00 MiB, K (f16): 192.00 MiB, V (f16): 192.00 MiB
llama_new_context_with_model: CPU output buffer size = 0.20 MiB
llama_new_context_with_model: CPU compute buffer size = 160.01 MiB
llama_new_context_with_model: graph nodes = 921
llama_new_context_with_model: graph splits = 1
fish: Job 1, './moondream2-q8.llamafile -ngl…' terminated by signal SIGILL (Illegal instruction)

And here's What I got On my CPU and the AI model 

system_info":"AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LAMMAFILE = 1 | ","tid":"9430432","timestamp":1714936052,"total_threads":8}
{"function":"load_model","level":"INFO","line":432,"msg":"Multi Modal Mode Enabled","tid":"9430432","timestamp":1714936052}
clip_model_load: model name: vikhyatk/moondream2
clip_model_load: description: image encoder for vikhyatk/moondream2
clip_model_load: GGUF version: 3