Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- .\gpt4all-lora-quantized-win64.exe -m ..\..\gpt4all-lora-unfiltered-quantized.bin -n 64 --color -p "Hello, is this thing on?"
- main: seed = 1680685727
- llama_model_load: loading model from '..\..\gpt4all-lora-unfiltered-quantized.bin' - please wait ...
- llama_model_load: ggml ctx size = 6065.35 MB
- llama_model_load: memory_size = 2048.00 MB, n_mem = 65536
- llama_model_load: loading model part 1/1 from '..\..\gpt4all-lora-unfiltered-quantized.bin'
- llama_model_load: .................................... done
- llama_model_load: model size = 4017.27 MB / num tensors = 291
- system_info: n_threads = 4 / 12 | AVX = 1 | AVX2 = 1 | AVX512 = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 0 | VSX = 0 |
- sampling parameters: temp = 0.100000, top_k = 40, top_p = 0.950000, repeat_last_n = 64, repeat_penalty = 1.300000
- Yes it's working fine!
- [end of text]
- main: mem per token = 14368648 bytes
- main: load time = 1357.89 ms
- main: sample time = 3.00 ms
- main: predict time = 7961.98 ms / 153.11 ms per token
- main: total time = 9898.32 ms
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement