Advertisement
Guest User

Untitled

a guest
Apr 5th, 2023
15
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 1.09 KB | None | 0 0
  1. .\gpt4all-lora-quantized-win64.exe -m ..\..\gpt4all-lora-unfiltered-quantized.bin -n 64 --color -p "Hello, is this thing on?"
  2. main: seed = 1680685727
  3. llama_model_load: loading model from '..\..\gpt4all-lora-unfiltered-quantized.bin' - please wait ...
  4. llama_model_load: ggml ctx size = 6065.35 MB
  5. llama_model_load: memory_size = 2048.00 MB, n_mem = 65536
  6. llama_model_load: loading model part 1/1 from '..\..\gpt4all-lora-unfiltered-quantized.bin'
  7. llama_model_load: .................................... done
  8. llama_model_load: model size = 4017.27 MB / num tensors = 291
  9.  
  10. system_info: n_threads = 4 / 12 | AVX = 1 | AVX2 = 1 | AVX512 = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 0 | VSX = 0 |
  11. sampling parameters: temp = 0.100000, top_k = 40, top_p = 0.950000, repeat_last_n = 64, repeat_penalty = 1.300000
  12.  
  13.  
  14. Yes it's working fine!
  15. [end of text]
  16.  
  17.  
  18. main: mem per token = 14368648 bytes
  19. main: load time = 1357.89 ms
  20. main: sample time = 3.00 ms
  21. main: predict time = 7961.98 ms / 153.11 ms per token
  22. main: total time = 9898.32 ms
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement