ffreemt commited on
Commit
adb51a2
·
1 Parent(s): c0f505d

Try Gemma-4-E4B-Uncensored-HauhauCS-Aggressive-Q5_K_P Dockfile --n-gpu-layers 99

Browse files
Files changed (1) hide show
  1. Dockerfile +3 -2
Dockerfile CHANGED
@@ -1,4 +1,4 @@
1
- FROM ghcr.io/ggml-org/llama.cpp:full
2
 
3
  WORKDIR /app
4
 
@@ -22,4 +22,5 @@ CMD ["--server", \
22
  "--cache-type-k", "q8_0", \
23
  "--cache-type-v", "iq4_nl", \
24
  "-c", "128000", \
25
- "-n", "38912"]
 
 
1
+ FROM ghcr.io/ggml-org/llama.cpp:full-cuda
2
 
3
  WORKDIR /app
4
 
 
22
  "--cache-type-k", "q8_0", \
23
  "--cache-type-v", "iq4_nl", \
24
  "-c", "128000", \
25
+ "-n", "38912", \
26
+ "--n-gpu-layers", "99"]