Yash Sachdeva commited on
Commit
38fc296
·
1 Parent(s): 52de7f3

smallest llama model

Browse files
Files changed (2) hide show
  1. Dockerfile +1 -1
  2. question_paper.py +1 -1
Dockerfile CHANGED
@@ -32,7 +32,7 @@ ENV HOME=/home/user \
32
 
33
  WORKDIR $HOME/app
34
 
35
- RUN huggingface-cli download TheBloke/Llama-2-7b-Chat-GGUF llama-2-7b-chat.Q3_K_M.gguf --local-dir . --local-dir-use-symlinks False
36
 
37
 
38
  COPY --chown=user . $HOME/app
 
32
 
33
  WORKDIR $HOME/app
34
 
35
+ RUN huggingface-cli download TheBloke/Llama-2-7b-Chat-GGUF llama-2-7b-chat.Q2_K.gguf --local-dir . --local-dir-use-symlinks False
36
 
37
 
38
  COPY --chown=user . $HOME/app
question_paper.py CHANGED
@@ -10,7 +10,7 @@ app = FastAPI()
10
  @app.get("/")
11
  def llama():
12
  llm = Llama(
13
- model_path="./llama-2-7b-chat.Q3_K_M.gguf"
14
  # n_gpu_layers=-1, # Uncomment to use GPU acceleration
15
  # seed=1337, # Uncomment to set a specific seed
16
  # n_ctx=2048, # Uncomment to increase the context window
 
10
  @app.get("/")
11
  def llama():
12
  llm = Llama(
13
+ model_path="./llama-2-7b-chat.Q2_K.gguf"
14
  # n_gpu_layers=-1, # Uncomment to use GPU acceleration
15
  # seed=1337, # Uncomment to set a specific seed
16
  # n_ctx=2048, # Uncomment to increase the context window