Skip to content

Commit

Permalink
Small fixes test llama-2-7B-GPTQ
Browse files Browse the repository at this point in the history
  • Loading branch information
lopozz committed Feb 29, 2024
1 parent 48cb08c commit bca1da4
Showing 1 changed file with 1 addition and 5 deletions.
6 changes: 1 addition & 5 deletions tests/configs/cuda_inference_pytorch_llama_gptq.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,16 +7,13 @@ defaults:
- override hydra/job_logging: colorlog # colorful logging
- override hydra/hydra_logging: colorlog # colorful logging

experiment_name: cuda_inference_pytorch_llama_gptq
experiment_name: cuda_inference_pytorch_llama_gptq

backend:
device: cuda
device_ids: 0
model: TheBloke/Llama-2-7B-GPTQ
# hub_kwargs:
# trust_remote_code: true
quantization_config:
# disable_exllama: true
exllama_config:
version: 2

Expand All @@ -26,7 +23,6 @@ launcher:
benchmark:
memory: true
latency: true
energy: true

# hydra/cli specific settings
hydra:
Expand Down

0 comments on commit bca1da4

Please sign in to comment.