From bca1da489f5920e5b4a1b99ed5e7927df97e5870 Mon Sep 17 00:00:00 2001 From: Lorenzo Pozzi Date: Thu, 29 Feb 2024 15:45:13 +0100 Subject: [PATCH] Small fixes test llama-2-7B-GPTQ --- tests/configs/cuda_inference_pytorch_llama_gptq.yaml | 6 +----- 1 file changed, 1 insertion(+), 5 deletions(-) diff --git a/tests/configs/cuda_inference_pytorch_llama_gptq.yaml b/tests/configs/cuda_inference_pytorch_llama_gptq.yaml index 126e5b3a..6eabf834 100644 --- a/tests/configs/cuda_inference_pytorch_llama_gptq.yaml +++ b/tests/configs/cuda_inference_pytorch_llama_gptq.yaml @@ -7,16 +7,13 @@ defaults: - override hydra/job_logging: colorlog # colorful logging - override hydra/hydra_logging: colorlog # colorful logging -experiment_name: cuda_inference_pytorch_llama_gptq +experiment_name: cuda_inference_pytorch_llama_gptq backend: device: cuda device_ids: 0 model: TheBloke/Llama-2-7B-GPTQ - # hub_kwargs: - # trust_remote_code: true quantization_config: - # disable_exllama: true exllama_config: version: 2 @@ -26,7 +23,6 @@ launcher: benchmark: memory: true latency: true - energy: true # hydra/cli specific settings hydra: