From d19bf86db2fcca87bd4156ea2b1edcd16b5fe811 Mon Sep 17 00:00:00 2001 From: Aakriti Kinra <52823721+AakritiKinra@users.noreply.github.com> Date: Wed, 1 Jan 2025 16:08:43 +0530 Subject: [PATCH] Update clm.py --- llments/eval/factscore/clm.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/llments/eval/factscore/clm.py b/llments/eval/factscore/clm.py index 92b590f..36e24e6 100644 --- a/llments/eval/factscore/clm.py +++ b/llments/eval/factscore/clm.py @@ -64,10 +64,10 @@ def load_model(self) -> None: def _generate( self, prompt: str, + prompts: Union[str, List[str]], sample_idx: int = 0, max_sequence_length: int = 2048, max_output_length: int = 128, - prompts: Union[str, List[str]], end_if_newline: bool = False, end_if_second_newline: bool = False, verbose: bool = False, @@ -76,8 +76,8 @@ def _generate( Args: prompt (str): The input prompt to generate text from. - sample_idx (int, optional): Index to differentiate between samples. Defaults to 0. prompts (Union[str, List[str]]): Single prompt string or a list of prompt strings. + sample_idx (int, optional): Index to differentiate between samples. Defaults to 0. max_sequence_length (int, optional): Maximum length of the input sequence. Defaults to 2048. max_output_length (int, optional): Maximum length of the generated output.