From ea981473f0c1cdf8c5deddf2c0202ac69961070e Mon Sep 17 00:00:00 2001 From: Andreea Popescu Date: Thu, 5 Sep 2024 21:19:53 +0800 Subject: [PATCH] wip --- src/compute_horde_prompt_gen/run.py | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/src/compute_horde_prompt_gen/run.py b/src/compute_horde_prompt_gen/run.py index c55d64d..b1e2272 100644 --- a/src/compute_horde_prompt_gen/run.py +++ b/src/compute_horde_prompt_gen/run.py @@ -36,6 +36,10 @@ def generate_prompts( max_new_tokens=max_new_tokens, temperature=temperature, ) + print("----------") + print(sequences) + print("----------") + seconds_taken = (datetime.datetime.now() - start_ts).total_seconds() log.info(f"{i=} generation took {seconds_taken:.2f}s") @@ -44,7 +48,7 @@ def generate_prompts( # output = model.decode(sequence) # log.info(f"\n\n{i=} output={output}\n\n") generated_prompts = parse_output(sequence) - log.debug(f"{i=} sequence={j} {generated_prompts=} from {output=}") + log.debug(f"{i=} sequence={j} {generated_prompts=} from {sequence=}") log.info(f"{i=} sequence={j} generated {len(generated_prompts)} prompts") new_prompts.extend(generated_prompts)