| import os |
| from time import sleep |
|
|
| try: |
| from google import genai |
| from google.genai.types import GenerateContentConfigDict, ThinkingConfig |
| except ImportError as e: |
| pass |
|
|
| from lcb_runner.runner.base_runner import BaseRunner |
| from lcb_runner.lm_styles import LMStyle |
|
|
|
|
| class GeminiRunner(BaseRunner): |
| client = genai.Client( |
| |
| vertexai=True, |
| project=os.getenv("VERTEX_GEMINI_PROJECT"), |
| location=os.getenv("VERTEX_GEMINI_LOCATION"), |
| ) |
| safety_settings = [ |
| { |
| "category": "HARM_CATEGORY_HARASSMENT", |
| "threshold": "BLOCK_NONE", |
| }, |
| { |
| "category": "HARM_CATEGORY_HATE_SPEECH", |
| "threshold": "BLOCK_NONE", |
| }, |
| { |
| "category": "HARM_CATEGORY_SEXUALLY_EXPLICIT", |
| "threshold": "BLOCK_NONE", |
| }, |
| { |
| "category": "HARM_CATEGORY_DANGEROUS_CONTENT", |
| "threshold": "BLOCK_NONE", |
| }, |
| ] |
|
|
| def __init__(self, args, model): |
| super().__init__(args, model) |
| self.args = args |
| self.model = model |
| if self.model.model_style == LMStyle.GeminiThinking: |
| self.generation_config = GenerateContentConfigDict( |
| |
| |
| |
| |
| |
| safety_settings=GeminiRunner.safety_settings, |
| thinking_config=ThinkingConfig(include_thoughts=True), |
| ) |
| print("GeminiThinking model") |
| else: |
| self.generation_config = GenerateContentConfigDict( |
| max_output_tokens=args.max_tokens, |
| temperature=args.temperature, |
| top_p=args.top_p, |
| safety_settings=GeminiRunner.safety_settings, |
| candidate_count=args.n, |
| ) |
|
|
| def _run_single(self, prompt: str) -> list[str]: |
|
|
| try: |
| outputs = self.client.models.generate_content( |
| model=self.model.model_name, |
| contents=prompt, |
| config=self.generation_config, |
| ).candidates |
|
|
| if outputs is None: |
| print("No outputs from Gemini") |
| return ["" for _ in range(self.args.n)] |
| except Exception as e: |
| print("Exception: ", repr(e)) |
| print("Sleeping for 30 seconds...") |
| print("Consider reducing the number of parallel processes.") |
| sleep(30) |
| return self._run_single(prompt) |
|
|
| new_outputs = [] |
| for output in outputs: |
| try: |
| texts = [part.text for part in output.content.parts] |
| texts = [ |
| "## Part " + str(i) + "\n" + text for i, text in enumerate(texts) |
| ] |
| text = "\n\n\n".join(texts) |
| if text == "": |
| print("Empty text for output") |
| print(output.__dict__) |
| new_outputs.append(text) |
| except Exception as e: |
| print("Cannot extract text exception: ", repr(e)) |
| print(output.__dict__) |
| new_outputs.append("") |
| outputs = new_outputs |
|
|
| return outputs |
|
|