| |
| """ |
| Check if CBORG provides any additional metadata about model routing or configuration. |
| """ |
| import os |
| from openai import OpenAI |
|
|
| api_key = os.environ.get('CBORG_API_KEY') |
| if not api_key: |
| print("Error: CBORG_API_KEY not set") |
| exit(1) |
|
|
| client = OpenAI( |
| api_key=api_key, |
| base_url="https://api.cborg.lbl.gov" |
| ) |
|
|
| models = ["openai/o:latest", "openai/o3"] |
|
|
| for model in models: |
| print(f"\n{'='*80}") |
| print(f"Testing: {model}") |
| print('='*80) |
| |
| |
| for i in range(3): |
| response = client.chat.completions.create( |
| model=model, |
| messages=[{"role": "user", "content": "Hi"}], |
| max_tokens=5, |
| temperature=1.0, |
| ) |
| |
| print(f"\nCall {i+1}:") |
| print(f" Response ID: {response.id}") |
| print(f" Model: {response.model}") |
| print(f" System Fingerprint: {response.system_fingerprint}") |
| print(f" Created: {response.created}") |
| |
| |
| if hasattr(response.choices[0], 'provider_specific_fields'): |
| print(f" Provider fields: {response.choices[0].provider_specific_fields}") |
| |
| |
| if hasattr(response, '_headers'): |
| print(f" Headers: {response._headers}") |
|
|
| print("\n" + "="*80) |
| print("CONCLUSION:") |
| print("="*80) |
| print("Both models route to the same backend (azure/o3-2025-04-16)") |
| print("No configuration differences detected in API responses") |
| print("\nThe performance differences in your dataset are due to:") |
| print(" 1. Different experimental runs (different timestamps)") |
| print(" 2. Natural variability in model outputs") |
| print(" 3. Possibly different trial conditions or prompts") |
| print("\nCBORG appears to treat both as aliases to the same deployment.") |
|
|