| |
| """ |
| Test Azure OpenAI LLM deployment with current API version. |
| """ |
| import os |
| from openai import AzureOpenAI |
| from dotenv import load_dotenv |
|
|
| load_dotenv() |
|
|
| def test_llm_deployment(): |
| """Test LLM deployment with current API version.""" |
| print("=" * 80) |
| print("Testing Azure OpenAI LLM Deployment") |
| print("=" * 80) |
| print() |
|
|
| endpoint = os.getenv("AZURE_OPENAI_ENDPOINT") |
| api_key = os.getenv("AZURE_OPENAI_API_KEY") |
| deployment_name = os.getenv("AZURE_OPENAI_DEPLOYMENT_NAME") |
| api_version = os.getenv("AZURE_OPENAI_API_VERSION", "2024-02-01") |
|
|
| print(f"Endpoint: {endpoint}") |
| print(f"Deployment: {deployment_name}") |
| print(f"API Version: {api_version}") |
| print() |
| print("Sending test request...") |
| print() |
|
|
| try: |
| client = AzureOpenAI( |
| api_key=api_key, |
| api_version=api_version, |
| azure_endpoint=endpoint |
| ) |
|
|
| response = client.chat.completions.create( |
| model=deployment_name, |
| messages=[ |
| {"role": "system", "content": "You are a helpful assistant."}, |
| {"role": "user", "content": "Say 'Hello, world!' if you can read this."} |
| ], |
| temperature=0, |
| max_tokens=50 |
| ) |
|
|
| message = response.choices[0].message.content |
| tokens_used = response.usage.total_tokens |
|
|
| print(f"β
SUCCESS: LLM responded successfully!") |
| print(f" Response: {message}") |
| print(f" Model: {deployment_name}") |
| print(f" Tokens used: {tokens_used}") |
| print(f" API Version: {api_version}") |
| print() |
| print("=" * 80) |
| print("β
LLM deployment works with API version:", api_version) |
| print("=" * 80) |
| return True |
|
|
| except Exception as e: |
| error_msg = str(e) |
| print(f"β ERROR: LLM request failed") |
| print() |
| print(f"Error message: {error_msg}") |
| print() |
|
|
| if "404" in error_msg or "Resource not found" in error_msg: |
| print("DIAGNOSIS: Deployment not found with API version", api_version) |
| print() |
| print("Possible solutions:") |
| print(" 1. Your LLM deployment might require a different API version") |
| print(" 2. Try API version 2024-07-18 for gpt-4o-mini") |
| print(" 3. You may need separate API versions for LLM vs embeddings") |
| print() |
| elif "401" in error_msg: |
| print("DIAGNOSIS: Authentication failed") |
| print() |
|
|
| print("=" * 80) |
| print("β LLM deployment test FAILED") |
| print("=" * 80) |
| return False |
|
|
| if __name__ == "__main__": |
| test_llm_deployment() |
|
|