|
31 | 31 | from nemoguardrails.logging.callbacks import LoggingCallbackHandler |
32 | 32 | from nemoguardrails.logging.explain import ExplainInfo, LLMCallInfo |
33 | 33 | from nemoguardrails.logging.stats import LLMStats |
| 34 | +from nemoguardrails.logging.utils import extract_model_name_and_base_url |
34 | 35 |
|
35 | 36 |
|
36 | 37 | @pytest.mark.asyncio |
@@ -261,3 +262,122 @@ def __init__(self, content, msg_type): |
261 | 262 | assert logged_prompt is not None |
262 | 263 | assert "[cyan]Custom[/]" in logged_prompt |
263 | 264 | assert "[cyan]Function[/]" in logged_prompt |
| 265 | + |
| 266 | + |
| 267 | +def test_extract_model_and_url_from_kwargs(): |
| 268 | + """Test extracting model_name and openai_api_base from kwargs (ChatOpenAI case).""" |
| 269 | + serialized = { |
| 270 | + "kwargs": { |
| 271 | + "model_name": "gpt-4", |
| 272 | + "openai_api_base": "https://api.openai.com/v1", |
| 273 | + "temperature": 0.7, |
| 274 | + } |
| 275 | + } |
| 276 | + |
| 277 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 278 | + |
| 279 | + assert model_name == "gpt-4" |
| 280 | + assert base_url == "https://api.openai.com/v1" |
| 281 | + |
| 282 | + |
| 283 | +def test_extract_model_and_url_from_repr(): |
| 284 | + """Test extracting from repr string (ChatNIM case).""" |
| 285 | + # Property values in single-quotes |
| 286 | + serialized = { |
| 287 | + "kwargs": {"temperature": 0.1}, |
| 288 | + "repr": "ChatNIM(model='meta/llama-3.3-70b-instruct', client=<openai.OpenAI object at 0x10d8e4e90>, endpoint_url='https://nim.int.aire.nvidia.com/v1')", |
| 289 | + } |
| 290 | + |
| 291 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 292 | + |
| 293 | + assert model_name == "meta/llama-3.3-70b-instruct" |
| 294 | + assert base_url == "https://nim.int.aire.nvidia.com/v1" |
| 295 | + |
| 296 | + # Property values in double-quotes |
| 297 | + serialized = { |
| 298 | + "repr": 'ChatOpenAI(model="gpt-3.5-turbo", base_url="https://custom.api.com/v1")' |
| 299 | + } |
| 300 | + |
| 301 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 302 | + |
| 303 | + assert model_name == "gpt-3.5-turbo" |
| 304 | + assert base_url == "https://custom.api.com/v1" |
| 305 | + |
| 306 | + # Model is stored in the `model_name` property |
| 307 | + serialized = { |
| 308 | + "repr": "SomeProvider(model_name='custom-model-v2', api_base='https://example.com')" |
| 309 | + } |
| 310 | + |
| 311 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 312 | + |
| 313 | + assert model_name == "custom-model-v2" |
| 314 | + assert base_url == "https://example.com" |
| 315 | + |
| 316 | + |
| 317 | +def test_extract_model_and_url_from_various_url_properties(): |
| 318 | + """Test extracting various URL property names.""" |
| 319 | + test_cases = [ |
| 320 | + ("api_base='https://api1.com'", "https://api1.com"), |
| 321 | + ("api_host='https://api2.com'", "https://api2.com"), |
| 322 | + ("azure_endpoint='https://azure.com'", "https://azure.com"), |
| 323 | + ("endpoint='https://endpoint.com'", "https://endpoint.com"), |
| 324 | + ("openai_api_base='https://openai.com'", "https://openai.com"), |
| 325 | + ] |
| 326 | + |
| 327 | + for url_pattern, expected_url in test_cases: |
| 328 | + serialized = {"repr": f"Provider(model='test-model', {url_pattern})"} |
| 329 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 330 | + assert base_url == expected_url, f"Failed for pattern: {url_pattern}" |
| 331 | + |
| 332 | + |
| 333 | +def test_extract_model_and_url_kwargs_priority_over_repr(): |
| 334 | + """Test that kwargs values, if present, take priority over repr values.""" |
| 335 | + serialized = { |
| 336 | + "kwargs": { |
| 337 | + "model_name": "gpt-4-from-kwargs", |
| 338 | + "openai_api_base": "https://kwargs.api.com", |
| 339 | + }, |
| 340 | + "repr": "ChatOpenAI(model='gpt-3.5-from-repr', base_url='https://repr.api.com')", |
| 341 | + } |
| 342 | + |
| 343 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 344 | + |
| 345 | + assert model_name == "gpt-4-from-kwargs" |
| 346 | + assert base_url == "https://kwargs.api.com" |
| 347 | + |
| 348 | + |
| 349 | +def test_extract_model_and_url_with_missing_values(): |
| 350 | + """Test extraction when values are missing.""" |
| 351 | + # No model or URL |
| 352 | + serialized = {"kwargs": {"temperature": 0.7}} |
| 353 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 354 | + assert model_name is None |
| 355 | + assert base_url is None |
| 356 | + |
| 357 | + # Only model, no URL |
| 358 | + serialized = {"kwargs": {"model_name": "gpt-4"}} |
| 359 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 360 | + assert model_name == "gpt-4" |
| 361 | + assert base_url is None |
| 362 | + |
| 363 | + # Only URL, no model |
| 364 | + serialized = {"repr": "Provider(endpoint_url='https://example.com')"} |
| 365 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 366 | + assert model_name is None |
| 367 | + assert base_url == "https://example.com" |
| 368 | + |
| 369 | + |
| 370 | +def test_extract_model_and_url_with_empty_values(): |
| 371 | + """Test extraction when values are empty strings.""" |
| 372 | + serialized = {"kwargs": {"model_name": "", "openai_api_base": ""}} |
| 373 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 374 | + assert model_name is None |
| 375 | + assert base_url is None |
| 376 | + |
| 377 | + |
| 378 | +def test_extract_model_and_url_with_empty_serialized_data(): |
| 379 | + """Test extraction with empty or minimal serialized dict.""" |
| 380 | + serialized = {} |
| 381 | + model_name, base_url = extract_model_name_and_base_url(serialized) |
| 382 | + assert model_name is None |
| 383 | + assert base_url is None |
0 commit comments