@@ -240,7 +240,7 @@ async def test_anthropic_vertex_ai_prompt_caching(anthropic_messages, sync_mode)
240240async def test_anthropic_api_prompt_caching_basic ():
241241 litellm .set_verbose = True
242242 response = await litellm .acompletion (
243- model = "anthropic/claude-3-5 -sonnet-20241022 " ,
243+ model = "anthropic/claude-3-7 -sonnet-20250219 " ,
244244 messages = [
245245 # System Message
246246 {
@@ -309,7 +309,7 @@ async def test_anthropic_api_prompt_caching_basic_with_cache_creation():
309309
310310 litellm .set_verbose = True
311311 response = await litellm .acompletion (
312- model = "anthropic/claude-3-5 -sonnet-20241022 " ,
312+ model = "anthropic/claude-3-7 -sonnet-20250219 " ,
313313 messages = [
314314 # System Message
315315 {
@@ -462,7 +462,7 @@ async def test_anthropic_api_prompt_caching_with_content_str():
462462async def test_anthropic_api_prompt_caching_no_headers ():
463463 litellm .set_verbose = True
464464 response = await litellm .acompletion (
465- model = "anthropic/claude-3-5 -sonnet-20241022 " ,
465+ model = "anthropic/claude-3-7 -sonnet-20250219 " ,
466466 messages = [
467467 # System Message
468468 {
@@ -522,7 +522,7 @@ async def test_anthropic_api_prompt_caching_no_headers():
522522@pytest .mark .asyncio ()
523523async def test_anthropic_api_prompt_caching_streaming ():
524524 response = await litellm .acompletion (
525- model = "anthropic/claude-3-5 -sonnet-20241022 " ,
525+ model = "anthropic/claude-3-7 -sonnet-20250219 " ,
526526 messages = [
527527 # System Message
528528 {
0 commit comments