|
15 | 15 | constructor = @azureChat; |
16 | 16 | defaultModel = azureChat; |
17 | 17 | visionModel = azureChat(Deployment="gpt-4o"); |
18 | | - structuredModel = azureChat("APIVersion","2024-08-01-preview",... |
19 | | - "Deployment","gpt-4o-2024-08-06"); |
| 18 | + structuredModel = azureChat("Deployment","gpt-4o-2024-08-06"); |
20 | 19 | end |
21 | 20 |
|
22 | 21 | methods(Test) |
@@ -63,7 +62,7 @@ function jsonFormatWithSystemPrompt(testCase) |
63 | 62 | end |
64 | 63 |
|
65 | 64 | function responseFormatRequiresNewAPI(testCase) |
66 | | - chat = azureChat; |
| 65 | + chat = azureChat(APIVersion="2024-02-01"); |
67 | 66 | testCase.verifyError(@() generate(chat, ... |
68 | 67 | "What is the smallest prime?", ... |
69 | 68 | ResponseFormat=struct("number",1)), ... |
@@ -127,7 +126,7 @@ function specialErrorForUnsupportedResponseFormat(testCase) |
127 | 126 | testCase.assumeFail("Disabled until `llms.internal.callAzureChat` is updated to use `max_completion_tokens` instead of the deprecated `max_tokens` in the OpenAI API.") |
128 | 127 |
|
129 | 128 | testCase.verifyError(@() generate(... |
130 | | - azureChat(APIVersion="2024-08-01-preview",DeploymentID="o1-mini"), ... |
| 129 | + azureChat(DeploymentID="o1-mini"), ... |
131 | 130 | "What is the smallest prime?", ... |
132 | 131 | ResponseFormat=struct("number",1)), ... |
133 | 132 | "llms:noStructuredOutputForAzureDeployment"); |
|
0 commit comments