1412
1412
"deprecation_date" : " 2025-03-31" ,
1413
1413
"supports_tool_choice" : true
1414
1414
},
1415
+ "azure/gpt-3.5-turbo-0125" : {
1416
+ "max_tokens" : 4096 ,
1417
+ "max_input_tokens" : 16384 ,
1418
+ "max_output_tokens" : 4096 ,
1419
+ "input_cost_per_token" : 0.0000005 ,
1420
+ "output_cost_per_token" : 0.0000015 ,
1421
+ "litellm_provider" : " azure" ,
1422
+ "mode" : " chat" ,
1423
+ "supports_function_calling" : true ,
1424
+ "supports_parallel_function_calling" : true ,
1425
+ "deprecation_date" : " 2025-03-31" ,
1426
+ "supports_tool_choice" : true
1427
+ },
1415
1428
"azure/gpt-35-turbo-16k" : {
1416
1429
"max_tokens" : 4096 ,
1417
1430
"max_input_tokens" : 16385 ,
1433
1446
"supports_function_calling" : true ,
1434
1447
"supports_tool_choice" : true
1435
1448
},
1449
+ "azure/gpt-3.5-turbo" : {
1450
+ "max_tokens" : 4096 ,
1451
+ "max_input_tokens" : 4097 ,
1452
+ "max_output_tokens" : 4096 ,
1453
+ "input_cost_per_token" : 0.0000005 ,
1454
+ "output_cost_per_token" : 0.0000015 ,
1455
+ "litellm_provider" : " azure" ,
1456
+ "mode" : " chat" ,
1457
+ "supports_function_calling" : true ,
1458
+ "supports_tool_choice" : true
1459
+ },
1436
1460
"azure/gpt-3.5-turbo-instruct-0914" : {
1437
1461
"max_tokens" : 4097 ,
1438
1462
"max_input_tokens" : 4097 ,
3634
3658
"source" : " https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models" ,
3635
3659
"supports_tool_choice" : true
3636
3660
},
3661
+ "gemini-2.0-pro-exp-02-05" : {
3662
+ "max_tokens" : 8192 ,
3663
+ "max_input_tokens" : 2097152 ,
3664
+ "max_output_tokens" : 8192 ,
3665
+ "max_images_per_prompt" : 3000 ,
3666
+ "max_videos_per_prompt" : 10 ,
3667
+ "max_video_length" : 1 ,
3668
+ "max_audio_length_hours" : 8.4 ,
3669
+ "max_audio_per_prompt" : 1 ,
3670
+ "max_pdf_size_mb" : 30 ,
3671
+ "input_cost_per_image" : 0 ,
3672
+ "input_cost_per_video_per_second" : 0 ,
3673
+ "input_cost_per_audio_per_second" : 0 ,
3674
+ "input_cost_per_token" : 0 ,
3675
+ "input_cost_per_character" : 0 ,
3676
+ "input_cost_per_token_above_128k_tokens" : 0 ,
3677
+ "input_cost_per_character_above_128k_tokens" : 0 ,
3678
+ "input_cost_per_image_above_128k_tokens" : 0 ,
3679
+ "input_cost_per_video_per_second_above_128k_tokens" : 0 ,
3680
+ "input_cost_per_audio_per_second_above_128k_tokens" : 0 ,
3681
+ "output_cost_per_token" : 0 ,
3682
+ "output_cost_per_character" : 0 ,
3683
+ "output_cost_per_token_above_128k_tokens" : 0 ,
3684
+ "output_cost_per_character_above_128k_tokens" : 0 ,
3685
+ "litellm_provider" : " vertex_ai-language-models" ,
3686
+ "mode" : " chat" ,
3687
+ "supports_system_messages" : true ,
3688
+ "supports_function_calling" : true ,
3689
+ "supports_vision" : true ,
3690
+ "supports_audio_input" : true ,
3691
+ "supports_video_input" : true ,
3692
+ "supports_pdf_input" : true ,
3693
+ "supports_response_schema" : true ,
3694
+ "supports_tool_choice" : true ,
3695
+ "source" : " https://cloud.google.com/vertex-ai/generative-ai/pricing"
3696
+ },
3637
3697
"gemini-2.0-flash-exp" : {
3638
3698
"max_tokens" : 8192 ,
3639
3699
"max_input_tokens" : 1048576 ,
7035
7095
"input_cost_per_token" : 0.00000072 ,
7036
7096
"output_cost_per_token" : 0.00000072 ,
7037
7097
"litellm_provider" : " bedrock_converse" ,
7038
- "mode" : " chat"
7098
+ "mode" : " chat" ,
7099
+ "supports_function_calling" : true ,
7100
+ "supports_tool_choice" : false
7039
7101
},
7040
7102
"meta.llama2-13b-chat-v1" : {
7041
7103
"max_tokens" : 4096 ,
7375
7437
"supports_function_calling" : true ,
7376
7438
"supports_tool_choice" : false
7377
7439
},
7440
+ "us.meta.llama3-3-70b-instruct-v1:0" : {
7441
+ "max_tokens" : 4096 ,
7442
+ "max_input_tokens" : 128000 ,
7443
+ "max_output_tokens" : 4096 ,
7444
+ "input_cost_per_token" : 0.00000072 ,
7445
+ "output_cost_per_token" : 0.00000072 ,
7446
+ "litellm_provider" : " bedrock_converse" ,
7447
+ "mode" : " chat" ,
7448
+ "supports_function_calling" : true ,
7449
+ "supports_tool_choice" : false
7450
+ },
7378
7451
"512-x-512/50-steps/stability.stable-diffusion-xl-v0" : {
7379
7452
"max_tokens" : 77 ,
7380
7453
"max_input_tokens" : 77 ,
8084
8157
"input_cost_per_token" : 0.00000035 ,
8085
8158
"output_cost_per_token" : 0.00000140 ,
8086
8159
"litellm_provider" : " perplexity" ,
8087
- "mode" : " chat" ,
8088
- "supports_tool_choice" : true
8160
+ "mode" : " chat"
8089
8161
},
8090
8162
"perplexity/codellama-70b-instruct" : {
8091
8163
"max_tokens" : 16384 ,
8094
8166
"input_cost_per_token" : 0.00000070 ,
8095
8167
"output_cost_per_token" : 0.00000280 ,
8096
8168
"litellm_provider" : " perplexity" ,
8097
- "mode" : " chat" ,
8098
- "supports_tool_choice" : true
8169
+ "mode" : " chat"
8099
8170
},
8100
8171
"perplexity/llama-3.1-70b-instruct" : {
8101
8172
"max_tokens" : 131072 ,
8104
8175
"input_cost_per_token" : 0.000001 ,
8105
8176
"output_cost_per_token" : 0.000001 ,
8106
8177
"litellm_provider" : " perplexity" ,
8107
- "mode" : " chat" ,
8108
- "supports_tool_choice" : true
8178
+ "mode" : " chat"
8109
8179
},
8110
8180
"perplexity/llama-3.1-8b-instruct" : {
8111
8181
"max_tokens" : 131072 ,
8114
8184
"input_cost_per_token" : 0.0000002 ,
8115
8185
"output_cost_per_token" : 0.0000002 ,
8116
8186
"litellm_provider" : " perplexity" ,
8117
- "mode" : " chat" ,
8118
- "supports_tool_choice" : true
8187
+ "mode" : " chat"
8119
8188
},
8120
8189
"perplexity/llama-3.1-sonar-huge-128k-online" : {
8121
8190
"max_tokens" : 127072 ,
8125
8194
"output_cost_per_token" : 0.000005 ,
8126
8195
"litellm_provider" : " perplexity" ,
8127
8196
"mode" : " chat" ,
8128
- "deprecation_date" : " 2025-02-22" ,
8129
- "supports_tool_choice" : true
8197
+ "deprecation_date" : " 2025-02-22"
8130
8198
},
8131
8199
"perplexity/llama-3.1-sonar-large-128k-online" : {
8132
8200
"max_tokens" : 127072 ,
8136
8204
"output_cost_per_token" : 0.000001 ,
8137
8205
"litellm_provider" : " perplexity" ,
8138
8206
"mode" : " chat" ,
8139
- "deprecation_date" : " 2025-02-22" ,
8140
- "supports_tool_choice" : true
8207
+ "deprecation_date" : " 2025-02-22"
8141
8208
},
8142
8209
"perplexity/llama-3.1-sonar-large-128k-chat" : {
8143
8210
"max_tokens" : 131072 ,
8147
8214
"output_cost_per_token" : 0.000001 ,
8148
8215
"litellm_provider" : " perplexity" ,
8149
8216
"mode" : " chat" ,
8150
- "deprecation_date" : " 2025-02-22" ,
8151
- "supports_tool_choice" : true
8217
+ "deprecation_date" : " 2025-02-22"
8152
8218
},
8153
8219
"perplexity/llama-3.1-sonar-small-128k-chat" : {
8154
8220
"max_tokens" : 131072 ,
8158
8224
"output_cost_per_token" : 0.0000002 ,
8159
8225
"litellm_provider" : " perplexity" ,
8160
8226
"mode" : " chat" ,
8161
- "deprecation_date" : " 2025-02-22" ,
8162
- "supports_tool_choice" : true
8227
+ "deprecation_date" : " 2025-02-22"
8163
8228
},
8164
8229
"perplexity/llama-3.1-sonar-small-128k-online" : {
8165
8230
"max_tokens" : 127072 ,
8169
8234
"output_cost_per_token" : 0.0000002 ,
8170
8235
"litellm_provider" : " perplexity" ,
8171
8236
"mode" : " chat" ,
8172
- "deprecation_date" : " 2025-02-22" ,
8173
- "supports_tool_choice" : true
8237
+ "deprecation_date" : " 2025-02-22"
8238
+ },
8239
+ "perplexity/sonar" : {
8240
+ "max_tokens" : 127072 ,
8241
+ "max_input_tokens" : 127072 ,
8242
+ "max_output_tokens" : 127072 ,
8243
+ "input_cost_per_token" : 0.000001 ,
8244
+ "output_cost_per_token" : 0.000001 ,
8245
+ "litellm_provider" : " perplexity" ,
8246
+ "mode" : " chat"
8247
+ },
8248
+ "perplexity/sonar-pro" : {
8249
+ "max_tokens" : 200000 ,
8250
+ "max_input_tokens" : 200000 ,
8251
+ "max_output_tokens" : 8096 ,
8252
+ "input_cost_per_token" : 0.000003 ,
8253
+ "output_cost_per_token" : 0.000015 ,
8254
+ "litellm_provider" : " perplexity" ,
8255
+ "mode" : " chat"
8256
+ },
8257
+ "perplexity/sonar" : {
8258
+ "max_tokens" : 127072 ,
8259
+ "max_input_tokens" : 127072 ,
8260
+ "max_output_tokens" : 127072 ,
8261
+ "input_cost_per_token" : 0.000001 ,
8262
+ "output_cost_per_token" : 0.000001 ,
8263
+ "litellm_provider" : " perplexity" ,
8264
+ "mode" : " chat"
8265
+ },
8266
+ "perplexity/sonar-pro" : {
8267
+ "max_tokens" : 200000 ,
8268
+ "max_input_tokens" : 200000 ,
8269
+ "max_output_tokens" : 8096 ,
8270
+ "input_cost_per_token" : 0.000003 ,
8271
+ "output_cost_per_token" : 0.000015 ,
8272
+ "litellm_provider" : " perplexity" ,
8273
+ "mode" : " chat"
8174
8274
},
8175
8275
"perplexity/pplx-7b-chat" : {
8176
8276
"max_tokens" : 8192 ,
8179
8279
"input_cost_per_token" : 0.00000007 ,
8180
8280
"output_cost_per_token" : 0.00000028 ,
8181
8281
"litellm_provider" : " perplexity" ,
8182
- "mode" : " chat" ,
8183
- "supports_tool_choice" : true
8282
+ "mode" : " chat"
8184
8283
},
8185
8284
"perplexity/pplx-70b-chat" : {
8186
8285
"max_tokens" : 4096 ,
8189
8288
"input_cost_per_token" : 0.00000070 ,
8190
8289
"output_cost_per_token" : 0.00000280 ,
8191
8290
"litellm_provider" : " perplexity" ,
8192
- "mode" : " chat" ,
8193
- "supports_tool_choice" : true
8291
+ "mode" : " chat"
8194
8292
},
8195
8293
"perplexity/pplx-7b-online" : {
8196
8294
"max_tokens" : 4096 ,
8200
8298
"output_cost_per_token" : 0.00000028 ,
8201
8299
"input_cost_per_request" : 0.005 ,
8202
8300
"litellm_provider" : " perplexity" ,
8203
- "mode" : " chat" ,
8204
- "supports_tool_choice" : true
8301
+ "mode" : " chat"
8205
8302
},
8206
8303
"perplexity/pplx-70b-online" : {
8207
8304
"max_tokens" : 4096 ,
8211
8308
"output_cost_per_token" : 0.00000280 ,
8212
8309
"input_cost_per_request" : 0.005 ,
8213
8310
"litellm_provider" : " perplexity" ,
8214
- "mode" : " chat" ,
8215
- "supports_tool_choice" : true
8311
+ "mode" : " chat"
8216
8312
},
8217
8313
"perplexity/llama-2-70b-chat" : {
8218
8314
"max_tokens" : 4096 ,
8221
8317
"input_cost_per_token" : 0.00000070 ,
8222
8318
"output_cost_per_token" : 0.00000280 ,
8223
8319
"litellm_provider" : " perplexity" ,
8224
- "mode" : " chat" ,
8225
- "supports_tool_choice" : true
8320
+ "mode" : " chat"
8226
8321
},
8227
8322
"perplexity/mistral-7b-instruct" : {
8228
8323
"max_tokens" : 4096 ,
8231
8326
"input_cost_per_token" : 0.00000007 ,
8232
8327
"output_cost_per_token" : 0.00000028 ,
8233
8328
"litellm_provider" : " perplexity" ,
8234
- "mode" : " chat" ,
8235
- "supports_tool_choice" : true
8329
+ "mode" : " chat"
8236
8330
},
8237
8331
"perplexity/mixtral-8x7b-instruct" : {
8238
8332
"max_tokens" : 4096 ,
8241
8335
"input_cost_per_token" : 0.00000007 ,
8242
8336
"output_cost_per_token" : 0.00000028 ,
8243
8337
"litellm_provider" : " perplexity" ,
8244
- "mode" : " chat" ,
8245
- "supports_tool_choice" : true
8338
+ "mode" : " chat"
8246
8339
},
8247
8340
"perplexity/sonar-small-chat" : {
8248
8341
"max_tokens" : 16384 ,
8251
8344
"input_cost_per_token" : 0.00000007 ,
8252
8345
"output_cost_per_token" : 0.00000028 ,
8253
8346
"litellm_provider" : " perplexity" ,
8254
- "mode" : " chat" ,
8255
- "supports_tool_choice" : true
8347
+ "mode" : " chat"
8256
8348
},
8257
8349
"perplexity/sonar-small-online" : {
8258
8350
"max_tokens" : 12000 ,
8262
8354
"output_cost_per_token" : 0.00000028 ,
8263
8355
"input_cost_per_request" : 0.005 ,
8264
8356
"litellm_provider" : " perplexity" ,
8265
- "mode" : " chat" ,
8266
- "supports_tool_choice" : true
8357
+ "mode" : " chat"
8267
8358
},
8268
8359
"perplexity/sonar-medium-chat" : {
8269
8360
"max_tokens" : 16384 ,
8272
8363
"input_cost_per_token" : 0.0000006 ,
8273
8364
"output_cost_per_token" : 0.0000018 ,
8274
8365
"litellm_provider" : " perplexity" ,
8275
- "mode" : " chat" ,
8276
- "supports_tool_choice" : true
8366
+ "mode" : " chat"
8277
8367
},
8278
8368
"perplexity/sonar-medium-online" : {
8279
8369
"max_tokens" : 12000 ,
8283
8373
"output_cost_per_token" : 0.0000018 ,
8284
8374
"input_cost_per_request" : 0.005 ,
8285
8375
"litellm_provider" : " perplexity" ,
8286
- "mode" : " chat" ,
8287
- "supports_tool_choice" : true
8376
+ "mode" : " chat"
8288
8377
},
8289
8378
"fireworks_ai/accounts/fireworks/models/llama-v3p2-1b-instruct" : {
8290
8379
"max_tokens" : 16384 ,
9044
9133
"output_cost_per_second" : 0.00 ,
9045
9134
"litellm_provider" : " assemblyai"
9046
9135
}
9047
- }
9136
+ }
0 commit comments