Changelog
 2024-06-19
 Added support for basic function calling
- Function calling is now supported on enabled models
 - Properties added on models page to show which models support function calling
 
 2024-06-18
 Native support for AI Gateways
Workers AI now natively supports AI Gateway.
 2024-06-11
Deprecation announcement for @cf/meta/llama-2-7b-chat-int8
We will be deprecating @cf/meta/llama-2-7b-chat-int8 on 2024-06-30.
Replace the model ID in your code with a new model of your choice:
@cf/meta/llama-3-8b-instructis the newest model in the Llama family (and is currently free for a limited time on Workers AI).@cf/meta/llama-3-8b-instruct-awqis the new Llama 3 in a similar precision to your currently selected model. This model is also currently free for a limited time.
If you do not switch to a different model by June 30th, we will automatically start returning inference from @cf/meta/llama-3-8b-instruct-awq.
 2024-05-29
 Add new public LoRAs and note on LoRA routing
- Added documentation on new public LoRAs.
 - Noted that you can now run LoRA inference with the base model rather than explicitly calling the 
-loraversion 
 2024-05-17
 Add OpenAI compatible API endpoints
Added OpenAI compatible API endpoints for /v1/chat/completions and /v1/embeddings. For more details, refer to Configurations.
 2024-04-11
 Add AI native binding
- Added new AI native binding, you can now run models with 
const resp = await env.AI.run(modelName, inputs) - Deprecated 
@cloudflare/ainpm package. While existing solutions using the @cloudflare/ai package will continue to work, no new Workers AI features will be supported. Moving to native AI bindings is highly recommended