Elastic Managed LLM
Stack Serverless
Elastic provides a default LLM vetted for GenAI product features across our platform. Details of the currently used model are available in the model card.
The default LLM may change in the future based on evaluations of performance, security, and accuracy.
- Requires the
manage_inference
cluster privilege (the built-ininference_admin
role grants this privilege)
The Elastic Managed LLM is currently proxying to AWS Bedrock in AWS US regions, beginning with us-east-1
.
Customer projects or deployments hosted in any cloud service provider or region have access to the Elastic Managed LLM in the AWS US region us-east-1
.
All data is encrypted in transit. The LLM is configured for zero data retention: none of the prompts or outputs are stored by the model.
Only request metadata is logged in AWS CloudWatch. No information related to prompts is retained. Logged metadata includes the timestamp, model used, region, and request status.
Read more at our AI Data FAQs to learn about our data practices for AI related features.
The Elastic Managed LLM incurs a cost per million tokens for input and output tokens. Refer to the Elastic Cloud pricing pages for details.