Elastic Introduces Native Inference Service in Elastic Cloud
9 oktober, 17:02
9 oktober, 17:02
Elastic (NYSE: ESTC), the Search AI Company, today announced the Elastic Inference Service (EIS), a GPU-accelerated inference-as-a-service for Elasticsearch semantic search, vector search, and generative AI workflows.
Every generative AI and vector search application relies on inference, and Elastic now delivers these capabilities natively as part of Elastic Cloud. As volumes grow, managing infrastructure, testing models, and handling integrations creates operational overhead that slows teams down. This has created a need for GPU-acceleration and an integrated workflow to provide speed, scalability, and cost efficiency.
“Inference at scale is incredibly important for vector search, semantic search and GenAI workflows,” said Steve Kearns, General Manager, Search at Elastic. “The Elastic Inference Service meets that challenge by providing our customers with an API-based inference service using NVIDIA GPUs with our best-in-class Elasticsearch vector database for low-latency, high-throughput inference.”
Elastic Learned Sparse EncodeR (ELSER) — Elastic’s built-in sparse vector model for state-of-the-art search relevance — is the first text-embedding model available on EIS in technical preview. Support for additional models for multilingual embeddings, reranking, and models from the recently announced Jina acquisition, will be available soon.
Some key benefits for developers who use EIS include:
For additional information on the Elastic Inference Service, read the Elastic blog.
Availability
The Elastic Inference Service is available to use on Serverless and Elastic Cloud Hosted deployments. All CSPs and regions can access the inference endpoints on EIS.
Additional models will be available soon to support a wider variety of search and inference needs.
About Elastic
Elastic (NYSE: ESTC), the Search AI Company, integrates its deep expertise in search technology with artificial intelligence to help everyone transform all of their data into answers, actions, and outcomes. Elastic's Search AI Platform — the foundation for its search, observability, and security solutions — is used by thousands of companies, including more than 50% of the Fortune 500. Learn more at elastic.co.
Elastic and associated marks are trademarks or registered trademarks of elasticsearch BV and its subsidiaries. All other company and product names may be trademarks of their respective owners. The release and timing of any features such as the additional models and region availability or functionality described in this post remain at Elastic's sole discretion. Any features or functionality not currently available may not be delivered on time or at all.
View source version on businesswire.com: https://www.businesswire.com/news/home/20251009383846/en/
9 oktober, 17:02
Elastic (NYSE: ESTC), the Search AI Company, today announced the Elastic Inference Service (EIS), a GPU-accelerated inference-as-a-service for Elasticsearch semantic search, vector search, and generative AI workflows.
Every generative AI and vector search application relies on inference, and Elastic now delivers these capabilities natively as part of Elastic Cloud. As volumes grow, managing infrastructure, testing models, and handling integrations creates operational overhead that slows teams down. This has created a need for GPU-acceleration and an integrated workflow to provide speed, scalability, and cost efficiency.
“Inference at scale is incredibly important for vector search, semantic search and GenAI workflows,” said Steve Kearns, General Manager, Search at Elastic. “The Elastic Inference Service meets that challenge by providing our customers with an API-based inference service using NVIDIA GPUs with our best-in-class Elasticsearch vector database for low-latency, high-throughput inference.”
Elastic Learned Sparse EncodeR (ELSER) — Elastic’s built-in sparse vector model for state-of-the-art search relevance — is the first text-embedding model available on EIS in technical preview. Support for additional models for multilingual embeddings, reranking, and models from the recently announced Jina acquisition, will be available soon.
Some key benefits for developers who use EIS include:
For additional information on the Elastic Inference Service, read the Elastic blog.
Availability
The Elastic Inference Service is available to use on Serverless and Elastic Cloud Hosted deployments. All CSPs and regions can access the inference endpoints on EIS.
Additional models will be available soon to support a wider variety of search and inference needs.
About Elastic
Elastic (NYSE: ESTC), the Search AI Company, integrates its deep expertise in search technology with artificial intelligence to help everyone transform all of their data into answers, actions, and outcomes. Elastic's Search AI Platform — the foundation for its search, observability, and security solutions — is used by thousands of companies, including more than 50% of the Fortune 500. Learn more at elastic.co.
Elastic and associated marks are trademarks or registered trademarks of elasticsearch BV and its subsidiaries. All other company and product names may be trademarks of their respective owners. The release and timing of any features such as the additional models and region availability or functionality described in this post remain at Elastic's sole discretion. Any features or functionality not currently available may not be delivered on time or at all.
View source version on businesswire.com: https://www.businesswire.com/news/home/20251009383846/en/
Aktieråd
Fondförvaltare
Sparande
Fondförvaltare
Sparande
1 DAG %
Senast
Stockholmsbörsen
10 oktober, 17:54
Trumps hot om Kinatullar sänkte Stockholmsbörsen
Intellego Technologies
10 oktober, 17:16
Blankarrapport om Intellego
Handelskrig
10 oktober, 17:14
Trump hotar Kina med höjda tullar
Saab
10 oktober, 15:26
Saab får order från Försvarsmakten värd 4 miljarder kronor
OMX Stockholm 30
1 DAG %
Senast
2 707,50