Elastic announced the Elasticsearch open Inference API now supports Cohere's text embedding models. This includes Elasticsearch native support for efficient int8 embeddings, which optimize performance and reduce memory cost for semantic search across the large datasets commonly found in enterprise scenarios. With this integration, Elasticsearch developers can experience immediate performance gains, including up to 4x memory savings and up to 30% faster search, without impacting search quality.