Oracle HeatWave GenAI provides integrated, automated, and secure generative AI with in-database large language models (LLMs); an automated, in-database vector store; scale-out vector processing; and the ability to have contextual conversations in natural language—letting you take advantage of generative AI without AI expertise, data movement, or additional cost.
Key features of HeatWave GenAI include
Quantized versions of the following in-database LLMs are currently available in HeatWave:
You can create a vector store for enterprise unstructured content with a single SQL command. All the steps to create a vector store and vector embeddings are automated and executed inside the database, including discovering the documents in object storage, parsing them, generating embeddings in a highly parallel and optimized way, and inserting them into the vector store, making HeatWave Vector Store efficient and easy to use.
HeatWave runs on commodity hardware. In-database LLMs don’t run on GPUs; they run on CPUs. As a result, you can reduce costs and don’t need to worry about the availability of LLMs in various data centers.
When using in-database LLMs and an in-database vector store, data doesn’t leave the database, helping increase data security.
Yes, HeatWave GenAI is available natively on AWS, along with other HeatWave capabilities, including HeatWave MySQL, HeatWave Lakehouse, HeatWave AutoML, and HeatWave Autopilot.
Yes, embeddings can be generated for text data in 27 languages.
Prompts can be issued in English. Prompts issued in other languages, such as Spanish and German, can be translated to English.
No, vector search is performed within the HeatWave cluster.
HeatWave runs on a MySQL node. We recommend a MySQL node with a MySQL.32 shape, plus HeatWave nodes using the HeatWave.512GB shape for a production environment. For development/testing, a smaller MySQL shape can be used. You can review supported MySQL shapes here. For HeatWave GenAI, the HeatWave.32GB shape isn’t supported.
PDF, text, PowerPoint, Word, and HTML are the supported formats.
There is no additional cost beyond the cost of the HeatWave cluster for using HeatWave GenAI. You can invoke in-database LLMs and embedding models provided with HeatWave GenAI at no additional charge. You can also invoke external LLMs available via OCI Generative AI on OCI and Amazon Bedrock on AWS and will then be charged for those services.
No, LLMs are pretrained models. Your data isn’t used for training LLMs.
No, in-database LLMs provided with HeatWave can’t be fine-tuned.
No, you can’t bring your own LLMs or embedding models. However, you can invoke the external LLMs or embedding models available via OCI Generative AI when running HeatWave GenAI on OCI and via Amazon Bedrock when running HeatWave GenAI on AWS.
Based on our testing, results are comparable to non-quantized LLMs for use cases that rely on HeatWave Vector Store. You can easily test the models to evaluate the performance and quality of results.
You need to generate embeddings only once and they will be stored in HeatWave Vector Store. Changes to unstructured documents in object storage will automatically trigger updates to associated vector embeddings.
Yes, optical character recognition support allows HeatWave Vector Store to convert scanned content saved as images into text data that can be analyzed, for example, to conduct similarity searches.