Enter a new era of productivity with generative AI solutions for your business. Leverage AI, embedded as you need it, across the full stack.
Learn how to build an AI chatbot with unstructured data using Oracle Database 23ai, OCI AI services, and RAG.
Learn to build a competency development system using OCI Generative AI to analyze, summarize, and create tips for employee growth.
Use AI to automate email tasks with real-time categorization, sentiment analysis, and instant replies for an efficient workflow.
Enhance PeopleSoft with PICASO and generative AI for instant document content retrieval.
Create an AI-powered chat interface for Fusion Financials using OCI, Google Cloud, and Gemini AI.
Learn how to create AI-powered chatbots using Oracle Digital Assistant, OCI Data Science, LangChain, and Oracle Database 23ai.
Learn how to deploy a RAG-based chatbot using Oracle Database 23ai and Google Vertex AI on GCP.
Learn how to create a photo analysis app using generative AI, Oracle Cloud Infrastructure, and Streamlit.
Enable real-time, natural language data interaction with AI-driven NL2SQL and Oracle Autonomous Database.
Build AI chatbots using Oracle 23AI Vector Search, Oracle OCI Generative AI Service, and LlamaIndex. Integrate private and public data sets for smarte...
Implement AI Vector Search in Oracle APEX using Oracle Database 23ai for context-aware, semantic similarity search.
Streamline construction with OCI Vision’s AI for early anomaly detection using drones. Save time and costs with automated quality control.
Learn to automate invoice extraction with OCI Document Understanding. Simplify document processing in ERP systems using AI.
Learn how OCI Vision and Oracle APEX can help in breast and lung cancer research through advanced AI and machine learning models.
Discover a GenAI-based app that transforms procurement by enabling natural language queries for real-time ERP system data.
Create a movie recommendation app using HeatWave AutoML, Oracle APEX, and machine learning for tailored suggestions and powerful admin dashboards.
Learn to create a smart chatbot using Oracle Database 23ai and OCI Generative AI with this comprehensive tutorial.
Learn to build chatbots with SQL dialog skills using AI and ML for natural language processing and database interactions.
Access valuable insights from PDF documents and unstructured manuals with natural language queries.
Learn to automate email invoice processing using OCI Document Understanding and Oracle Integration Cloud, and free up your staff for key tasks.
Use AI to automate email tasks with real-time categorization, sentiment analysis, and instant replies for an efficient workflow.
Discover how to convert documents into actionable insights using OCI Generative AI, OpenSearch, and RAG in our interactive demo.
Discover OCI Vision for object detection in manufacturing, retail, and other industries. Improve quality control and analysis with advanced AI technology.
Learn to use ONNX models to vectorize PDF content and build an AI search engine with Oracle APEX.
Deploy NVIDIA NIM on OCI Kubernetes Engine for scalable, efficient inference using OCI Object Storage and NVIDIA GPUs for optimal performance.
Explore OCI Language for large-scale text analysis and translation to enhance apps with powerful AI.
Explore how to automate tasks securely using RAG and a choice of LLMs on a single GPU, helping ensure data privacy and efficiency.
Use OCI Language for real-time text analysis and translation across 30 languages without needing AI expertise.
Integrate OCI Speech and Siebel CRM to efficiently transcribe service calls, enhancing compliance and customization capabilities.
Use OCI Generative AI to create a draft job description based on the job title, company, and division.
Create a movie recommendation app using HeatWave AutoML, Oracle APEX, and machine learning for tailored suggestions and powerful admin dashboards.
Access valuable insights from PDF documents and unstructured manuals with natural language queries.
Discover how you can use HeatWave GenAI to easily build a chatbot letting you have contextual conversations with your unstructured data using natural language.
Use AI to automate email tasks with real-time categorization, sentiment analysis, and instant replies for an efficient workflow.
Automatically identify damaged packages using AI services, including OCI Vision, Oracle Digital Assistant, and Oracle Analytics Cloud, for efficient logistics management.
Explore how Autonomous Database Select AI simplifies data queries using natural language, enhancing productivity and analysis.
Accelerate secure document analysis and compliance for investment data with GenAI, OCI, and OpenSearch.
Learn how to deploy a RAG-based chatbot using Oracle Database 23ai and Google Vertex AI on GCP.
Learn how to create a photo analysis app using generative AI, Oracle Cloud Infrastructure, and Streamlit.
Explore automated Q&A generation using OCI Generative AI and RAG for efficient enterprise solutions.
Learn to deploy an AI chatbot on an Ampere A1 compute instance using minikube and Kubernetes, enhancing your deployment skills in this hands-on worksh...
Learn to set up Kubeflow on OCI Container Engine for Kubernetes for machine learning and MLOps.
Learn to deploy LLMs using Hugging Face and Kubernetes on OCI for scalable and secure AI deployments.
Help speed up your AI application deployment using Oracle Cloud and Kubernetes, enhancing scalability and reliability with cloud native strategies.
Discover how to convert documents into actionable insights using OCI Generative AI, OpenSearch, and RAG in our interactive demo.
Create an AI-powered chat interface for Fusion Financials using OCI, Google Cloud, and Gemini AI.
Learn how to create AI-powered chatbots using Oracle Digital Assistant, OCI Data Science, LangChain, and Oracle Database 23ai.
Learn how to deploy a RAG-based chatbot using Oracle Database 23ai and Google Vertex AI on GCP.
Learn how to create a photo analysis app using generative AI, Oracle Cloud Infrastructure, and Streamlit.
Create a movie recommendation app using HeatWave AutoML, Oracle APEX, and machine learning for tailored suggestions and powerful admin dashboards.
Automate Jira subtask creation to help save time, reduce errors, and effortlessly scale agile workflows.
Enhance PeopleSoft with PICASO and generative AI for instant document content retrieval.
Create an AI-powered chat interface for Fusion Financials using OCI, Google Cloud, and Gemini AI.
Learn how to create AI-powered chatbots using Oracle Digital Assistant, OCI Data Science, LangChain, and Oracle Database 23ai.
Learn how to create a photo analysis app using generative AI, Oracle Cloud Infrastructure, and Streamlit.
Enable real-time, natural language data interaction with AI-driven NL2SQL and Oracle Autonomous Database.
Access valuable insights from PDF documents and unstructured manuals with natural language queries.
Help speed up your AI application deployment using Oracle Cloud and Kubernetes, enhancing scalability and reliability with cloud native strategies.
Create an AI-powered chat interface for Fusion Financials using OCI, Google Cloud, and Gemini AI.
Automate email triage with OCI Language to analyze and identify emails and then respond accurately, streamlining business processes and improving outc...
Automate Jira subtask creation to help save time, reduce errors, and effortlessly scale agile workflows.
Use generative AI in Oracle APEX to create apps with natural language prompts and build blueprints, pages, and features based on your needs.
Explore image intelligence with OCI Vision. Use pretrained or custom models for object detection and text extraction in your apps.
Automatically identify damaged packages using AI services, including OCI Vision, Oracle Digital Assistant, and Oracle Analytics Cloud, for efficient l...
Deploy NVIDIA NIM on OCI Kubernetes Engine for scalable, efficient inference using OCI Object Storage and NVIDIA GPUs for optimal performance.
Create a movie recommendation app using HeatWave AutoML, Oracle APEX, and machine learning for tailored suggestions and powerful admin dashboards.
Quickly generate summaries of product reviews for ecommerce sites using HeatWave GenAI. It’s scalable and easy to implement.
Discover how to quickly answer domain-specific questions using HeatWave GenAI and vector store with this detailed tutorial.
Discover how to convert documents into actionable insights using OCI Generative AI, OpenSearch, and RAG in our interactive demo.
Create an AI-powered chat interface for Fusion Financials using OCI, Google Cloud, and Gemini AI.
Learn how to create AI-powered chatbots using Oracle Digital Assistant, OCI Data Science, LangChain, and Oracle Database 23ai.
Learn how to deploy a RAG-based chatbot using Oracle Database 23ai and Google Vertex AI on GCP.
Enable real-time, natural language data interaction with AI-driven NL2SQL and Oracle Autonomous Database.
Automate Jira subtask creation to help save time, reduce errors, and effortlessly scale agile workflows.
Build AI chatbots using Oracle 23AI Vector Search, Oracle OCI Generative AI Service, and LlamaIndex. Integrate private and public data sets for smarte...
Automatically identify damaged packages using AI services, including OCI Vision, Oracle Digital Assistant, and Oracle Analytics Cloud, for efficient logistics management.
Ingest live data into a RAG-based knowledge search engine store using an Oracle low-code modular LLM app engine.
Use OCI Language to detect PII in service requests on Siebel CRM, addressing privacy and compliance in your call center operations.
Integrate OCI Speech and Siebel CRM to efficiently transcribe service calls, enhancing compliance and customization capabilities.
Help speed up your AI application deployment using Oracle Cloud and Kubernetes, enhancing scalability and reliability with cloud native strategies.
Analyze specs, help check RFP compliance, and scan emails with OCI AI, Oracle Integration Cloud, and a tool that’s adaptable for any industry, includi...
In the fast-paced world of software development, staying informed is crucial. Imagine having an AI assistant that can help quickly transform a complex webpage into content that’s bite-sized, easily consumable, and shareable. This is one of many things that Oracle Cloud Infrastructure (OCI) Generative AI can help you do.
Below is an example of how you can build such an AI assistant with OCI Generative AI.
The AI-powered GitHub trending projects summarizer is a personal content generation engine that automatically retrieves and summarizes the top 25 trending GitHub projects. OCI Generative AI helps extract, read, and compile each project’s README file into a concise, engaging, and informative summary that can be shared with others.
Try it out, with detailed steps and sample code on GitHub.
You can easily switch between multiple LLMs offered through OCI Generative AI simply by modifying the model_id
variable in summarize_llm.py
.
The above is a subset of available models. We’re constantly making newer models available.
Below is a code snippet to call OCI Generative AI:
content.text = """Generate an abstractive summary of the given Markdown contents. Here are the contents to summarize: {}""".format(summary_txt)
chat_detail.content = content.text
chat_detail.serving_mode = oci.generative_ai_inference.models.OnDemandServingMode(model_id="meta.llama-3.1-405b-instruct") # configurable model chat_response = generative_ai_inference_client.chat(chat_detail)
Retrieval-augmented generation (RAG) is one of the most important use cases for AI. RAG lets you augment the knowledge of an LLM without retraining it. It’s a way for the LLM to extract new information, from a database or elsewhere, and quickly present it to the end user.
This allows the LLM to acquire up-to-date knowledge regardless of when the LLM was trained and when inference was run. As a result, the updated data can make your LLM more intelligent with little to no effort.
After uploading documents to Oracle Cloud Infrastructure (OCI) GenAI Agents, the service will process the data and provide a way to consume it through a chatbot.
Try it out, with detailed steps and sample code on GitHub.
Below is a code snippet for using the RAG agent in OCI:
# ask a question to RAG agent question = "What steps do I take if I have a new patient under the patient admission recommendations?" # Initialize service client with default config file agent_runtime_client = GenerativeAiAgentRuntimeClient(config)
chat_response = agent_runtime_client.chat( agent_endpoint_id="ocid1.test.oc1..<id>", chat_details=ChatDetails(user_message=question))
# Get the data from response print(chat_response.data)
Oracle Database 23ai supports all modern data types and workloads, including vectors, and incorporates AI and machine learning capabilities directly within the database. By generating and storing vector embeddings for the data in question, developers can enable searches for semantic similarities using mathematical calculations. This technology allows for combining similarity searches with searches on business data using simple SQL, so anyone with a basic understanding of SQL can tap into its power.
By running Oracle APEX (a low-code application platform) on top of Oracle Database 23ai, the Oracle AI Vector Search capabilities are natively available at no additional cost. APEX developers can seamlessly include these advanced search functionalities in their applications, creating more accurate and context-aware results.
This scenario will implement a semantic search in Oracle APEX using AI Vector Search in Oracle Database 23ai.
Read the blog post about the scenario.
Below is code to convert image descriptions into vectors and store them in the database:
UPDATE SM_POSTS
SET
AI_IMAGE_VECTOR = TO_VECTOR(VECTOR_EMBEDDING ( DOC_MODEL
USING AI_IMAGE_DESCRIPTION AS DATA
));
Now that we have vectors, we can use them to perform semantic searches. In this demo, we can do that in the source query of the Cards Region:
SELECT A.*, TO_CHAR(ROUND(VECTOR_DISTANCE,2), '0.99')AS VECTOR_DISTANCE_DISPLAY FROM
(SELECT
p.id,
p.user_name,
p.comment_text,
p.file_blob,
p.file_mime,
p.post_date,
p.REACTIONS,
p.USER_REACTION_CSS,
p.CREATED,
(
CASE
WHEN :P1_SEARCH IS NOT NULL AND :P1_VECTOR_SEARCH = 'Y'
THEN VECTOR_DISTANCE (
TO_VECTOR(VECTOR_EMBEDDING (doc_model USING :P1_SEARCH AS data)),
ai_image_vector
)
ELSE NULL
END
) AS vector_distance,
ai_image_description
FROM
mv_SM_POSTS p
WHERE
(:P1_VECTOR_SEARCH <> 'Y' AND :P1_SEARCH IS NOT NULL AND UPPER(ai_image_description) LIKE UPPER('%'||:P1_SEARCH||'%'))
OR :P1_VECTOR_SEARCH = 'Y'
OR :P1_SEARCH IS NULL
ORDER BY
vector_distance ASC, p.CREATED asc) A
Oracle Database 23ai with AI Vector Search, combined with Oracle APEX, enables developers to rapidly build context-aware apps with improved search capability.
Generative AI can be especially good at helping to summarize sentiment, as this scenario shows. An ecommerce site may have hundreds of stock-keeping units, or SKUs, with dozens of reviews for each one. To help quickly summarize product reviews, developers can tap into HeatWave GenAI’s integrated capabilities, using in-database large language models and an automated, in-database vector store.
HeatWave GenAI can also help translate and analyze sentiment on demand. All operations can be automated with HeatWave GenAI, keeping summaries up-to-date as new reviews are added.
By keeping the data and processing within HeatWave, developers can scale solutions with their GenAI needs, making AI as simple as a database query.
Try it out, with detailed steps and sample code on GitHub.
Below is a code snippet illustrating how to summarize positive reviews:
SELECT "################### Computing summaries for EXISTING reviews on a product ###################" AS "";
SELECT "" AS "";
CALL SUMMARIZE_TRANSLATE(1, "POSITIVE", "en", @positive_english_summary);
SELECT @positive_english_summary AS "--- English summary of positive reviews on the T-Shirt ---";
Open source LLMs, such as those created by Hugging Face, are powerful tools that let developers try out GenAI solutions relatively quickly. Kubernetes, combined with Oracle Cloud Infrastructure (OCI), enables GenAI solutions to scale, while also providing flexibility, portability and resilience.
In this demo, you’ll see how easy it can be to deploy fine-tuned LLM inference containers on OCI Kubernetes Engine, a managed Kubernetes service that simplifies deployments and operations at scale for enterprises. The service enables developers to retain the custom model and data sets within their own tenancy without relying on a third-party inference API.
We’ll use Text Generation Inference as the inference framework to expose the LLMs.
Try it out, with detailed steps and sample code on GitHub.
Below is a code snippet illustrating how to deploy an open source LLM:
# select model from HuggingFace
model=HuggingFaceH4/zephyr-7b-beta
# deploy selected model
docker run ghcr.io/huggingface/text-generation-inference:2.0 --model-id $model
# invoke the deployed model
curl IP_address:port/generate_stream \
-X POST \
-d '{"inputs":"What is Deep Learning?","parameters":{"max_new_tokens":50}}' \
-H 'Content-Type: application/json'
Oracle Code Assist is an AI code companion designed to help boost developer velocity and enhance code consistency. Powered by large language models (LLMs) on Oracle Cloud Infrastructure (OCI) and fine-tuned and optimized for Java, SQL, and application development on OCI, Oracle Code Assist provides developers with context-specific suggestions. You can tailor it to your organization’s best practices and codebases.
Currently available in beta for JetBrains IntelliJ IDEA and Microsoft Visual Studio Code, the plugin can assist with documentation, legacy code comprehension, and code completion.
To learn how to join the beta program and get started, visit our GitHub repository.
Deploy, scale, and monitor GenAI workloads in minutes with Oracle Cloud Infrastructure (OCI) AI Blueprints, complete with hardware recommendations, software components, and out-of-the-box monitoring.
Effectively deploy and scale LLMs with vLLM—lightning-fast inference, seamless integration, and zero hassle.
Choose from custom models or a variety of open source models on Hugging Face.
Automatically provision GPU nodes and store models in OCI Object Storage.
Get a ready-to-use API endpoint for instant model inference.
Enable autoscaling based on inference latency for mission-critical applications.
Easily integrate and scale inference workloads without deep technical expertise.
Monitor performance with built-in observability tools, such as Prometheus and Grafana.
Fine-tune smarter, not harder—benchmark performance and optimize AI training with data-driven insights.
Benchmark fine-tuning performance using the MLCommons methodology.
Fine-tune a quantized Llama 2 70B model with a standardized data set.
Track training time, resource utilization, and performance metrics.
Automatically log results in MLflow and visualize insights in Grafana.
Make data-driven infrastructure decisions for optimized fine-tuning jobs.
Supercharge LLM fine-tuning with low-rank adaptation (LoRA)—faster, more efficient, and ready for deployment.
Utilize LoRA for efficient fine-tuning of LLMs with minimal computational overhead.
Leverage your custom data sets or publicly available data sets from Hugging Face for training.
Track and analyze detailed training metrics logged in MLflow throughout the fine-tuning process.
Store the fine-tuned model and training results in an object storage bucket for seamless deployment.
Optimize performance with a design that helps ensure quick, effective model adaptation without heavy resource usage.
Scale the solution as needed, from small data sets to large-scale model fine-tuning.