Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

update language and embedding model #17484

Merged
merged 3 commits into from
Jan 11, 2025
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -10,9 +10,9 @@
"\n",
"Deploy with Launchables. Launchables are pre-configured, fully optimized environments that users can deploy with a single click.\n",
"\n",
"In this notebook, you will use NVIDIA NIM Microservices for the LLM, llama-3.3-70b, to generate a report on a given topic. You will also use a NIM for an NVIDIA text embedding model, nv-embedqa-e5-v5. Given a set of documents, LlamaIndex will create an Index which it can run queries against. \n",
"In this notebook, you will use an NVIDIA LLM NIM microservice (llama-3.3-70b) to generate a report on a given topic, and an NVIDIA NeMo Retriever embedding NIM (llama-3.2-nv-embedqa-1b-v2) for optimized text question-answering retrieval. Given a set of documents, LlamaIndex will create an Index which it can run queries against. \n",
"\n",
"You can get started by leveraging NVIDIA API Catalog and call a hosted model's NIM API Endpoint. Once you familiarize yourself with this blueprint, you may want to self-host models with NIM Microservices.\n",
"You can get started by calling a hosted model's NIM API endpoint from the NVIDIA API catalog. Once you familiarize yourself with this blueprint, you may want to self-host models with NVIDIA NIM.\n",
"\n",
"The Blueprint provides a workflow architecture for automating and orchestrating the creation of well-researched, high-quality content. \n",
"\n",
Expand Down Expand Up @@ -259,7 +259,7 @@
"### Set LLM and Embedding Model\n",
"\n",
"In this notebook, you will use the newest llama model, llama-3.3-70b-instruct, as the LLM.\n",
"You will also use NVIDIA's embedding model, nv-embedqa-e5-v5."
"You will also use NVIDIA's embedding model, llama-3.2-nv-embedqa-1b-v2."
]
},
{
Expand All @@ -272,7 +272,7 @@
"\n",
"Settings.llm = NVIDIA(model=\"meta/llama-3.3-70b-instruct\")\n",
"Settings.embed_model = NVIDIAEmbedding(\n",
" model=\"nvidia/nv-embedqa-e5-v5\", truncate=\"END\"\n",
" model=\"nvidia/llama-3.2-nv-embedqa-1b-v2\", truncate=\"END\"\n",
")"
]
},
Expand Down
Loading