In this module, we will learn what LLM and RAG are and implement a simple RAG pipeline to answer questions about the FAQ Documents from our Zoomcamp courses
What we will do:
- Index Zoomcamp FAQ documents
- Create a Q&A system for answering questions about these documents
- LLM
- RAG
- RAG architecture
- Course outcome
- Installing libraries
- Alternative: installing anaconda or miniconda
pip install tqdm notebook==7.1.2 openai elasticsearch pandas scikit-learn
- We will use the search engine we build in the build-your-own-search-engine workshop: minsearch
- Indexing the documents
- Peforming the search
- Invoking OpenAI API
- Building the prompt
- Getting the answer
If you don't want to use a service, you can run an LLM locally refer to module 2 for more details.
In particular, check "2.7 Ollama - Running LLMs on a CPU" - it can work with OpenAI API, so to make the example from 1.4 work locally, you only need to change a few lines of code.
- Cleaning the code we wrote so far
- Making it modular
- Run ElasticSearch with Docker
- Index the documents
- Replace MinSearch with ElasticSearch
Running ElasticSearch:
docker run -it \
--rm \
--name elasticsearch \
-p 9200:9200 \
-p 9300:9300 \
-e "discovery.type=single-node" \
-e "xpack.security.enabled=false" \
docker.elastic.co/elasticsearch/elasticsearch:8.4.3
Index settings:
{
"settings": {
"number_of_shards": 1,
"number_of_replicas": 0
},
"mappings": {
"properties": {
"text": {"type": "text"},
"section": {"type": "text"},
"question": {"type": "text"},
"course": {"type": "keyword"}
}
}
}
Query:
{
"size": 5,
"query": {
"bool": {
"must": {
"multi_match": {
"query": query,
"fields": ["question^3", "text", "section"],
"type": "best_fields"
}
},
"filter": {
"term": {
"course": "data-engineering-zoomcamp"
}
}
}
}
}
- Replace it with a link
- Did you take notes? Add them above this line (Send a PR with links to your notes)