Skip to content

v7.3.0

Compare
Choose a tag to compare
@davidmezzetti davidmezzetti released this 15 Jul 16:06
· 182 commits to master since this release

This release adds a new RAG front-end application template, streaming LLM and streaming RAG support along with significant text extraction improvements

See below for full details on the new features, improvements and bug fixes.

New Features

  • Add support for streaming LLM generation (#680)
  • Add RAG API endpoint (#735)
  • Add RAG deepdive notebook (#737)
  • Add RAG example application (#743)

Improvements

  • Improve textractor pipeline (#748)
  • Can't specify embedding model via API? (#632)
  • Configuration documentation update request (#705)
  • RAG alias for Extractor (#732)
  • Rename Extractor pipeline to RAG (#736)
  • Support max_seq_length parameter with model pooling (#746)

Bug Fixes

  • Fix issue with max tokens for llama.cpp components (#733)
  • Fix issue with loading non-transformer LLM models in Extractor/RAG pipeline (#734)
  • Fix issue with setting quantize=False in HFTrainer pipeline (#747)