diff --git a/CHANGELOG.md b/CHANGELOG.md index 285a798d9c0c2..aa9f216837182 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,64 @@ # ChangeLog +## [2024-08-12] + +### `llama-index-core` [0.10.65] + +- chore: bump nltk version (#15277) + +### `llama-index-tools-box` [0.1.0] + +- Box tools for AI Agents (#15236) + +### `llama-index-multi-modal-llms-gemini` [0.1.8] + +- feat: add default_headers to Gemini multi-model (#15296) + +### `llama-index-vector-stores-clickhouse` [0.2.0] + +- chore: stop using ServiceContext from the clickhouse integration (#15300) + +### `llama-index-experimental` [0.2.0] + +- chore: remove ServiceContext usage from experimental package (#15301) + +### `llama-index-extractors-marvin` [0.1.4] + +- fix: MarvinMetadataExtractor functionality and apply async support (#15247) + +### `llama-index-utils-workflow` [0.1.1] + +- chore: bump black version (#15288) +- chore: bump nltk version (#15277) + +### `llama-index-readers-microsoft-onedrive` [0.1.9] + +- chore: bump nltk version (#15277) + +### `llama-index-embeddings-upstage` [0.1.3] + +- chore: bump nltk version (#15277) + +### `llama-index-embeddings-nvidia` [0.1.5] + +- chore: bump nltk version (#15277) + +### `llama-index-embeddings-litellm` [0.1.1] + +- chore: bump nltk version (#15277) + +### `llama-index-legacy` [0.9.48post1] + +- chore: bump nltk version (#15277) + +### `llama-index-packs-streamlit-chatbot` [0.1.5] + +- chore: bump nltk version (#15277) + +### `llama-index-embeddings-huggingface` [0.2.3] + +- Feature: added multiprocessing for creating hf embedddings (#15260) + ## [2024-08-09] ### `llama-index-core` [0.10.64] diff --git a/docs/docs/CHANGELOG.md b/docs/docs/CHANGELOG.md index 285a798d9c0c2..e88e2510a7429 100644 --- a/docs/docs/CHANGELOG.md +++ b/docs/docs/CHANGELOG.md @@ -1,5 +1,60 @@ # ChangeLog +## [2024-08-12] + +### `llama-index-core` [0.10.65] + +- chore: bump nltk version (#15277) + +### `llama-index-multi-modal-llms-gemini` [0.1.8] + +- feat: add default_headers to Gemini multi-model (#15296) + +### `llama-index-vector-stores-clickhouse` [0.2.0] + +- chore: stop using ServiceContext from the clickhouse integration (#15300) + +### `llama-index-experimental` [0.2.0] + +- chore: remove ServiceContext usage from experimental package (#15301) + +### `llama-index-extractors-marvin` [0.1.4] + +- fix: MarvinMetadataExtractor functionality and apply async support (#15247) + +### `llama-index-utils-workflow` [0.1.1] + +- chore: bump black version (#15288) +- chore: bump nltk version (#15277) + +### `llama-index-readers-microsoft-onedrive` [0.1.9] + +- chore: bump nltk version (#15277) + +### `llama-index-embeddings-upstage` [0.1.3] + +- chore: bump nltk version (#15277) + +### `llama-index-embeddings-nvidia` [0.1.5] + +- chore: bump nltk version (#15277) + +### `llama-index-embeddings-litellm` [0.1.1] + +- chore: bump nltk version (#15277) + +### `llama-index-legacy` [0.9.48post1] + +- chore: bump nltk version (#15277) + +### `llama-index-packs-streamlit-chatbot` [0.1.5] + +- chore: bump nltk version (#15277) + +### `llama-index-embeddings-huggingface` [0.2.3] + +- Feature: added multiprocessing for creating hf embedddings (#15260) + ## [2024-08-09] ### `llama-index-core` [0.10.64] diff --git a/docs/mkdocs.yml b/docs/mkdocs.yml index c0e1078594bac..2a1cc9c5c9d07 100644 --- a/docs/mkdocs.yml +++ b/docs/mkdocs.yml @@ -638,8 +638,8 @@ nav: - ./examples/workflow/rag.ipynb - ./examples/workflow/react_agent.ipynb - ./examples/workflow/reflection.ipynb - - ./examples/workflow/workflows_cookbook.ipynb - ./examples/workflow/sub_question_query_engine.ipynb + - ./examples/workflow/workflows_cookbook.ipynb - Component Guides: - ./module_guides/index.md - Models: diff --git a/llama-index-core/llama_index/core/__init__.py b/llama-index-core/llama_index/core/__init__.py index 7a4211acad390..81020526eb955 100644 --- a/llama-index-core/llama_index/core/__init__.py +++ b/llama-index-core/llama_index/core/__init__.py @@ -1,6 +1,6 @@ """Init file of LlamaIndex.""" -__version__ = "0.10.64" +__version__ = "0.10.65" import logging from logging import NullHandler diff --git a/llama-index-core/pyproject.toml b/llama-index-core/pyproject.toml index 74e231cc2a81c..5d12d22960dd4 100644 --- a/llama-index-core/pyproject.toml +++ b/llama-index-core/pyproject.toml @@ -43,7 +43,7 @@ name = "llama-index-core" packages = [{include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.10.64" +version = "0.10.65" [tool.poetry.dependencies] SQLAlchemy = {extras = ["asyncio"], version = ">=1.4.49"} diff --git a/llama-index-integrations/embeddings/llama-index-embeddings-litellm/pyproject.toml b/llama-index-integrations/embeddings/llama-index-embeddings-litellm/pyproject.toml index 645f157e67ee2..bee66a5d98ff1 100644 --- a/llama-index-integrations/embeddings/llama-index-embeddings-litellm/pyproject.toml +++ b/llama-index-integrations/embeddings/llama-index-embeddings-litellm/pyproject.toml @@ -27,7 +27,7 @@ exclude = ["**/BUILD"] license = "MIT" name = "llama-index-embeddings-litellm" readme = "README.md" -version = "0.1.0" +version = "0.1.1" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" diff --git a/llama-index-integrations/embeddings/llama-index-embeddings-nvidia/pyproject.toml b/llama-index-integrations/embeddings/llama-index-embeddings-nvidia/pyproject.toml index c7a30dfe0c801..1636731fec9d6 100644 --- a/llama-index-integrations/embeddings/llama-index-embeddings-nvidia/pyproject.toml +++ b/llama-index-integrations/embeddings/llama-index-embeddings-nvidia/pyproject.toml @@ -27,7 +27,7 @@ exclude = ["**/BUILD"] license = "MIT" name = "llama-index-embeddings-nvidia" readme = "README.md" -version = "0.1.4" +version = "0.1.5" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" diff --git a/llama-index-integrations/embeddings/llama-index-embeddings-upstage/pyproject.toml b/llama-index-integrations/embeddings/llama-index-embeddings-upstage/pyproject.toml index 881cda0225589..7cb7c0e063b58 100644 --- a/llama-index-integrations/embeddings/llama-index-embeddings-upstage/pyproject.toml +++ b/llama-index-integrations/embeddings/llama-index-embeddings-upstage/pyproject.toml @@ -30,7 +30,7 @@ license = "MIT" name = "llama-index-embeddings-upstage" packages = [{include = "llama_index/"}] readme = "README.md" -version = "0.1.2" +version = "0.1.3" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" diff --git a/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-gemini/pyproject.toml b/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-gemini/pyproject.toml index 3f4b144816eaf..af3f5910d65c0 100644 --- a/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-gemini/pyproject.toml +++ b/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-gemini/pyproject.toml @@ -27,7 +27,7 @@ exclude = ["**/BUILD"] license = "MIT" name = "llama-index-multi-modal-llms-gemini" readme = "README.md" -version = "0.1.7" +version = "0.1.8" [tool.poetry.dependencies] python = ">=3.9,<4.0" diff --git a/llama-index-integrations/readers/llama-index-readers-microsoft-onedrive/pyproject.toml b/llama-index-integrations/readers/llama-index-readers-microsoft-onedrive/pyproject.toml index b52cf305dd22f..a6e3f9412c1f7 100644 --- a/llama-index-integrations/readers/llama-index-readers-microsoft-onedrive/pyproject.toml +++ b/llama-index-integrations/readers/llama-index-readers-microsoft-onedrive/pyproject.toml @@ -29,7 +29,7 @@ license = "MIT" maintainers = ["godwin3737"] name = "llama-index-readers-microsoft-onedrive" readme = "README.md" -version = "0.1.8" +version = "0.1.9" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" diff --git a/llama-index-legacy/pyproject.toml b/llama-index-legacy/pyproject.toml index ae984f9fe66fa..f607c44aa3c6b 100644 --- a/llama-index-legacy/pyproject.toml +++ b/llama-index-legacy/pyproject.toml @@ -42,7 +42,7 @@ name = "llama-index-legacy" packages = [{include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.9.48" +version = "0.9.48post1" [tool.poetry.dependencies] SQLAlchemy = {extras = ["asyncio"], version = ">=1.4.49"} diff --git a/llama-index-packs/llama-index-packs-streamlit-chatbot/pyproject.toml b/llama-index-packs/llama-index-packs-streamlit-chatbot/pyproject.toml index a45a3f8783cf8..fba2249f99430 100644 --- a/llama-index-packs/llama-index-packs-streamlit-chatbot/pyproject.toml +++ b/llama-index-packs/llama-index-packs-streamlit-chatbot/pyproject.toml @@ -29,7 +29,7 @@ license = "MIT" maintainers = ["carolinedlu"] name = "llama-index-packs-streamlit-chatbot" readme = "README.md" -version = "0.1.4" +version = "0.1.5" [tool.poetry.dependencies] python = ">=3.8.1,<3.9.7 || >3.9.7,<4.0" diff --git a/llama-index-utils/llama-index-utils-workflow/pyproject.toml b/llama-index-utils/llama-index-utils-workflow/pyproject.toml index 49e95b8824695..b2ee3635301e3 100644 --- a/llama-index-utils/llama-index-utils-workflow/pyproject.toml +++ b/llama-index-utils/llama-index-utils-workflow/pyproject.toml @@ -24,7 +24,7 @@ exclude = ["**/BUILD"] license = "MIT" name = "llama-index-utils-workflow" readme = "README.md" -version = "0.1.0" +version = "0.1.1" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" diff --git a/poetry.lock b/poetry.lock index 76dd343a37a2b..8f7eeef45841e 100644 --- a/poetry.lock +++ b/poetry.lock @@ -1,4 +1,4 @@ -# This file is automatically @generated by Poetry 1.8.3 and should not be changed by hand. +# This file is automatically @generated by Poetry 1.7.1 and should not be changed by hand. [[package]] name = "aiohappyeyeballs" @@ -1644,13 +1644,13 @@ llama-index-llms-openai = ">=0.1.1,<0.2.0" [[package]] name = "llama-index-core" -version = "0.10.64" +version = "0.10.65" description = "Interface between LLMs and your data" optional = false python-versions = "<4.0,>=3.8.1" files = [ - {file = "llama_index_core-0.10.64-py3-none-any.whl", hash = "sha256:03a22f8bbace4ec92a191d606fb01d44809982a854073a1092b8d7d9fe31749c"}, - {file = "llama_index_core-0.10.64.tar.gz", hash = "sha256:8f2599bfcc00efd7fb525e255f7d0610b02f0d06e2050a20cee5c0139171e3e6"}, + {file = "llama_index_core-0.10.65-py3-none-any.whl", hash = "sha256:bd9b6e213d2df75cc9bd45c0ee8fb5c01539e05a6d39635f664858adc45cce2c"}, + {file = "llama_index_core-0.10.65.tar.gz", hash = "sha256:4a7007ab0c0f4aa7b2e8788e206e9ad7ea3a7029b32309b9f2fed551dabbd3ba"}, ] [package.dependencies] @@ -1662,7 +1662,7 @@ fsspec = ">=2023.5.0" httpx = "*" nest-asyncio = ">=1.5.8,<2.0.0" networkx = ">=3.0" -nltk = ">=3.8.1,<4.0.0" +nltk = ">=3.8.2" numpy = "<2.0.0" openai = ">=1.1.0" pandas = "*" @@ -4589,4 +4589,4 @@ test = ["big-O", "importlib-resources", "jaraco.functools", "jaraco.itertools", [metadata] lock-version = "2.0" python-versions = ">=3.8.1,<4.0" -content-hash = "0ef4142d4a7a34de5701e81185fff4bac00d2d6cec12d3d27656e3c4adec7ef1" +content-hash = "a4bd1435ea7367067e89639fc7f5346bd7ce1b91295c0271def5ea461ecfd806" diff --git a/pyproject.toml b/pyproject.toml index ed619f93c3ef5..978f3ae425712 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -44,7 +44,7 @@ name = "llama-index" packages = [{from = "_llama-index", include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.10.64" +version = "0.10.65" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" @@ -57,7 +57,7 @@ llama-index-agent-openai = ">=0.1.4,<0.3.0" llama-index-readers-file = "^0.1.4" llama-index-readers-llama-parse = ">=0.1.2" llama-index-indices-managed-llama-cloud = ">=0.2.0" -llama-index-core = "^0.10.64" +llama-index-core = "^0.10.65" llama-index-multi-modal-llms-openai = "^0.1.3" llama-index-cli = "^0.1.2"