Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

chore(deps): update python #1465

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

renovate-bot
Copy link
Contributor

@renovate-bot renovate-bot commented Sep 23, 2024

This PR contains the following updates:

Package Change Age Adoption Passing Confidence Update Pending
Werkzeug (changelog) ==3.0.6 -> ==3.1.3 age adoption passing confidence minor
accelerate ==v1.0.1 -> ==1.2.1 age adoption passing confidence minor
bitsandbytes ==0.44.1 -> ==0.45.0 age adoption passing confidence minor
click (changelog) ==8.1.7 -> ==8.1.8 age adoption passing confidence patch
cryptography (changelog) ==43.0.1 -> ==43.0.3 age adoption passing confidence patch
datasets ==3.1.0 -> ==3.2.0 age adoption passing confidence minor
elasticsearch ==8.15.0 -> ==8.17.0 age adoption passing confidence minor
fastapi (changelog) ==0.114.0 -> ==0.115.6 age adoption passing confidence minor
gcsfs ==2024.9.0 -> ==2024.12.0 age adoption passing confidence minor
google-auth ~=2.34.0 -> ~=2.37.0 age adoption passing confidence minor
google-cloud-aiplatform ==1.65.0 -> ==1.75.0 age adoption passing confidence minor 1.76.0
google-cloud-pubsub ~=2.23.0 -> ~=2.27.1 age adoption passing confidence minor 2.27.2
google-cloud-storage ==2.18.2 -> ==2.19.0 age adoption passing confidence minor
gradio ==5.8.0 -> ==5.9.1 age adoption passing confidence minor 5.12.0 (+2)
langchain (changelog) ==0.2.16 -> ==0.3.14 age adoption passing confidence minor
langchain-weaviate ==0.0.2 -> ==0.0.3 age adoption passing confidence patch
mlflow ==2.17.2 -> ==2.19.0 age adoption passing confidence minor
numpy (changelog) ==2.1.1 -> ==2.2.1 age adoption passing confidence minor
pandas (source) ==2.2.2 -> ==2.2.3 age adoption passing confidence patch
peft ==v0.13.2 -> ==0.14.0 age adoption passing confidence minor
pgvector ==0.3.2 -> ==0.3.6 age adoption passing confidence patch
psutil ==6.1.0 -> ==6.1.1 age adoption passing confidence patch
psycopg2-binary (source, changelog) ==2.9.9 -> ==2.9.10 age adoption passing confidence patch
pymupdf (changelog) ==1.24.10 -> ==1.25.1 age adoption passing confidence minor
qdrant-client ==1.11.1 -> ==1.12.2 age adoption passing confidence minor
redis (changelog) ==5.0.8 -> ==5.2.1 age adoption passing confidence minor
ruff (source, changelog) >=0.6,<=0.6.4 -> >=0.8,<=0.8.6 age adoption passing confidence minor 0.9.1 (+1)
scikit-learn (changelog) ==1.5.1 -> ==1.6.0 age adoption passing confidence minor 1.6.1
scipy ==1.14.1 -> ==1.15.0 age adoption passing confidence minor 1.15.1
six ==1.16.0 -> ==1.17.0 age adoption passing confidence minor
streamlit (source, changelog) ==1.38.0 -> ==1.41.1 age adoption passing confidence minor
tensorflow ==2.17.0 -> ==2.18.0 age adoption passing confidence minor
thejsonlogger ==0.0.3 -> ==0.0.5 age adoption passing confidence patch
transformers ==4.44.2 -> ==4.47.1 age adoption passing confidence minor 4.48.0
transformers ==v4.46.1 -> ==4.47.1 age adoption passing confidence minor 4.48.0
trl ==v0.11.4 -> ==0.13.0 age adoption passing confidence minor
tzdata ==2024.1 -> ==2024.2 age adoption passing confidence minor
uvicorn (changelog) ==0.30.6 -> ==0.34.0 age adoption passing confidence minor
weaviate-client ==4.7.1 -> ==4.10.2 age adoption passing confidence minor
All locks refreshed lockFileMaintenance

Warning

Some dependencies could not be looked up. Check the Dependency Dashboard for more information.


Release Notes

huggingface/accelerate (accelerate)

v1.2.1: : Patchfix

Compare Source

Full Changelog: huggingface/accelerate@v1.2.0...v1.2.1

v1.2.0: : Bug Squashing & Fixes across the board

Compare Source

Core

Big Modeling

DeepSpeed

Documentation

New Contributors

Full Changelog

Code Diff

Release diff: huggingface/accelerate@v1.1.1...v1.2.0

v1.1.1

Compare Source

v1.1.0: : Python 3.9 minimum, torch dynamo deepspeed support, and bug fixes

Compare Source

Internals:

DeepSpeed

Megatron

Big Model Inference

Examples

Full Changelog

New Contributors

Full Changelog: huggingface/accelerate@v1.0.1...v1.1.0

bitsandbytes-foundation/bitsandbytes (bitsandbytes)

v0.45.0: : LLM.int8() support for H100; faster 4-bit/8-bit inference

Compare Source

Highlights

H100 Support for LLM.int8()

PR #​1401 brings full LLM.int8() support for NVIDIA Hopper GPUs such as the H100, H200, and H800!

As part of the compatibility enhancements, we've rebuilt much of the LLM.int8() code in order to simplify for future compatibility and maintenance. We no longer use the col32 or architecture-specific tensor layout formats while maintaining backwards compatibility. We additionally bring performance improvements targeted for inference scenarios.

Performance Improvements

This release includes broad performance improvements for a wide variety of inference scenarios. See this X thread for a detailed explanation.

The improvements were measured using the 🤗optimum-benchmark tool.

For more benchmark results, see benchmarking/README.md.

LLM.int8()
  • Turing/Ampere/Ada: The observed per-token throughput is improved by 60-85%, while latency is decreased by 40-45%.
  • H100: With our benchmarking of Llama 3.1 70B, we observed the new LLM.int8() to consistently outperform NF4 at batch size >= 8.

Example throughput improvement for Qwen 2.5 14B Instruct on RTX 4090:

  • Batch size = 1: 9.05 tokens/s => 15.44 tokens/s
  • Batch size = 8: 66.62 tokens/s => 110.95 tokens/s

Example throughput improvement for Qwen 2.5 3B Instruct on T4:

  • Batch size = 1: 3.34 tokens/s => 5.98 tokens/s
  • Batch size = 8: 24.28 tokens/s => 44.15 tokens/s
NF4/FP4
  • Turing/Ampere/Ada: With batch size of 1, per-token throughput is improved by 10-25% and per-token latency is decreased by 10-20%.
  • H100: Across all batch sizes, per-token throughput is improved by up to 28% and per-token latency is decreased by up to 22%.

Example throughput improvement for Qwen 2.5 14B Instruct on RTX 4090:

  • Batch size = 1: 31.46 tokens/s => 39.03 tokens/s
  • Batch size = 8: 110.70 tokens/s => 111.29 tokens/s

Example throughput improvement for Qwen 2.5 3B Instruct on T4:

  • Batch size = 1: 11.05 tokens/s => 13.58 tokens/s
  • Batch size = 8: 69.8 tokens/s => 76.80 tokens/s

Changes

Packaging Changes

The size of our wheel has been reduced by ~43.5% from 122.4 MB to 69.1 MB! This results in an on-disk size decrease from ~396MB to ~224MB.

CUDA Toolkit Versions
  • Binaries built with CUDA Toolkit 12.6.2 are now included in the PyPI distribution.
  • The CUDA 12.5.0 build has been updated to CUDA Toolkit 12.5.1.
Breaking

🤗PEFT users wishing to merge adapters with 8-bit weights will need to upgrade to peft>=0.14.0.

New

Deprecations

A number of public API functions have been marked for deprecation and will emit FutureWarning when used. These functions will become unavailable in future releases. This should have minimal impact on most end-users.

k-bit quantization

The k-bit quantization features are deprecated in favor of blockwise quantization. For all optimizers, using block_wise=False is not recommended and support will be removed in a future release.

LLM.int8() deprecations:

As part of the refactoring process, we've implemented many new 8bit operations. These operations no longer use specialized data layouts.

The following relevant functions from bitsandbytes.functional are now deprecated :

  • dequant_min_max
  • dequantize_no_absmax
  • extract_outliers
  • get_special_format_str
  • get_transform_buffer
  • get_transform_func
  • mm_dequant (replacement: int8_mm_dequant)
  • igemmlt (replacement: int8_linear_matmul)
  • nvidia_transform
  • transform
  • quantize_no_absmax
  • vectorwise_dequant
  • vectorwise_quant (~replacement: int8_vectorwise_quant)
  • vectorwise_mm_dequant (~replacement: int8_mm_dequant)
General Deprecations

Additionally the following functions from bitsandbytes.functional are deprecated:

  • _mul
  • arange
  • post_call
  • pre_call

What's Changed


Configuration

📅 Schedule: Branch creation - "* 0-3 * * 1" (UTC), Automerge - At any time (no schedule defined).

🚦 Automerge: Enabled.

Rebasing: Whenever PR is behind base branch, or you tick the rebase/retry checkbox.

👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired.


  • If you want to rebase/retry this PR, check this box

This PR was generated by Mend Renovate. View the repository job log.

@renovate-bot renovate-bot requested a review from a team as a code owner September 23, 2024 00:47
@forking-renovate forking-renovate bot added the lang: python Issues specific to Python. label Sep 23, 2024
@bourgeoisor
Copy link
Member

Looks like most of the issues look like this:

6.059 The conflict is caused by:
6.059     langchain 0.3.0 depends on langchain-core<0.4.0 and >=0.3.0
6.059     langchain-community 0.3.0 depends on langchain-core<0.4.0 and >=0.3.0
6.059     langchain-google-vertexai 1.0.10 depends on langchain-core<0.3 and >=0.2.33

Might need to manually bump langchain-google-vertexai, latest is 2.0.1

@renovate-bot renovate-bot force-pushed the renovate/python branch 7 times, most recently from 2e4015a to 9815aef Compare September 26, 2024 18:40
@renovate-bot renovate-bot force-pushed the renovate/python branch 15 times, most recently from 89c302d to b8a5279 Compare October 4, 2024 03:20
@renovate-bot renovate-bot force-pushed the renovate/python branch 9 times, most recently from 083a305 to bd397ab Compare December 18, 2024 14:56
@renovate-bot renovate-bot force-pushed the renovate/python branch 8 times, most recently from aae76a1 to 21195db Compare December 28, 2024 04:59
@renovate-bot renovate-bot force-pushed the renovate/python branch 5 times, most recently from d6b96d4 to 0461f8a Compare January 3, 2025 20:05
@renovate-bot renovate-bot force-pushed the renovate/python branch 6 times, most recently from c6454f6 to 33e0acd Compare January 11, 2025 00:27
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
dependencies lang: python Issues specific to Python.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants