Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Updating swift-transformer #247

Open
BrandonWeng opened this issue Nov 4, 2024 · 5 comments
Open

Updating swift-transformer #247

BrandonWeng opened this issue Nov 4, 2024 · 5 comments

Comments

@BrandonWeng
Copy link

Hey folks, was wondering what it would take to upgrade swift-transformer to the latest version?

Apologies, totally new to Swift.. Happy to make the PR if there's no known blockers

@ZachNagengast
Copy link
Contributor

@BrandonWeng Is there a specific feature you're looking for in the latest version? We upgraded to the point right before jinja was added to avoid another dependency that we don't have much use for, but would consider upgrading there's a need.

@BrandonWeng
Copy link
Author

I think we found a way around this.

We were trying to get MLX running, but the examples required > 0.1.12

https://github.com/ml-explore/mlx-swift-examples/tree/main

@BrandonWeng
Copy link
Author

I'll just leave this here: #249

Happy to close the issue + PR if you don't think that its necessary. Just wanted to leave it here in case other folks run into the same issue. Spent several hours trying to work around it but this turned out to be the simplest solution for us

@ZachNagengast
Copy link
Contributor

Thanks! Curious to hear more about the approach you're taking with MLX, we have a PR in progress that still needs a couple perf improvements #200

@BrandonWeng
Copy link
Author

Unfortunately, I'm pretty new to Swift and its ecosystem as a whole. I'm just trying out a bunch of different things right now. Will report back once I have a better understanding!

For now, I've only been comparing the mlx models, the quantized models use significantly less memory:
mlx-community/Llama-3.2-1B-Instruct-bf16 uses around 2.5GB of memory and mlx-community/Llama-3.2-1B-Instruct-8bit is around 1.5GB. Performance wise bf16 isn't too far off from 8bit

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants