-
Notifications
You must be signed in to change notification settings - Fork 6
Issues: foundation-model-stack/fms-acceleration
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Slowdown and Higher Memory Consumption for GPTQ-LoRA with Bfloat16
#84
opened Sep 12, 2024 by
achew010
Distributed Training Problems for QLoRA models with Transformers pre-release 4.45
#83
opened Sep 11, 2024 by
achew010
Ensure Model is Correct Loaded Depending on Plugin For Augmentation Purposes
#77
opened Aug 29, 2024 by
fabianlim
Inconsistency in Padding-Free Benchmarks with Different Transformers Versions
#70
opened Aug 19, 2024 by
achew010
Introduce a Better Dequantization Fix on Triton Function for FOAK Plugin's GPTQ Fused Operations
#52
opened Jul 15, 2024 by
achew010
BNB Benchmark Experiments Run Out of Memory with Non-Zero Lora Dropout
#50
opened Jul 12, 2024 by
achew010
Release Upper Limit on Torch, Transformers, Accelerate
dependency
Issue arises because of a dependency
#17
opened May 23, 2024 by
fabianlim
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.