-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update Benchmarks and Documentation for GraniteCausalLM #86
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Signed-off-by: Yu Chin Fabian Lim <flim@sg.ibm.com>
0e23809
to
76b6685
Compare
Signed-off-by: Yu Chin Fabian Lim <flim@sg.ibm.com>
Signed-off-by: Yu Chin Fabian Lim <flim@sg.ibm.com>
a035913
to
0258544
Compare
awesome, great results @fabianlim |
Indeed, awesome results @fabianlim ! |
@wynterl @raghukiran1224 the loss for BNB + fused ops looks problematic. |
Signed-off-by: Yu Chin Fabian Lim <flim@sg.ibm.com>
Signed-off-by: Yu Chin Fabian Lim <flim@sg.ibm.com>
36c7fa7
to
e87f351
Compare
Signed-off-by: Yu Chin Fabian Lim <flim@sg.ibm.com>
Signed-off-by: Yu Chin Fabian Lim <flim@sg.ibm.com>
0a451c1
to
426a4a5
Compare
This was referenced Oct 8, 2024
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
In this PR we update the benchmarks for
GraniteCausalLM
Note this PR requires the following dependency updates
transformers>=4.45
: forGraniteCausalLM
accelerate>=0.34.1
: required fortransformers>=4.45
ifGraniteCausalLM
is needed.trl > 0.11.1
: when using baseline bnb, requires this fix for a bug that was introduced intransformers==4.45
Fix Inconsistency with IsShardedQLoRA Setting huggingface/trl#2089bitsandtbyes==0.43.3
: it seems that the later versions give segmentation fault errorsKnown issues with quant peft
fused lora dequant problem(this is an issue with the compiled binaries in bitsandbytes 0.43.3, that is not compatible with maybe the CUDA toolkit or torch version)rank 1 stuck at prepare_model(this is resolved by disablinglow_cpu_mem_mode
)meta device problem(see 2 in Distributed Training Problems for QLoRA models with Transformers pre-release 4.45 #83) (this is resolved by disablinglow_cpu_mem_mode
)Performance
Overall impressive improvements with kernels.
FULL FT

PEFT

Quantized Peft (BNB)
