FSDP Compatible, Sequential SparseGPT #1947
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Adding back the
sequential_update
flag toSparseGPTModifier
andWandaModifier
. Previously this flag affected whether or not we calibrate modules within a transformers block sequentially. Regardless of this flag we would always perform OBCQ sequentially across different transformer blocks.In the updated FSDP compatible implementation, transformers blocks are calibrated in parallel with no option for sequential calibration. The new
sequential_update
flag affects whether transformer blocks are processed sequentially, modules within blocks are always calibrated in parallel. Note that running sequential updates requires a lot more computation, as the whole model forward pass it runnum_calibration_samples
times for each transformer block