This repository was archived by the owner on Dec 14, 2023. It is now read-only.
File tree Expand file tree Collapse file tree 4 files changed +4
-1
lines changed
Expand file tree Collapse file tree 4 files changed +4
-1
lines changed Original file line number Diff line number Diff line change @@ -37,6 +37,7 @@ trainable_modules:
3737seed : 64
3838mixed_precision : " fp16"
3939use_8bit_adam : False # This seems to be incompatible at the moment.
40+ gradient_checkpointing : False
4041enable_xformers_memory_efficient_attention : False
4142
4243# Use scaled dot product attention (Only available with >= Torch 2.0)
Original file line number Diff line number Diff line change @@ -37,6 +37,6 @@ trainable_modules:
3737seed : 64
3838mixed_precision : " fp16"
3939use_8bit_adam : False # This seems to be incompatible at the moment.
40-
40+ gradient_checkpointing : False
4141# Xformers must be installed
4242enable_xformers_memory_efficient_attention : True
Original file line number Diff line number Diff line change @@ -42,6 +42,7 @@ trainable_modules:
4242seed : 64
4343mixed_precision : " fp16"
4444use_8bit_adam : False # This seems to be incompatible at the moment.
45+ gradient_checkpointing : False
4546
4647# Xformers must be installed
4748enable_xformers_memory_efficient_attention : True
Original file line number Diff line number Diff line change @@ -37,6 +37,7 @@ trainable_modules:
3737seed : 64
3838mixed_precision : " fp16"
3939use_8bit_adam : False # This seems to be incompatible at the moment.
40+ gradient_checkpointing : False
4041enable_xformers_memory_efficient_attention : False
4142
4243# Use scaled dot product attention (Only available with >= Torch 2.0)
You can’t perform that action at this time.
0 commit comments