Skip to content

Conversation

@jithunnair-amd
Copy link
Owner

Automated changes by create-pull-request GitHub action

tjruwase and others added 6 commits May 10, 2021 15:23
* Validate env; remove write size from logs

* Performance scripts for auto-tunining/auto-generating aio params of
deepspeed config.

* Formatting fixes

* Address feedback
…deepspeedai#1066)

* rename train_step_batch_size to train_micro_batch_size_per_gpu

* clarify batch_size related doc
* use the original function's name as the key to old_functions dict

* update profile output format

* print at global rank 0

* add flops calculation in bwd pass using time from ds timers

* improve aggregated profiling out to show all depth

* print samples/second

* update readme and examples

* update docs

* fix typo and reorder printing

* fix format
* Fix docstring

* Make screenshots clickable for easier viewing

* Navigation menu in alphabetical order; More clicable screenshots

* Rename 1Cycle doc

* Tweak naming

* Remove no longer used flag

* ZeRO3 Offload release

* Single GPU results

* Rearrange figures

* Single GPU text

* tweak intro

* zero3-offload section

* Add asynchronous i/o docs
* [configure_distributed_model] improve assert

This PR changes the 2 asserts to actually print the names of the params that are wrong. e.g.:
```
fp16 is enabled but the following parameters have dtype that is not fp16: wav2vec2.masked_spec_embed
```

* style

Co-authored-by: Olatunji Ruwase <olruwase@microsoft.com>
@jithunnair-amd jithunnair-amd self-assigned this May 13, 2021
@jithunnair-amd jithunnair-amd deleted the IFU-$GITHUB_REF-2021-05-13 branch May 13, 2021 18:23
jithunnair-amd added a commit that referenced this pull request May 28, 2021
* Update skipIfRocm to add customizable reason string; update skipped unit tests

* Don't skip test_stage_to_global for now
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants