Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Small (<5%) regression in functorch_maml_omniglot_cuda model from 0.2.1 to latest #1040

Open
zou3519 opened this issue Oct 10, 2022 · 3 comments

Comments

@zou3519
Copy link
Contributor

zou3519 commented Oct 10, 2022

On my machine with v100 P100 GPUs, the runtime goes from 286ms to 316ms

To repro:

# setup pytorch/benchmark
git clone https://github.com/pytorch/benchmark
cd benchmark
# this doesn't need to complete successfully -- we just need to install torchbenchmark's basic dependencies.
python setup.py install

python run_benchmark.py functorch
@samdow
Copy link
Contributor

samdow commented Oct 10, 2022

On A100s, seeing 107ms to 111ms ~4% regression

@samdow
Copy link
Contributor

samdow commented Oct 10, 2022

On AWS V100s, I'm seeing 169ms to 174ms which is ~3% regression

@zou3519
Copy link
Contributor Author

zou3519 commented Oct 10, 2022

With V100 on FAIR cluster I see 190ms to 211ms which is roughly 5%, so I can repro your v100 numbers @samdow. These numbers aren't crazy enough to investigate for 1.13, but might be worth looking into if this is overhead or something else

@zou3519 zou3519 changed the title 10% regression in functorch_maml_omniglot_cuda model from 0.2.1 to latest Small (<5%) regression in functorch_maml_omniglot_cuda model from 0.2.1 to latest Oct 10, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants