EvalDNN is an open-source toolbox for model evaluation of deep learning systems, supporting multiple frameworks and metrics.
Author: Yongqiang Tian*, Zhihua Zeng*, Ming Wen, Yepang Liu, Tzu-yang Kuo, and Shing-Chi, Cheung.
*The first two author contribute equally.
This project is mainly supported by Microsoft Asia Cloud Research Software Fellow Award 2019
A video is here: https://youtu.be/v69bNJN2bJc
A paper to inroduce this tool is in submit and will be released soon.
The paper is accepted by ICSE'20 Demo Track
We have collected some feedback from users and we are preparing new version with more functionality.
EvalDNN supports the model based on following frameworks:
- TensorFlow
- PyTorch
- Keras
- MXNet
EvalDNN supports the model based on following metrics:
- Top-K accuracy
- Neuron Coverage
- Robustness
pip install EvalDNN
Check demo/demo.ipynb
.
More examples are avaiable to the evaldnn/benchmarks/
and evaldnn/tests
The examples covers all 4 frameworks and 3 metrics.
Create a new .py under evaldnn.models
then follow the exising implementation in evaldnn.models
Create a new .py under evaldnn.metrics
then follow the exising implementation in evaldnn.metrics
The full benchmark is available here: https://yqtianust.github.io/EvalDNN-benchmark/index.html
The code to reproduce the results in benchmark is in evaldnn/benchmarks/
.
For example, run
python3 evaldnn/benchmarks/eval_keras
To evaluate the model using ImageNet dataset, please download it and put it into /EvalDNN-data/ILSVRC2012_img_val
.
The file ILSVRC2012_validation_ground_truth.txt
in github release should also be put into above folder.