From 537f005266cca3742b751013bb222f85c0a35c67 Mon Sep 17 00:00:00 2001 From: Ldpe2G Date: Thu, 14 Apr 2022 16:09:56 +0800 Subject: [PATCH 1/7] skip some benchmark tests --- benchmark/__init__.py | 2 ++ benchmark/test_alexnet.py | 8 +++----- benchmark/test_convnext.py | 6 +++--- benchmark/test_crossformer.py | 6 +++--- benchmark/test_cswin.py | 6 +++--- benchmark/test_densenet.py | 6 +++--- benchmark/test_ghostnet.py | 6 +++--- benchmark/test_inception.py | 6 +++--- benchmark/test_mlp_mixer.py | 6 +++--- benchmark/test_mnasnet.py | 6 +++--- benchmark/test_mobilenet.py | 12 ++++++------ benchmark/test_pvt.py | 12 ++++++------ benchmark/test_res2net.py | 6 +++--- benchmark/test_res_mlp.py | 6 +++--- benchmark/test_resnet.py | 18 +++++++++--------- benchmark/test_rexnet.py | 12 ++++++------ benchmark/test_shufflenet.py | 6 +++--- benchmark/test_squeezenet.py | 6 +++--- benchmark/test_swin_transformer.py | 18 +++++++++--------- benchmark/test_uniformer.py | 24 ++++++++++++------------ benchmark/test_vision_transformer.py | 18 +++++++++--------- 21 files changed, 98 insertions(+), 98 deletions(-) diff --git a/benchmark/__init__.py b/benchmark/__init__.py index d4a5f1a2..4e64445e 100644 --- a/benchmark/__init__.py +++ b/benchmark/__init__.py @@ -1,4 +1,6 @@ import numpy as np +import os +import unittest import oneflow as flow diff --git a/benchmark/test_alexnet.py b/benchmark/test_alexnet.py index a6d8984d..76cfc77b 100644 --- a/benchmark/test_alexnet.py +++ b/benchmark/test_alexnet.py @@ -2,31 +2,29 @@ import oneflow_benchmark from flowvision.models.alexnet import alexnet - @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size1(benchmark, net=alexnet, input_shape=[1, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size2(benchmark, net=alexnet, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size4(benchmark, net=alexnet, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size8(benchmark, net=alexnet, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size16(benchmark, net=alexnet, input_shape=[16, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) diff --git a/benchmark/test_convnext.py b/benchmark/test_convnext.py index e671eabf..c9d6220f 100644 --- a/benchmark/test_convnext.py +++ b/benchmark/test_convnext.py @@ -10,7 +10,7 @@ def test_convnext_tiny_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_convnext_tiny_224_batch_size2( benchmark, net=convnext_tiny_224, input_shape=[2, 3, 224, 224] @@ -18,7 +18,7 @@ def test_convnext_tiny_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_convnext_tiny_224_batch_size4( benchmark, net=convnext_tiny_224, input_shape=[4, 3, 224, 224] @@ -26,7 +26,7 @@ def test_convnext_tiny_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_convnext_tiny_224_batch_size8( benchmark, net=convnext_tiny_224, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_crossformer.py b/benchmark/test_crossformer.py index e9bbb7ef..50821f28 100644 --- a/benchmark/test_crossformer.py +++ b/benchmark/test_crossformer.py @@ -10,7 +10,7 @@ def test_crossformer_tiny_patch4_group7_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_crossformer_tiny_patch4_group7_224_batch_size2( benchmark, net=crossformer_tiny_patch4_group7_224, input_shape=[2, 3, 224, 224] @@ -18,7 +18,7 @@ def test_crossformer_tiny_patch4_group7_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_crossformer_tiny_patch4_group7_224_batch_size4( benchmark, net=crossformer_tiny_patch4_group7_224, input_shape=[4, 3, 224, 224] @@ -26,7 +26,7 @@ def test_crossformer_tiny_patch4_group7_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_crossformer_tiny_patch4_group7_224_batch_size8( benchmark, net=crossformer_tiny_patch4_group7_224, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_cswin.py b/benchmark/test_cswin.py index 2b4cf8ef..a8fbc05e 100644 --- a/benchmark/test_cswin.py +++ b/benchmark/test_cswin.py @@ -10,7 +10,7 @@ def test_cswin_tiny_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_cswin_tiny_224_batch_size2( benchmark, net=cswin_tiny_224, input_shape=[2, 3, 224, 224] @@ -18,7 +18,7 @@ def test_cswin_tiny_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_cswin_tiny_224_batch_size4( benchmark, net=cswin_tiny_224, input_shape=[4, 3, 224, 224] @@ -26,7 +26,7 @@ def test_cswin_tiny_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_cswin_tiny_224_batch_size8( benchmark, net=cswin_tiny_224, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_densenet.py b/benchmark/test_densenet.py index a95bbae4..5ab0a5af 100644 --- a/benchmark/test_densenet.py +++ b/benchmark/test_densenet.py @@ -10,7 +10,7 @@ def test_densenet121_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size2( benchmark, net=densenet121, input_shape=[2, 3, 224, 224] @@ -18,7 +18,7 @@ def test_densenet121_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size4( benchmark, net=densenet121, input_shape=[4, 3, 224, 224] @@ -26,7 +26,7 @@ def test_densenet121_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size8( benchmark, net=densenet121, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_ghostnet.py b/benchmark/test_ghostnet.py index 805c481d..37264aca 100644 --- a/benchmark/test_ghostnet.py +++ b/benchmark/test_ghostnet.py @@ -8,19 +8,19 @@ def test_ghostnet_batch_size1(benchmark, net=ghostnet, input_shape=[1, 3, 224, 2 model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_ghostnet_batch_size2(benchmark, net=ghostnet, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_ghostnet_batch_size4(benchmark, net=ghostnet, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_ghostnet_batch_size8(benchmark, net=ghostnet, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) diff --git a/benchmark/test_inception.py b/benchmark/test_inception.py index 2f65d01c..25f29a9e 100644 --- a/benchmark/test_inception.py +++ b/benchmark/test_inception.py @@ -10,7 +10,7 @@ def test_inception_v3_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_inception_v3_batch_size2( benchmark, net=inception_v3, input_shape=[2, 3, 299, 299] @@ -18,7 +18,7 @@ def test_inception_v3_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_inception_v3_batch_size4( benchmark, net=inception_v3, input_shape=[4, 3, 299, 299] @@ -26,7 +26,7 @@ def test_inception_v3_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_inception_v3_batch_size8( benchmark, net=inception_v3, input_shape=[8, 3, 299, 299] diff --git a/benchmark/test_mlp_mixer.py b/benchmark/test_mlp_mixer.py index b5f833e7..361b311a 100644 --- a/benchmark/test_mlp_mixer.py +++ b/benchmark/test_mlp_mixer.py @@ -10,7 +10,7 @@ def test_mlp_mixer_b16_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mlp_mixer_b16_224_batch_size2( benchmark, net=mlp_mixer_b16_224, input_shape=[2, 3, 224, 224] @@ -18,7 +18,7 @@ def test_mlp_mixer_b16_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mlp_mixer_b16_224_batch_size4( benchmark, net=mlp_mixer_b16_224, input_shape=[4, 3, 224, 224] @@ -26,7 +26,7 @@ def test_mlp_mixer_b16_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mlp_mixer_b16_224_batch_size8( benchmark, net=mlp_mixer_b16_224, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_mnasnet.py b/benchmark/test_mnasnet.py index 132e49b2..a01af049 100644 --- a/benchmark/test_mnasnet.py +++ b/benchmark/test_mnasnet.py @@ -10,7 +10,7 @@ def test_mnasnet0_5_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mnasnet0_5_batch_size2( benchmark, net=mnasnet0_5, input_shape=[2, 3, 224, 224] @@ -18,7 +18,7 @@ def test_mnasnet0_5_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mnasnet0_5_batch_size4( benchmark, net=mnasnet0_5, input_shape=[4, 3, 224, 224] @@ -26,7 +26,7 @@ def test_mnasnet0_5_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mnasnet0_5_batch_size8( benchmark, net=mnasnet0_5, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_mobilenet.py b/benchmark/test_mobilenet.py index 0b0b6cf6..c0d0cb15 100644 --- a/benchmark/test_mobilenet.py +++ b/benchmark/test_mobilenet.py @@ -11,7 +11,7 @@ def test_mobilenet_v3_large_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v3_large_batch_size2( benchmark, net=mobilenet_v3_large, input_shape=[2, 3, 224, 224] @@ -19,7 +19,7 @@ def test_mobilenet_v3_large_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v3_large_batch_size4( benchmark, net=mobilenet_v3_large, input_shape=[4, 3, 224, 224] @@ -27,7 +27,7 @@ def test_mobilenet_v3_large_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v3_large_batch_size8( benchmark, net=mobilenet_v3_large, input_shape=[8, 3, 224, 224] @@ -51,7 +51,7 @@ def test_mobilenet_v2_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size2( benchmark, net=mobilenet_v2, input_shape=[2, 3, 224, 224] @@ -59,7 +59,7 @@ def test_mobilenet_v2_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size4( benchmark, net=mobilenet_v2, input_shape=[4, 3, 224, 224] @@ -67,7 +67,7 @@ def test_mobilenet_v2_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size8( benchmark, net=mobilenet_v2, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_pvt.py b/benchmark/test_pvt.py index 0f758132..e9057807 100644 --- a/benchmark/test_pvt.py +++ b/benchmark/test_pvt.py @@ -9,19 +9,19 @@ def test_pvt_tiny_batch_size1(benchmark, net=pvt_tiny, input_shape=[1, 3, 224, 2 model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_tiny_batch_size2(benchmark, net=pvt_tiny, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_tiny_batch_size4(benchmark, net=pvt_tiny, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_tiny_batch_size8(benchmark, net=pvt_tiny, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) @@ -39,19 +39,19 @@ def test_pvt_small_batch_size1(benchmark, net=pvt_small, input_shape=[1, 3, 224, model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_small_batch_size2(benchmark, net=pvt_small, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_small_batch_size4(benchmark, net=pvt_small, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_small_batch_size8(benchmark, net=pvt_small, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) diff --git a/benchmark/test_res2net.py b/benchmark/test_res2net.py index 35a75bc1..b2014723 100644 --- a/benchmark/test_res2net.py +++ b/benchmark/test_res2net.py @@ -10,7 +10,7 @@ def test_res2net50_26w_4s_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_res2net50_26w_4s_batch_size2( benchmark, net=res2net50_26w_4s, input_shape=[2, 3, 224, 224] @@ -18,7 +18,7 @@ def test_res2net50_26w_4s_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_res2net50_26w_4s_batch_size4( benchmark, net=res2net50_26w_4s, input_shape=[4, 3, 224, 224] @@ -26,7 +26,7 @@ def test_res2net50_26w_4s_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_res2net50_26w_4s_batch_size8( benchmark, net=res2net50_26w_4s, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_res_mlp.py b/benchmark/test_res_mlp.py index beb74d51..2b11fe83 100644 --- a/benchmark/test_res_mlp.py +++ b/benchmark/test_res_mlp.py @@ -10,7 +10,7 @@ def test_resmlp_12_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resmlp_12_224_batch_size2( benchmark, net=resmlp_12_224, input_shape=[2, 3, 224, 224] @@ -18,7 +18,7 @@ def test_resmlp_12_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resmlp_12_224_batch_size4( benchmark, net=resmlp_12_224, input_shape=[4, 3, 224, 224] @@ -26,7 +26,7 @@ def test_resmlp_12_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resmlp_12_224_batch_size8( benchmark, net=resmlp_12_224, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_resnet.py b/benchmark/test_resnet.py index bdb9ecc0..50a29fed 100644 --- a/benchmark/test_resnet.py +++ b/benchmark/test_resnet.py @@ -10,19 +10,19 @@ def test_resnet50_batch_size1(benchmark, net=resnet50, input_shape=[1, 3, 224, 2 model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnet50_batch_size2(benchmark, net=resnet50, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnet50_batch_size4(benchmark, net=resnet50, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnet50_batch_size8(benchmark, net=resnet50, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) @@ -42,7 +42,7 @@ def test_resnext50_32x4d_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size2( benchmark, net=resnext50_32x4d, input_shape=[2, 3, 224, 224] @@ -50,7 +50,7 @@ def test_resnext50_32x4d_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size4( benchmark, net=resnext50_32x4d, input_shape=[4, 3, 224, 224] @@ -58,7 +58,7 @@ def test_resnext50_32x4d_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size8( benchmark, net=resnext50_32x4d, input_shape=[8, 3, 224, 224] @@ -82,7 +82,7 @@ def test_wide_resnet50_2_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size2( benchmark, net=wide_resnet50_2, input_shape=[2, 3, 224, 224] @@ -90,7 +90,7 @@ def test_wide_resnet50_2_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size4( benchmark, net=wide_resnet50_2, input_shape=[4, 3, 224, 224] @@ -98,7 +98,7 @@ def test_wide_resnet50_2_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size8( benchmark, net=wide_resnet50_2, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_rexnet.py b/benchmark/test_rexnet.py index e041ac8e..861027d7 100644 --- a/benchmark/test_rexnet.py +++ b/benchmark/test_rexnet.py @@ -11,7 +11,7 @@ def test_rexnetv1_1_0_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnetv1_1_0_batch_size2( benchmark, net=rexnetv1_1_0, input_shape=[2, 3, 224, 224] @@ -19,7 +19,7 @@ def test_rexnetv1_1_0_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnetv1_1_0_batch_size4( benchmark, net=rexnetv1_1_0, input_shape=[4, 3, 224, 224] @@ -27,7 +27,7 @@ def test_rexnetv1_1_0_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnetv1_1_0_batch_size8( benchmark, net=rexnetv1_1_0, input_shape=[8, 3, 224, 224] @@ -51,7 +51,7 @@ def test_rexnet_lite_1_0_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnet_lite_1_0_batch_size2( benchmark, net=rexnet_lite_1_0, input_shape=[2, 3, 224, 224] @@ -59,7 +59,7 @@ def test_rexnet_lite_1_0_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnet_lite_1_0_batch_size4( benchmark, net=rexnet_lite_1_0, input_shape=[4, 3, 224, 224] @@ -67,7 +67,7 @@ def test_rexnet_lite_1_0_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnet_lite_1_0_batch_size8( benchmark, net=rexnet_lite_1_0, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_shufflenet.py b/benchmark/test_shufflenet.py index 5b7f774e..15462ef5 100644 --- a/benchmark/test_shufflenet.py +++ b/benchmark/test_shufflenet.py @@ -10,7 +10,7 @@ def test_shufflenet_v2_x0_5_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_shufflenet_v2_x0_5_batch_size2( benchmark, net=shufflenet_v2_x0_5, input_shape=[2, 3, 224, 224] @@ -18,7 +18,7 @@ def test_shufflenet_v2_x0_5_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_shufflenet_v2_x0_5_batch_size4( benchmark, net=shufflenet_v2_x0_5, input_shape=[4, 3, 224, 224] @@ -26,7 +26,7 @@ def test_shufflenet_v2_x0_5_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_shufflenet_v2_x0_5_batch_size8( benchmark, net=shufflenet_v2_x0_5, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_squeezenet.py b/benchmark/test_squeezenet.py index 34a61ffd..d8e3eb88 100644 --- a/benchmark/test_squeezenet.py +++ b/benchmark/test_squeezenet.py @@ -10,7 +10,7 @@ def test_squeezenet1_0_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_squeezenet1_0_batch_size2( benchmark, net=squeezenet1_0, input_shape=[2, 3, 224, 224] @@ -18,7 +18,7 @@ def test_squeezenet1_0_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_squeezenet1_0_batch_size4( benchmark, net=squeezenet1_0, input_shape=[4, 3, 224, 224] @@ -26,7 +26,7 @@ def test_squeezenet1_0_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_squeezenet1_0_batch_size8( benchmark, net=squeezenet1_0, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_swin_transformer.py b/benchmark/test_swin_transformer.py index 469a0346..b4fe8754 100644 --- a/benchmark/test_swin_transformer.py +++ b/benchmark/test_swin_transformer.py @@ -12,7 +12,7 @@ def test_swin_tiny_patch4_window7_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_tiny_patch4_window7_224_batch_size2( benchmark, net=swin_tiny_patch4_window7_224, input_shape=[2, 3, 224, 224] @@ -20,7 +20,7 @@ def test_swin_tiny_patch4_window7_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_tiny_patch4_window7_224_batch_size4( benchmark, net=swin_tiny_patch4_window7_224, input_shape=[4, 3, 224, 224] @@ -28,7 +28,7 @@ def test_swin_tiny_patch4_window7_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_tiny_patch4_window7_224_batch_size8( benchmark, net=swin_tiny_patch4_window7_224, input_shape=[8, 3, 224, 224] @@ -52,7 +52,7 @@ def test_swin_small_patch4_window7_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_small_patch4_window7_224_batch_size2( benchmark, net=swin_small_patch4_window7_224, input_shape=[2, 3, 224, 224] @@ -60,7 +60,7 @@ def test_swin_small_patch4_window7_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_small_patch4_window7_224_batch_size4( benchmark, net=swin_small_patch4_window7_224, input_shape=[4, 3, 224, 224] @@ -68,7 +68,7 @@ def test_swin_small_patch4_window7_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_small_patch4_window7_224_batch_size8( benchmark, net=swin_small_patch4_window7_224, input_shape=[8, 3, 224, 224] @@ -92,7 +92,7 @@ def test_swin_base_patch4_window7_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size2( benchmark, net=swin_base_patch4_window7_224, input_shape=[2, 3, 224, 224] @@ -100,7 +100,7 @@ def test_swin_base_patch4_window7_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size4( benchmark, net=swin_base_patch4_window7_224, input_shape=[4, 3, 224, 224] @@ -108,7 +108,7 @@ def test_swin_base_patch4_window7_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size8( benchmark, net=swin_base_patch4_window7_224, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_uniformer.py b/benchmark/test_uniformer.py index b360d4fc..6267436a 100644 --- a/benchmark/test_uniformer.py +++ b/benchmark/test_uniformer.py @@ -13,7 +13,7 @@ def test_uniformer_base_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size2( benchmark, net=uniformer_base, input_shape=[2, 3, 224, 224] @@ -21,7 +21,7 @@ def test_uniformer_base_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size4( benchmark, net=uniformer_base, input_shape=[4, 3, 224, 224] @@ -29,7 +29,7 @@ def test_uniformer_base_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size8( benchmark, net=uniformer_base, input_shape=[8, 3, 224, 224] @@ -53,7 +53,7 @@ def test_uniformer_base_ls_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_ls_batch_size2( benchmark, net=uniformer_base_ls, input_shape=[2, 3, 224, 224] @@ -61,7 +61,7 @@ def test_uniformer_base_ls_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_ls_batch_size4( benchmark, net=uniformer_base_ls, input_shape=[4, 3, 224, 224] @@ -69,7 +69,7 @@ def test_uniformer_base_ls_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_ls_batch_size8( benchmark, net=uniformer_base_ls, input_shape=[8, 3, 224, 224] @@ -93,7 +93,7 @@ def test_uniformer_small_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_batch_size2( benchmark, net=uniformer_small, input_shape=[2, 3, 224, 224] @@ -101,7 +101,7 @@ def test_uniformer_small_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_batch_size4( benchmark, net=uniformer_small, input_shape=[4, 3, 224, 224] @@ -109,7 +109,7 @@ def test_uniformer_small_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_batch_size8( benchmark, net=uniformer_small, input_shape=[8, 3, 224, 224] @@ -133,7 +133,7 @@ def test_uniformer_small_plus_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_plus_batch_size2( benchmark, net=uniformer_small_plus, input_shape=[2, 3, 224, 224] @@ -141,7 +141,7 @@ def test_uniformer_small_plus_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_plus_batch_size4( benchmark, net=uniformer_small_plus, input_shape=[4, 3, 224, 224] @@ -149,7 +149,7 @@ def test_uniformer_small_plus_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_plus_batch_size8( benchmark, net=uniformer_small_plus, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_vision_transformer.py b/benchmark/test_vision_transformer.py index fef6cc70..0c09106a 100644 --- a/benchmark/test_vision_transformer.py +++ b/benchmark/test_vision_transformer.py @@ -12,7 +12,7 @@ def test_vit_tiny_patch16_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_tiny_patch16_224_batch_size2( benchmark, net=vit_tiny_patch16_224, input_shape=[2, 3, 224, 224] @@ -20,7 +20,7 @@ def test_vit_tiny_patch16_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_tiny_patch16_224_batch_size4( benchmark, net=vit_tiny_patch16_224, input_shape=[4, 3, 224, 224] @@ -28,7 +28,7 @@ def test_vit_tiny_patch16_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_tiny_patch16_224_batch_size8( benchmark, net=vit_tiny_patch16_224, input_shape=[8, 3, 224, 224] @@ -52,7 +52,7 @@ def test_vit_small_patch16_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_small_patch16_224_batch_size2( benchmark, net=vit_small_patch16_224, input_shape=[2, 3, 224, 224] @@ -60,7 +60,7 @@ def test_vit_small_patch16_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_small_patch16_224_batch_size4( benchmark, net=vit_small_patch16_224, input_shape=[4, 3, 224, 224] @@ -68,7 +68,7 @@ def test_vit_small_patch16_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_small_patch16_224_batch_size8( benchmark, net=vit_small_patch16_224, input_shape=[8, 3, 224, 224] @@ -92,7 +92,7 @@ def test_vit_base_patch16_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size2( benchmark, net=vit_base_patch16_224, input_shape=[2, 3, 224, 224] @@ -100,7 +100,7 @@ def test_vit_base_patch16_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size4( benchmark, net=vit_base_patch16_224, input_shape=[4, 3, 224, 224] @@ -108,7 +108,7 @@ def test_vit_base_patch16_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size8( benchmark, net=vit_base_patch16_224, input_shape=[8, 3, 224, 224] From ba8c719ae80854b64132af2ef9c9028085484e97 Mon Sep 17 00:00:00 2001 From: Ldpe2G Date: Thu, 14 Apr 2022 17:08:46 +0800 Subject: [PATCH 2/7] update --- benchmark/test_alexnet.py | 2 ++ benchmark/test_convnext.py | 4 ++-- benchmark/test_crossformer.py | 4 ++-- benchmark/test_cswin.py | 4 ++-- benchmark/test_densenet.py | 4 ---- benchmark/test_ghostnet.py | 4 ++-- benchmark/test_inception.py | 3 --- benchmark/test_mlp_mixer.py | 4 ++-- benchmark/test_mnasnet.py | 4 ++-- benchmark/test_mobilenet.py | 7 ------- benchmark/test_pvt.py | 8 ++++---- benchmark/test_res2net.py | 4 ++-- benchmark/test_res_mlp.py | 4 ++-- benchmark/test_resnet.py | 11 ----------- benchmark/test_rexnet.py | 8 ++++---- benchmark/test_shufflenet.py | 3 --- benchmark/test_squeezenet.py | 4 ++-- benchmark/test_swin_transformer.py | 13 ++++--------- benchmark/test_uniformer.py | 17 ++++++----------- benchmark/test_vision_transformer.py | 12 ++++-------- 20 files changed, 42 insertions(+), 82 deletions(-) diff --git a/benchmark/test_alexnet.py b/benchmark/test_alexnet.py index 76cfc77b..2e368a71 100644 --- a/benchmark/test_alexnet.py +++ b/benchmark/test_alexnet.py @@ -2,6 +2,7 @@ import oneflow_benchmark from flowvision.models.alexnet import alexnet +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size1(benchmark, net=alexnet, input_shape=[1, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) @@ -25,6 +26,7 @@ def test_alexnet_batch_size8(benchmark, net=alexnet, input_shape=[8, 3, 224, 224 model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size16(benchmark, net=alexnet, input_shape=[16, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) diff --git a/benchmark/test_convnext.py b/benchmark/test_convnext.py index c9d6220f..8f7b547a 100644 --- a/benchmark/test_convnext.py +++ b/benchmark/test_convnext.py @@ -2,7 +2,7 @@ import oneflow_benchmark from flowvision.models.convnext import convnext_tiny_224 - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_convnext_tiny_224_batch_size1( benchmark, net=convnext_tiny_224, input_shape=[1, 3, 224, 224] @@ -34,7 +34,7 @@ def test_convnext_tiny_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_convnext_tiny_224_batch_size16( benchmark, net=convnext_tiny_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_crossformer.py b/benchmark/test_crossformer.py index 50821f28..1c3e8b6a 100644 --- a/benchmark/test_crossformer.py +++ b/benchmark/test_crossformer.py @@ -2,7 +2,7 @@ import oneflow_benchmark from flowvision.models.crossformer import crossformer_tiny_patch4_group7_224 - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_crossformer_tiny_patch4_group7_224_batch_size1( benchmark, net=crossformer_tiny_patch4_group7_224, input_shape=[1, 3, 224, 224] @@ -34,7 +34,7 @@ def test_crossformer_tiny_patch4_group7_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_crossformer_tiny_patch4_group7_224_batch_size16( benchmark, net=crossformer_tiny_patch4_group7_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_cswin.py b/benchmark/test_cswin.py index a8fbc05e..f9765334 100644 --- a/benchmark/test_cswin.py +++ b/benchmark/test_cswin.py @@ -2,7 +2,7 @@ import oneflow_benchmark from flowvision.models.cswin import cswin_tiny_224 - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_cswin_tiny_224_batch_size1( benchmark, net=cswin_tiny_224, input_shape=[1, 3, 224, 224] @@ -34,7 +34,7 @@ def test_cswin_tiny_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_cswin_tiny_224_batch_size16( benchmark, net=cswin_tiny_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_densenet.py b/benchmark/test_densenet.py index 5ab0a5af..0310617a 100644 --- a/benchmark/test_densenet.py +++ b/benchmark/test_densenet.py @@ -3,14 +3,12 @@ from flowvision.models.densenet import densenet121 -@oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size1( benchmark, net=densenet121, input_shape=[1, 3, 224, 224] ): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size2( benchmark, net=densenet121, input_shape=[2, 3, 224, 224] @@ -18,7 +16,6 @@ def test_densenet121_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size4( benchmark, net=densenet121, input_shape=[4, 3, 224, 224] @@ -26,7 +23,6 @@ def test_densenet121_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size8( benchmark, net=densenet121, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_ghostnet.py b/benchmark/test_ghostnet.py index 37264aca..7a014d7b 100644 --- a/benchmark/test_ghostnet.py +++ b/benchmark/test_ghostnet.py @@ -2,7 +2,7 @@ import oneflow_benchmark from flowvision.models.ghostnet import ghostnet - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_ghostnet_batch_size1(benchmark, net=ghostnet, input_shape=[1, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) @@ -26,7 +26,7 @@ def test_ghostnet_batch_size8(benchmark, net=ghostnet, input_shape=[8, 3, 224, 2 model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_ghostnet_batch_size16(benchmark, net=ghostnet, input_shape=[16, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) diff --git a/benchmark/test_inception.py b/benchmark/test_inception.py index 25f29a9e..63fc962f 100644 --- a/benchmark/test_inception.py +++ b/benchmark/test_inception.py @@ -10,7 +10,6 @@ def test_inception_v3_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_inception_v3_batch_size2( benchmark, net=inception_v3, input_shape=[2, 3, 299, 299] @@ -18,7 +17,6 @@ def test_inception_v3_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_inception_v3_batch_size4( benchmark, net=inception_v3, input_shape=[4, 3, 299, 299] @@ -26,7 +24,6 @@ def test_inception_v3_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_inception_v3_batch_size8( benchmark, net=inception_v3, input_shape=[8, 3, 299, 299] diff --git a/benchmark/test_mlp_mixer.py b/benchmark/test_mlp_mixer.py index 361b311a..06e010cc 100644 --- a/benchmark/test_mlp_mixer.py +++ b/benchmark/test_mlp_mixer.py @@ -2,7 +2,7 @@ import oneflow_benchmark from flowvision.models.mlp_mixer import mlp_mixer_b16_224 - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mlp_mixer_b16_224_batch_size1( benchmark, net=mlp_mixer_b16_224, input_shape=[1, 3, 224, 224] @@ -34,7 +34,7 @@ def test_mlp_mixer_b16_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mlp_mixer_b16_224_batch_size16( benchmark, net=mlp_mixer_b16_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_mnasnet.py b/benchmark/test_mnasnet.py index a01af049..46648dc8 100644 --- a/benchmark/test_mnasnet.py +++ b/benchmark/test_mnasnet.py @@ -2,7 +2,7 @@ import oneflow_benchmark from flowvision.models.mnasnet import mnasnet0_5 - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mnasnet0_5_batch_size1( benchmark, net=mnasnet0_5, input_shape=[1, 3, 224, 224] @@ -34,7 +34,7 @@ def test_mnasnet0_5_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mnasnet0_5_batch_size16( benchmark, net=mnasnet0_5, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_mobilenet.py b/benchmark/test_mobilenet.py index c0d0cb15..26acc70e 100644 --- a/benchmark/test_mobilenet.py +++ b/benchmark/test_mobilenet.py @@ -11,7 +11,6 @@ def test_mobilenet_v3_large_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v3_large_batch_size2( benchmark, net=mobilenet_v3_large, input_shape=[2, 3, 224, 224] @@ -19,7 +18,6 @@ def test_mobilenet_v3_large_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v3_large_batch_size4( benchmark, net=mobilenet_v3_large, input_shape=[4, 3, 224, 224] @@ -27,7 +25,6 @@ def test_mobilenet_v3_large_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v3_large_batch_size8( benchmark, net=mobilenet_v3_large, input_shape=[8, 3, 224, 224] @@ -44,14 +41,12 @@ def test_mobilenet_v3_large_batch_size16( benchmark(run, model, x, optimizer) -@oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size1( benchmark, net=mobilenet_v2, input_shape=[1, 3, 224, 224] ): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size2( benchmark, net=mobilenet_v2, input_shape=[2, 3, 224, 224] @@ -59,7 +54,6 @@ def test_mobilenet_v2_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size4( benchmark, net=mobilenet_v2, input_shape=[4, 3, 224, 224] @@ -67,7 +61,6 @@ def test_mobilenet_v2_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size8( benchmark, net=mobilenet_v2, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_pvt.py b/benchmark/test_pvt.py index e9057807..7d296b57 100644 --- a/benchmark/test_pvt.py +++ b/benchmark/test_pvt.py @@ -3,7 +3,7 @@ from flowvision.models.pvt import pvt_tiny from flowvision.models.pvt import pvt_small - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_tiny_batch_size1(benchmark, net=pvt_tiny, input_shape=[1, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) @@ -27,13 +27,13 @@ def test_pvt_tiny_batch_size8(benchmark, net=pvt_tiny, input_shape=[8, 3, 224, 2 model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_tiny_batch_size16(benchmark, net=pvt_tiny, input_shape=[16, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_small_batch_size1(benchmark, net=pvt_small, input_shape=[1, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) @@ -57,7 +57,7 @@ def test_pvt_small_batch_size8(benchmark, net=pvt_small, input_shape=[8, 3, 224, model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_small_batch_size16( benchmark, net=pvt_small, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_res2net.py b/benchmark/test_res2net.py index b2014723..5d3d036e 100644 --- a/benchmark/test_res2net.py +++ b/benchmark/test_res2net.py @@ -2,7 +2,7 @@ import oneflow_benchmark from flowvision.models.res2net import res2net50_26w_4s - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_res2net50_26w_4s_batch_size1( benchmark, net=res2net50_26w_4s, input_shape=[1, 3, 224, 224] @@ -34,7 +34,7 @@ def test_res2net50_26w_4s_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_res2net50_26w_4s_batch_size16( benchmark, net=res2net50_26w_4s, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_res_mlp.py b/benchmark/test_res_mlp.py index 2b11fe83..3e870425 100644 --- a/benchmark/test_res_mlp.py +++ b/benchmark/test_res_mlp.py @@ -2,7 +2,7 @@ import oneflow_benchmark from flowvision.models.res_mlp import resmlp_12_224 - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resmlp_12_224_batch_size1( benchmark, net=resmlp_12_224, input_shape=[1, 3, 224, 224] @@ -34,7 +34,7 @@ def test_resmlp_12_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resmlp_12_224_batch_size16( benchmark, net=resmlp_12_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_resnet.py b/benchmark/test_resnet.py index 50a29fed..b8967d05 100644 --- a/benchmark/test_resnet.py +++ b/benchmark/test_resnet.py @@ -10,19 +10,16 @@ def test_resnet50_batch_size1(benchmark, net=resnet50, input_shape=[1, 3, 224, 2 model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnet50_batch_size2(benchmark, net=resnet50, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnet50_batch_size4(benchmark, net=resnet50, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnet50_batch_size8(benchmark, net=resnet50, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) @@ -42,7 +39,6 @@ def test_resnext50_32x4d_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size2( benchmark, net=resnext50_32x4d, input_shape=[2, 3, 224, 224] @@ -50,7 +46,6 @@ def test_resnext50_32x4d_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size4( benchmark, net=resnext50_32x4d, input_shape=[4, 3, 224, 224] @@ -58,7 +53,6 @@ def test_resnext50_32x4d_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size8( benchmark, net=resnext50_32x4d, input_shape=[8, 3, 224, 224] @@ -66,7 +60,6 @@ def test_resnext50_32x4d_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size16( benchmark, net=resnext50_32x4d, input_shape=[16, 3, 224, 224] @@ -74,7 +67,6 @@ def test_resnext50_32x4d_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size1( benchmark, net=wide_resnet50_2, input_shape=[1, 3, 224, 224] @@ -82,7 +74,6 @@ def test_wide_resnet50_2_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size2( benchmark, net=wide_resnet50_2, input_shape=[2, 3, 224, 224] @@ -90,7 +81,6 @@ def test_wide_resnet50_2_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size4( benchmark, net=wide_resnet50_2, input_shape=[4, 3, 224, 224] @@ -98,7 +88,6 @@ def test_wide_resnet50_2_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size8( benchmark, net=wide_resnet50_2, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_rexnet.py b/benchmark/test_rexnet.py index 861027d7..f30970db 100644 --- a/benchmark/test_rexnet.py +++ b/benchmark/test_rexnet.py @@ -3,7 +3,7 @@ from flowvision.models.rexnet import rexnetv1_1_0 from flowvision.models.rexnet_lite import rexnet_lite_1_0 - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnetv1_1_0_batch_size1( benchmark, net=rexnetv1_1_0, input_shape=[1, 3, 224, 224] @@ -35,7 +35,7 @@ def test_rexnetv1_1_0_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnetv1_1_0_batch_size16( benchmark, net=rexnetv1_1_0, input_shape=[16, 3, 224, 224] @@ -43,7 +43,7 @@ def test_rexnetv1_1_0_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnet_lite_1_0_batch_size1( benchmark, net=rexnet_lite_1_0, input_shape=[1, 3, 224, 224] @@ -75,7 +75,7 @@ def test_rexnet_lite_1_0_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnet_lite_1_0_batch_size16( benchmark, net=rexnet_lite_1_0, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_shufflenet.py b/benchmark/test_shufflenet.py index 15462ef5..88bed665 100644 --- a/benchmark/test_shufflenet.py +++ b/benchmark/test_shufflenet.py @@ -10,7 +10,6 @@ def test_shufflenet_v2_x0_5_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_shufflenet_v2_x0_5_batch_size2( benchmark, net=shufflenet_v2_x0_5, input_shape=[2, 3, 224, 224] @@ -18,7 +17,6 @@ def test_shufflenet_v2_x0_5_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_shufflenet_v2_x0_5_batch_size4( benchmark, net=shufflenet_v2_x0_5, input_shape=[4, 3, 224, 224] @@ -26,7 +24,6 @@ def test_shufflenet_v2_x0_5_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_shufflenet_v2_x0_5_batch_size8( benchmark, net=shufflenet_v2_x0_5, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_squeezenet.py b/benchmark/test_squeezenet.py index d8e3eb88..5a539c4a 100644 --- a/benchmark/test_squeezenet.py +++ b/benchmark/test_squeezenet.py @@ -2,7 +2,7 @@ import oneflow_benchmark from flowvision.models.squeezenet import squeezenet1_0 - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_squeezenet1_0_batch_size1( benchmark, net=squeezenet1_0, input_shape=[1, 3, 224, 224] @@ -34,7 +34,7 @@ def test_squeezenet1_0_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_squeezenet1_0_batch_size16( benchmark, net=squeezenet1_0, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_swin_transformer.py b/benchmark/test_swin_transformer.py index b4fe8754..aa460dab 100644 --- a/benchmark/test_swin_transformer.py +++ b/benchmark/test_swin_transformer.py @@ -4,7 +4,7 @@ from flowvision.models.swin_transformer import swin_small_patch4_window7_224 from flowvision.models.swin_transformer import swin_base_patch4_window7_224 - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_tiny_patch4_window7_224_batch_size1( benchmark, net=swin_tiny_patch4_window7_224, input_shape=[1, 3, 224, 224] @@ -36,7 +36,7 @@ def test_swin_tiny_patch4_window7_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_tiny_patch4_window7_224_batch_size16( benchmark, net=swin_tiny_patch4_window7_224, input_shape=[16, 3, 224, 224] @@ -44,7 +44,7 @@ def test_swin_tiny_patch4_window7_224_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_small_patch4_window7_224_batch_size1( benchmark, net=swin_small_patch4_window7_224, input_shape=[1, 3, 224, 224] @@ -76,7 +76,7 @@ def test_swin_small_patch4_window7_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_small_patch4_window7_224_batch_size16( benchmark, net=swin_small_patch4_window7_224, input_shape=[16, 3, 224, 224] @@ -85,14 +85,12 @@ def test_swin_small_patch4_window7_224_batch_size16( benchmark(run, model, x, optimizer) -@oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size1( benchmark, net=swin_base_patch4_window7_224, input_shape=[1, 3, 224, 224] ): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size2( benchmark, net=swin_base_patch4_window7_224, input_shape=[2, 3, 224, 224] @@ -100,7 +98,6 @@ def test_swin_base_patch4_window7_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size4( benchmark, net=swin_base_patch4_window7_224, input_shape=[4, 3, 224, 224] @@ -108,7 +105,6 @@ def test_swin_base_patch4_window7_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size8( benchmark, net=swin_base_patch4_window7_224, input_shape=[8, 3, 224, 224] @@ -116,7 +112,6 @@ def test_swin_base_patch4_window7_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size16( benchmark, net=swin_base_patch4_window7_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_uniformer.py b/benchmark/test_uniformer.py index 6267436a..0a28ea69 100644 --- a/benchmark/test_uniformer.py +++ b/benchmark/test_uniformer.py @@ -5,7 +5,6 @@ from flowvision.models.uniformer import uniformer_base_ls from flowvision.models.uniformer import uniformer_small - @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size1( benchmark, net=uniformer_base, input_shape=[1, 3, 224, 224] @@ -13,7 +12,6 @@ def test_uniformer_base_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size2( benchmark, net=uniformer_base, input_shape=[2, 3, 224, 224] @@ -21,7 +19,6 @@ def test_uniformer_base_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size4( benchmark, net=uniformer_base, input_shape=[4, 3, 224, 224] @@ -29,7 +26,6 @@ def test_uniformer_base_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size8( benchmark, net=uniformer_base, input_shape=[8, 3, 224, 224] @@ -37,7 +33,6 @@ def test_uniformer_base_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size16( benchmark, net=uniformer_base, input_shape=[16, 3, 224, 224] @@ -45,7 +40,7 @@ def test_uniformer_base_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_ls_batch_size1( benchmark, net=uniformer_base_ls, input_shape=[1, 3, 224, 224] @@ -77,7 +72,7 @@ def test_uniformer_base_ls_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_ls_batch_size16( benchmark, net=uniformer_base_ls, input_shape=[16, 3, 224, 224] @@ -85,7 +80,7 @@ def test_uniformer_base_ls_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_batch_size1( benchmark, net=uniformer_small, input_shape=[1, 3, 224, 224] @@ -117,7 +112,7 @@ def test_uniformer_small_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_batch_size16( benchmark, net=uniformer_small, input_shape=[16, 3, 224, 224] @@ -125,7 +120,7 @@ def test_uniformer_small_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_plus_batch_size1( benchmark, net=uniformer_small_plus, input_shape=[1, 3, 224, 224] @@ -157,7 +152,7 @@ def test_uniformer_small_plus_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_plus_batch_size16( benchmark, net=uniformer_small_plus, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_vision_transformer.py b/benchmark/test_vision_transformer.py index 0c09106a..77eeb878 100644 --- a/benchmark/test_vision_transformer.py +++ b/benchmark/test_vision_transformer.py @@ -4,7 +4,7 @@ from flowvision.models.vision_transformer import vit_small_patch16_224 from flowvision.models.vision_transformer import vit_base_patch16_224 - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_tiny_patch16_224_batch_size1( benchmark, net=vit_tiny_patch16_224, input_shape=[1, 3, 224, 224] @@ -36,7 +36,7 @@ def test_vit_tiny_patch16_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_tiny_patch16_224_batch_size16( benchmark, net=vit_tiny_patch16_224, input_shape=[16, 3, 224, 224] @@ -44,7 +44,7 @@ def test_vit_tiny_patch16_224_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_small_patch16_224_batch_size1( benchmark, net=vit_small_patch16_224, input_shape=[1, 3, 224, 224] @@ -76,7 +76,7 @@ def test_vit_small_patch16_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - +@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_small_patch16_224_batch_size16( benchmark, net=vit_small_patch16_224, input_shape=[16, 3, 224, 224] @@ -92,7 +92,6 @@ def test_vit_base_patch16_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size2( benchmark, net=vit_base_patch16_224, input_shape=[2, 3, 224, 224] @@ -100,7 +99,6 @@ def test_vit_base_patch16_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size4( benchmark, net=vit_base_patch16_224, input_shape=[4, 3, 224, 224] @@ -108,7 +106,6 @@ def test_vit_base_patch16_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size8( benchmark, net=vit_base_patch16_224, input_shape=[8, 3, 224, 224] @@ -116,7 +113,6 @@ def test_vit_base_patch16_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size16( benchmark, net=vit_base_patch16_224, input_shape=[16, 3, 224, 224] From 92bddb45ad5d903abfdbf24394aa9197b7af1fec Mon Sep 17 00:00:00 2001 From: oneflow-ci-bot Date: Thu, 14 Apr 2022 09:13:05 +0000 Subject: [PATCH 3/7] auto format by CI --- benchmark/test_alexnet.py | 30 +++++++-- benchmark/test_convnext.py | 30 +++++++-- benchmark/test_crossformer.py | 30 +++++++-- benchmark/test_cswin.py | 30 +++++++-- benchmark/test_densenet.py | 3 + benchmark/test_ghostnet.py | 30 +++++++-- benchmark/test_inception.py | 3 + benchmark/test_mlp_mixer.py | 30 +++++++-- benchmark/test_mnasnet.py | 30 +++++++-- benchmark/test_mobilenet.py | 6 ++ benchmark/test_pvt.py | 60 +++++++++++++++--- benchmark/test_res2net.py | 30 +++++++-- benchmark/test_res_mlp.py | 30 +++++++-- benchmark/test_resnet.py | 11 ++++ benchmark/test_rexnet.py | 60 +++++++++++++++--- benchmark/test_shufflenet.py | 3 + benchmark/test_squeezenet.py | 30 +++++++-- benchmark/test_swin_transformer.py | 64 ++++++++++++++++--- benchmark/test_uniformer.py | 95 +++++++++++++++++++++++----- benchmark/test_vision_transformer.py | 64 ++++++++++++++++--- 20 files changed, 564 insertions(+), 105 deletions(-) diff --git a/benchmark/test_alexnet.py b/benchmark/test_alexnet.py index 2e368a71..da614d4c 100644 --- a/benchmark/test_alexnet.py +++ b/benchmark/test_alexnet.py @@ -2,31 +2,51 @@ import oneflow_benchmark from flowvision.models.alexnet import alexnet -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size1(benchmark, net=alexnet, input_shape=[1, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size2(benchmark, net=alexnet, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size4(benchmark, net=alexnet, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size8(benchmark, net=alexnet, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_alexnet_batch_size16(benchmark, net=alexnet, input_shape=[16, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) diff --git a/benchmark/test_convnext.py b/benchmark/test_convnext.py index 8f7b547a..488a4e4d 100644 --- a/benchmark/test_convnext.py +++ b/benchmark/test_convnext.py @@ -2,7 +2,11 @@ import oneflow_benchmark from flowvision.models.convnext import convnext_tiny_224 -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_convnext_tiny_224_batch_size1( benchmark, net=convnext_tiny_224, input_shape=[1, 3, 224, 224] @@ -10,7 +14,11 @@ def test_convnext_tiny_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_convnext_tiny_224_batch_size2( benchmark, net=convnext_tiny_224, input_shape=[2, 3, 224, 224] @@ -18,7 +26,11 @@ def test_convnext_tiny_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_convnext_tiny_224_batch_size4( benchmark, net=convnext_tiny_224, input_shape=[4, 3, 224, 224] @@ -26,7 +38,11 @@ def test_convnext_tiny_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_convnext_tiny_224_batch_size8( benchmark, net=convnext_tiny_224, input_shape=[8, 3, 224, 224] @@ -34,7 +50,11 @@ def test_convnext_tiny_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_convnext_tiny_224_batch_size16( benchmark, net=convnext_tiny_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_crossformer.py b/benchmark/test_crossformer.py index 1c3e8b6a..5de651ea 100644 --- a/benchmark/test_crossformer.py +++ b/benchmark/test_crossformer.py @@ -2,7 +2,11 @@ import oneflow_benchmark from flowvision.models.crossformer import crossformer_tiny_patch4_group7_224 -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_crossformer_tiny_patch4_group7_224_batch_size1( benchmark, net=crossformer_tiny_patch4_group7_224, input_shape=[1, 3, 224, 224] @@ -10,7 +14,11 @@ def test_crossformer_tiny_patch4_group7_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_crossformer_tiny_patch4_group7_224_batch_size2( benchmark, net=crossformer_tiny_patch4_group7_224, input_shape=[2, 3, 224, 224] @@ -18,7 +26,11 @@ def test_crossformer_tiny_patch4_group7_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_crossformer_tiny_patch4_group7_224_batch_size4( benchmark, net=crossformer_tiny_patch4_group7_224, input_shape=[4, 3, 224, 224] @@ -26,7 +38,11 @@ def test_crossformer_tiny_patch4_group7_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_crossformer_tiny_patch4_group7_224_batch_size8( benchmark, net=crossformer_tiny_patch4_group7_224, input_shape=[8, 3, 224, 224] @@ -34,7 +50,11 @@ def test_crossformer_tiny_patch4_group7_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_crossformer_tiny_patch4_group7_224_batch_size16( benchmark, net=crossformer_tiny_patch4_group7_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_cswin.py b/benchmark/test_cswin.py index f9765334..8742a372 100644 --- a/benchmark/test_cswin.py +++ b/benchmark/test_cswin.py @@ -2,7 +2,11 @@ import oneflow_benchmark from flowvision.models.cswin import cswin_tiny_224 -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_cswin_tiny_224_batch_size1( benchmark, net=cswin_tiny_224, input_shape=[1, 3, 224, 224] @@ -10,7 +14,11 @@ def test_cswin_tiny_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_cswin_tiny_224_batch_size2( benchmark, net=cswin_tiny_224, input_shape=[2, 3, 224, 224] @@ -18,7 +26,11 @@ def test_cswin_tiny_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_cswin_tiny_224_batch_size4( benchmark, net=cswin_tiny_224, input_shape=[4, 3, 224, 224] @@ -26,7 +38,11 @@ def test_cswin_tiny_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_cswin_tiny_224_batch_size8( benchmark, net=cswin_tiny_224, input_shape=[8, 3, 224, 224] @@ -34,7 +50,11 @@ def test_cswin_tiny_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_cswin_tiny_224_batch_size16( benchmark, net=cswin_tiny_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_densenet.py b/benchmark/test_densenet.py index 0310617a..9576fb43 100644 --- a/benchmark/test_densenet.py +++ b/benchmark/test_densenet.py @@ -9,6 +9,7 @@ def test_densenet121_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size2( benchmark, net=densenet121, input_shape=[2, 3, 224, 224] @@ -16,6 +17,7 @@ def test_densenet121_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size4( benchmark, net=densenet121, input_shape=[4, 3, 224, 224] @@ -23,6 +25,7 @@ def test_densenet121_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size8( benchmark, net=densenet121, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_ghostnet.py b/benchmark/test_ghostnet.py index 7a014d7b..e2373b55 100644 --- a/benchmark/test_ghostnet.py +++ b/benchmark/test_ghostnet.py @@ -2,31 +2,51 @@ import oneflow_benchmark from flowvision.models.ghostnet import ghostnet -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_ghostnet_batch_size1(benchmark, net=ghostnet, input_shape=[1, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_ghostnet_batch_size2(benchmark, net=ghostnet, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_ghostnet_batch_size4(benchmark, net=ghostnet, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_ghostnet_batch_size8(benchmark, net=ghostnet, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_ghostnet_batch_size16(benchmark, net=ghostnet, input_shape=[16, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) diff --git a/benchmark/test_inception.py b/benchmark/test_inception.py index 63fc962f..2f65d01c 100644 --- a/benchmark/test_inception.py +++ b/benchmark/test_inception.py @@ -10,6 +10,7 @@ def test_inception_v3_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_inception_v3_batch_size2( benchmark, net=inception_v3, input_shape=[2, 3, 299, 299] @@ -17,6 +18,7 @@ def test_inception_v3_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_inception_v3_batch_size4( benchmark, net=inception_v3, input_shape=[4, 3, 299, 299] @@ -24,6 +26,7 @@ def test_inception_v3_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_inception_v3_batch_size8( benchmark, net=inception_v3, input_shape=[8, 3, 299, 299] diff --git a/benchmark/test_mlp_mixer.py b/benchmark/test_mlp_mixer.py index 06e010cc..bfea08b3 100644 --- a/benchmark/test_mlp_mixer.py +++ b/benchmark/test_mlp_mixer.py @@ -2,7 +2,11 @@ import oneflow_benchmark from flowvision.models.mlp_mixer import mlp_mixer_b16_224 -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mlp_mixer_b16_224_batch_size1( benchmark, net=mlp_mixer_b16_224, input_shape=[1, 3, 224, 224] @@ -10,7 +14,11 @@ def test_mlp_mixer_b16_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mlp_mixer_b16_224_batch_size2( benchmark, net=mlp_mixer_b16_224, input_shape=[2, 3, 224, 224] @@ -18,7 +26,11 @@ def test_mlp_mixer_b16_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mlp_mixer_b16_224_batch_size4( benchmark, net=mlp_mixer_b16_224, input_shape=[4, 3, 224, 224] @@ -26,7 +38,11 @@ def test_mlp_mixer_b16_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mlp_mixer_b16_224_batch_size8( benchmark, net=mlp_mixer_b16_224, input_shape=[8, 3, 224, 224] @@ -34,7 +50,11 @@ def test_mlp_mixer_b16_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mlp_mixer_b16_224_batch_size16( benchmark, net=mlp_mixer_b16_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_mnasnet.py b/benchmark/test_mnasnet.py index 46648dc8..00ef3128 100644 --- a/benchmark/test_mnasnet.py +++ b/benchmark/test_mnasnet.py @@ -2,7 +2,11 @@ import oneflow_benchmark from flowvision.models.mnasnet import mnasnet0_5 -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mnasnet0_5_batch_size1( benchmark, net=mnasnet0_5, input_shape=[1, 3, 224, 224] @@ -10,7 +14,11 @@ def test_mnasnet0_5_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mnasnet0_5_batch_size2( benchmark, net=mnasnet0_5, input_shape=[2, 3, 224, 224] @@ -18,7 +26,11 @@ def test_mnasnet0_5_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mnasnet0_5_batch_size4( benchmark, net=mnasnet0_5, input_shape=[4, 3, 224, 224] @@ -26,7 +38,11 @@ def test_mnasnet0_5_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mnasnet0_5_batch_size8( benchmark, net=mnasnet0_5, input_shape=[8, 3, 224, 224] @@ -34,7 +50,11 @@ def test_mnasnet0_5_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mnasnet0_5_batch_size16( benchmark, net=mnasnet0_5, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_mobilenet.py b/benchmark/test_mobilenet.py index 26acc70e..914eca82 100644 --- a/benchmark/test_mobilenet.py +++ b/benchmark/test_mobilenet.py @@ -11,6 +11,7 @@ def test_mobilenet_v3_large_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v3_large_batch_size2( benchmark, net=mobilenet_v3_large, input_shape=[2, 3, 224, 224] @@ -18,6 +19,7 @@ def test_mobilenet_v3_large_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v3_large_batch_size4( benchmark, net=mobilenet_v3_large, input_shape=[4, 3, 224, 224] @@ -25,6 +27,7 @@ def test_mobilenet_v3_large_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v3_large_batch_size8( benchmark, net=mobilenet_v3_large, input_shape=[8, 3, 224, 224] @@ -47,6 +50,7 @@ def test_mobilenet_v2_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size2( benchmark, net=mobilenet_v2, input_shape=[2, 3, 224, 224] @@ -54,6 +58,7 @@ def test_mobilenet_v2_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size4( benchmark, net=mobilenet_v2, input_shape=[4, 3, 224, 224] @@ -61,6 +66,7 @@ def test_mobilenet_v2_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size8( benchmark, net=mobilenet_v2, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_pvt.py b/benchmark/test_pvt.py index 7d296b57..2cf37322 100644 --- a/benchmark/test_pvt.py +++ b/benchmark/test_pvt.py @@ -3,61 +3,101 @@ from flowvision.models.pvt import pvt_tiny from flowvision.models.pvt import pvt_small -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_tiny_batch_size1(benchmark, net=pvt_tiny, input_shape=[1, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_tiny_batch_size2(benchmark, net=pvt_tiny, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_tiny_batch_size4(benchmark, net=pvt_tiny, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_tiny_batch_size8(benchmark, net=pvt_tiny, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_tiny_batch_size16(benchmark, net=pvt_tiny, input_shape=[16, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_small_batch_size1(benchmark, net=pvt_small, input_shape=[1, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_small_batch_size2(benchmark, net=pvt_small, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_small_batch_size4(benchmark, net=pvt_small, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_small_batch_size8(benchmark, net=pvt_small, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_pvt_small_batch_size16( benchmark, net=pvt_small, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_res2net.py b/benchmark/test_res2net.py index 5d3d036e..b754e9fd 100644 --- a/benchmark/test_res2net.py +++ b/benchmark/test_res2net.py @@ -2,7 +2,11 @@ import oneflow_benchmark from flowvision.models.res2net import res2net50_26w_4s -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_res2net50_26w_4s_batch_size1( benchmark, net=res2net50_26w_4s, input_shape=[1, 3, 224, 224] @@ -10,7 +14,11 @@ def test_res2net50_26w_4s_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_res2net50_26w_4s_batch_size2( benchmark, net=res2net50_26w_4s, input_shape=[2, 3, 224, 224] @@ -18,7 +26,11 @@ def test_res2net50_26w_4s_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_res2net50_26w_4s_batch_size4( benchmark, net=res2net50_26w_4s, input_shape=[4, 3, 224, 224] @@ -26,7 +38,11 @@ def test_res2net50_26w_4s_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_res2net50_26w_4s_batch_size8( benchmark, net=res2net50_26w_4s, input_shape=[8, 3, 224, 224] @@ -34,7 +50,11 @@ def test_res2net50_26w_4s_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_res2net50_26w_4s_batch_size16( benchmark, net=res2net50_26w_4s, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_res_mlp.py b/benchmark/test_res_mlp.py index 3e870425..b2ef7de1 100644 --- a/benchmark/test_res_mlp.py +++ b/benchmark/test_res_mlp.py @@ -2,7 +2,11 @@ import oneflow_benchmark from flowvision.models.res_mlp import resmlp_12_224 -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resmlp_12_224_batch_size1( benchmark, net=resmlp_12_224, input_shape=[1, 3, 224, 224] @@ -10,7 +14,11 @@ def test_resmlp_12_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resmlp_12_224_batch_size2( benchmark, net=resmlp_12_224, input_shape=[2, 3, 224, 224] @@ -18,7 +26,11 @@ def test_resmlp_12_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resmlp_12_224_batch_size4( benchmark, net=resmlp_12_224, input_shape=[4, 3, 224, 224] @@ -26,7 +38,11 @@ def test_resmlp_12_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resmlp_12_224_batch_size8( benchmark, net=resmlp_12_224, input_shape=[8, 3, 224, 224] @@ -34,7 +50,11 @@ def test_resmlp_12_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resmlp_12_224_batch_size16( benchmark, net=resmlp_12_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_resnet.py b/benchmark/test_resnet.py index b8967d05..bdb9ecc0 100644 --- a/benchmark/test_resnet.py +++ b/benchmark/test_resnet.py @@ -10,16 +10,19 @@ def test_resnet50_batch_size1(benchmark, net=resnet50, input_shape=[1, 3, 224, 2 model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnet50_batch_size2(benchmark, net=resnet50, input_shape=[2, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnet50_batch_size4(benchmark, net=resnet50, input_shape=[4, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnet50_batch_size8(benchmark, net=resnet50, input_shape=[8, 3, 224, 224]): model, x, optimizer = fetch_args(net, input_shape) @@ -39,6 +42,7 @@ def test_resnext50_32x4d_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size2( benchmark, net=resnext50_32x4d, input_shape=[2, 3, 224, 224] @@ -46,6 +50,7 @@ def test_resnext50_32x4d_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size4( benchmark, net=resnext50_32x4d, input_shape=[4, 3, 224, 224] @@ -53,6 +58,7 @@ def test_resnext50_32x4d_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size8( benchmark, net=resnext50_32x4d, input_shape=[8, 3, 224, 224] @@ -60,6 +66,7 @@ def test_resnext50_32x4d_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_resnext50_32x4d_batch_size16( benchmark, net=resnext50_32x4d, input_shape=[16, 3, 224, 224] @@ -67,6 +74,7 @@ def test_resnext50_32x4d_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size1( benchmark, net=wide_resnet50_2, input_shape=[1, 3, 224, 224] @@ -74,6 +82,7 @@ def test_wide_resnet50_2_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size2( benchmark, net=wide_resnet50_2, input_shape=[2, 3, 224, 224] @@ -81,6 +90,7 @@ def test_wide_resnet50_2_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size4( benchmark, net=wide_resnet50_2, input_shape=[4, 3, 224, 224] @@ -88,6 +98,7 @@ def test_wide_resnet50_2_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_wide_resnet50_2_batch_size8( benchmark, net=wide_resnet50_2, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_rexnet.py b/benchmark/test_rexnet.py index f30970db..2c71a51c 100644 --- a/benchmark/test_rexnet.py +++ b/benchmark/test_rexnet.py @@ -3,7 +3,11 @@ from flowvision.models.rexnet import rexnetv1_1_0 from flowvision.models.rexnet_lite import rexnet_lite_1_0 -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnetv1_1_0_batch_size1( benchmark, net=rexnetv1_1_0, input_shape=[1, 3, 224, 224] @@ -11,7 +15,11 @@ def test_rexnetv1_1_0_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnetv1_1_0_batch_size2( benchmark, net=rexnetv1_1_0, input_shape=[2, 3, 224, 224] @@ -19,7 +27,11 @@ def test_rexnetv1_1_0_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnetv1_1_0_batch_size4( benchmark, net=rexnetv1_1_0, input_shape=[4, 3, 224, 224] @@ -27,7 +39,11 @@ def test_rexnetv1_1_0_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnetv1_1_0_batch_size8( benchmark, net=rexnetv1_1_0, input_shape=[8, 3, 224, 224] @@ -35,7 +51,11 @@ def test_rexnetv1_1_0_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnetv1_1_0_batch_size16( benchmark, net=rexnetv1_1_0, input_shape=[16, 3, 224, 224] @@ -43,7 +63,11 @@ def test_rexnetv1_1_0_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnet_lite_1_0_batch_size1( benchmark, net=rexnet_lite_1_0, input_shape=[1, 3, 224, 224] @@ -51,7 +75,11 @@ def test_rexnet_lite_1_0_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnet_lite_1_0_batch_size2( benchmark, net=rexnet_lite_1_0, input_shape=[2, 3, 224, 224] @@ -59,7 +87,11 @@ def test_rexnet_lite_1_0_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnet_lite_1_0_batch_size4( benchmark, net=rexnet_lite_1_0, input_shape=[4, 3, 224, 224] @@ -67,7 +99,11 @@ def test_rexnet_lite_1_0_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnet_lite_1_0_batch_size8( benchmark, net=rexnet_lite_1_0, input_shape=[8, 3, 224, 224] @@ -75,7 +111,11 @@ def test_rexnet_lite_1_0_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_rexnet_lite_1_0_batch_size16( benchmark, net=rexnet_lite_1_0, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_shufflenet.py b/benchmark/test_shufflenet.py index 88bed665..5b7f774e 100644 --- a/benchmark/test_shufflenet.py +++ b/benchmark/test_shufflenet.py @@ -10,6 +10,7 @@ def test_shufflenet_v2_x0_5_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_shufflenet_v2_x0_5_batch_size2( benchmark, net=shufflenet_v2_x0_5, input_shape=[2, 3, 224, 224] @@ -17,6 +18,7 @@ def test_shufflenet_v2_x0_5_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_shufflenet_v2_x0_5_batch_size4( benchmark, net=shufflenet_v2_x0_5, input_shape=[4, 3, 224, 224] @@ -24,6 +26,7 @@ def test_shufflenet_v2_x0_5_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_shufflenet_v2_x0_5_batch_size8( benchmark, net=shufflenet_v2_x0_5, input_shape=[8, 3, 224, 224] diff --git a/benchmark/test_squeezenet.py b/benchmark/test_squeezenet.py index 5a539c4a..07a95507 100644 --- a/benchmark/test_squeezenet.py +++ b/benchmark/test_squeezenet.py @@ -2,7 +2,11 @@ import oneflow_benchmark from flowvision.models.squeezenet import squeezenet1_0 -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_squeezenet1_0_batch_size1( benchmark, net=squeezenet1_0, input_shape=[1, 3, 224, 224] @@ -10,7 +14,11 @@ def test_squeezenet1_0_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_squeezenet1_0_batch_size2( benchmark, net=squeezenet1_0, input_shape=[2, 3, 224, 224] @@ -18,7 +26,11 @@ def test_squeezenet1_0_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_squeezenet1_0_batch_size4( benchmark, net=squeezenet1_0, input_shape=[4, 3, 224, 224] @@ -26,7 +38,11 @@ def test_squeezenet1_0_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_squeezenet1_0_batch_size8( benchmark, net=squeezenet1_0, input_shape=[8, 3, 224, 224] @@ -34,7 +50,11 @@ def test_squeezenet1_0_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_squeezenet1_0_batch_size16( benchmark, net=squeezenet1_0, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_swin_transformer.py b/benchmark/test_swin_transformer.py index aa460dab..875ecf60 100644 --- a/benchmark/test_swin_transformer.py +++ b/benchmark/test_swin_transformer.py @@ -4,7 +4,11 @@ from flowvision.models.swin_transformer import swin_small_patch4_window7_224 from flowvision.models.swin_transformer import swin_base_patch4_window7_224 -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_tiny_patch4_window7_224_batch_size1( benchmark, net=swin_tiny_patch4_window7_224, input_shape=[1, 3, 224, 224] @@ -12,7 +16,11 @@ def test_swin_tiny_patch4_window7_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_tiny_patch4_window7_224_batch_size2( benchmark, net=swin_tiny_patch4_window7_224, input_shape=[2, 3, 224, 224] @@ -20,7 +28,11 @@ def test_swin_tiny_patch4_window7_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_tiny_patch4_window7_224_batch_size4( benchmark, net=swin_tiny_patch4_window7_224, input_shape=[4, 3, 224, 224] @@ -28,7 +40,11 @@ def test_swin_tiny_patch4_window7_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_tiny_patch4_window7_224_batch_size8( benchmark, net=swin_tiny_patch4_window7_224, input_shape=[8, 3, 224, 224] @@ -36,7 +52,11 @@ def test_swin_tiny_patch4_window7_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_tiny_patch4_window7_224_batch_size16( benchmark, net=swin_tiny_patch4_window7_224, input_shape=[16, 3, 224, 224] @@ -44,7 +64,11 @@ def test_swin_tiny_patch4_window7_224_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_small_patch4_window7_224_batch_size1( benchmark, net=swin_small_patch4_window7_224, input_shape=[1, 3, 224, 224] @@ -52,7 +76,11 @@ def test_swin_small_patch4_window7_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_small_patch4_window7_224_batch_size2( benchmark, net=swin_small_patch4_window7_224, input_shape=[2, 3, 224, 224] @@ -60,7 +88,11 @@ def test_swin_small_patch4_window7_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_small_patch4_window7_224_batch_size4( benchmark, net=swin_small_patch4_window7_224, input_shape=[4, 3, 224, 224] @@ -68,7 +100,11 @@ def test_swin_small_patch4_window7_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_small_patch4_window7_224_batch_size8( benchmark, net=swin_small_patch4_window7_224, input_shape=[8, 3, 224, 224] @@ -76,7 +112,11 @@ def test_swin_small_patch4_window7_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_small_patch4_window7_224_batch_size16( benchmark, net=swin_small_patch4_window7_224, input_shape=[16, 3, 224, 224] @@ -91,6 +131,7 @@ def test_swin_base_patch4_window7_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size2( benchmark, net=swin_base_patch4_window7_224, input_shape=[2, 3, 224, 224] @@ -98,6 +139,7 @@ def test_swin_base_patch4_window7_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size4( benchmark, net=swin_base_patch4_window7_224, input_shape=[4, 3, 224, 224] @@ -105,6 +147,7 @@ def test_swin_base_patch4_window7_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size8( benchmark, net=swin_base_patch4_window7_224, input_shape=[8, 3, 224, 224] @@ -112,6 +155,7 @@ def test_swin_base_patch4_window7_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size16( benchmark, net=swin_base_patch4_window7_224, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_uniformer.py b/benchmark/test_uniformer.py index 0a28ea69..5b38a15c 100644 --- a/benchmark/test_uniformer.py +++ b/benchmark/test_uniformer.py @@ -5,6 +5,7 @@ from flowvision.models.uniformer import uniformer_base_ls from flowvision.models.uniformer import uniformer_small + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size1( benchmark, net=uniformer_base, input_shape=[1, 3, 224, 224] @@ -12,6 +13,7 @@ def test_uniformer_base_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size2( benchmark, net=uniformer_base, input_shape=[2, 3, 224, 224] @@ -19,6 +21,7 @@ def test_uniformer_base_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size4( benchmark, net=uniformer_base, input_shape=[4, 3, 224, 224] @@ -26,6 +29,7 @@ def test_uniformer_base_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size8( benchmark, net=uniformer_base, input_shape=[8, 3, 224, 224] @@ -33,6 +37,7 @@ def test_uniformer_base_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_batch_size16( benchmark, net=uniformer_base, input_shape=[16, 3, 224, 224] @@ -40,7 +45,11 @@ def test_uniformer_base_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_ls_batch_size1( benchmark, net=uniformer_base_ls, input_shape=[1, 3, 224, 224] @@ -48,7 +57,11 @@ def test_uniformer_base_ls_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_ls_batch_size2( benchmark, net=uniformer_base_ls, input_shape=[2, 3, 224, 224] @@ -56,7 +69,11 @@ def test_uniformer_base_ls_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_ls_batch_size4( benchmark, net=uniformer_base_ls, input_shape=[4, 3, 224, 224] @@ -64,7 +81,11 @@ def test_uniformer_base_ls_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_ls_batch_size8( benchmark, net=uniformer_base_ls, input_shape=[8, 3, 224, 224] @@ -72,7 +93,11 @@ def test_uniformer_base_ls_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_base_ls_batch_size16( benchmark, net=uniformer_base_ls, input_shape=[16, 3, 224, 224] @@ -80,7 +105,11 @@ def test_uniformer_base_ls_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_batch_size1( benchmark, net=uniformer_small, input_shape=[1, 3, 224, 224] @@ -88,7 +117,11 @@ def test_uniformer_small_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_batch_size2( benchmark, net=uniformer_small, input_shape=[2, 3, 224, 224] @@ -96,7 +129,11 @@ def test_uniformer_small_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_batch_size4( benchmark, net=uniformer_small, input_shape=[4, 3, 224, 224] @@ -104,7 +141,11 @@ def test_uniformer_small_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_batch_size8( benchmark, net=uniformer_small, input_shape=[8, 3, 224, 224] @@ -112,7 +153,11 @@ def test_uniformer_small_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_batch_size16( benchmark, net=uniformer_small, input_shape=[16, 3, 224, 224] @@ -120,7 +165,11 @@ def test_uniformer_small_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_plus_batch_size1( benchmark, net=uniformer_small_plus, input_shape=[1, 3, 224, 224] @@ -128,7 +177,11 @@ def test_uniformer_small_plus_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_plus_batch_size2( benchmark, net=uniformer_small_plus, input_shape=[2, 3, 224, 224] @@ -136,7 +189,11 @@ def test_uniformer_small_plus_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_plus_batch_size4( benchmark, net=uniformer_small_plus, input_shape=[4, 3, 224, 224] @@ -144,7 +201,11 @@ def test_uniformer_small_plus_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_plus_batch_size8( benchmark, net=uniformer_small_plus, input_shape=[8, 3, 224, 224] @@ -152,7 +213,11 @@ def test_uniformer_small_plus_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_uniformer_small_plus_batch_size16( benchmark, net=uniformer_small_plus, input_shape=[16, 3, 224, 224] diff --git a/benchmark/test_vision_transformer.py b/benchmark/test_vision_transformer.py index 77eeb878..695230b4 100644 --- a/benchmark/test_vision_transformer.py +++ b/benchmark/test_vision_transformer.py @@ -4,7 +4,11 @@ from flowvision.models.vision_transformer import vit_small_patch16_224 from flowvision.models.vision_transformer import vit_base_patch16_224 -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_tiny_patch16_224_batch_size1( benchmark, net=vit_tiny_patch16_224, input_shape=[1, 3, 224, 224] @@ -12,7 +16,11 @@ def test_vit_tiny_patch16_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_tiny_patch16_224_batch_size2( benchmark, net=vit_tiny_patch16_224, input_shape=[2, 3, 224, 224] @@ -20,7 +28,11 @@ def test_vit_tiny_patch16_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_tiny_patch16_224_batch_size4( benchmark, net=vit_tiny_patch16_224, input_shape=[4, 3, 224, 224] @@ -28,7 +40,11 @@ def test_vit_tiny_patch16_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_tiny_patch16_224_batch_size8( benchmark, net=vit_tiny_patch16_224, input_shape=[8, 3, 224, 224] @@ -36,7 +52,11 @@ def test_vit_tiny_patch16_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_tiny_patch16_224_batch_size16( benchmark, net=vit_tiny_patch16_224, input_shape=[16, 3, 224, 224] @@ -44,7 +64,11 @@ def test_vit_tiny_patch16_224_batch_size16( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_small_patch16_224_batch_size1( benchmark, net=vit_small_patch16_224, input_shape=[1, 3, 224, 224] @@ -52,7 +76,11 @@ def test_vit_small_patch16_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_small_patch16_224_batch_size2( benchmark, net=vit_small_patch16_224, input_shape=[2, 3, 224, 224] @@ -60,7 +88,11 @@ def test_vit_small_patch16_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_small_patch16_224_batch_size4( benchmark, net=vit_small_patch16_224, input_shape=[4, 3, 224, 224] @@ -68,7 +100,11 @@ def test_vit_small_patch16_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_small_patch16_224_batch_size8( benchmark, net=vit_small_patch16_224, input_shape=[8, 3, 224, 224] @@ -76,7 +112,11 @@ def test_vit_small_patch16_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) -@unittest.skipUnless(os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test") + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_small_patch16_224_batch_size16( benchmark, net=vit_small_patch16_224, input_shape=[16, 3, 224, 224] @@ -92,6 +132,7 @@ def test_vit_base_patch16_224_batch_size1( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size2( benchmark, net=vit_base_patch16_224, input_shape=[2, 3, 224, 224] @@ -99,6 +140,7 @@ def test_vit_base_patch16_224_batch_size2( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size4( benchmark, net=vit_base_patch16_224, input_shape=[4, 3, 224, 224] @@ -106,6 +148,7 @@ def test_vit_base_patch16_224_batch_size4( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size8( benchmark, net=vit_base_patch16_224, input_shape=[8, 3, 224, 224] @@ -113,6 +156,7 @@ def test_vit_base_patch16_224_batch_size8( model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) + @oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_vit_base_patch16_224_batch_size16( benchmark, net=vit_base_patch16_224, input_shape=[16, 3, 224, 224] From 8f3c2c38e4bcdeb0d1043e3b8e3e27a21d2cfcc1 Mon Sep 17 00:00:00 2001 From: Shenghang Tsai Date: Thu, 14 Apr 2022 20:39:18 +0800 Subject: [PATCH 4/7] fix --- benchmark/test_mobilenet.py | 8 ++++++-- benchmark/test_swin_transformer.py | 8 ++++++-- 2 files changed, 12 insertions(+), 4 deletions(-) diff --git a/benchmark/test_mobilenet.py b/benchmark/test_mobilenet.py index 914eca82..8e110bb8 100644 --- a/benchmark/test_mobilenet.py +++ b/benchmark/test_mobilenet.py @@ -42,8 +42,12 @@ def test_mobilenet_v3_large_batch_size16( ): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - - + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) +@oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_mobilenet_v2_batch_size1( benchmark, net=mobilenet_v2, input_shape=[1, 3, 224, 224] ): diff --git a/benchmark/test_swin_transformer.py b/benchmark/test_swin_transformer.py index 875ecf60..c0648f77 100644 --- a/benchmark/test_swin_transformer.py +++ b/benchmark/test_swin_transformer.py @@ -123,8 +123,12 @@ def test_swin_small_patch4_window7_224_batch_size16( ): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - - + +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) +@oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_swin_base_patch4_window7_224_batch_size1( benchmark, net=swin_base_patch4_window7_224, input_shape=[1, 3, 224, 224] ): From e134bf39514dbb5d94831946fe004bc6bc113e86 Mon Sep 17 00:00:00 2001 From: oneflow-ci-bot Date: Thu, 14 Apr 2022 12:40:18 +0000 Subject: [PATCH 5/7] auto format by CI --- benchmark/test_mobilenet.py | 3 ++- benchmark/test_swin_transformer.py | 3 ++- 2 files changed, 4 insertions(+), 2 deletions(-) diff --git a/benchmark/test_mobilenet.py b/benchmark/test_mobilenet.py index 8e110bb8..d00e40bf 100644 --- a/benchmark/test_mobilenet.py +++ b/benchmark/test_mobilenet.py @@ -42,7 +42,8 @@ def test_mobilenet_v3_large_batch_size16( ): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - + + @unittest.skipUnless( os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test", diff --git a/benchmark/test_swin_transformer.py b/benchmark/test_swin_transformer.py index c0648f77..57a3953b 100644 --- a/benchmark/test_swin_transformer.py +++ b/benchmark/test_swin_transformer.py @@ -123,7 +123,8 @@ def test_swin_small_patch4_window7_224_batch_size16( ): model, x, optimizer = fetch_args(net, input_shape) benchmark(run, model, x, optimizer) - + + @unittest.skipUnless( os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test", From 2c6f7bf265da787c7fc2177f6e33b213eefe264e Mon Sep 17 00:00:00 2001 From: Shenghang Tsai Date: Thu, 14 Apr 2022 20:41:26 +0800 Subject: [PATCH 6/7] quick fix --- benchmark/test_densenet.py | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/benchmark/test_densenet.py b/benchmark/test_densenet.py index 9576fb43..97e182a0 100644 --- a/benchmark/test_densenet.py +++ b/benchmark/test_densenet.py @@ -2,7 +2,11 @@ import oneflow_benchmark from flowvision.models.densenet import densenet121 - +@unittest.skipUnless( + os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", + "set ONEFLOW_BENCHMARK_ALL=1 to run this test", +) +@oneflow_benchmark.ci_settings(compare={"median": "5%"}) def test_densenet121_batch_size1( benchmark, net=densenet121, input_shape=[1, 3, 224, 224] ): From ca8ebc663b58667cf8cd1b6ef0c861522780b7bb Mon Sep 17 00:00:00 2001 From: oneflow-ci-bot Date: Thu, 14 Apr 2022 12:42:11 +0000 Subject: [PATCH 7/7] auto format by CI --- benchmark/test_densenet.py | 1 + 1 file changed, 1 insertion(+) diff --git a/benchmark/test_densenet.py b/benchmark/test_densenet.py index 97e182a0..68fcbfa1 100644 --- a/benchmark/test_densenet.py +++ b/benchmark/test_densenet.py @@ -2,6 +2,7 @@ import oneflow_benchmark from flowvision.models.densenet import densenet121 + @unittest.skipUnless( os.getenv("ONEFLOW_BENCHMARK_ALL") == "1", "set ONEFLOW_BENCHMARK_ALL=1 to run this test",