diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_distilled_patch16_224/SIR_4.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_distilled_patch16_224/SIR_4.py index 35cd88be74..53555f906a 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_distilled_patch16_224/SIR_4.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_distilled_patch16_224/SIR_4.py @@ -35,6 +35,7 @@ def forward( self, var_0, # (shape: [86, 198, 192], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[192], weight=self.parameter_0, bias=self.parameter_1, epsilon=1e-06) var_2 = paddle.nn.functional.common.linear(x=var_1, weight=self.parameter_5, bias=self.parameter_3, name=None) var_3 = var_2.reshape((-1, 198, 3, 3, 64,)) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_distilled_patch16_224/SIR_56.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_distilled_patch16_224/SIR_56.py index 387e1a8151..1ca00a2704 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_distilled_patch16_224/SIR_56.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_distilled_patch16_224/SIR_56.py @@ -35,6 +35,7 @@ def forward( self, var_0, # (shape: [54, 198, 192], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[192], weight=self.parameter_3, bias=self.parameter_4, epsilon=1e-06) var_2 = paddle.nn.functional.common.linear(x=var_1, weight=self.parameter_0, bias=self.parameter_5, name=None) var_3 = var_2.reshape((-1, 198, 3, 3, 64,)) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_patch16_224/SIR_4.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_patch16_224/SIR_4.py index 494b38151a..a54e05ddde 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_patch16_224/SIR_4.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_patch16_224/SIR_4.py @@ -35,6 +35,7 @@ def forward( self, var_0, # (shape: [86, 197, 192], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[192], weight=self.parameter_1, bias=self.parameter_4, epsilon=1e-06) var_2 = paddle.nn.functional.common.linear(x=var_1, weight=self.parameter_2, bias=self.parameter_0, name=None) var_3 = var_2.reshape((-1, 197, 3, 3, 64,)) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_patch16_224/SIR_56.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_patch16_224/SIR_56.py index 4d2c9d2a1a..a15e7b6858 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_patch16_224/SIR_56.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/DeiT_DeiT_tiny_patch16_224/SIR_56.py @@ -35,6 +35,7 @@ def forward( self, var_0, # (shape: [54, 197, 192], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[192], weight=self.parameter_4, bias=self.parameter_1, epsilon=1e-06) var_2 = paddle.nn.functional.common.linear(x=var_1, weight=self.parameter_3, bias=self.parameter_2, name=None) var_3 = var_2.reshape((-1, 197, 3, 3, 64,)) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/EfficientNet_EfficientNetB0/SIR_140.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/EfficientNet_EfficientNetB0/SIR_140.py index eb657a7d66..d359417040 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/EfficientNet_EfficientNetB0/SIR_140.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/EfficientNet_EfficientNetB0/SIR_140.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [11, 1280, 7, 7], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.pooling.adaptive_avg_pool2d(var_0, output_size=1, data_format='NCHW', name=None) var_2 = paddle.nn.functional.common.dropout(var_1, p=0.2, axis=None, training=True, mode='upscale_in_train', name=None) var_3 = paddle.tensor.manipulation.squeeze(var_2, axis=[2, 3]) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/EfficientNet_EfficientNetB0/SIR_96.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/EfficientNet_EfficientNetB0/SIR_96.py index 3e3f8ac8fe..781956af76 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/EfficientNet_EfficientNetB0/SIR_96.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/EfficientNet_EfficientNetB0/SIR_96.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [43, 1280, 7, 7], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.pooling.adaptive_avg_pool2d(var_0, output_size=1, data_format='NCHW', name=None) var_2 = paddle.nn.functional.common.dropout(var_1, p=0.2, axis=None, training=True, mode='upscale_in_train', name=None) var_3 = paddle.tensor.manipulation.squeeze(var_2, axis=[2, 3]) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Inception_InceptionV4/SIR_39.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Inception_InceptionV4/SIR_39.py index b6e14e98ef..6e6a9b8a56 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Inception_InceptionV4/SIR_39.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Inception_InceptionV4/SIR_39.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [22, 1536, 8, 8], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.pooling.adaptive_avg_pool2d(var_0, output_size=1, data_format='NCHW', name=None) var_2 = paddle.tensor.manipulation.squeeze(var_1, axis=[2, 3]) var_3 = paddle.nn.functional.common.dropout(var_2, p=0.2, axis=None, training=True, mode='downscale_in_infer', name=None) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Inception_InceptionV4/SIR_40.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Inception_InceptionV4/SIR_40.py index 0241671b64..7262a08d42 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Inception_InceptionV4/SIR_40.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Inception_InceptionV4/SIR_40.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [10, 1536, 8, 8], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.pooling.adaptive_avg_pool2d(var_0, output_size=1, data_format='NCHW', name=None) var_2 = paddle.tensor.manipulation.squeeze(var_1, axis=[2, 3]) var_3 = paddle.nn.functional.common.dropout(var_2, p=0.2, axis=None, training=True, mode='downscale_in_infer', name=None) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/MobileViT_MobileViT_XXS/SIR_26.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/MobileViT_MobileViT_XXS/SIR_26.py index d1678f6bdd..8c16e59c62 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/MobileViT_MobileViT_XXS/SIR_26.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/MobileViT_MobileViT_XXS/SIR_26.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [43, 320, 8, 8], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.pooling.adaptive_avg_pool2d(var_0, output_size=1, data_format='NCHW', name=None) var_2 = var_1.reshape([43, 320]) var_3 = paddle.nn.functional.common.dropout(var_2, p=0.1, axis=None, training=True, mode='upscale_in_train', name=None) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/MobileViT_MobileViT_XXS/SIR_27.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/MobileViT_MobileViT_XXS/SIR_27.py index 1e92ae3eb7..bb9fb3dc8a 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/MobileViT_MobileViT_XXS/SIR_27.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/MobileViT_MobileViT_XXS/SIR_27.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [11, 320, 8, 8], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.pooling.adaptive_avg_pool2d(var_0, output_size=1, data_format='NCHW', name=None) var_2 = var_1.reshape([11, 320]) var_3 = paddle.nn.functional.common.dropout(var_2, p=0.1, axis=None, training=True, mode='upscale_in_train', name=None) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/PeleeNet_PeleeNet/SIR_41.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/PeleeNet_PeleeNet/SIR_41.py index afd4f505d7..4e0c49bada 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/PeleeNet_PeleeNet/SIR_41.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/PeleeNet_PeleeNet/SIR_41.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [43, 704, 7, 7], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.pooling.avg_pool2d(var_0, kernel_size=[7, 7]) var_2 = var_1.flatten(1) var_3 = paddle.nn.functional.common.dropout(var_2, p=0.05, training=True) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/PeleeNet_PeleeNet/SIR_42.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/PeleeNet_PeleeNet/SIR_42.py index 5aafbb2da7..de558608ea 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/PeleeNet_PeleeNet/SIR_42.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/PeleeNet_PeleeNet/SIR_42.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [11, 704, 7, 7], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.pooling.avg_pool2d(var_0, kernel_size=[7, 7]) var_2 = var_1.flatten(1) var_3 = paddle.nn.functional.common.dropout(var_2, p=0.05, training=True) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/ReXNet_ReXNet_3_0/SIR_60.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/ReXNet_ReXNet_3_0/SIR_60.py index 3fa9b429fa..a9851348c7 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/ReXNet_ReXNet_3_0/SIR_60.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/ReXNet_ReXNet_3_0/SIR_60.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [22, 3840, 1, 1], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout(var_0, p=0.2, axis=None, training=True, mode='upscale_in_train', name=None) var_2 = paddle.nn.functional.conv._conv_nd(var_1, self.parameter_1, bias=self.parameter_0, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_3 = var_2.squeeze(axis=-1) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/ReXNet_ReXNet_3_0/SIR_61.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/ReXNet_ReXNet_3_0/SIR_61.py index a29d40c6db..42ed5ee84d 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/ReXNet_ReXNet_3_0/SIR_61.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/ReXNet_ReXNet_3_0/SIR_61.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [10, 3840, 1, 1], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout(var_0, p=0.2, axis=None, training=True, mode='upscale_in_train', name=None) var_2 = paddle.nn.functional.conv._conv_nd(var_1, self.parameter_0, bias=self.parameter_1, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_3 = var_2.squeeze(axis=-1) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_0.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_0.py index bfce4b074d..16a7c5d6e9 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_0.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_0.py @@ -219,6 +219,7 @@ def forward( self, var_0, # (shape: [43, 3, 224, 224], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_13, bias=self.parameter_10, stride=[2, 2], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = paddle.nn.functional.activation.relu(var_1) var_3 = paddle.nn.functional.pooling.max_pool2d(var_2, kernel_size=3, stride=2, padding=0, return_mask=False, ceil_mode=False, data_format='NCHW', name=None) @@ -322,7 +323,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_1.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_1.py index 8b40cdc01e..0ca56bfe59 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_1.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_1.py @@ -219,6 +219,7 @@ def forward( self, var_0, # (shape: [11, 3, 224, 224], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_34, bias=self.parameter_2, stride=[2, 2], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = paddle.nn.functional.activation.relu(var_1) var_3 = paddle.nn.functional.pooling.max_pool2d(var_2, kernel_size=3, stride=2, padding=0, return_mask=False, ceil_mode=False, data_format='NCHW', name=None) @@ -322,7 +323,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_2.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_2.py index 107a91c7f7..56a491f7eb 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_2.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_2.py @@ -219,6 +219,7 @@ def forward( self, var_0, # (shape: [22, 3, 224, 224], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_1, bias=self.parameter_42, stride=[2, 2], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = paddle.nn.functional.activation.relu(var_1) var_3 = paddle.nn.functional.pooling.max_pool2d(var_2, kernel_size=3, stride=2, padding=0, return_mask=False, ceil_mode=False, data_format='NCHW', name=None) @@ -322,7 +323,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_3.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_3.py index c51366e2ca..f9c962fdb5 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_3.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/SqueezeNet_SqueezeNet1_0/SIR_3.py @@ -219,6 +219,7 @@ def forward( self, var_0, # (shape: [10, 3, 224, 224], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_24, bias=self.parameter_46, stride=[2, 2], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = paddle.nn.functional.activation.relu(var_1) var_3 = paddle.nn.functional.pooling.max_pool2d(var_2, kernel_size=3, stride=2, padding=0, return_mask=False, ceil_mode=False, data_format='NCHW', name=None) @@ -322,7 +323,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_108.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_108.py index 9c36a67b7d..1e3eaa1721 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_108.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_108.py @@ -27,6 +27,7 @@ def forward( self, var_0, # (shape: [11, 192, 28, 28], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_0, bias=self.parameter_3, stride=[2, 2], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = var_1.flatten(2) var_3 = var_2.transpose([0, 2, 1]) @@ -69,7 +70,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_136.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_136.py index 3fe32b2856..9c4d75e67d 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_136.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_136.py @@ -27,6 +27,7 @@ def forward( self, var_0, # (shape: [11, 384, 14, 14], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_1, bias=self.parameter_2, stride=[2, 2], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = var_1.flatten(2) var_3 = var_2.transpose([0, 2, 1]) @@ -69,7 +70,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_14.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_14.py index 2957cbe252..f45422bef1 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_14.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_14.py @@ -27,6 +27,7 @@ def forward( self, var_0, # (shape: [43, 96, 56, 56], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_1, bias=self.parameter_0, stride=[2, 2], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = var_1.flatten(2) var_3 = var_2.transpose([0, 2, 1]) @@ -69,7 +70,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_2.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_2.py index a03cc2c61a..4b6dad9154 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_2.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_2.py @@ -27,6 +27,7 @@ def forward( self, var_0, # (shape: [43, 3, 224, 224], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_3, bias=self.parameter_2, stride=[4, 4], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = var_1.flatten(2) var_3 = var_2.transpose([0, 2, 1]) @@ -69,7 +70,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_26.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_26.py index d7278db900..c171476c89 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_26.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_26.py @@ -27,6 +27,7 @@ def forward( self, var_0, # (shape: [43, 192, 28, 28], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_3, bias=self.parameter_1, stride=[2, 2], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = var_1.flatten(2) var_3 = var_2.transpose([0, 2, 1]) @@ -69,7 +70,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_70.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_70.py index 25f0d55b4c..9a4e976ff5 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_70.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_70.py @@ -27,6 +27,7 @@ def forward( self, var_0, # (shape: [43, 384, 14, 14], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_1, bias=self.parameter_0, stride=[2, 2], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = var_1.flatten(2) var_3 = var_2.transpose([0, 2, 1]) @@ -69,7 +70,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_84.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_84.py index 0f71b8ec45..93e93954c5 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_84.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_84.py @@ -27,6 +27,7 @@ def forward( self, var_0, # (shape: [11, 3, 224, 224], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_1, bias=self.parameter_3, stride=[4, 4], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = var_1.flatten(2) var_3 = var_2.transpose([0, 2, 1]) @@ -69,7 +70,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_96.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_96.py index 6ab271812e..8a66e139e3 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_96.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Twins_alt_gvt_base/SIR_96.py @@ -27,6 +27,7 @@ def forward( self, var_0, # (shape: [11, 96, 56, 56], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_3, bias=self.parameter_2, stride=[2, 2], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = var_1.flatten(2) var_3 = var_2.transpose([0, 2, 1]) @@ -69,7 +70,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Xception_Xception65_deeplab/SIR_35.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Xception_Xception65_deeplab/SIR_35.py index 2c64b37969..4bfec374f2 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Xception_Xception65_deeplab/SIR_35.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Xception_Xception65_deeplab/SIR_35.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [22, 2048, 10, 10], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout(var_0, p=0.5, axis=None, training=True, mode='downscale_in_infer', name=None) var_2 = paddle.nn.functional.pooling.adaptive_avg_pool2d(var_1, output_size=1, data_format='NCHW', name=None) var_3 = paddle.tensor.manipulation.squeeze(var_2, axis=[2, 3]) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Xception_Xception65_deeplab/SIR_55.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Xception_Xception65_deeplab/SIR_55.py index f74397f710..994ccab514 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Xception_Xception65_deeplab/SIR_55.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Clas_cases/Xception_Xception65_deeplab/SIR_55.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [10, 2048, 10, 10], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout(var_0, p=0.5, axis=None, training=True, mode='downscale_in_infer', name=None) var_2 = paddle.nn.functional.pooling.adaptive_avg_pool2d(var_1, output_size=1, data_format='NCHW', name=None) var_3 = paddle.tensor.manipulation.squeeze(var_2, axis=[2, 3]) @@ -60,7 +61,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/faster_rcnn_faster_rcnn_swin_tiny_fpn_1x_coco/SIR_3.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/faster_rcnn_faster_rcnn_swin_tiny_fpn_1x_coco/SIR_3.py index e51d1a1279..0ea90a31ef 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/faster_rcnn_faster_rcnn_swin_tiny_fpn_1x_coco/SIR_3.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/faster_rcnn_faster_rcnn_swin_tiny_fpn_1x_coco/SIR_3.py @@ -27,6 +27,7 @@ def forward( self, var_0, # (shape: [1, 3, 800, 1216], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_0, bias=self.parameter_2, stride=[4, 4], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = var_1.flatten(2) var_3 = var_2.transpose([0, 2, 1]) @@ -73,7 +74,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/faster_rcnn_faster_rcnn_swin_tiny_fpn_2x_coco/SIR_3.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/faster_rcnn_faster_rcnn_swin_tiny_fpn_2x_coco/SIR_3.py index 8579d7c484..7f8aeb5de0 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/faster_rcnn_faster_rcnn_swin_tiny_fpn_2x_coco/SIR_3.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/faster_rcnn_faster_rcnn_swin_tiny_fpn_2x_coco/SIR_3.py @@ -27,6 +27,7 @@ def forward( self, var_0, # (shape: [1, 3, 544, 640], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_1 = paddle.nn.functional.conv._conv_nd(var_0, self.parameter_3, bias=self.parameter_2, stride=[4, 4], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) var_2 = var_1.flatten(2) var_3 = var_2.transpose([0, 2, 1]) @@ -73,7 +74,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/smalldet_ppyoloe_plus_sod_crn_l_80e_coco/SIR_127.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/smalldet_ppyoloe_plus_sod_crn_l_80e_coco/SIR_127.py index 1935bb11ee..4f08d13720 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/smalldet_ppyoloe_plus_sod_crn_l_80e_coco/SIR_127.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/smalldet_ppyoloe_plus_sod_crn_l_80e_coco/SIR_127.py @@ -44,6 +44,7 @@ def forward( var_0, # (shape: [1, 169, 1024], dtype: paddle.float32, stop_gradient: False) var_1, # (shape: [1, 169, 1024], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_2 = paddle.nn.functional.common.dropout(var_0, p=0.1, axis=None, training=True, mode='upscale_in_train', name=None) var_3 = var_1.__add__(var_2) var_4 = paddle.nn.functional.norm.layer_norm(var_3, normalized_shape=[1024], weight=self.parameter_5, bias=self.parameter_3, epsilon=1e-05) @@ -94,7 +95,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/vitdet_ppyoloe_vit_base_csppan_cae_36e_coco/SIR_6.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/vitdet_ppyoloe_vit_base_csppan_cae_36e_coco/SIR_6.py index d343a9f079..89f59aeadf 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/vitdet_ppyoloe_vit_base_csppan_cae_36e_coco/SIR_6.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Det_cases/vitdet_ppyoloe_vit_base_csppan_cae_36e_coco/SIR_6.py @@ -43,6 +43,7 @@ def forward( self, var_0, # (shape: [1, 577, 768], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[768], weight=self.parameter_3, bias=self.parameter_1, epsilon=1e-06) var_2 = var_1.shape var_3 = var_2.__getitem__(1) @@ -104,7 +105,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_30.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_30.py index 553d9084e2..a1109994e7 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_30.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_30.py @@ -36,6 +36,7 @@ def forward( var_0, # (shape: [10, 200, 64], dtype: paddle.float32, stop_gradient: False) var_1, # (shape: [1, 1, 200, 200], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_2 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[64], weight=self.parameter_0, bias=self.parameter_2, epsilon=1e-06) var_3 = paddle.nn.functional.common.linear(x=var_2, weight=self.parameter_1, bias=self.parameter_5, name=None) var_4 = var_3.reshape((0, -1, 3, 2, 32,)) @@ -94,7 +95,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_43.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_43.py index 6e99d2aa17..191fd0ef15 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_43.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_43.py @@ -36,6 +36,7 @@ def forward( var_0, # (shape: [10, 100, 128], dtype: paddle.float32, stop_gradient: False) var_1, # (shape: [1, 1, 100, 100], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_2 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[128], weight=self.parameter_2, bias=self.parameter_4, epsilon=1e-06) var_3 = paddle.nn.functional.common.linear(x=var_2, weight=self.parameter_5, bias=self.parameter_1, name=None) var_4 = var_3.reshape((0, -1, 3, 4, 32,)) @@ -94,7 +95,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_57.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_57.py index 55f1e6dd4f..427e9ed33d 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_57.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_57.py @@ -35,6 +35,7 @@ def forward( self, var_0, # (shape: [10, 100, 128], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[128], weight=self.parameter_5, bias=self.parameter_3, epsilon=1e-06) var_2 = paddle.nn.functional.common.linear(x=var_1, weight=self.parameter_1, bias=self.parameter_0, name=None) var_3 = var_2.reshape((0, -1, 3, 4, 32,)) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_72.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_72.py index 15b1b7b770..bc00da2df7 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_72.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet/SIR_72.py @@ -35,6 +35,7 @@ def forward( self, var_0, # (shape: [10, 50, 256], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[256], weight=self.parameter_4, bias=self.parameter_1, epsilon=1e-06) var_2 = paddle.nn.functional.common.linear(x=var_1, weight=self.parameter_0, bias=self.parameter_3, name=None) var_3 = var_2.reshape((0, -1, 3, 8, 32,)) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_13.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_13.py index c897f9d76d..63b041378a 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_13.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_13.py @@ -36,6 +36,7 @@ def forward( var_0, # (shape: [10, 640, 64], dtype: paddle.float32, stop_gradient: False) var_1, # (shape: [1, 1, 640, 640], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_2 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[64], weight=self.parameter_3, bias=self.parameter_4, epsilon=1e-06) var_3 = paddle.nn.functional.common.linear(x=var_2, weight=self.parameter_5, bias=self.parameter_0, name=None) var_4 = var_3.reshape((0, -1, 3, 2, 32,)) @@ -94,7 +95,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_26.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_26.py index 20bd907619..11ba3955c9 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_26.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_26.py @@ -36,6 +36,7 @@ def forward( var_0, # (shape: [10, 320, 128], dtype: paddle.float32, stop_gradient: False) var_1, # (shape: [1, 1, 320, 320], dtype: paddle.float32, stop_gradient: True) ): + paddle.seed(33) var_2 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[128], weight=self.parameter_0, bias=self.parameter_3, epsilon=1e-06) var_3 = paddle.nn.functional.common.linear(x=var_2, weight=self.parameter_2, bias=self.parameter_1, name=None) var_4 = var_3.reshape((0, -1, 3, 4, 32,)) @@ -94,7 +95,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_40.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_40.py index 9b44197580..d6b8dc3a04 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_40.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_40.py @@ -35,6 +35,7 @@ def forward( self, var_0, # (shape: [10, 320, 128], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[128], weight=self.parameter_2, bias=self.parameter_3, epsilon=1e-06) var_2 = paddle.nn.functional.common.linear(x=var_1, weight=self.parameter_0, bias=self.parameter_1, name=None) var_3 = var_2.reshape((0, -1, 3, 4, 32,)) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_55.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_55.py index dc15e76f89..09eab4e63e 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_55.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Ocr_cases/rec_rec_svtrnet_ch/SIR_55.py @@ -35,6 +35,7 @@ def forward( self, var_0, # (shape: [10, 160, 256], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[256], weight=self.parameter_4, bias=self.parameter_0, epsilon=1e-06) var_2 = paddle.nn.functional.common.linear(x=var_1, weight=self.parameter_5, bias=self.parameter_2, name=None) var_3 = var_2.reshape((0, -1, 3, 8, 32,)) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/danet_danet_resnet50_os8_cityscapes_1024x512_80k/SIR_34.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/danet_danet_resnet50_os8_cityscapes_1024x512_80k/SIR_34.py index d4189271cf..4bbe67c27e 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/danet_danet_resnet50_os8_cityscapes_1024x512_80k/SIR_34.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/danet_danet_resnet50_os8_cityscapes_1024x512_80k/SIR_34.py @@ -37,6 +37,7 @@ def forward( var_1, # (shape: [1, 512, 64, 128], dtype: paddle.float32, stop_gradient: False) var_2, # (shape: [1, 2048, 64, 128], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_3 = var_0.__add__(var_1) var_4 = paddle.nn.functional.common.dropout2d(var_3, p=0.1, training=True, data_format='NCHW', name=None) var_5 = paddle.nn.functional.conv._conv_nd(var_4, self.parameter_2, bias=self.parameter_0, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/danet_danet_resnet50_os8_voc12aug_512x512_40k/SIR_34.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/danet_danet_resnet50_os8_voc12aug_512x512_40k/SIR_34.py index 09315b966c..a441be2c45 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/danet_danet_resnet50_os8_voc12aug_512x512_40k/SIR_34.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/danet_danet_resnet50_os8_voc12aug_512x512_40k/SIR_34.py @@ -37,6 +37,7 @@ def forward( var_1, # (shape: [1, 512, 64, 64], dtype: paddle.float32, stop_gradient: False) var_2, # (shape: [1, 2048, 64, 64], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_3 = var_0.__add__(var_1) var_4 = paddle.nn.functional.common.dropout2d(var_3, p=0.1, training=True, data_format='NCHW', name=None) var_5 = paddle.nn.functional.conv._conv_nd(var_4, self.parameter_5, bias=self.parameter_0, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/dnlnet_dnlnet_resnet101_os8_cityscapes_1024x512_80k/SIR_34.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/dnlnet_dnlnet_resnet101_os8_cityscapes_1024x512_80k/SIR_34.py index ef7b514ddc..a85af7d67c 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/dnlnet_dnlnet_resnet101_os8_cityscapes_1024x512_80k/SIR_34.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/dnlnet_dnlnet_resnet101_os8_cityscapes_1024x512_80k/SIR_34.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [1, 512, 64, 128], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout2d(var_0, p=0.1, training=True, data_format='NCHW', name=None) var_2 = paddle.nn.functional.conv._conv_nd(var_1, self.parameter_1, bias=self.parameter_0, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) return var_2 @@ -58,7 +59,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/emanet_emanet_resnet101_os8_cityscapes_1024x512_80k/SIR_35.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/emanet_emanet_resnet101_os8_cityscapes_1024x512_80k/SIR_35.py index d7cd4ed268..630ea7fb36 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/emanet_emanet_resnet101_os8_cityscapes_1024x512_80k/SIR_35.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/emanet_emanet_resnet101_os8_cityscapes_1024x512_80k/SIR_35.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [1, 256, 64, 128], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout2d(var_0, p=0.1, training=True, data_format='NCHW', name=None) var_2 = paddle.nn.functional.conv._conv_nd(var_1, self.parameter_1, bias=self.parameter_0, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) return var_2 @@ -58,7 +59,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/emanet_emanet_resnet101_os8_voc12aug_512x512_40k/SIR_35.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/emanet_emanet_resnet101_os8_voc12aug_512x512_40k/SIR_35.py index 409d673020..e76ab9ea62 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/emanet_emanet_resnet101_os8_voc12aug_512x512_40k/SIR_35.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/emanet_emanet_resnet101_os8_voc12aug_512x512_40k/SIR_35.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [1, 256, 64, 64], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout2d(var_0, p=0.1, training=True, data_format='NCHW', name=None) var_2 = paddle.nn.functional.conv._conv_nd(var_1, self.parameter_1, bias=self.parameter_0, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) return var_2 @@ -58,7 +59,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_10.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_10.py index 67f478ef0a..9811b6d1a5 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_10.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_10.py @@ -11,6 +11,7 @@ def forward( self, var_0, # (shape: [1, 64, 128, 256], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout2d(var_0, p=0.01, training=True, data_format='NCHW', name=None) return var_1 @@ -49,7 +50,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_20.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_20.py index af3ad3a1a8..61dcb2448f 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_20.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_20.py @@ -12,6 +12,7 @@ def forward( var_0, # (shape: [1, 64, 128, 256], dtype: paddle.float32, stop_gradient: False) var_1, # (shape: [1, 64, 128, 256], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_2 = paddle.nn.functional.common.dropout2d(var_0, p=0.01, training=True, data_format='NCHW', name=None) var_3 = var_1.__add__(var_2) return var_3 @@ -54,7 +55,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_40.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_40.py index de7fdefa5a..759ca9ccb4 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_40.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_40.py @@ -11,6 +11,7 @@ def forward( self, var_0, # (shape: [1, 128, 64, 128], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout2d(var_0, p=0.1, training=True, data_format='NCHW', name=None) return var_1 @@ -49,7 +50,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_46.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_46.py index f7eba71488..6dfa5ac413 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_46.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/enet_enet_cityscapes_1024x512_80k/SIR_46.py @@ -12,6 +12,7 @@ def forward( var_0, # (shape: [1, 128, 64, 128], dtype: paddle.float32, stop_gradient: False) var_1, # (shape: [1, 128, 64, 128], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_2 = paddle.nn.functional.common.dropout2d(var_0, p=0.1, training=True, data_format='NCHW', name=None) var_3 = var_1.__add__(var_2) return var_3 @@ -54,7 +55,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/espnet_espnet_cityscapes_1024x512_120k/SIR_143.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/espnet_espnet_cityscapes_1024x512_120k/SIR_143.py index 9eda8da478..17e1a3501f 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/espnet_espnet_cityscapes_1024x512_120k/SIR_143.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/espnet_espnet_cityscapes_1024x512_120k/SIR_143.py @@ -17,6 +17,7 @@ def forward( var_1, # (shape: [1, 32, 256, 512], dtype: paddle.float32, stop_gradient: False) var_2, # (shape: [1, 19, 64, 128], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_3 = paddle.nn.functional.common.interpolate(var_0, scale_factor=2, mode='bilinear', align_corners=True) var_4 = paddle.tensor.manipulation.concat([var_1, var_3], axis=1) var_5 = paddle.nn.functional.common.dropout2d(var_4, p=0.0, training=True, data_format='NCHW', name=None) @@ -68,7 +69,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/isanet_isanet_resnet101_os8_voc12aug_512x512_40k/SIR_38.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/isanet_isanet_resnet101_os8_voc12aug_512x512_40k/SIR_38.py index 6be07d8645..55ea96421c 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/isanet_isanet_resnet101_os8_voc12aug_512x512_40k/SIR_38.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/isanet_isanet_resnet101_os8_voc12aug_512x512_40k/SIR_38.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [1, 512, 64, 64], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout2d(var_0, p=0.1, training=True, data_format='NCHW', name=None) var_2 = paddle.nn.functional.conv._conv_nd(var_1, self.parameter_0, bias=self.parameter_1, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) return var_2 @@ -58,7 +59,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/isanet_isanet_resnet50_os8_cityscapes_769x769_80k/SIR_36.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/isanet_isanet_resnet50_os8_cityscapes_769x769_80k/SIR_36.py index 42f4aae362..5141193507 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/isanet_isanet_resnet50_os8_cityscapes_769x769_80k/SIR_36.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/isanet_isanet_resnet50_os8_cityscapes_769x769_80k/SIR_36.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [1, 512, 97, 97], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout2d(var_0, p=0.1, training=True, data_format='NCHW', name=None) var_2 = paddle.nn.functional.conv._conv_nd(var_1, self.parameter_1, bias=self.parameter_0, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) return var_2 @@ -58,7 +59,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/pfpn_pfpn_resnet101_os8_cityscapes_512x1024_40k/SIR_68.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/pfpn_pfpn_resnet101_os8_cityscapes_512x1024_40k/SIR_68.py index 54d92a9174..686495bd72 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/pfpn_pfpn_resnet101_os8_cityscapes_512x1024_40k/SIR_68.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/pfpn_pfpn_resnet101_os8_cityscapes_512x1024_40k/SIR_68.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [1, 256, 256, 128], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout2d(var_0, p=0.1, training=True, data_format='NCHW', name=None) var_2 = paddle.nn.functional.conv._conv_nd(var_1, self.parameter_1, bias=self.parameter_0, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) return var_2 @@ -58,7 +59,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_14.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_14.py index 642e2b6643..f70d30b572 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_14.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_14.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[64], weight=self.parameter_7, bias=self.parameter_10, epsilon=1e-06 ) @@ -146,7 +147,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_26.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_26.py index 9e067eda4c..871ad397eb 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_26.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_26.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[160], weight=self.parameter_6, bias=self.parameter_11, epsilon=1e-06 ) @@ -146,7 +147,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_38.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_38.py index 7b01cef15a..17814f9da3 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_38.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_38.py @@ -43,6 +43,7 @@ def forward( self, var_0, # (shape: [1, 1024, 256], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[256], weight=self.parameter_4, bias=self.parameter_5, epsilon=1e-06) var_2 = var_1.shape var_3 = var_2.__getitem__(0) @@ -102,7 +103,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_5.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_5.py index bcab4791a0..70423da935 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_5.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x1024_160k/SIR_5.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[32], weight=self.parameter_0, bias=self.parameter_8, epsilon=1e-06 ) @@ -146,7 +147,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_14.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_14.py index ff6aa280d3..a41547d248 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_14.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_14.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[64], weight=self.parameter_7, bias=self.parameter_1, epsilon=1e-06 ) @@ -146,7 +147,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_26.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_26.py index 2bfe99868f..c65e0a0a0e 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_26.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_26.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[160], weight=self.parameter_2, bias=self.parameter_0, epsilon=1e-06 ) @@ -148,7 +149,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_38.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_38.py index b6037b18b4..4bb0c30654 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_38.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_38.py @@ -43,6 +43,7 @@ def forward( self, var_0, # (shape: [1, 512, 256], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[256], weight=self.parameter_4, bias=self.parameter_6, epsilon=1e-06) var_2 = var_1.shape var_3 = var_2.__getitem__(0) @@ -102,7 +103,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_5.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_5.py index 2de5ba4195..60d342e046 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_5.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b0_cityscapes_1024x512_160k/SIR_5.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[32], weight=self.parameter_2, bias=self.parameter_9, epsilon=1e-06 ) @@ -148,7 +149,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_106.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_106.py index 8620b5088e..902a7ee875 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_106.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_106.py @@ -43,6 +43,7 @@ def forward( self, var_0, # (shape: [1, 1024, 512], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[512], weight=self.parameter_6, bias=self.parameter_2, epsilon=1e-06) var_2 = var_1.shape var_3 = var_2.__getitem__(0) @@ -102,7 +103,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_18.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_18.py index 3f1cea55eb..5c3656936a 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_18.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_18.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[128], weight=self.parameter_9, bias=self.parameter_3, epsilon=1e-06 ) @@ -150,7 +151,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_38.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_38.py index 24bc068284..dbe50dfc27 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_38.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_38.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[320], weight=self.parameter_6, bias=self.parameter_3, epsilon=1e-06 ) @@ -146,7 +147,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_5.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_5.py index ba6f81f570..93d7801e7f 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_5.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x1024_160k/SIR_5.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[64], weight=self.parameter_2, bias=self.parameter_11, epsilon=1e-06 ) @@ -148,7 +149,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_106.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_106.py index 6988b0c60e..319fc60348 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_106.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_106.py @@ -43,6 +43,7 @@ def forward( self, var_0, # (shape: [1, 512, 512], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[512], weight=self.parameter_3, bias=self.parameter_5, epsilon=1e-06) var_2 = var_1.shape var_3 = var_2.__getitem__(0) @@ -102,7 +103,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_18.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_18.py index 7de1c1387b..9727460f65 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_18.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_18.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[128], weight=self.parameter_11, bias=self.parameter_6, epsilon=1e-06 ) @@ -146,7 +147,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_38.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_38.py index a9c9d9410f..08ea52b8a8 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_38.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_38.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[320], weight=self.parameter_1, bias=self.parameter_0, epsilon=1e-06 ) @@ -148,7 +149,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_5.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_5.py index 1d469c8707..0982bd4f1e 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_5.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segformer_segformer_b3_cityscapes_1024x512_160k/SIR_5.py @@ -62,6 +62,7 @@ def forward( var_1, # (shape: [], dtype: paddle.int32, stop_gradient: True) var_2, # (shape: [], dtype: paddle.int32, stop_gradient: True) ): + paddle.seed(33) var_3 = paddle.nn.functional.norm.layer_norm( var_0, normalized_shape=[64], weight=self.parameter_4, bias=self.parameter_2, epsilon=1e-06 ) @@ -150,7 +151,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_base_mask_ade20k_512x512_160k/SIR_5.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_base_mask_ade20k_512x512_160k/SIR_5.py index 3dfe1ec2aa..f616070009 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_base_mask_ade20k_512x512_160k/SIR_5.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_base_mask_ade20k_512x512_160k/SIR_5.py @@ -35,6 +35,7 @@ def forward( self, var_0, # (shape: [1, 1025, 768], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[768], weight=self.parameter_0, bias=self.parameter_1, epsilon=1e-05) var_2 = var_1.shape var_3 = var_2.__getitem__(1) @@ -93,7 +94,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_base_mask_ade20k_512x512_160k/SIR_66.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_base_mask_ade20k_512x512_160k/SIR_66.py index 91d771e26d..c54c7799a9 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_base_mask_ade20k_512x512_160k/SIR_66.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_base_mask_ade20k_512x512_160k/SIR_66.py @@ -31,6 +31,7 @@ def forward( self, var_0, # (shape: [1, 1174, 768], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[768], weight=self.parameter_3, bias=self.parameter_4, epsilon=1e-05) var_2 = var_1.shape var_3 = var_2.__getitem__(1) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_small_mask_ade20k_512x512_160k/SIR_5.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_small_mask_ade20k_512x512_160k/SIR_5.py index 2fac395b48..43e06f15ea 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_small_mask_ade20k_512x512_160k/SIR_5.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_small_mask_ade20k_512x512_160k/SIR_5.py @@ -35,6 +35,7 @@ def forward( self, var_0, # (shape: [1, 1025, 384], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[384], weight=self.parameter_0, bias=self.parameter_1, epsilon=1e-05) var_2 = var_1.shape var_3 = var_2.__getitem__(1) @@ -93,7 +94,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_small_mask_ade20k_512x512_160k/SIR_66.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_small_mask_ade20k_512x512_160k/SIR_66.py index 6521bea416..ff892627a2 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_small_mask_ade20k_512x512_160k/SIR_66.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/segmenter_segmenter_vit_small_mask_ade20k_512x512_160k/SIR_66.py @@ -31,6 +31,7 @@ def forward( self, var_0, # (shape: [1, 1174, 384], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.norm.layer_norm(var_0, normalized_shape=[384], weight=self.parameter_1, bias=self.parameter_2, epsilon=1e-05) var_2 = var_1.shape var_3 = var_2.__getitem__(1) @@ -89,7 +90,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/topformer_topformer_small_ade20k_512x512_160k/SIR_128.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/topformer_topformer_small_ade20k_512x512_160k/SIR_128.py index 5b09ac5c06..b74b065c5f 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/topformer_topformer_small_ade20k_512x512_160k/SIR_128.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/topformer_topformer_small_ade20k_512x512_160k/SIR_128.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [1, 192, 64, 64], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout2d(var_0, p=0.1, training=True, data_format='NCHW', name=None) var_2 = paddle.nn.functional.conv._conv_nd(var_1, self.parameter_1, bias=self.parameter_0, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) return var_2 @@ -58,7 +59,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self): diff --git a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/topformer_topformer_tiny_ade20k_512x512_160k/SIR_126.py b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/topformer_topformer_tiny_ade20k_512x512_160k/SIR_126.py index b08582afc2..833bdd4069 100644 --- a/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/topformer_topformer_tiny_ade20k_512x512_160k/SIR_126.py +++ b/framework/e2e/PaddleLT_new/layercase/sublayer1000/Seg_cases/topformer_topformer_tiny_ade20k_512x512_160k/SIR_126.py @@ -19,6 +19,7 @@ def forward( self, var_0, # (shape: [1, 128, 64, 64], dtype: paddle.float32, stop_gradient: False) ): + paddle.seed(33) var_1 = paddle.nn.functional.common.dropout2d(var_0, p=0.1, training=True, data_format='NCHW', name=None) var_2 = paddle.nn.functional.conv._conv_nd(var_1, self.parameter_1, bias=self.parameter_0, stride=[1, 1], padding=[0, 0], padding_algorithm='EXPLICIT', dilation=[1, 1], groups=1, data_format='NCHW', channel_dim=1, op_type='conv2d', use_cudnn=True) return var_2 @@ -58,7 +59,7 @@ def train(self, net, to_static, with_prim=False, with_cinn=False): net = paddle.jit.to_static(net, build_strategy=build_strategy, full_graph=True) else: net = paddle.jit.to_static(net, full_graph=True) - paddle.seed(123) + paddle.seed(33) outs = net(*self.inputs) return outs def test_ast_prim_cinn(self):