From 8d8af8aae28efb826de7c4986baf003d41d49b39 Mon Sep 17 00:00:00 2001 From: kangguangli Date: Mon, 12 Dec 2022 08:39:51 +0000 Subject: [PATCH] remove reset reference in unittest for `fluid.layers.cross_entropy` --- .../slim/tests/test_imperative_qat_matmul.py | 4 ++- ...est_fleet_fp16_allreduce_meta_optimizer.py | 7 +++-- ...st_fleet_graph_execution_meta_optimizer.py | 28 +++++++++++++------ .../fleet/test_fleet_graph_executor.py | 7 +++-- .../fleet/test_fleet_lamb_meta_optimizer.py | 11 +++++--- .../fleet/test_fleet_lars_meta_optimizer.py | 11 +++++--- .../fleet/test_fleet_meta_optimizer_base.py | 7 +++-- .../test_fleet_pipeline_meta_optimizer.py | 7 +++-- ..._pipeline_meta_optimizer_with_recompute.py | 7 +++-- .../test_fleet_raw_program_meta_optimizer.py | 4 +-- .../dygraph_to_static/test_mobile_net.py | 7 +++-- .../unittests/dygraph_to_static/test_tsm.py | 8 ++++-- .../ipu/test_cross_entropy2_op_ipu.py | 8 ++++-- .../tests/unittests/ipu/test_dy2static_ipu.py | 11 +++----- .../fluid/tests/unittests/test_layers.py | 24 +++++++++++++--- 15 files changed, 105 insertions(+), 46 deletions(-) diff --git a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_matmul.py b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_matmul.py index 482f4a49efbe0..b584204f4f180 100644 --- a/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_matmul.py +++ b/python/paddle/fluid/contrib/slim/tests/test_imperative_qat_matmul.py @@ -173,7 +173,9 @@ def func_qat(self): label = fluid.dygraph.to_variable(y_data) out = lenet(img) acc = paddle.static.accuracy(out, label) - loss = fluid.layers.cross_entropy(out, label) + loss = paddle.nn.functional.cross_entropy( + out, label, reduction='none', use_softmax=False + ) avg_loss = paddle.mean(loss) avg_loss.backward() diff --git a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_fp16_allreduce_meta_optimizer.py b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_fp16_allreduce_meta_optimizer.py index 498c946e95f57..595b0508d57cf 100644 --- a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_fp16_allreduce_meta_optimizer.py +++ b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_fp16_allreduce_meta_optimizer.py @@ -42,8 +42,11 @@ def net(self, main_prog, startup_prog, dtype='float32'): prediction = paddle.fluid.layers.fc( input=[fc_2], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) diff --git a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_graph_execution_meta_optimizer.py b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_graph_execution_meta_optimizer.py index c6c8d9f2b98c9..f9271832cf27b 100644 --- a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_graph_execution_meta_optimizer.py +++ b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_graph_execution_meta_optimizer.py @@ -72,8 +72,11 @@ def node_func(): prediction = paddle.fluid.layers.fc( input=[fc_2], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) @@ -135,8 +138,11 @@ def node_func(): prediction = paddle.fluid.layers.fc( input=[fc_2], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) @@ -210,8 +216,11 @@ def node_func(): prediction = paddle.fluid.layers.fc( input=[fc_2], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) @@ -272,8 +281,11 @@ def node_func(): prediction = paddle.fluid.layers.fc( input=[fc_2], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) diff --git a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_graph_executor.py b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_graph_executor.py index 6951d4c40a672..8ddfb7b40610c 100644 --- a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_graph_executor.py +++ b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_graph_executor.py @@ -59,8 +59,11 @@ def node_func(): prediction = paddle.fluid.layers.fc( input=[fc_2], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) diff --git a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_lamb_meta_optimizer.py b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_lamb_meta_optimizer.py index f8ba968f553f3..1a20f09a1689d 100755 --- a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_lamb_meta_optimizer.py +++ b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_lamb_meta_optimizer.py @@ -47,8 +47,11 @@ def net(self, main_prog, startup_prog): prediction = paddle.fluid.layers.fc( input=[fc_2], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) @@ -122,8 +125,8 @@ def test_lamb_apply_with_amp(self): fc_1 = paddle.fluid.layers.fc(input=input_x, size=64, act='tanh') fc_2 = paddle.fluid.layers.fc(input=fc_1, size=64, act='tanh') prediction = paddle.fluid.layers.fc(input=[fc_2], size=2, act='softmax') - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, label=input_y, reduction='none', use_softmax=False ) avg_cost = paddle.mean(x=cost) diff --git a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_lars_meta_optimizer.py b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_lars_meta_optimizer.py index 2789c921dd999..f0e5f2517dc0f 100755 --- a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_lars_meta_optimizer.py +++ b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_lars_meta_optimizer.py @@ -47,8 +47,11 @@ def net(self, main_prog, startup_prog): prediction = paddle.fluid.layers.fc( input=[fc_2], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) @@ -127,8 +130,8 @@ def test_lars_apply_with_amp(self): fc_1 = paddle.fluid.layers.fc(input=input_x, size=64, act='tanh') fc_2 = paddle.fluid.layers.fc(input=fc_1, size=64, act='tanh') prediction = paddle.fluid.layers.fc(input=[fc_2], size=2, act='softmax') - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, label=input_y, reduction='none', use_softmax=False ) avg_cost = paddle.mean(x=cost) diff --git a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_meta_optimizer_base.py b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_meta_optimizer_base.py index 2dad7b56c26a7..5d545e2268a00 100755 --- a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_meta_optimizer_base.py +++ b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_meta_optimizer_base.py @@ -43,8 +43,11 @@ def net(main_prog, startup_prog): prediction = paddle.fluid.layers.fc( input=[fc_2], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) diff --git a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_pipeline_meta_optimizer.py b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_pipeline_meta_optimizer.py index 564b8c02fd653..900e78675a713 100644 --- a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_pipeline_meta_optimizer.py +++ b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_pipeline_meta_optimizer.py @@ -56,8 +56,11 @@ def net(self): prediction = paddle.fluid.layers.fc( input=[fc_2], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) return avg_cost diff --git a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_pipeline_meta_optimizer_with_recompute.py b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_pipeline_meta_optimizer_with_recompute.py index 85456cc96c00a..c31698ba5ffa6 100644 --- a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_pipeline_meta_optimizer_with_recompute.py +++ b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_pipeline_meta_optimizer_with_recompute.py @@ -52,8 +52,11 @@ def test_pipeline_optimizer(self): prediction = paddle.fluid.layers.fc( input=[fc_7], size=2, act='softmax' ) - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, + label=input_y, + reduction='none', + use_softmax=False, ) avg_cost = paddle.mean(x=cost) diff --git a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_raw_program_meta_optimizer.py b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_raw_program_meta_optimizer.py index 2b08af3d33c70..32e4bfb62f989 100644 --- a/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_raw_program_meta_optimizer.py +++ b/python/paddle/fluid/tests/unittests/collective/fleet/test_fleet_raw_program_meta_optimizer.py @@ -41,8 +41,8 @@ def test_pipeline_optimizer(self): fc_2 = paddle.fluid.layers.fc(input=fc_1, size=64, act='tanh') prediction = paddle.fluid.layers.fc(input=[fc_2], size=2, act='softmax') - cost = paddle.fluid.layers.cross_entropy( - input=prediction, label=input_y + cost = paddle.nn.functional.cross_entropy( + input=prediction, label=input_y, reduction='none', use_softmax=False ) avg_cost = paddle.mean(x=cost) diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_mobile_net.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_mobile_net.py index d5a4ae996d68b..9b803ff1130f2 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_mobile_net.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_mobile_net.py @@ -531,8 +531,11 @@ def train_mobilenet(args, to_static): t_end = time.time() softmax_out = paddle.nn.functional.softmax(out) - loss = fluid.layers.cross_entropy( - input=softmax_out, label=label + loss = paddle.nn.functional.cross_entropy( + input=softmax_out, + label=label, + reduction='none', + use_softmax=False, ) avg_loss = paddle.mean(x=loss) acc_top1 = paddle.static.accuracy(input=out, label=label, k=1) diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_tsm.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_tsm.py index 805e42a03a10e..4353a8a72b60b 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_tsm.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_tsm.py @@ -329,8 +329,12 @@ def train(args, fake_data_reader, to_static): labels = to_variable(y_data) labels.stop_gradient = True outputs = video_model(imgs) - loss = fluid.layers.cross_entropy( - input=outputs, label=labels, ignore_index=-1 + loss = paddle.nn.functional.cross_entropy( + input=outputs, + label=labels, + ignore_index=-1, + reduction='none', + use_softmax=False, ) avg_loss = paddle.mean(loss) acc_top1 = paddle.static.accuracy( diff --git a/python/paddle/fluid/tests/unittests/ipu/test_cross_entropy2_op_ipu.py b/python/paddle/fluid/tests/unittests/ipu/test_cross_entropy2_op_ipu.py index 25c39bd9612ad..77785d7359e84 100644 --- a/python/paddle/fluid/tests/unittests/ipu/test_cross_entropy2_op_ipu.py +++ b/python/paddle/fluid/tests/unittests/ipu/test_cross_entropy2_op_ipu.py @@ -63,8 +63,12 @@ def build_model(self, on_ipu): label = paddle.static.data( name=self.feed_list[1], shape=self.feed_shape[1], dtype='int64' ) - out = paddle.fluid.layers.cross_entropy( - input=x, label=label, **self.attrs + out = paddle.nn.functional.cross_entropy( + input=x, + label=label, + reduction='none', + use_softmax=False, + **self.attrs ) self.fetch_list = [out.name] diff --git a/python/paddle/fluid/tests/unittests/ipu/test_dy2static_ipu.py b/python/paddle/fluid/tests/unittests/ipu/test_dy2static_ipu.py index 69630e67a77e1..c3c1dd891480c 100644 --- a/python/paddle/fluid/tests/unittests/ipu/test_dy2static_ipu.py +++ b/python/paddle/fluid/tests/unittests/ipu/test_dy2static_ipu.py @@ -49,12 +49,9 @@ def forward(self, x, target=None): if target is not None: if self.use_softmax: x = paddle.nn.functional.softmax(x) - if self.loss_op: - loss = self.loss_op(x, target) - else: - loss = paddle.paddle.nn.functional.cross_entropy( - x, target, reduction='none', use_softmax=False - ) + loss = paddle.paddle.nn.functional.cross_entropy( + x, target, reduction='none', use_softmax=False + ) if self.use_reduction: loss = paddle.mean(loss) if self.use_identity_loss: @@ -69,7 +66,7 @@ def setUp(self): self.set_data_feed() def set_op_attrs(self): - self.loss_op = paddle.fluid.layers.cross_entropy + pass def set_data_feed(self): self.data = paddle.uniform((8, 3, 10, 10), dtype='float32') diff --git a/python/paddle/fluid/tests/unittests/test_layers.py b/python/paddle/fluid/tests/unittests/test_layers.py index 121a8f8b9d47f..2659a875058d0 100644 --- a/python/paddle/fluid/tests/unittests/test_layers.py +++ b/python/paddle/fluid/tests/unittests/test_layers.py @@ -2022,7 +2022,9 @@ def make_recognize_digits_mlp(self): act='softmax', param_attr=["sftmax.w1", "sftmax.w2"], ) - cost = layers.cross_entropy(input=predict, label=label) + cost = paddle.nn.functional.cross_entropy( + input=predict, label=label, reduction='none', use_softmax=False + ) avg_cost = paddle.mean(cost) return avg_cost @@ -2061,7 +2063,9 @@ def make_recognize_digits_conv(self): ) predict = layers.fc(input=conv_pool_2, size=10, act="softmax") - cost = layers.cross_entropy(input=predict, label=label) + cost = paddle.nn.functional.cross_entropy( + input=predict, label=label, reduction='none', use_softmax=False + ) avg_cost = paddle.mean(cost) return avg_cost @@ -2114,7 +2118,12 @@ def make_word_embedding(self): predict_word = layers.fc( input=hidden1, size=dict_size, act='softmax' ) - cost = layers.cross_entropy(input=predict_word, label=next_word) + cost = paddle.nn.functional.cross_entropy( + input=predict_word, + label=next_word, + reduction='none', + use_softmax=False, + ) avg_cost = paddle.mean(cost) return avg_cost @@ -2346,7 +2355,14 @@ def make_cross_entropy(self): x = self._get_data(name="x", shape=[30, 10], dtype="float32") label = self._get_data(name="label", shape=[30, 1], dtype="int64") mode = 'channel' - out = layers.cross_entropy(x, label, False, 4) + out = paddle.nn.functional.cross_entropy( + x, + label, + soft_label=False, + ignore_index=4, + reduction='none', + use_softmax=False, + ) return out def make_uniform_random_batch_size_like(self):