From d0971f9447ed9806022a3d47587b62394682bc51 Mon Sep 17 00:00:00 2001 From: serhii-karelov <4457579+serhii-karelov@users.noreply.github.com> Date: Sun, 27 Aug 2023 08:33:32 +0700 Subject: [PATCH] Update the short ReLU description don't call it one-to-one operation --- src/article/Article.svelte | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/article/Article.svelte b/src/article/Article.svelte index b4af39d3..e17aba47 100644 --- a/src/article/Article.svelte +++ b/src/article/Article.svelte @@ -178,7 +178,7 @@ Neural networks are extremely prevalent in modern technology—because they are so accurate! The highest performing CNNs today consist of an absurd amount of layers, which are able to learn more and more features. Part of the reason these groundbreaking CNNs are able to achieve such tremendous accuracies is because of their non-linearity. ReLU applies much-needed non-linearity into the model. Non-linearity is necessary to produce non-linear decision boundaries, so that the output cannot be written as a linear combination of the inputs. If a non-linear activation function was not present, deep CNN architectures would devolve into a single, equivalent convolutional layer, which would not perform nearly as well. The ReLU activation function is specifically used as a non-linear activation function, as opposed to other non-linear functions such as Sigmoid because it has been empirically observed that CNNs using ReLU are faster to train than their counterparts.
- The ReLU activation function is a one-to-one mathematical operation: {reluEquation} + The ReLU activation function is an elementwise mathematical operation: {reluEquation}