In fact, the absolute value function is not differentiable over the entire set of real numbers because of its sharp point at x=0. However, excluding just one point from its domain, the function becomes differentiable as shown in R∖{0}. In other words, as opposed to f, the function defined as g has a derivativeg′.
f(x):=∣x∣,x∈Rg(x):=∣x∣,x∈R∖{0}
In many cases, g′ can be regarded as the derivative of f, and this is referred to as the weak derivative of f. This is precisely why activation functions like ReLU, which are used in deep learning, can be utilized even though they are not differentiable at x=0.