From Normalized Augmented Feature Vector, we get the value
In which we want to find the
Obviously, we do differentiation wrt
As such, we get
In traditional perceptron, we update the learning rate with the variable increment rule:
Perceptron Convergence Theorem
If training samples are linearly separable, the Perceptron Algorithm will converge to a solution vector in a finite number of updates.