This commit is contained in:
		@@ -99,7 +99,7 @@ The Softmax function @softmax #cite(<liang2017soft>) converts $n$ numbers of a v
 | 
				
			|||||||
Its a generalization of the Sigmoid function and often used as an Activation Layer in neural networks.
 | 
					Its a generalization of the Sigmoid function and often used as an Activation Layer in neural networks.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
$
 | 
					$
 | 
				
			||||||
sigma(bold(z))_j = (e^(z_j)) / (sum_(k=1)^k e^(z_k)) "for" j=(1,...,k)
 | 
					sigma(bold(z))_j = (e^(z_j)) / (sum_(k=1)^k e^(z_k)) "for" j:={1,...,k}
 | 
				
			||||||
$ <softmax>
 | 
					$ <softmax>
 | 
				
			||||||
 | 
					
 | 
				
			||||||
The softmax function has high similarities with the Boltzmann distribution and was first introduced in the 19th century #cite(<Boltzmann>).
 | 
					The softmax function has high similarities with the Boltzmann distribution and was first introduced in the 19th century #cite(<Boltzmann>).
 | 
				
			||||||
@@ -112,7 +112,7 @@ And equation~\eqref{eq:crelbinary} is the special case of the general Cross Entr
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
$
 | 
					$
 | 
				
			||||||
H(p,q) &= -sum_(x in cal(X)) p(x) log q(x)\
 | 
					H(p,q) &= -sum_(x in cal(X)) p(x) log q(x)\
 | 
				
			||||||
H(p,q) &= -p log(q) + (1-p) log(1-q)\
 | 
					H(p,q) &= -(p log(q) + (1-p) log(1-q))\
 | 
				
			||||||
cal(L)(p,q) &= -1/N sum_(i=1)^(cal(B)) (p_i log(q_i) + (1-p_i) log(1-q_i))
 | 
					cal(L)(p,q) &= -1/N sum_(i=1)^(cal(B)) (p_i log(q_i) + (1-p_i) log(1-q_i))
 | 
				
			||||||
$
 | 
					$
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user