add some math formulation of label set selection
This commit is contained in:
parent
8da13101f8
commit
ec2de9a5c7
@ -28,3 +28,39 @@ That means taking the absolute value of the prediction minus the class center re
|
|||||||
\end{align}
|
\end{align}
|
||||||
|
|
||||||
\cite{activelearning}
|
\cite{activelearning}
|
||||||
|
|
||||||
|
\begin{equation}\label{eq:minnot}
|
||||||
|
\text{min}_n(S) \coloneqq a \subset S \mid \text{where a are the n smallest numbers of S}
|
||||||
|
\end{equation}
|
||||||
|
|
||||||
|
\begin{equation}\label{eq:maxnot}
|
||||||
|
\text{max}_n(S) \coloneqq a \subset S \mid \text{where a are the n largest numbers of S}
|
||||||
|
\end{equation}
|
||||||
|
|
||||||
|
\subsection{Low certainty first}
|
||||||
|
We take the samples with the lowest certainty score first and give it to the user for labeling.
|
||||||
|
|
||||||
|
\begin{equation}
|
||||||
|
\text{min}_\mathcal{B}(S(z))
|
||||||
|
\end{equation}
|
||||||
|
|
||||||
|
\subsection{High certainty first}
|
||||||
|
We take the samples with the highest certainty score first and give it to the user for labeling.
|
||||||
|
|
||||||
|
\begin{equation}
|
||||||
|
\text{max}_\mathcal{B}(S(z))
|
||||||
|
\end{equation}
|
||||||
|
|
||||||
|
\subsection{Low and High certain first}
|
||||||
|
|
||||||
|
We take half the batch-size $\mathcal{B}$ of low certainty and the other half with high certainty samples.
|
||||||
|
|
||||||
|
\begin{equation}
|
||||||
|
\text{max}_{\mathcal{B}/2}(S(z)) \cup \text{max}_{\mathcal{B}/2}(S(z))
|
||||||
|
\end{equation}
|
||||||
|
|
||||||
|
\subsection{Mid certain first}
|
||||||
|
|
||||||
|
\begin{equation}
|
||||||
|
S(z) \setminus (\text{min}_{\mathcal{S}/2 - \mathcal{B}/2}(S(z)) \cup \text{max}_{\mathcal{S}/2 - \mathcal{B}/2}(S(z)))
|
||||||
|
\end{equation}
|
13
src/main.tex
13
src/main.tex
@ -75,19 +75,6 @@
|
|||||||
\input{experimentalresults}
|
\input{experimentalresults}
|
||||||
\input{conclusionandoutlook}
|
\input{conclusionandoutlook}
|
||||||
|
|
||||||
\section{Semi-Supervised learning}\label{sec:semi-supervised-learning}
|
|
||||||
In traditional supervised learning we have a labeled dataset.
|
|
||||||
Each datapoint is associated with a corresponding target label.
|
|
||||||
The goal is to fit a model to predict the labels from datapoints.
|
|
||||||
|
|
||||||
In traditional unsupervised learning there are also datapoints but no labels are known.
|
|
||||||
The goal is to find patterns or structures in the data.
|
|
||||||
Moreover, it can be used for clustering or downprojection.
|
|
||||||
|
|
||||||
Those two techniques combined yield semi-supervised learning.
|
|
||||||
Some of the labels are known, but for most of the data we have only the raw datapoints.
|
|
||||||
The basic idea is that the unlabeled data can significantly improve the model performance when used in combination with the labeled data.
|
|
||||||
|
|
||||||
\section{FixMatch}\label{sec:fixmatch}
|
\section{FixMatch}\label{sec:fixmatch}
|
||||||
There is an already existing approach called FixMatch.
|
There is an already existing approach called FixMatch.
|
||||||
This was introduced in a Google Research paper from 2020~\cite{fixmatch}.
|
This was introduced in a Google Research paper from 2020~\cite{fixmatch}.
|
||||||
|
@ -13,6 +13,19 @@
|
|||||||
\subsection{Methods}\label{subsec:methods}
|
\subsection{Methods}\label{subsec:methods}
|
||||||
|
|
||||||
\subsubsection{Active-Learning}
|
\subsubsection{Active-Learning}
|
||||||
|
\subsubsection{Semi-Supervised learning}
|
||||||
|
In traditional supervised learning we have a labeled dataset.
|
||||||
|
Each datapoint is associated with a corresponding target label.
|
||||||
|
The goal is to fit a model to predict the labels from datapoints.
|
||||||
|
|
||||||
|
In traditional unsupervised learning there are also datapoints but no labels are known.
|
||||||
|
The goal is to find patterns or structures in the data.
|
||||||
|
Moreover, it can be used for clustering or downprojection.
|
||||||
|
|
||||||
|
Those two techniques combined yield semi-supervised learning.
|
||||||
|
Some of the labels are known, but for most of the data we have only the raw datapoints.
|
||||||
|
The basic idea is that the unlabeled data can significantly improve the model performance when used in combination with the labeled data.
|
||||||
|
|
||||||
\subsubsection{ROC and AUC}
|
\subsubsection{ROC and AUC}
|
||||||
\subsubsection{RESNet}
|
\subsubsection{RESNet}
|
||||||
\subsubsection{CNN}
|
\subsubsection{CNN}
|
||||||
@ -26,6 +39,7 @@ Pooling layers sample down the feature maps created by the convolutional layers.
|
|||||||
This helps reducing the computational complexity of the overall network and help with overfitting.
|
This helps reducing the computational complexity of the overall network and help with overfitting.
|
||||||
Common pooling layers include average- and max pooling.
|
Common pooling layers include average- and max pooling.
|
||||||
Finally, after some convolution layers the feature map is flattened and passed to a network of fully connected layers to perform a classification or regression task.
|
Finally, after some convolution layers the feature map is flattened and passed to a network of fully connected layers to perform a classification or regression task.
|
||||||
|
\ref{fig:cnn-architecture} shows a typical binary classification task.
|
||||||
|
|
||||||
\begin{figure}[h]
|
\begin{figure}[h]
|
||||||
\centering
|
\centering
|
||||||
|
Loading…
Reference in New Issue
Block a user