add stuff for CAML
All checks were successful
Build Typst document / build_typst_documents (push) Successful in 27s
All checks were successful
Build Typst document / build_typst_documents (push) Successful in 27s
This commit is contained in:
parent
a358401ffb
commit
1805bc2d78
@ -195,20 +195,44 @@ There are several different ResNet architectures, the most common are ResNet-18,
|
|||||||
|
|
||||||
For this bachelor theis the ResNet-50 architecture was used to predict the corresponding embeddings for the few-shot learning methods.
|
For this bachelor theis the ResNet-50 architecture was used to predict the corresponding embeddings for the few-shot learning methods.
|
||||||
|
|
||||||
|
=== P$>$M$>$F
|
||||||
|
Todo
|
||||||
|
|
||||||
=== CAML
|
=== CAML
|
||||||
// https://arxiv.org/pdf/2310.10971v2
|
// https://arxiv.org/pdf/2310.10971v2
|
||||||
CAML (Context aware meta learning) is one of the state-of-the-art methods for few-shot learning.
|
CAML (Context aware meta learning) is one of the state-of-the-art methods for few-shot learning.
|
||||||
#todo[Here we should describe in detail how caml works]
|
It consists of three different components: a frozen pre-trained image encoder, a fixed Equal Length and Maximally Equiangular Set (ELMES) class encoder and a non-causal sequence model.
|
||||||
|
|
||||||
|
*Architecture:* CAML first encodes the query and support set images using the fronzen pre-trained feature extractor as shown in @camlarchitecture.
|
||||||
|
This step brings the images into a low dimensional space where similar images are encoded into similar embeddings.
|
||||||
|
The class labels are encoded with the ELMES class encoder.
|
||||||
|
Since the class of the query image is unknown in this stage we add a special learnable "unknown token" to the encoder.
|
||||||
|
This embedding is learned during pre-training.
|
||||||
|
Afterwards each image embedding is concatenated with the corresponding class embedding.
|
||||||
|
|
||||||
|
#todo[We should add stuff here why we have a max amount of shots bc. of pretrained model]
|
||||||
|
|
||||||
|
*ELMES Encoder:* The ELMES (Equal Length and Maximally Equiangular Set) encoder encodes the class labels to vectors of equal length.
|
||||||
|
The encoder is a bijective mapping between the labels and set of vectors that are equal length and maximally equiangular.
|
||||||
|
#todo[Describe what equiangular and bijective means]
|
||||||
|
Similar to one-hot encoding but with some advantages.
|
||||||
|
|
||||||
|
*Non-causal sequence model:*
|
||||||
|
#todo[Desc. what this is]
|
||||||
|
|
||||||
|
*Large-Scale Pre-Training:*
|
||||||
|
#todo[Desc. what this is]
|
||||||
|
|
||||||
|
*Theoretical Analysis:*
|
||||||
|
#todo[Mybe not that important?]
|
||||||
|
|
||||||
|
*Results:*
|
||||||
|
|
||||||
#figure(
|
#figure(
|
||||||
image("rsc/caml_architecture.png", width: 80%),
|
image("rsc/caml_architecture.png", width: 80%),
|
||||||
caption: [Architecture of CAML. #cite(<caml_paper>)],
|
caption: [Architecture of CAML. #cite(<caml_paper>)],
|
||||||
) <camlarchitecture>
|
) <camlarchitecture>
|
||||||
|
|
||||||
=== P$>$M$>$F
|
|
||||||
Todo
|
|
||||||
|
|
||||||
=== Softmax
|
=== Softmax
|
||||||
#todo[Maybe remove this section]
|
#todo[Maybe remove this section]
|
||||||
The Softmax function @softmax #cite(<liang2017soft>) converts $n$ numbers of a vector into a probability distribution.
|
The Softmax function @softmax #cite(<liang2017soft>) converts $n$ numbers of a vector into a probability distribution.
|
||||||
|
Loading…
Reference in New Issue
Block a user