fix caml stuff and add things to last sec
	
		
			
	
		
	
	
		
	
		
			All checks were successful
		
		
	
	
		
			
				
	
				Build Typst document / build_typst_documents (push) Successful in 11s
				
			
		
		
	
	
				
					
				
			
		
			All checks were successful
		
		
	
	Build Typst document / build_typst_documents (push) Successful in 11s
				
			This commit is contained in:
		| @@ -6,14 +6,17 @@ The only benefit of Few-Shot learning is that it can be used in environments whe | |||||||
| But this should not be the case in most scenarios. | But this should not be the case in most scenarios. | ||||||
| Most of the time plenty of good samples are available and in this case Patchcore or EfficientAD should perform great. | Most of the time plenty of good samples are available and in this case Patchcore or EfficientAD should perform great. | ||||||
|  |  | ||||||
| The only case where Few-Shot learning could be used is in a scenario where one wants to detect the anomaly class itself. | The only case where Few-Shot learning could be used is in a scenarios where one wants to detect the anomaly class itself. | ||||||
| Patchcore and EfficientAD can only detect if an anomaly is present or not but not what the anomaly is. | Patchcore and EfficientAD can only detect if an anomaly is present or not but not what type of anomaly it actually is. | ||||||
| So chaining a Few-Shot learner after Patchcore or EfficientAD could be a good idea to use the best of both worlds. | So chaining a Few-Shot learner after Patchcore or EfficientAD could be a good idea to use the best of both worlds. | ||||||
|  |  | ||||||
| In most of the tests performed P>M>F performed the best. | In most of the tests P>M>F performed the best. | ||||||
| But also the simple ResNet50 method performed better than expected in most cases and can be considered if the computational resources are limited and if a simple architecture is enough. | But also the simple ResNet50 method performed better than expected in most cases and can be considered if the computational resources are limited and if a simple architecture is enough. | ||||||
|  |  | ||||||
| == Outlook | == Outlook | ||||||
| In the future when new Few-Shot learning methods evolve it could be interesting to test again how they perform in anomaly detection tasks. | In the future when new Few-Shot learning methods evolve it could be interesting to test again how they perform in anomaly detection tasks. | ||||||
| There might be a lack of research in the area where the classes to detect are very similar to each other | There might be a lack of research in the area where the classes to detect are very similar to each other | ||||||
| and when building a few-shot learning algorithm tailored specifically for very similar classes this could boost the performance by a large margin. | and when building a few-shot learning algorithm tailored specifically for very similar classes this could boost the performance by a large margin. | ||||||
|  |  | ||||||
|  | It might be interesting to test the SOT method (see @SOT) with a ResNet50 feature extractor similar as proposed in this thesis but with SOT for embedding comparison. | ||||||
|  | Moreover, TRIDENT (see @TRIDENT) could achive promising results in a anomaly detection scenario. | ||||||
|   | |||||||
| @@ -183,7 +183,9 @@ So it is clearly a bad idea to add more good shots to the support set. | |||||||
| == CAML | == CAML | ||||||
| === Approach | === Approach | ||||||
| For the CAML implementation the pretrained model weights from the original paper were used. | For the CAML implementation the pretrained model weights from the original paper were used. | ||||||
| This brings the limitation of a maximum squence length to the non-causal sequence model. | The non-causal sequence model (transformer) is pretrained with every class having the same number of shots. | ||||||
|  | This brings the limitation that it can only process default few-shot learning tasks in the n-way k-shots fashion. | ||||||
|  | Since it expects the input sequence to be distributed with the same number of shots per class. | ||||||
| This is the reason why for this method the two imbalanced test cases couldn't be conducted. | This is the reason why for this method the two imbalanced test cases couldn't be conducted. | ||||||
|  |  | ||||||
| As a feture extractor a ViT-B/16 model was used, which is a Vision Transformer with a patch size of 16. | As a feture extractor a ViT-B/16 model was used, which is a Vision Transformer with a patch size of 16. | ||||||
|   | |||||||
| @@ -392,7 +392,7 @@ If the pre-trained model lacks relevant information for the task, SgVA-CLIP migh | |||||||
| This might be a no-go for anomaly detection tasks because the images in such tasks are often very task-specific and not covered by general pre-trained models. | This might be a no-go for anomaly detection tasks because the images in such tasks are often very task-specific and not covered by general pre-trained models. | ||||||
| Also, fine-tuning the model can require considerable computational resources, which might be a limitation in some cases.~#cite(<peng2023sgvaclipsemanticguidedvisualadapting>) | Also, fine-tuning the model can require considerable computational resources, which might be a limitation in some cases.~#cite(<peng2023sgvaclipsemanticguidedvisualadapting>) | ||||||
|  |  | ||||||
| === TRIDENT (Transductive Decoupled Variational Inference for Few-Shot Classification) | === TRIDENT (Transductive Decoupled Variational Inference for Few-Shot Classification) <TRIDENT> | ||||||
| // https://arxiv.org/pdf/2208.10559v1 | // https://arxiv.org/pdf/2208.10559v1 | ||||||
| // https://arxiv.org/abs/2208.10559v1 | // https://arxiv.org/abs/2208.10559v1 | ||||||
|  |  | ||||||
| @@ -406,7 +406,7 @@ This feature extractor dynamically aligns features from both the support and the | |||||||
| This model is specifically designed for few-shot classification tasks but might also work well for anomaly detection. | This model is specifically designed for few-shot classification tasks but might also work well for anomaly detection. | ||||||
| Its ability to isolate critical features while droping irellevant context aligns with requirements needed for anomaly detection. | Its ability to isolate critical features while droping irellevant context aligns with requirements needed for anomaly detection. | ||||||
|  |  | ||||||
| === SOT (Self-Optimal-Transport Feature Transform) | === SOT (Self-Optimal-Transport Feature Transform) <SOT> | ||||||
| // https://arxiv.org/pdf/2204.03065v1 | // https://arxiv.org/pdf/2204.03065v1 | ||||||
| // https://arxiv.org/abs/2204.03065v1 | // https://arxiv.org/abs/2204.03065v1 | ||||||
|  |  | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user