Learning to Balance: Bayesian Meta-Learning for Imbalanced and Out-of-distrib...MLAI2
?
While tasks could come with varying the number of instances and classes in realistic settings, the existing meta-learning approaches for few-shot classification assume that number of instances per task and class is fixed. Due to such restriction, they learn to equally utilize the meta-knowledge across all the tasks, even when the number of instances per task and class largely varies. Moreover, they do not consider distributional difference in unseen tasks, on which the meta-knowledge may have less usefulness depending on the task relatedness. To overcome these limitations, we propose a novel meta-learning model that adaptively balances the effect of the meta-learning and task-specific learning within each task. Through the learning of the balancing variables, we can decide whether to obtain a solution by relying on the meta-knowledge or task-specific learning. We formulate this objective into a Bayesian inference framework and tackle it using variational inference. We validate our Bayesian Task-Adaptive Meta-Learning (Bayesian TAML) on two realistic task- and class-imbalanced datasets, on which it significantly outperforms existing meta-learning approaches. Further ablation study confirms the effectiveness of each balancing component and the Bayesian learning framework.
Picked-up lists of GAN variants which provided insights to the community. (GANs-Improved GANs-DCGAN-Unrolled GAN-InfoGAN-f-GAN-EBGAN-WGAN)
After short introduction to GANs, we look through the remaining difficulties of standard GANs and their temporary solutions (Improved GANs). By following the slides, we can see the other solutions which tried to resolve the problems in various ways, e.g. careful architecture selection (DCGAN), slight change in update (Unrolled GAN), additional constraint (InfoGAN), generalization of the loss function using various divergence (f-GAN), providing new framework of energy based model (EBGAN), another step of generalization of the loss function (WGAN).
Unsupervised Data Augmentation for Consistency TrainingSungchul Kim
?
This document discusses semi-supervised learning and unsupervised data augmentation (UDA). It begins by explaining techniques in semi-supervised learning like entropy minimization and consistency regularization. It then introduces UDA, which trains models to be less sensitive to noise by minimizing the divergence between predictions on original and augmented data. The document reports on experiments applying UDA and comparing it to other methods on image datasets, finding it achieves better performance. It also explores techniques like training signal annealing and discusses ablation studies.
Emerging Properties in Self-Supervised Vision TransformersSungchul Kim
?
The document summarizes the DINO self-supervised learning approach for vision transformers. DINO uses a teacher-student framework where the teacher's predictions are used to supervise the student through knowledge distillation. Two global and several local views of an image are passed through the student, while only global views are passed through the teacher. The student is trained to match the teacher's predictions for local views. DINO achieves state-of-the-art results on ImageNet with linear evaluation and transfers well to downstream tasks. It also enables vision transformers to discover object boundaries and semantic layouts.
Introduction to MAML (Model Agnostic Meta Learning) with DiscussionsJoonyoung Yi
?
The document describes Model-Agnostic Meta-Learning (MAML), an algorithm for fast adaptation of neural networks to new tasks. MAML learns model parameters that can quickly be fine-tuned to new tasks using only a small number of gradient steps. The meta-learner optimizes the model's initialization such that a single gradient update on new tasks minimizes loss. MAML is model-agnostic, requiring no specific architecture, and can be used for classification, regression and reinforcement learning tasks.
1. Two papers on unsupervised domain adaptation were presented at ICML2018: "Learning Semantic Representations for Unsupervised Domain Adaptation" and "CyCADA: Cycle-Consistent Adversarial Domain Adaptation".
2. The CyCADA paper uses cycle-consistent adversarial domain adaptation with cycle GAN to translate images at the pixel level while also aligning representations at the semantic level.
3. The semantic representation paper uses semantic alignment and introduces techniques like adding noise to improve over previous semantic alignment methods.
Dsh data sensitive hashing for high dimensional k-nn searchWooSung Choi
?
Gao, Jinyang, et al. "Dsh: data sensitive hashing for high-dimensional k-nnsearch." Proceedings of the 2014 ACM SIGMOD international conference on Management of data. ACM, 2014.
Unsupervised Data Augmentation for Consistency TrainingSungchul Kim
?
This document discusses semi-supervised learning and unsupervised data augmentation (UDA). It begins by explaining techniques in semi-supervised learning like entropy minimization and consistency regularization. It then introduces UDA, which trains models to be less sensitive to noise by minimizing the divergence between predictions on original and augmented data. The document reports on experiments applying UDA and comparing it to other methods on image datasets, finding it achieves better performance. It also explores techniques like training signal annealing and discusses ablation studies.
Emerging Properties in Self-Supervised Vision TransformersSungchul Kim
?
The document summarizes the DINO self-supervised learning approach for vision transformers. DINO uses a teacher-student framework where the teacher's predictions are used to supervise the student through knowledge distillation. Two global and several local views of an image are passed through the student, while only global views are passed through the teacher. The student is trained to match the teacher's predictions for local views. DINO achieves state-of-the-art results on ImageNet with linear evaluation and transfers well to downstream tasks. It also enables vision transformers to discover object boundaries and semantic layouts.
Introduction to MAML (Model Agnostic Meta Learning) with DiscussionsJoonyoung Yi
?
The document describes Model-Agnostic Meta-Learning (MAML), an algorithm for fast adaptation of neural networks to new tasks. MAML learns model parameters that can quickly be fine-tuned to new tasks using only a small number of gradient steps. The meta-learner optimizes the model's initialization such that a single gradient update on new tasks minimizes loss. MAML is model-agnostic, requiring no specific architecture, and can be used for classification, regression and reinforcement learning tasks.
1. Two papers on unsupervised domain adaptation were presented at ICML2018: "Learning Semantic Representations for Unsupervised Domain Adaptation" and "CyCADA: Cycle-Consistent Adversarial Domain Adaptation".
2. The CyCADA paper uses cycle-consistent adversarial domain adaptation with cycle GAN to translate images at the pixel level while also aligning representations at the semantic level.
3. The semantic representation paper uses semantic alignment and introduces techniques like adding noise to improve over previous semantic alignment methods.
Dsh data sensitive hashing for high dimensional k-nn searchWooSung Choi
?
Gao, Jinyang, et al. "Dsh: data sensitive hashing for high-dimensional k-nnsearch." Proceedings of the 2014 ACM SIGMOD international conference on Management of data. ACM, 2014.
Vector Optimization (by Jinhwan Seok. M.S student at KAIST)
The concept of vector optimization and its applications
-Regularized least squares
-Smoothing approximation
-Reconstruction
Reference)
convex optimization, Boyd (2004)
Generative model is nowadays a very good tool for Anomaly Detection. Thus I bring a interesting generative model 'Diffusion' for solving the anomaly detection task. Presentation consists of the concept of diffusion and method to use diffusion for anomaly detection.