2020-2023小样本学习(FSL)顶会论文及开源代码合集(已分类整理)

文章总结了2020-2023年间160+篇小样本学习(FSL)论文,涵盖数据、模型、算法和技术等多个方面,重点介绍了对比学习、自监督学习、元学习等方法,并探讨了其在不同领域的应用。

摘要生成于 C知道 ,由 DeepSeek-R1 满血版支持, 前往体验 >

这次分享的是近四年(2020-2023)各大顶会中的小样本学习(FSL)论文,有160+篇,涵盖了FSL三大类方法:数据、模型、算法,以及FSL的应用、技术、理论等领域

由于论文数量太多,我就不一一分析总结了,建议大家收藏下慢慢研读。

全部160+篇论文原文及开源代码文末直接领取。

数据(12篇)

Towards better understanding and better generalization of low-shot classification in histology images with contrastive learning

标题:通过对比学习来更好地理解和提高组织病理图像中的小样本分类的泛化能力

方法介绍:本文通过设置三个跨域任务来推动组织病理图像的小样本学习研究,模拟实际临床问题。为实现高效标注和更好的泛化能力,作者提出结合对比学习和潜在增强来构建小样本系统。对比学习可以在无手动标注下学习有用表示,而潜在增强以非监督方式传递基数据集的语义变化。这两者可充分利用无标注训练数据,可扩展到其他数据饥渴问题。

  1. FlipDA: Effective and robust data augmentation for few-shot learning

  2. PromDA: Prompt-based data augmentation for low-resource NLU tasks

  3. Generating representative samples for few-shot classification

  4. FeLMi : Few shot learning with hard mixup

  5. Understanding cross-domain few-shot learning based on domain similarity and few-shot difficulty

  6. Label hallucination for few-shot classification

  7. STUNT: Few-shot tabular learning with self-generated tasks from unlabeled tables

  8. Unsupervised meta-learning via few-shot pseudo-supervised contrastive learning

  9. Progressive mix-up for few-shot supervised multi-source domain transfer

  10. Cross-level distillation and feature denoising for cross-domain few-shot classification

  11. Tuning language models as training data generators for augmentation-enhanced few-shot learning

模型(35篇)

多任务学习

When does self-supervision improve few-shot learning?

标题:自监督学习在什么情况下可以改进小样本学习?

方法介绍:虽然自监督学习的收益可能随着更大的训练数据集而增加,但我们也观察到,当用于元学习和自监督的图像分布不同时,自监督学习实际上可能会损害性能。通过系统地变化域移度和在多个域上分析几种元学习算法的性能,作者进行了详细的分析研究。基于这一分析,作者提出了一种从大规模通用无标注图像池中自动选择适合特定数据集的自监督学习图像的技术,可以进一步改进性能。

  1. Pareto self-supervised training for few-shot learning

  2. Bridging multi-task learning and meta-learning: Towards efficient training and effective adaptation

  3. Task-level self-supervision for cross-domain few-shot learning

嵌入/度量学习

Few-shot learning as cluster-induced voronoi diagrams: A geometric approach

标题:将小样本学习视为由簇诱导的Voronoi图:一种几何方法

方法介绍:小样本学习仍面临泛化能力不足的挑战,本文从几何视角出发,发现流行的 ProtoNet 模型本质上是特征空间中的 Voronoi 图。通过利用“由簇诱导的 Voronoi 图”的技术,可以逐步改进空间分割,在小样本学习的多个阶段提升准确率和鲁棒性。这一基于该图的框架数学优雅、几何可解释,可以补偿极端数据不足,防止过拟合,并实现快速几何推理。

  1. Few-shot learning with siamese networks and label tuning

  2. Matching feature sets for few-shot image classification

  3. EASE: Unsupervised discriminant subspace learning for transductive few-shot learning

  4. Cross-domain few-shot learning with task-specific adapters

  5. Rethinking generalization in few-shot classification

  6. Hybrid graph neural networks for few-shot learning

  7. Hubs and hyperspheres: Reducing hubness and improving transductive few-shot learning with hyperspherical embeddings

  8. Revisiting prototypical network for cross domain few-shot learning

  9. Transductive few-shot learning with prototype-based label propagation by iterative graph refinement

  10. Few-sample feature selection via feature manifold learning

  11. Interval bound interpolation for few-shot learning with few tasks

  12. A closer look at few-shot classification again

  13. TART: Improved few-shot text classification using task-adaptive reference transformation

外部存储器辅助学习

Dynamic memory induction networks for few-shot text classification

标题:动态记忆诱导网络用于短文本分类

方法介绍:本文提出了动态记忆诱导网络,用于短文本的少样本分类。该模型利用动态路由为基于记忆的少样本学习提供更大灵活性,以便更好地适应支持集,这是少样本分类模型的关键能力。在此基础上,作者进一步开发了包含查询信息的诱导模型,旨在增强元学习的泛化能力。

  1. Few-shot visual learning with contextual memory and fine-grained calibration

  2. Learn from concepts: Towards the purified memory for few-shot learning

  3. Prototype memory and attention mechanisms for few shot image generation

  4. Hierarchical variational memory for few-shot learning across domains

  5. Remember the difference: Cross-domain few-shot semantic segmentation via meta-memory transfer

  6. Consistent prototype learning for few-shot continual relation extraction

生成式建模

Few-shot relation extraction via bayesian meta-learning on relation graphs

标题:通过关系图上的贝叶斯元学习实现短文本关系提取

方法介绍:作者提出了一种新的贝叶斯元学习方法,用于有效学习关系原型向量的后验分布,其中关系原型向量的先验由定义在全局关系图上的图神经网络参数化。此外,为了有效优化原型向量的后验分布,作者使用了相关于MAML算法的随机梯度兰weibo乎动力学,它可以处理原型向量的不确定性,整个框架可以端到端高效优化。

  1. Interventional few-shot learning

  2. Modeling the probabilistic distribution of unlabeled data for one-shot medical image segmentation

  3. SCHA-VAE: Hierarchical context aggregation for few-shot generation

  4. Diversity vs. Recognizability: Human-like generalization in one-shot generative models

  5. Generalized one-shot domain adaptation of generative adversarial networks

  6. Towards diverse and faithful one-shot adaption of generative adversarial networks

  7. Few-shot cross-domain image generation via inference-time latent-code learning

  8. Adaptive IMLE for few-shot pretraining-free generative modelling

  9. MetaModulation: Learning variational feature hierarchies for few-shot learning with fewer tasks

算法(24篇)

优化已有参数

  1. Revisit finetuning strategy for few-shot learning to transfer the emdeddings

  2. Prototypical calibration for few-shot learning of language models

  3. Prompt, generate, then cache: Cascade of foundation models makes strong few-shot learners

  4. Supervised masked knowledge distillation for few-shot transformers

  5. Hint-Aug: Drawing hints from foundation vision transformers towards boosted few-shot parameter-efficient tuning

  6. Few-shot learning with visual distribution calibration and cross-modal distribution alignment

  7. MetricPrompt: Prompting model as a relevance metric for few-shot text classification

  8. Multitask pre-training of modular prompt for chinese few-shot learning

  9. Cold-start data selection for better few-shot language model fine-tuning: A prompt-based uncertainty propagation approach

  10. Instruction induction: From few examples to natural language task descriptions

  11. Hierarchical verbalizer for few-shot hierarchical text classification

优化元学习中的参数

  1. How to train your MAML to excel in few-shot classification

  2. Meta-learning with fewer tasks through task interpolation

  3. Dynamic kernel selection for improved generalization and memory efficiency in meta-learning

  4. What matters for meta-learning vision regression tasks?

  5. Stochastic deep networks with linear competing units for model-agnostic meta-learning

  6. Robust meta-learning with sampling noise and label noise via Eigen-Reptile

  7. Attentional meta-learners for few-shot polythetic classification

  8. PLATINUM: Semi-supervised model agnostic meta-learning using submodular mutual information

  9. FAITH: Few-shot graph classification with hierarchical task graphs

  10. A contrastive rule for meta-learning

  11. Meta-ticket: Finding optimal subnetworks for few-shot learning within randomly initialized neural networks

学习搜索步骤

  1. Optimization as a model for few-shot learning

  2. Meta Navigator: Search for a good adaptation policy for few-shot learning

应用(57篇)

计算机视觉

  1. Analogy-forming transformers for few-shot 3D parsing

  2. Universal few-shot learning of dense prediction tasks with visual token matching

  3. Meta learning to bridge vision and language models for multimodal few-shot learning

  4. Few-shot geometry-aware keypoint localization

  5. AsyFOD: An asymmetric adaptation paradigm for few-shot domain adaptive object detection

  6. A strong baseline for generalized few-shot semantic segmentation

  7. StyleAdv: Meta style adversarial training for cross-domain few-shot learning

  8. DiGeo: Discriminative geometry-aware learning for generalized few-shot object detection

  9. Hierarchical dense correlation distillation for few-shot segmentation

  10. CF-Font: Content fusion for few-shot font generation

  11. MoLo: Motion-augmented long-short contrastive learning for few-shot action recognition

  12. MIANet: Aggregating unbiased instance and general information for few-shot semantic segmentation

  13. FreeNeRF: Improving few-shot neural rendering with free frequency regularization

  14. Exploring incompatible knowledge transfer in few-shot image generation

  15. Where is my spot? few-shot image generation via latent subspace optimization

  16. FGNet: Towards filling the intra-class and inter-class gaps for few-shot segmentation

  17. GeCoNeRF: Few-shot neural radiance fields via geometric consistency

机器人技术

  1. One solution is not all you need: Few-shot extrapolation via structured MaxEnt RL

  2. Bowtie networks: Generative modeling for joint few-shot recognition and novel-view synthesis

  3. Demonstration-conditioned reinforcement learning for few-shot imitation

  4. Hierarchical few-shot imitation with skill transition models

  5. Prompting decision transformer for few-shot policy generalization

  6. Stage conscious attention network (SCAN): A demonstration-conditioned policy for few-shot imitation

  7. Online prototype alignment for few-shot policy transfer

自然语言处理

  1. A dual prompt learning framework for few-shot dialogue state tracking

  2. CLUR: Uncertainty estimation for few-shot text classification with contrastive learning

  3. Few-shot document-level event argument extraction

  4. MetaAdapt: Domain adaptive few-shot misinformation detection via meta learning

  5. Code4Struct: Code generation for few-shot event structure prediction

  6. MANNER: A variational memory-augmented model for cross domain few-shot named entity recognition

  7. Few-shot event detection: An empirical study and a unified view

  8. CodeIE: Large code generation models are better few-shot information extractors

  9. Few-shot in-context learning on knowledge base question answering

  10. Linguistic representations for fewer-shot relation extraction across domains

  11. Few-shot reranking for multi-hop QA via language model prompting

知识图谱

  1. Adaptive attentional network for few-shot knowledge graph completion

  2. Learning inter-entity-interaction for few-shot knowledge graph completion

  3. Few-shot relational reasoning via connection subgraph pretraining

  4. Hierarchical relational learning for few-shot knowledge graph completion

  5. The unreasonable effectiveness of few-shot learning for machine translation

声音信号处理

  1. Audio2Head: Audio-driven one-shot talking-head generation with natural head motion

  2. Few-shot low-resource knowledge graph completion with multi-view task representation generation

  3. Normalizing flow-based neural process for few-shot knowledge graph completion

推荐系统

  1. Few-shot news recommendation via cross-lingual transfer

  2. ColdNAS: Search to modulate for user cold-start recommendation

  3. Contrastive collaborative filtering for cold-start item recommendation

  4. SMINet: State-aware multi-aspect interests representation network for cold-start users recommendation

  5. Multimodality helps unimodality: Cross-modal few-shot learning with multimodal models

  6. M2EU: Meta learning for cold-start recommendation via enhancing user preference estimation

  7. Aligning distillation for cold-start item recommendation

其他

  1. Context-enriched molecule representations improve few-shot drug discovery

  2. Sequential latent variable models for few-shot high-dimensional time-series forecasting

  3. Transfer NAS with meta-learned Bayesian surrogates

  4. Few-shot domain adaptation for end-to-end communication

  5. Contrastive meta-learning for few-shot node classification

  6. Task-equivariant graph few-shot learning

  7. Leveraging transferable knowledge concept graph embedding for cold-start cognitive diagnosis

理论(7篇)

  1. Bridging the gap between practice and PAC-Bayes theory in few-shot meta-learning

  2. bounds for meta-learning: An information-theoretic analysis

  3. Generalization bounds for meta-learning via PAC-Bayes and uniform stability

  4. Unraveling model-agnostic meta-learning via the adaptation learning rate

  5. On the importance of firth bias reduction in few-shot classification

  6. Global convergence of MAML and theory-inspired neural architecture search for few-shot learning

  7. Smoothed embeddings for certified few-shot learning

小样本/零样本学习(15篇)

  1. Finetuned language models are zero-shot learners

  2. Zero-shot stance detection via contrastive learning

  3. JointCL: A joint contrastive learning framework for zero-shot stance detection

  4. Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification

  5. Nearest neighbor zero-shot inference

  6. Continued pretraining for better zero- and few-shot promptability

  7. InstructDial: Improving zero and few-shot generalization in dialogue through instruction tuning

  8. Prompt-and-Rerank: A method for zero-shot and few-shot arbitrary textual style transfer with small language models

  9. Learning instructions with unlabeled data for zero-shot cross-task generalization

  10. Zero-shot cross-lingual transfer of prompt-based tuning with a unified multilingual prompt

  11. Finetune like you pretrain: Improved finetuning of zero-shot vision models

  12. SemSup-XC: Semantic supervision for zero and few-shot extreme classification

  13. Zero- and few-shot event detection via prompt-based meta learning

  14. HINT: Hypernetwork instruction tuning for efficient zero- and few-shot generalisation

  15. What does the failure to reason with "respectively" in zero/few-shot settings tell us about language models? acl 2023

小样本学习变体(12篇)

  1. FiT: Parameter efficient few-shot transfer learning for personalized and federated image classification

  2. Towards addressing label skews in one-shot federated learning

  3. Data-free one-shot federated learning under very high statistical heterogeneity

  4. Contrastive meta-learning for partially observable few-shot learning

  5. On the soft-subnetwork for few-shot class incremental learning

  6. Warping the space: Weight space rotation for class-incremental few-shot learning

  7. Neural collapse inspired feature-classifier alignment for few-shot class-incremental learning

  8. Learning with fantasy: Semantic-aware virtual contrastive constraint for few-shot class-incremental learning

  9. Few-shot class-incremental learning via class-aware bilateral distillation

  10. Glocal energy-based learning for few-shot open-set recognition

  11. Open-set likelihood maximization for few-shot learning

  12. Federated few-shot learning

数据集/基准(5篇)

  1. FewNLU: Benchmarking state-of-the-art methods for few-shot natural language understanding

  2. Bongard-HOI: Benchmarking few-shot visual reasoning for human-object interactions

  3. Hard-Meta-Dataset++: Towards understanding few-shot performance on difficult tasks

  4. MEWL: Few-shot multimodal word learning with referential uncertainty

  5. UNISUMM and SUMMZOO: Unified model and diverse benchmark for few-shot summarization

关注下方【学姐带你玩AI】🚀🚀🚀

回复“小样本160”获取论文+源代码合集

码字不易,欢迎大家点赞评论收藏!

评论 3
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值