Workshop
The 2nd Learning from Limited Labeled Data (LLD) Workshop: Representation Learning for Weak Supervision and Beyond
Isabelle Augenstein · Stephen Bach · Matthew Blaschko · Eugene Belilovsky · Edouard Oyallon · Anthony Platanios · Alex Ratner · Christopher Re · Xiang Ren · Paroma Varma
Room R01
Mon 6 May, 7:45 a.m. PDT
Modern representation learning techniques like deep neural networks have had a major impact on a wide range of tasks, achieving new state-of-the-art performances on benchmarks using little or no feature engineering. However, these gains are often difficult to translate into real-world settings because they usually require massive hand-labeled training sets. Collecting such training sets by hand is often infeasible due to the time and expense of labeling data; moreover, hand-labeled training sets are static and must be completely relabeled when real-world modeling goals change.
Increasingly popular approaches for addressing this labeled data scarcity include using weak supervision---higher-level approaches to labeling training data that are cheaper and/or more efficient, such as distant or heuristic supervision, constraints, or noisy labels; multi-task learning, to effectively pool limited supervision signal; data augmentation strategies to express class invariances; and introduction of other forms of structured prior knowledge. An overarching goal of such approaches is to use domain knowledge and data resources provided by subject matter experts, but to solicit it in higher-level, lower-fidelity, or more opportunistic ways.
In this workshop, we examine these increasingly popular and critical techniques in the context of representation learning. While approaches for representation learning in the large labeled sample setting have become increasingly standardized and powerful, the same is not the case in the limited labeled data and/or weakly supervised case. Developing new representation learning techniques that address these challenges is an exciting emerging direction for research [e.g., 1, 2]. Learned representations have been shown to lead to models robust to noisy inputs, and are an effective way of exploiting unlabeled data and transferring knowledge to new tasks where labeled data is sparse.
In this workshop, we aim to bring together researchers approaching these challenges from a variety of angles. Specifically this includes:
Learning representations to reweight and de-bias weak supervision
Representations to enforce structured prior knowledge (e.g. invariances, logic constraints).
Learning representations for higher-level supervision from subject matter experts
Representations for zero and few shot learning
Representation learning for multi-task learning in the limited labeled setting
Representation learning for data augmentation
Theoretical or empirically observed properties of representations in the above contexts
The second LLD workshop continues the conversation from the 2017 NIPS Workshop on Learning with Limited Labeled Data (http://lld-workshop.github.io). LLD 2017 received 65 submissions, of which 44 were accepted and was one of the largest workshops at NIPS 2017. Our goal is to once again bring together researchers interested in this growing field. With funding support, we are excited to again organize best paper awards for the most outstanding submitted papers. We also will have seven distinguished and diverse speakers from a range of machine learning perspectives, a panel on where the most promising directions for future research are, and a discussion session on developing new benchmarks and other evaluations for these techniques.
The LLD workshop organizers are also committed to fostering a strong sense of inclusion for all groups at this workshop, and to help this concretely, aside from $$1K for the paper awards, the remainder of the funding (both current and pending) will sponsor several travel awards specifically for traditionally underrepresented groups. We will also post a code of conduct emphasizing our commitment to inclusion, which we will expect all attendees to uphold.
[1] Norouzi et al. “Zero-Shot Learning by Convex Combination of Semantic Embeddings.” ICLR 2014.
[2] Liu et al. “Heterogeneous Supervision for Relation Extraction: A Representation Learning Approach.” EMNLP 2017.
Increasingly popular approaches for addressing this labeled data scarcity include using weak supervision---higher-level approaches to labeling training data that are cheaper and/or more efficient, such as distant or heuristic supervision, constraints, or noisy labels; multi-task learning, to effectively pool limited supervision signal; data augmentation strategies to express class invariances; and introduction of other forms of structured prior knowledge. An overarching goal of such approaches is to use domain knowledge and data resources provided by subject matter experts, but to solicit it in higher-level, lower-fidelity, or more opportunistic ways.
In this workshop, we examine these increasingly popular and critical techniques in the context of representation learning. While approaches for representation learning in the large labeled sample setting have become increasingly standardized and powerful, the same is not the case in the limited labeled data and/or weakly supervised case. Developing new representation learning techniques that address these challenges is an exciting emerging direction for research [e.g., 1, 2]. Learned representations have been shown to lead to models robust to noisy inputs, and are an effective way of exploiting unlabeled data and transferring knowledge to new tasks where labeled data is sparse.
In this workshop, we aim to bring together researchers approaching these challenges from a variety of angles. Specifically this includes:
Learning representations to reweight and de-bias weak supervision
Representations to enforce structured prior knowledge (e.g. invariances, logic constraints).
Learning representations for higher-level supervision from subject matter experts
Representations for zero and few shot learning
Representation learning for multi-task learning in the limited labeled setting
Representation learning for data augmentation
Theoretical or empirically observed properties of representations in the above contexts
The second LLD workshop continues the conversation from the 2017 NIPS Workshop on Learning with Limited Labeled Data (http://lld-workshop.github.io). LLD 2017 received 65 submissions, of which 44 were accepted and was one of the largest workshops at NIPS 2017. Our goal is to once again bring together researchers interested in this growing field. With funding support, we are excited to again organize best paper awards for the most outstanding submitted papers. We also will have seven distinguished and diverse speakers from a range of machine learning perspectives, a panel on where the most promising directions for future research are, and a discussion session on developing new benchmarks and other evaluations for these techniques.
The LLD workshop organizers are also committed to fostering a strong sense of inclusion for all groups at this workshop, and to help this concretely, aside from $$1K for the paper awards, the remainder of the funding (both current and pending) will sponsor several travel awards specifically for traditionally underrepresented groups. We will also post a code of conduct emphasizing our commitment to inclusion, which we will expect all attendees to uphold.
[1] Norouzi et al. “Zero-Shot Learning by Convex Combination of Semantic Embeddings.” ICLR 2014.
[2] Liu et al. “Heterogeneous Supervision for Relation Extraction: A Representation Learning Approach.” EMNLP 2017.
Live content is unavailable. Log in and register to view live content