• 沒有找到結果。

MICK: A Meta-Learning Framework for Few-shot Relation Classification with Small Training Data

N/A
N/A
Protected

Academic year: 2022

Share "MICK: A Meta-Learning Framework for Few-shot Relation Classification with Small Training Data"

Copied!
10
0
0

加載中.... (立即查看全文)

全文

(1)

MICK: A Meta-Learning Framework for Few-shot Relation Classification with Small Training Data

Xiaoqing Geng

gxq961127@sjtu.edu.cn Shanghai Jiao Tong University

Xiwen Chen

victoria-x@sjtu.edu.cn Shanghai Jiao Tong University

Kenny Q. Zhu

kzhu@cs.sjtu.edu.cn Shanghai Jiao Tong University

Libin Shen

libin@leyantech.com Leyan Tech

Yinggong Zhao

ygzhao@leyantech.com Leyan Tech

ABSTRACT

Few-shot relation classification seeks to classify incoming query instances after meeting only few support instances. This ability is gained by training with large amount of in-domain annotated data. In this paper, we tackle an even harder problem by further limiting the amount of data available at training time. We propose a few-shot learning framework for relation classification, which is particularly powerful when the training data is very small. In this framework, models not only strive to classify query instances, but also seek underlying knowledge about the support instances to obtain better instance representations. The framework also includes a method for aggregating cross-domain knowledge into models by open-source task enrichment. Additionally, we construct a brand new dataset: the TinyRel-CM dataset, a few-shot relation classifica- tion dataset in health domain with purposely small training data and challenging relation classes. Experimental results demonstrate that our framework brings performance gains for most underly- ing classification models, outperforms the state-of-the-art results given small training data, and achieves competitive results with sufficiently large training data.

KEYWORDS

few-shot learning, meta-learning, small training data, relation clas- sification

ACM Reference Format:

Xiaoqing Geng, Xiwen Chen, Kenny Q. Zhu, Libin Shen, and Yinggong Zhao. 2020. MICK: A Meta-Learning Framework for Few-shot Relation Classification with Small Training Data. In Proceedings of the 29th ACM International Conference on Information and Knowledge Management (CIKM

’20), October 19–23, 2020, Virtual Event, Ireland.ACM, New York, NY, USA, 10 pages. https://doi.org/10.1145/3340531.3411858

Corresponding Author.

Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from permissions@acm.org.

CIKM ’20, October 19–23, 2020, Virtual Event, Ireland

© 2020 Association for Computing Machinery.

ACM ISBN 978-1-4503-6859-9/20/10...$15.00 https://doi.org/10.1145/3340531.3411858

Table 1: An example of 5-way 5-shot relation classification scenario from the FewRel validation set.Entity 1 marks the head entity, andentity 2 marks the tail entity. The query in- stance is of Class 1: mother. The support instances of Classes 2-5 are omitted.

Support Set Class 1 mother:

Instance 1[Emmy Achté]ent ity 1was the mother of the interna- tionally famouse opera singers [Aino Ackté]ent ity 2and Irma Tervani.

Instance 2He was deposed in 922, and [Eadgifu]ent ity 1sent their son, [Louis]ent ity 2to safety in England.

Instance 3Jinnah and his wife [Rattanbai Petit]ent ity 1had sep- arated soon after their daughter, [Dina Wadia]ent ity 2was born.

Instance 4 Ariston had three other children by [Perictione]ent ity 1: Glaucon, [Adeimantus]ent ity 2, and Potone.

Instance 5She married (and murdered) [Polyctor]ent ity 2, son of Aegyptus and [Caliadne]ent ity 1. Apollodorus.

Class 2 part_of: ...

Class 3 military_rank: ...

Class 4 follows: ...

Class 5 crosses: ...

Query Instance

Dylan and [Caitlin]ent ity1 brought up their three children, [Aeronwy]ent ity2, Llewellyn and Colm.

1 INTRODUCTION

Relation classification (RC) is an indispensable problem in infor- mation extraction and knowledge discovery. Given a sentence (e.g., Washington is the capital of the United States) containing two tar- get entities (e.g., Washington and the United States), an RC model aims to distinguish the semantic relation between the entities men- tioned by the sentence (e.g., capital of ) among all candidate relation classes.

Conventional relation classification has been extensively inves- tigated. Recent approaches [24, 27, 30] train models with large amount of annotated data and achieve satisfactory results. How- ever, it is often costly to obtain necessary amount of labeled data by human-annotation, without which a decrease in the performance of the RC models is inevitable. In order to obtain sufficient labeled data, distant supervised methods [18] have been adopted to en- large annotation quantity by utilizing existing knowledge bases to perform auto-labeling on massive raw corpus. However, long-tail problems [11, 25, 28] and noise issues occur.

(2)

Few-shot relation classification is a particular RC task under minimum annotated data of concerned relation classes, i.e., a model is required to classify an incoming query instance given only few support instances (e.g., 1 or 5) during testing. An example is given in Table 1. It is worth noting that, in previous few-shot RC tasks, although models see only few support instances during testing, they are trained with a large amount of data (e.g., tens of thousands of instances in total), labeled with relation classes different from but in the same domain of the concerned relation classes. This situation arises when instances of several relation classes are hard to find while those of others are abundant. But in practice, it is often hard to get so many annotated training data as well. The difficulty lies in 2 aspects:

(1) Expertise required. Labeling is hard when it comes to pro- fessional fields such as bio-medical and risk management instead of general corpus. The relation classes in professional fields tend to be confusing and expertise is required. Thus the cost of labeling increases.

(2) Useless samples. Overwhelming majority (above 90% in some of our experiments) of instances are of relation class NA (no relation). This situation is particularly true in domains such as bio-medical and risk management, where a large proportion of instances are ordinary samples but only the samples with health issues or risks are of our concern. This makes labeling very unproductive.

In this paper, we highlight the situation where in-domain training data is hard to obtain as is illustrated above, and restrict the size of training data in few-shot relation classification.

Meta-learning is a popular method for few-shot learning circum- stances and is broadly studied in computer vision (CV) [14, 19, 21].

Instead of training a neural network to learn a specific task, in meta- learning, the model is trained with a variety of similar but different tasks to gain the ability of quick adaptation to new tasks without meeting a ton of data. One typical framework of meta-learning is to train an additional meta-learner which sets and adjusts the update rules for the conventional learner [1, 6, 10]. Another framework is based on metric learning and is aimed at learning the distance distribution among the relation classes [13, 21, 23]. In recent years, meta-learning has been shown to help in few-shot NLP tasks, in- cluding few-shot relation classification, which is the focus of this paper. Han et al. [11] constructed the FewRel dataset, a few-shot relation classification dataset and applied distinct meta-learning frameworks intended for CV tasks on the FewRel dataset. Ye and Ling [28], Gao et al. [7] and Gao et al. [8] improved the models specifically for relation classification task and achieved better per- formance.

While meta-learning frameworks outperform conventional meth- ods in few-shot RC, they are not applicable to situations where only small amount of annotated training data is available, which does happen in practice. In previous work, although only few support instances are needed during testing, the training set still must be suf- ficiently large (e.g., the FewRel dataset [11] contains 700 instances per class). Performance drops significantly when the training data size is restricted (e.g., to tens of instances per relation). Strong base- lines such as MLMAN [28] and Bert-Pair [8] achieves satisfactory performance (about 80% accuracy on 5-way 1-shot tasks) with full

x y

Support instance of class A Support instance of class B

Support instance of class C Query instance of class A

x y

x y

Previous Approaches Our Approach

Update Update

Original Representations

Previous Approaches Our Approach

Focus Focus

x y

x y

Figure 1: The general idea of model’s updating process of one iteration. Previous approaches update the representa- tions merely by the classification results on query instances.

Our approach updates the representations by classification results on both query instances and support instances. Solid lines indicate the intention of making representations of the query instance and each support instance of the same class closer to each other. Dotted lines indicate the intention of making representations of support instances within each class closer.

FewRel training data, but accuracy decreases by 20% if 10 relation classes with only 10 instances per class are given as training data (see Section 4.4). Besides, previous meta-learning methods concern much about computations on query instances but lose sight of knowledge within the support instances. Typical methods regard the features of support instances as standards to classify query in- stances during training iterations. We hold the view that improving

(3)

the quality of the standards itself by extracting knowledge within support instances is important but overlooked by previous work.

To enhance models and make them better cater to small train- ing data, we propose Meta-learning using Intra-support and Cross- domain Knowledge(MICK) framework. First, we utilize cross-domain knowledge by enriching the training tasks with cross-domain re- lation classification dataset (e.g., adding relation class mother to a medical relation classification training set, and forming a 3-way training task with relation classes Disease-cause-Disease, Disease- have-Symptom, and mother) (Section 3.3). The only requirement on the cross-domain dataset is to share the same language with the original dataset, thus it is easy to find. Although differences exist in distributions of data from distinct domains, basic knowledge such as language principles and grammar are shared, and thus compensate for the insufficient learning of basic knowledge due to lack of train- ing data. Moreover, even adding only a few cross-domain relation classes brings about a huge increase in the amount of feasible tasks.

Task enrichment makes the model more reliable by forcing it to solve extensive and diverse training tasks. Second, inspired by Chen et al. [4], we exploit underlying knowledge within support instances using a support classifier scheduled by a fast-slow learner (Section 3.4). Instead of updating the instance representations merely by classifying the query instances according to the support instances, we also use the support classifier to classify the support instances and update the model so that representations of support instances of the same class are closer to each other (see Figure 1). This makes the instance representations within the same class more compact, thus the classification of query instances becomes more accurate.

Additionally, we propose our own dataset, TinyRel-CM dataset, a Chinese few-shot relation classification dataset in health domain.

Different from previous few-shot RC dataset which contains abun- dant training data, we purposely limit the size of training data of our dataset. The TinyRel-CM dataset contains 27 relation classes with 50 instances per class, which sets up the challenge of few-shot relation classification under small training data. This dataset is also challenging because the relation classes in the test set are all very similar to each other. Experiments are conducted on both our proposed TinyRel-CM dataset and the FewRel dataset [11]. Experi- mental results show the strengths of our proposed framework.

In summary, our contributions include: (1) We propose a meta- learning framework that achieves state-of-the-art performance with the limitation of small training data and competitive results with sufficient training data (Section 4.4). (2) We utilize a support clas- sifier to extract intra-support knowledge and obtain more reason- able instance representations and demonstrate the improvement (Section 4.4). (3) We propose a task enrichment method to utilize cross-domain implicit knowledge, which is extremely useful un- der small training data (Section 4.4). (4) We propose TinyRel-CM dataset, the second and a challenging dataset for few-shot relation classification task with small training data (Section 4.1 and Section 4.4).

2 PROBLEM FORMULATION

We add limitation on the size of training data compared with previ- ous few-shot relation classification tasks. Just like in conventional few-shot relation classification, there is a training set Dtrainand

a test set Dtest. Each instance in both sets can be represented as a triple (s, e,r ), where s is a sentence of length T , e = (e1, e2)is the head and tail entities and r is the semantic relation between e1and e2conveyed by s. r ∈ R, where R = {r1, ..., rN}is the set of all candidate relation classes. Dtrainand Dtesthave disjoint relation sets, i.e., if a relation r appears in a triple of the training set, it must not appear in any triples of the test set and vice versa. Dtestis further split into a support set Dtest-sand a query set Dtest-q. The problem is to predict the classes of instances in Dtest-qgiven Dtest-s and Dtrain. While no restrictions are lied on how to use Dtrain, it is conventionally splited into a support set and a query set to train models. In a N -way K-shot scenario, Dtest-scontains N relation classes and K instances for each class. Both N and K are supposed to be small (e.g., 5-way 1-shot, 10-way 5-shot). Particularly, we limit the size of training data (i.e., Dtrainis also small). The difficulty of the task lies in not only the small size of Dtest-s(totally N × K instances) but also the small training data size.

3 METHODOLOGY

The proposed MICK framework contains a task enrichment method to aggregate cross-domain knowledge, and a support classifier scheduled by a fast-slow learner strategy to extract intra-support knowledge (illustrated in Figure 2). Next we present the structure and the learning process of MICK.

3.1 Overview of MICK

As is shown in Figure 2, the structure of our framework consists of four main parts: task enrichment, context encoder, class matching and support classifier.

During training with task enrichment, for each episode, we ran- domly select a N -way K-shot task composed of N × K support instances and several query instances extracted from both the orig- inal training set Dtrainand supplementary cross-domain dataset.

Task enrichment enables the context encoder to learn cross-domain underlying knowledge. The instances are first fed into the context encoder, which generates representation vectors for each instance.

Then we forward the encoded support and query vectors to class matching. Class matching aims to classify the query instances ac- cording to the representations of support instances. Additionally, the support vectors are fed into a support classifier, which is a N - way linear classifier. The support classifier aims to extract knowl- edge within support instances to facilitate the context encoder.

The back propagation process is scheduled by a fast-slow learner.

Fast-slow learner scheme is motivated by meta-learner based meta learning methods [1, 6, 10] where the traditional learner and meta- learner learn with different speed. Task-specific parameters in the support classifier update every episode with a fast learner. Task- agnostic parameters in the context encoder update every ϵ episodes with a slow learner. With the fast-slow learner, we obtain a support classifier which can quickly adapt to new classification tasks and a global context encoder that fits all tasks.

During testing, we only use the context encoder and class match- ing method to make predictions on Dtest.

(4)

Context

Encoder Class

Matching Support Classifier

match

L

pred Lsup

Dataset Task

enrichment Forward Pass

Backward Pass of Slow Learner Backward Pass of Fast Learner

Class A Class B Class C

Figure 2: The structure and learning process of the MICK framework (under a 3-way 3-shot example). Modules with black bor- der are typical meta-learning components. Modules with red border are our improvement. Cells of different colors represent instances from different classes. Light colors represent support instances. Dark colors represent query instances.

3.2 Preliminaries

Modules in black boxes in Figure 2 compose an ordinary meta- learning framework.

For an instance (s, e,r ), sentence s = {c0, ..., ci, ...cT −1}is padded to a predefined length T , where ci represents the one-hot vector of the ithword. In context encoder, each ciis mapped to an embedding xi ∈ Rdc+2dp, where dcis the word embedding1size and dpis the position embedding size. The representation matrix of sentence s is the concatenation of xi: X = [x0, ..., xi, ..., xT −1]. X ∈ RT ×(dc+2dp). Xis further fed into a sentence encoder (e.g., CNN or LSTM) to extract semantics of sentence s and get representation vector E ∈ Rdh, where dhis the dimension of hidden states. Thus, an instance can be represented as a pair (E,r ), where E is the representation vector and r is the relation class.

Given an encoded query instance Q = (Eq, rq)where rqis to be predicted, class matching aims to match rqwith some relation class ri ∈ R. Conventionally, a function F is adopted to measure the distance between Eqand Ei, where Ei is the representation vector of relation ri and is calculated with all the representation vectors of support instances belonging to class ri. F can be a function either without parameters (e.g., Euclidean distance function) or with parameters (e.g., a linear classifier). Relation class riis chosen as the prediction class if Eihas the closest distance to Eq.

3.3 Task Enrichment

In order to expand the size of training data and enrich training tasks, we utilize cross-domain relation classification datasets in the same language. These datasets are obtained from released data of other works or online resources. This task enrichment step is necessary under the circumstance of a tiny training dataset. With cross-domain datasets, training tasks are randomly extracted from (1) original data, (2) cross-domain data and original data, (3) origi- nal data, sequentially. This three-phase training scheme simulates the process of students learning from the simple to the deep, and reviewing before exams.

1Character embedding for Chinese corpus.

3.4 Support classifier and fast-slow learner

To further explore knowledge within support instances, we in- troduce a support classifier. The classifier receives representation vector E of each support instance as input and outputs the the prob- ability of the support instance belonging to each relation class. The output O equals to

O = exp(M[i])

PN −1

j=0 exp(M[j]), (1)

M = WE + b, (2)

where W and b are parameters to be trained, and M[i] represents the ithelement of M.

Algorithm 1Meta-learning with Support Classifier

Require:distribution over relation classes in training set p(R), context encoder Eθslow, class matching function F , support classifier Gθf ast, fast learner learning rate α, slow learner learning rate β, step size ϵ,

#classes per task N

1: Randomly initialize task-specific parameters θf ast 2: Randomly initialize task-agnostic parameters θslow 3: whilenot done do

4: Initialize slow loss Lslow= 0 5: forj = 1 to ϵ do

6: Sample N classes Ri∼ p ( R ) from training set Dtrain 7: Sample instances H = (s(i ), e(i ), r(i ))from Ri 8: Compute Lsup( Eθslow, Gθf ast)using H 9: Compute Lmat ch( Eθslow, F ) using H 10: Fast loss Lf ast= Lsup

11: Lslow= Lslow+ Lsup+ Lmat ch 12: θf ast= θf astα ▽θf ast Lf ast 13: end for

14: θslow= θslowβ ▽θslow Lslow 15: end while

16: Use Eθslow and F to perform classification of test set Dtest. The learning process with the support classifier is scheduled by a fast-slow learner (see Algorithm 1). The training process contains multiple episodes. For each episode, a task is randomly generated from the training data (line 6, 7). During training, we train two learners: a fast learner with learning rate α and a slow learner with learning rate β.

(5)

The fast learner learns parameters of the support classifier which are task-specific, and updates after each episode (line 12). The objective function of the fast learner is cross entropy loss (line 8, 10):

Lf ast = Lsup= −N −1X

i=0

r[i]loд(O[i]), (3) where N is the number of classes, r[i] is the ithelement of ground truth one-hot vector r, and O[i] is the ithelement of the output of the support classifier O.

The slow learner learns parameters of the context encoder (line 11) with objective function:

Lslow= Lsup+ Lmatch, (4)

where Lmatch is the objective function inherited from the core model which provides the context encoder and the class matching function. Lslowaccumulates during every ϵ episodes and then back propagates (line 11, 14).

4 EXPERIMENTS

In Section 4, we explain the experiment process in detail, including datasets, experimental setup, implementation details and evaluation results.

4.1 Dataset

Experiments are done on two datasets: FewRel dataset [11] and our proposed TinyRel-CM dataset.

FewRel DatasetFewRel dataset [11] is a few-shot relation classifi- cation dataset constructed through distant supervision and human annotation. It consists of 100 relation classes with 700 instances per class. The relation classes are split into subsets of size 64, 16 and 20 for training, validation and testing, respectively. The average length of a sentence in FewRel dataset is 24.99, and there are 124,577 unique tokens in total. At the time of writing, FewRel is the only few-shot relation classification dataset available.

TinyRel-CM Dataset2TinyRel-CM dataset is our proposed Chi- nese few-shot relation classification dataset in health domain with small training data. The TinyRel-CM dataset is constructed through the following steps: (1) Crawl data from Chinese health-related web- sites3to form a large corpus and an entity dictionary. (2) Automati- cally align entities in the corpus with the entity dictionary, forming a large candidate-sentence set. (3) 5 Chinese medical students man- ually filter out the unqualified candidate sentences and tag qualified ones with corresponding class labels to form an instance. An in- stance is added to the dataset only if 3 or more annotators make consistent decisions. This process costs 4 days.

The TinyRel-CM Dataset consists of 27 relation classes with 50 instances per class. The 27 relation classes cover binary relations among 4 entity types, and are grouped into 6 categories according to the entity types, forming 6 tasks with one group being the test set and other 5 groups serving as training set (see Table 2). Grouping makes TinyRel-CM dataset more challenging because all candidate

2Code and dataset released in https://github.com/XiaoqingGeng/MICK.

3www.9939.com, www.39.net, and www.xywy.com

relation classes during testing are highly similar. An example in- stance in the TinyRel-CM dataset is shown in Tabel 3. The average length of a sentence in TinyRel-CM dataset is 67.31 characters, and there are 2,197 unique characters in total. Comparison between TinyRel-CM dataset and FewRel dataset is shown in Table 4.

Table 2: Entity groups in TinyRel-CM dataset. D,S,F, and U stand for Disease, Symptom, Food, and nUtrient, respec- tively.

Group Classes

D-D complication, cause, is, include, NA D-S have, NA

D-F positive, negative, forbid, prevent, cause, NA D-U positive, negative, prevent, lack, cause, NA F-U contain, NA

S-F forbid, cause, positive, negative, prevent, NA

Table 3: An example instance in the TinyRel-CM dataset.

Group D-D

Class complication

Explanation Entity 2 is a complication of entity 1.

Example [子宫肌瘤]ent ity1出现了[慢性盆腔炎]ent ity2 发症,导致月经量过多。

Once [Hysteromyoma]ent ity1is complicated with [chronic pelvic inflammation]ent ity2, menstruation increases.

Table 4: Comparison of TinyRel-CM dataset to FewRel dataset.

Dataset #cls. #inst./cls. #inst.

FewRel 100 700 70,000

TinyRel-CM 27 50 1,350

4.2 Experimental Setup

We first conduct experiments with small training data. On the TinyRel-CM dataset, for each group of relation classes, we adopt N - way 5-shot, N -way 10-shot and N -way 15-shot test configurations, where N is the number of classes within the group. During training episodes, we conduct 5-way 15-shot training tasks. Thus totally 6 experiments are done on TinyRel-CM dataset. On the FewRel dataset, we modify the training set by shrinking the number of relation classes and instances per class to various extent. This aims to show not only the effect of our framework under small training data but also the performance trends of models with the change of data size. For each shrunken training set, we conduct different training task settings (shown in Table 5) and test with 4 config- urations: 5-way 1-shot, 5-way 5-shot, 10-way 1-shot and 10-way 5-shot.

Second, we experiment with sufficient training data. On the FewRel dataset, following Han et al. [11] and Ye and Ling [28], we train the model with 20-way 10-shot training tasks and test with

(6)

Table 5: Training task settings over shrunken training set.

% of full training set #cls. #inst./cls. Training task

7.00 30 100 5-way 15-shot

2.23 20 50 5-way 15-shot

1.00 15 30 5-way 10-shot

0.22 10 10 5-way 5-shot

4 configurations: 5-way 1-shot, 5-way 5-shot, 10-way 1-shot and 10-way 5-shot.

For the ablation tests, in addition to applying the whole MICK framework, we also apply the two proposed methods, support clas- sifier and task enrichment, individually on baseline models.

For all experiments, we randomly pick 2000 tasks and calculate the average accuracy in testing.

4.3 Implementation Details

During implementation, we apply our framework and data aug- mentation method to the following baselines: GNN [20], SNAIL [16], prototypical networks [21], proto-HATT [7], MLMAN [28], and Bert-Pair [8].

Codes for GNN, SNAIL and Bert-Pair are provided by Gao et al.

[8]. Prototypical network uses our own implementation. Codes for proto-HATT and MLMAN are provided in the original paper.

Due to the particularity of the Bert-Pair model, the support clas- sifier applied on Bert-Pair receives support instance pairs as input and computes the probability of the pairs belonging to the same class, different from other baselines. Although the distinct imple- mentation of support classifier, we keep our intention to extract knowledge within support instances.

GNN, SNAIL, and proto-HATT require the number of classes while training and testing to be equal. So a model is trained with N - way tasks to perform N -way test tasks. In SNAIL and proto-HATT, the number of instances per class while training and testing need to be equal. So a model is trained with K-shot tasks to perform K-shot test tasks.

We keep the original hyper parameters for each baseline, and set the learning rate of the fast learner 0.1. The cross-domain data for TinyRel-CM dataset are from Chinese Literature NER RE dataset [26] (13,297 instances covering 10 classes in general corpus in- cluding part_whole, near, etc.) and Chinese Information Extraction dataset [9] (1,100 instances covering 12 classes between persons including parent_of, friend_of, etc.). Cross-domain data for FewRel dataset is from NYT-10 dataset [18] which contains 143,391 in- stances over 57 classes in general courpus including contain, nation- ality, etc. (class NA is removed during task enrichment). The only requirement on the supplementary dataset is to share the common language with the original dataset.

4.4 Results and Analysis

Here, we show the experimental results and analyze them from different aspects.

4.4.1 With small training data.To show the effectiveness of our MICK framework under small training data, we apply it on each

baseline model on (1) FewRel dataset with training set shrunken to different extents, and (2) our proposed TinyRel-CM dataset.

Figure 3 shows the performance comparison between baseline models and our framework given different amount of training data on FewRel dataset. Each subgraph contains 6 groups, one group for each baseline model. For each group, the training data size increases from the left to right. Here, we focus on the blue curves which present baseline accuracies, the red curves which present the MICK enhanced model accuracies, and the gray bars which present the performance gains that MICK brings. As is illustrated, performance of models deteriorates with the decrease of training data size. For example, the prototypical networks achieves 57.53%

accuracy with full training data under 10-way 1-shot test tasks, but performs poorly with 22.30% accuracy given only 0.22% of full training data. Our framework fits in situations where only small training data is available. As is shown in Figure 3, with our methods, the less the training data, the more improvement the model tends to gain. This indicates the effectiveness of our framework under extremely small training data. While our methods only improve prototypical networks by 1.27% accuracy with full training data under 10-way 1-shot test tasks, it leads to 22.74% improvement given 0.22% of full training data.

Table 6 shows the experimental results on the TinyRel-CM dataset. Our framework considerably improves model performance in most cases. Strong baselines on FewRel dataset such as Bert-Pair do not perform well, partially because the TinyRel-CM dataset is more verbal and informal, thus quite distinct from BERT-Chinese’s pretraining data (Chinese Wikipedia).

On both datasets, when the input sentence contains multiple relation classes, model performance drops. E.g., under 2.23% full FewRel training data under 5-way 5-shot tasks using MLMAN, classification accuracy on relation class mother decreases by 14.14%

when spouse or child appears as interference.

4.4.2 With sufficient training data. We use full training data from FewRel dataset and apply MICK framework on each baseline model.

Here, we focus on the rightmost blue points, red points and gray bars of each group in Figure 3, which present the baseline performance, MICK performance and performance gains brought by MICK under full FewRel training data. As is shown in Figure 3, on full FewRel dataset, in most cases, our framework achieves better performance than baseline models. The framework brings more improvement for relatively poor baselines such as GNN and SNAIL (about 5% accuracy) than strong baselines such as MLMAN and Bert-Pair (about 1% accuracy). This is because our framework aims to help models learn better and doesn’t change the core part of the models. Full training data is sufficient for strong baselines to train a good model, leaving limited room for improvement.

4.4.3 Ablation tests.Here, we focus on the individual effects of support classifier and task enrichment on baseline models.

As is shown in Figure 3 and Table 6, in most cases, either adding support classifier or task enrichment improves performance for baseline models. This illustrates that both support classifier and task enrichment contribute to better performance.

On FewRel dataset, we focus on the green and orange curves in Figure 3, which present the accuracies of applying support classi- fier or task enrichment individually. Support classifier generally

(7)

Gains Baseline SC TE

GNN SNAIL Proto HATT MM BP MICK

0%

40%

80%

(a) 5-way 1-shot

Gains Baseline SC TE

GNN SNAIL Proto HATT MM BP MICK

0%

40%

80%

(b) 5-way 5-shot

Gains Baseline SC TE

GNN SNAIL Proto HATT MM BP MICK

0%

40%

80%

(c) 10-way 1-shot

Gains Baseline SC TE

GNN SNAIL Proto HATT MM BP MICK

0%

40%

80%

(d) 10-way 5-shot

Figure 3: Classification accuracy on FewRel validation set underN -way K-shot configurations. Gains is the difference between MICK accuracy and baseline accuracy. For each group, from the left to the right, training data is shrunken to 0.22%, 1.00%, 2.23%, 7.00%, and 100.00% of full training set size, respectively. For each shrunken training set, we apply baseline models and models with SC (Support Classifier individually), TE (Task Enrichment individually) and the whole MICK framework. Proto, HATT, MM, and BP stand for Prototypical Networks, Proto-HATT, MLMAN, and Bert-Pair, respectively.

brings more performance gains when less training data is given.

This is because with small training data, baseline models fail to extract adequate knowledge while the support classifier helps with the knowledge learning process. In some rare cases such as SNAIL under 5-way 1-shot and 5-way 5-shot scenarios with 0.22% training data, adding support classifier leads to not much improvement. This is because support classifier guides models to extract more knowl- edge from very limited resources, while the extractable knowledge is restricted by the training set size and some of the knowledge is even useless. Support classifier improves poor baselines such

as GNN and SNAIL to a larger extent than strong ones such as MLMAN and Bert-Pair because the support classifier makes up for the insufficient learning ability for poor models while is icing on the cake for strong models.

Task enrichment is much more helpful than support classifier under small training data (when training data is shrunken to less than 1.00%). This is because task enrichment compensates for the lack of learning sources of basic knowledge such as basic rules and grammar when training data is quite limited. Improvement of task enrichment is also more obvious on poor baselines than strong

(8)

Table 6: Classification accuracy (%) on TinyRel-CM dataset underN -way K-shot configuration. D, S, F, and U stand for Disease, Symptom, Food, and nUtrient, respectively. Proto, HATT, MM, and BP stand for Prototypical Networks, Proto-HATT, MLMAN, and Bert-Pair, respectively. Gray numbers indicate the accuracy is lower than baseline.

Method Baseline +SupportClassifier

Group D-D D-S D-F D-U F-U S-F D-D D-S D-F D-U F-U S-F

(N) (5) (2) (6) (6) (2) (6) (5) (2) (6) (6) (2) (6)

GNN (K= 5 ) 24.04 53.62 34.15 38.90 54.60 32.45 25.71 57.12 33.11 40.23 53.20 34.51 (K=10) 26.05 53.73 37.54 44.02 55.53 35.45 27.96 57.35 37.44 43.87 54.85 38.39 (K=15) 26.94 54.30 38.36 45.12 55.45 37.43 28.36 61.25 39.18 47.57 55.10 40.37 SNAIL (K= 5 ) 21.74 50.00 17.78 25.98 47.96 25.38 23.64 53.90 27.28 27.60 53.15 25.88 (K=10) 19.49 53.57 27.55 22.96 51.63 22.51 23.98 56.65 27.76 30.10 54.60 26.10 (K=15) 21.94 51.62 20.03 18.21 50.33 21.38 23.12 57.40 27.72 31.43 56.35 25.40 Proto (K= 5 ) 30.03 63.27 31.35 33.95 54.99 26.97 28.99 69.56 33.57 43.78 56.49 36.81 (K=10) 31.68 67.22 34.09 36.27 56.78 29.18 32.35 72.54 36.10 47.78 58.13 39.04 (K=15) 33.80 70.29 34.57 37.41 57.03 29.72 34.82 72.79 36.96 50.03 60.09 40.09 HATT (K= 5 ) 29.67 59.27 40.13 40.19 65.20 41.67 36.39 71.45 40.83 54.17 70.09 43.33 (K=10) 34.50 59.77 39.17 48.33 67.50 43.57 40.10 71.24 45.34 50.83 73.33 49.17 (K=15) 39.01 63.75 37.41 56.67 66.25 37.13 43.05 77.09 44.17 51.67 63.75 45.92 MM (K= 5 ) 31.22 60.58 44.97 47.11 61.04 44.78 33.48 62.34 44.93 49.49 62.40 45.90 (K=10) 36.66 64.96 49.44 52.49 66.68 50.57 38.89 67.30 50.09 54.09 68.39 50.77 (K=15) 40.65 68.14 52.44 56.00 70.32 53.13 43.05 69.41 53.82 57.23 72.02 52.87 BP (K= 5 ) 23.61 53.48 46.02 42.03 54.17 46.18 24.60 53.98 50.49 43.28 63.33 48.75 (K=10) 24.17 56.95 49.52 44.37 59.72 48.96 26.69 55.05 53.60 46.13 66.70 51.63 (K=15) 25.80 54.15 50.38 45.16 59.33 48.78 27.29 56.65 54.22 47.81 67.55 52.48

Method +TaskEnrich +SupportClassifier&TaskEnrich

Group D-D D-S D-F D-U F-U S-F D-D D-S D-F D-U F-U S-F

(N) (5) (2) (6) (6) (2) (6) (5) (2) (6) (6) (2) (6)

GNN (K= 5 ) 24.59 51.20 33.28 40.76 54.62 35.36 26.02 66.22 37.48 44.47 55.02 37.13 (K=10) 26.85 56.05 37.56 44.90 54.12 37.69 27.66 58.75 37.88 45.15 53.42 37.85 (K=15) 27.71 58.55 40.42 49.43 56.35 40.13 27.40 70.10 43.25 51.40 56.45 40.03 SNAIL (K= 5 ) 22.43 53.95 25.17 29.05 51.30 27.27 24.64 58.55 28.57 27.08 53.30 27.88 (K=10) 22.79 51.98 30.67 22.13 53.15 23.03 27.38 57.12 27.85 33.50 55.60 30.52 (K=15) 21.54 53.10 20.77 20.35 52.00 23.30 23.84 56.70 30.63 35.75 54.30 30.43 Proto (K= 5 ) 31.40 69.22 32.13 34.22 55.34 30.34 31.50 68.79 34.36 44.76 57.43 40.36 (K=10) 34.80 72.34 34.51 36.58 57.29 33.81 35.45 71.37 37.11 48.28 58.71 43.55 (K=15) 36.97 71.62 35.76 37.79 57.74 34.49 38.28 72.62 37.90 49.58 60.00 44.97 HATT (K= 5 ) 33.70 67.50 41.67 43.33 65.62 38.33 38.21 75.00 46.67 49.17 73.75 44.17 (K=10) 38.98 68.75 40.13 48.75 60.92 46.67 44.45 70.23 49.37 51.67 75.00 44.47 (K=15) 40.39 72.31 43.84 55.83 70.51 43.33 49.61 80.41 44.33 55.81 72.58 49.17 MM (K= 5 ) 36.05 64.98 45.24 50.03 61.34 46.82 38.17 64.68 45.32 51.30 64.24 48.34 (K=10) 42.04 70.02 50.83 54.85 67.57 52.47 44.89 69.93 50.75 56.60 70.79 53.69 (K=15) 46.55 71.95 54.18 58.19 72.01 55.28 49.23 72.41 54.25 59.67 74.58 56.67 BP (K= 5 ) 25.25 59.52 48.49 40.32 55.73 48.68 26.87 58.35 52.54 39.48 62.52 49.95 (K=10) 27.54 61.48 51.11 44.33 60.15 51.56 28.19 62.32 55.86 43.25 66.38 52.92 (K=15) 28.29 63.98 52.18 47.32 61.02 53.36 29.25 63.15 57.60 43.26 68.83 54.16 ones. Since poor models fail to master some of the basic knowledge

brought in original training data, the introduction of cross-domain data not only brings extra basic knowledge but also helps models to master knowledge in original training data by providing more related data.

Applying both support classifier and task enrichment achieves best performance in most cases, because the two methods comple- ment each other. Support classifier extract useful information from original training data and cross-domain data, and task enrichment provide extra sources for the support classifier. In some cases, al- though adding either support classifier or task enrichment does not

affect much, applying them together leads to a great improvement (e.g., HATT under 0.22% training data).

On TinyRel-CM dataset (Table 6), task enrichment brings about similar improvements for all baselines due to small training data.

Support classifier leads to more performance gains than task en- richment individually, indicating TinyRel-CM dataset is more chal- lenging (the relation classes are much more similar than FewRel dataset) and baseline models fail to extract sufficient useful knowl- edge. Although task enrichment brings not much improvement or even negative effects in some cases, adding support classifier simultaneously tend to raise the accuracy to a large extent (up to 10%). The reason is that with insufficient learning ability (although

(9)

baselines such as MLMAN and Bert-Pair are relatively strong, their learning ability is still insufficient with small training data), cross- domain data alone sometimes introduces noise and thus confuses the model. But with the additional support classifier that improves learning ability, models are able to learn more useful knowledge from cross-domain data.

On TinyRel-CM dataset (Table 6), few cases occur where adding both support classifier and task enrichment performs worse than adding only one of them (e.g., Bert-Pair under group D-U, and pro- totypical networks under group D-S in TinyRel-CM dataset). The main reason is that while adding the support classifier helps models learn more and better, the model occasionally tend to pay excessive attention to the distinctive distribution of the cross-domain data.

As is shown in the corresponding results, in the vast majority of cases, while adding support classifier on baseline models raises accuracy, adding support classifier on task enriched models brings less improvement or even bad effect. This indicates stronger learn- ing ability owing to the support classifier, and the distraction caused by cross-domain data.

Table 7: Human evaluation result on TinyRel-CM dataset un- der 5-shot scenario.

Group D-D D-S D-F D-U F-U S-F

(N) (5) (2) (6) (6) (2) (6)

Acc(%) 92.65 96.53 86.93 91.77 96.71 84.97

4.4.4 Dataset analysis. Our TinyRel-CM dataset is a challenging task. We human evaluate the dataset and results are illustrated in Table 7. During the human-evaluation process, under N -way K-shot scenario, we provide instances of N relation classes with K instances per relation class. We use labels 1 to N to name the relation classes instead of their real names. A person is required to classify a new coming instance into one of the N classes. We only evaluate the TinyRel-CM dataset under 5-shot scenarios because 10-shot and 15-shot tasks are too easy for human. Three volunteers participated in the human evaluation process and we take the average accuracy as the final result.

Comparing Table 6 and Table 7, the performance of state-of-the- art models is still far worse than human performance, indicating the TinyRel-CM dataset is a challenging task.

5 RELATED WORK

Relation classification task aims to categorize the semantic relation between two entities conveyed by a given sentence into a relation class. In recent years, deep learning has become a major method for relation classification. Zeng et al. [30] utilized a convolutional deep neural network (DNN) during relation classification to ex- tract lexical and sentence-level features. Vu et al. [24] employed a voting scheme by aggregating a convolutional neural network with a recurrent neural network. Traditional relation classifica- tion models suffer from lack of data. To eliminate this deficiency, distant-supervised approaches are proposed, which take advantage of knowledge bases to supervise the auto-annotation on massive raw data to form large datasets. Riedel et al. [18] constructed the

NYT-10 dataset with Freebase [3] as the distant supervision knowl- edge base and sentences in the New York Times (2005-2007) as raw text. Distant-supervised methods suffer from long tail problems and excessive noise. Zeng et al. [29] proposed piecewise convolutional neural networks (PCNNs) with instance-level attention to eliminate the negative effect caused by the wrongly labeled instances on NYT- 10 dataset. Liu et al. [15] further introduced soft label mechanism to automatically correct not only the wrongly labeled instances but also the original noise from the distant supervision knowledge base.

The lack of annotated data leads to the emergence of few-shot learning, where models need to perform classification tasks without seeing much data. Meta-learning is a popular method for few-shot learning and is widely investigated in computer vision (CV). Lake et al. [14] proposed Omniglot, a few-shot image classification dataset and put forward the idea of learning to learn, which is the essence of meta-learning. Memory-Augmented Neural Networks [19] uti- lized a recurrent neural network with augmented memory to store information for the instances the model has encountered. Meta Networks [17] implemented a high-level meta learner based on the conventional learner to control the update steps of the con- ventional learner. GNN [20] regarded support and query instances as nodes in a graph where information propagates among nodes, and classified a query node with the information of support nodes.

SNAIL [16] aggregated attention into the meta learner. Prototypi- cal networks [21] assumed that each relation has a prototype and classified a query instance into the relation of the closest prototype.

Image deformation meta-networks [4] utilized a image deformation sub-net to generate more training instances for one-shot image classification.

Few-shot relation classification is a newly-born task that requires models to do relation classification under merely a few support instances. Han et al. [11] proposed the FewRel dataset for few-shot relation classification and applied meta-learning methods intended for CV, including Meta Networks [17], GNN [20], SNAIL [16], and prototypical networks [21], on the FewRel dataset. Prototypical networks with CNN core turned out to have the best test accuracy among the reported results. Models that are more applicable for relation classification tasks are further proposed. ProtoHATT [7]

reinforced the prototypical networks with hybrid attention mecha- nism. MLMAN [28] improved the prototypical networks by adding mutual information between support instances and query instances.

Bert-Pair [8] adopted BERT [5] to conduct binary relation classifi- cations between a query instance and each support instance and fine-tuned on FewRel dataset. Baldini Soares et al. [2] trained a BERT-like language model on huge open-source data with a Match- ing the Blanks task and applied the trained model on few-shot relation classification tasks.

Meta networks perform worst among all the methods [11] and is time consuming (about 10 times slower than other methods with even better results). Matching the Blanks is high-resource and not comparable to other methods. So we do not adopt these two methods as baselines.

Previous methods lose sight of the significant knowledge within the support instances. Chen et al. [4] conducted one-shot image classification task with an image deformation sub-net. The sub-net is designed to generate more support instances to augment labeled

(10)

data and is trained by a prototype classifier. The prototype classifier updates the deformation sub-net according to the classification results on prototypes of generated support instances, and aims to improve the deformation process. Inspired by this work, we add a support classifier over each support instance during the training process. The support classifier helps with the update process of the parameters in both the support classifier and the encoder, and is scheduled with a fast-slow learner scheme. The support classifier aims to utilize knowledge within support instances to obtain better instance representations. Additionally, previous few-shot relation classification models are trained with sufficient training data despite the small support set size during testing, we put forward a new challenge on few-shot relation classification by limiting the training data size.

Relation classification datasets have been released in past decades.

Conventional ones include SemEval-2010 Task 8 dataset [12], ACE 2003-2004 dataset [22], TACRED dataset [31] and NYT-10 dataset [18]. All these datasets encompass sufficient data to train a strong model. Han et al. [11] released the first few-shot relation classifica- tion dataset, the FewRel dataset, which contains 100 relation classes with 700 instances per class. Although only few support instances are provided in each test task, the training data is sufficiently large.

We propose TinyRel-CM dataset, a brand new few-shot relation classification dataset with purposely small training data and chal- lenging relation classes. The TinyRel-CM dataset is the second and first Chinese few-shot relation classification dataset.

6 CONCLUSION

In this paper, we propose a few-shot learning framework for relation classification that aims to (1) extract intra-support knowledge by classifying both support and query instances, and (2) bring external implicit knowledge from cross-domain corpus by task enrichment.

Our framework is particularly powerful when only small amount of training data is available. Additionally, we construct our own dataset, the TinyRel-CM dataset, a Chinese few-shot relation clas- sification dataset in medical domain. The small training data size and highly similar relation classes make the TinyRel-CM dataset a challenging task. As for future work, we intend to futher investi- gate whether the proposed framework is able to handle zero-shot learning tasks.

ACKNOWLEDGEMENT

This work was partially supported by NSFC grant 91646205, SJTU Medicine-Engineering Cross-disciplinary Research Scheme and SJTU-Leyan Joint Research Scheme.

REFERENCES

[1] Marcin Andrychowicz, Misha Denil, Sergio Gomez, Matthew W. Hoffman, David Pfau, Tom Schaul, Brendan Shillingford, and Nando De Freitas. 2016. Learning to Learn by Gradient Descent by Gradient Descent. In NIPS.

[2] Livio Baldini Soares, Nicholas FitzGerald, Jeffrey Ling, and Tom Kwiatkowski.

2019. Matching the Blanks: Distributional Similarity for Relation Learning. In Proceedings of ACL. 2895–2905.

[3] Kurt Bollacker, Colin Evans, Praveen Paritosh, and Jamie and Taylor. 2008. Free- base: a Collaboratively Created Graph Database for Structuring Human Knowl- edge. In Proceedings of the 2008 ACM SIGMOD international conference on Man- agement of data. 1247–1250.

[4] Zitian Chen, Yanwei Fu, Yu-Xiong Wang, Lin Ma, Wei Liu, and Martial Hebert.

2019. Image Deformation Meta-Networks for One-Shot Learning. In Proceedings of CVPR. 8680–8689.

[5] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT:

Pre-training of Deep Bidirectional Transformers for Language Understanding.

arXiv preprint arXiv:1810.04805(2018).

[6] Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-Agnostic Meta- Learning for Fast Adaptation of Deep Networks. In Proceedings of ICML, Vol. 70.

1126–1135.

[7] Tianyu Gao, Xu Han, Zhiyuan Liu, and Maosong Sun. 2019. Hybrid Attention- Based Prototypical Networks for Noisy Few-Shot Relation Classification.

[8] Tianyu Gao, Xu Han, Hao Zhu, Zhiyuan Liu, Peng Li, Maosong Sun, and Jie Zhou.

2019. FewRel 2.0: Towards More Challenging Few-Shot Relation Classification.

In Proceedings of EMNLP-IJCNLP. 6249–6254.

[9] Wang Guan. 2017. Information-Extraction-Chinese. https://github.com/

crownpku/Information-Extraction-Chinese/tree/master/RE_BGRU_2ATT/

origin_data.

[10] David Ha, Andrew Dai, and Quoc V. Le. 2017. HyperNetworks. (2017).

[11] Xu Han, Hao Zhu, Pengfei Yu, Ziyun Wang, Yuan Yao, Zhiyuan Liu, and Maosong Sun. 2018. FewRel: A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation. In Proceedings of EMNLP. 4803–4809.

[12] Iris Hendrickx, Su Nam Kim, Zornitsa Kozareva, Preslav Nakov, Diarmuid Ó Séaghdha, Padó Sebastian, Marco Pennacchiotti, Lorenza Romano, and Stan Szpakowicz. 2009. Semeval-2010 Task 8: Multi-Way Classification of Semantic Relations Between Pairs of Nominals. In Proceedings of SemEval@ACL.

[13] Gregory Koch, Richard Zemel, and Ruslan Salakhutdinov. 2015. Siamese Neural Networks for One-Shot Image Recognition. In ICML Deep Learning Workshop, Vol. 2.

[14] Brenden M. Lake, Ruslan Salakhutdinov, and Joshua B. Tenenbaum. 2015. Human- level Concept Learning Through Probabilistic Program Induction. Science 350 (2015).

[15] Tianyu Liu, Kexiang Wang, Baobao Chang, and Zhifang Sui. 2017. A Soft-label Method for Noise-tolerant Distantly Supervised Relation Extraction. In Proceed- ings of EMNLP. 1790–1795.

[16] Nikhil Mishra, Mostafa Rohaninejad, Xi Chen, and Pieter Abbeel. 2018. A Simple Neural Attentive Meta-Learner. In Proceedings of ICLR.

[17] Tsendsuren Munkhdalai and Hong Yu. 2017. Meta networks. In Proceedings of ICML.

[18] Sebastian Riedel, Limin Yao, and Andrew D McCallum. 2010. Modeling Relations and Their Mentions without Labeled Text. In Proceedings of ECML-PKDD.

[19] Adam Santoro, Sergey Bartunov, Matthew Botvinick, Daan Wierstra, and Timothy Lillicrap. 2016. Meta-learning with Memory-Augmented Neural Networks. In ICML. 1842–1850.

[20] Victor Garcia Satorras and Joan Bruna Estrach. 2018. Few-shot Learning with Graph Nerual Networks. In Proceedings of ICLR.

[21] Jake Snell, Kevin Swersky, and Richard Zemel. 2017. Prototypical Networks for Few-shot Learning. In NIPS. 4077–4087.

[22] Stephanie Strassel, Mark A. Przybocki, Kay Peterson, Zhiyi Song, and Kazuaki Maeda. 2008. Linguistic Resources and Evaluation Techniques for Evaluation of Cross-Document Automatic Content Extraction. In Proceedings of LREC.

[23] Oriol Vinyals, Charles Blundell, Timothy Lillicrap, and Daan Wierstra. 2016.

Matching Networks for One Shot Learning. In NIPS. 2680–3638.

[24] Ngoc Thang Vu, Heike Adel, Pankaj Gupta, and Hinrich Schütze. 2016. Combining Recurrent and Convolutional Neural Networks for Relation Classification. In Proceedings of NAACL: Human Language Technologies. 534–539.

[25] Wenhan Xiong, Mo Yu, Shiyu Chang, Xiaoxiao Guo, and William Yang Wang.

2018. One-Shot Relational Learning for Knowledge Graphs. In Proceedings of EMNLP. 1980–1990.

[26] Jingjing Xu, Ji Wen, Xu Sun, and Qi Su. 2017. A Discourse-Level Named Entity Recognition and Relation Extraction Dataset for Chinese Literature Text. CoRR abs/1711.07010.

[27] Yan Xu, Lili Mou, Ge Li, Yunchuan Chen, Hao Peng, and Zhi Jin. 2015. Classifying Relations via Long Short Term Memory Networks along Shortest Dependency Paths. In Proceedings of EMNLP. 1785–1794.

[28] Zhi-Xiu Ye and Zhen-Hua Ling. 2019. Multi-Level Matching and Aggregation Network for Few-Shot Relation Classification. In Proceedings of ACL. 2872–2881.

[29] Daojian Zeng, Kang Liu, Yubo Chen, and Jun Zhao. 2015. Distant Supervision for Relation Extraction via Piecewise Convolutional Neural Networks. In Proceedings of EMNLP. 1753–1762.

[30] Daojian Zeng, Kang Liu, Siwei Lai, Guangyou Zhou, and Jun Zhao. 2014. Relation Classification via Convolutional Deep Neural Network. In Proceedings of COLING.

2335–2344.

[31] Yuhao Zhang, Victor Zhong, Danqi Chen, Gabor Angeli, and Christopher D.

Manning. 2017. Position-aware Attention and Supervised Data Improve Slot Filling. In Proceedings of EMNLP. 35–45.

數據

Figure 1: The general idea of model’s updating process of one iteration. Previous approaches update the  representa-tions merely by the classification results on query instances.
Figure 2: The structure and learning process of the MICK framework (under a 3-way 3-shot example)
Table 2: Entity groups in TinyRel-CM dataset. D,S,F, and U stand for Disease, Symptom, Food, and nUtrient,  respec-tively.
Table 5: Training task settings over shrunken training set.
+3

參考文獻

相關文件

(c) If the minimum energy required to ionize a hydrogen atom in the ground state is E, express the minimum momentum p of a photon for ionizing such a hydrogen atom in terms of E

Following the supply by the school of a copy of personal data in compliance with a data access request, the requestor is entitled to ask for correction of the personal data

what is the most sophisticated machine learning model for (my precious big) data. • myth: my big data work best with most

An instance associated with ≥ 2 labels e.g., a video shot includes several concepts Large-scale Data. SVM cannot handle large sets if using kernels

“Since our classification problem is essentially a multi-label task, during the prediction procedure, we assume that the number of labels for the unlabeled nodes is already known

The remaining positions contain //the rest of the original array elements //the rest of the original array elements.

Advantages of linear: easier feature engineering We expect that linear classification can be widely used in situations ranging from small-model to big-data classification. Chih-Jen

• If we want analysis with amortized costs to show that in the worst cast the average cost per operation is small, the total amortized cost of a sequence of operations must be