Top Banner
Cost-effective Interactive Attention Learning with Neural Attention Process Jay Heo 1 Junhyeon Park 1 Hyewon Jeong 1 Kwang joon Kim 2 Juho Lee 3 Eunho Yang 13 Sung Ju Hwang 13 Abstract We propose a novel interactive learning frame- work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated attentions, to correct the model’s behavior by updating the attention-generating network. However, such a model is prone to overfitting due to scarcity of human annotations, and requires costly retraining. Moreover, it is almost infeasible for the human annotators to examine attentions on tons of in- stances and features. We tackle these challenges by proposing a sample-efficient attention mecha- nism and a cost-effective reranking algorithm for instances and features. First, we propose Neural Attention Process (NAP), which is an attention generator that can update its behavior by incor- porating new attention-level supervisions with- out any retraining. Secondly, we propose an al- gorithm which prioritizes the instances and the features by their negative impacts, such that the model can yield large improvements with mini- mal human feedback. We validate IAL on vari- ous time-series datasets from multiple domains (healthcare, real-estate, and computer vision) on which it significantly outperforms baselines with conventional attention mechanisms, or without cost-effective reranking, with substantially less retraining and human-model interaction cost. 1. Introduction Deep neural networks are arguably the most prevalent tools for predictive modeling tasks nowadays, thanks to their ability to learn complex functions with multiple layers of non-linear transformations. However, the complex nature 1 Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea 2 Yonsei University College of Medicine, Seoul, South Korea 3 AITRICS, Seoul, South Korea. Correspon- dence to: Jay Heo <[email protected]>, Sung Ju Hwang <sjh- [email protected]>. Proceedings of the 37 th International Conference on Machine Learning, Vienna, Austria, PMLR 108, 2020. Copyright 2020 by the author(s). of the model, at the same time, makes it difficult to interpret what they have learned, which has led to the recent surge of interest in interpretable models that are capable of providing interpretations of the model and the prediction in human- understandable forms (Gilpin et al., 2018). Although recent works propose diverse solutions to inter- pretability (Choi et al., 2016; Ahmad et al., 2018; Lage et al., 2018), including attention mechanisms, activation visualization, and optimization for human-interpretability under human-in-the-loop, we face yet another challenge: not all machine-generated interpretations are correct or hu- man understandable. This is mainly due to two reasons: 1) correctness and reliability of a learning model heavily depends on the quantity and quality of the training data. 2) neural networks tend to learn non-robust features that help with predictions but are not human-perceptible (Ilyas et al., 2019). Such unreliability of the interpretations is highly problematic for safety-critical applications such as clinical risk predictions (Ahmad et al., 2018; Sankar et al., 2019) or autonomous driving (Chi & Mu, 2017). The main limitation of the existing models is that they mostly only consider passive roles for human supervisors, where they simply take the provided interpretations as is. Yet, a more effective way to use the interpretations is to use them as channels for human-model communications, such that the models learn by continuously interacting with the human supervisors, where they iteratively correct the model-generated interpretations. From a cognitive science perspective, human learning is done by internal reflection (back-propagation) and external explanation (human feed- back) during social interactions (Clark et al., 2015). Based on this motivation, we propose an interactive learning framework, where the model learns by iteratively interacting with the human supervisors who manipulate the model by adjusting the provided interpretations, which is depicted in Figure 1. The specific interpretation mechanism we consider in this work is the attention mechanism (Bahdanau et al., 2014). While active learning asks for supervision at the instance level, in our interactive learning model, it asks for supervision at the attention level. However, this leads to multiple challenges regarding efficiency, which hinders their applications to practical scenarios: Model retraining cost and overfitting: To reflect hu- arXiv:2006.05419v1 [cs.LG] 9 Jun 2020
10

Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Jul 04, 2020

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Cost-effective Interactive Attention Learning with Neural Attention Process

Jay Heo 1 Junhyeon Park 1 Hyewon Jeong 1 Kwang joon Kim 2 Juho Lee 3 Eunho Yang 1 3 Sung Ju Hwang 1 3

AbstractWe propose a novel interactive learning frame-work which we refer to as Interactive AttentionLearning (IAL), in which the human supervisorsinteractively manipulate the allocated attentions,to correct the model’s behavior by updating theattention-generating network. However, such amodel is prone to overfitting due to scarcity ofhuman annotations, and requires costly retraining.Moreover, it is almost infeasible for the humanannotators to examine attentions on tons of in-stances and features. We tackle these challengesby proposing a sample-efficient attention mecha-nism and a cost-effective reranking algorithm forinstances and features. First, we propose NeuralAttention Process (NAP), which is an attentiongenerator that can update its behavior by incor-porating new attention-level supervisions with-out any retraining. Secondly, we propose an al-gorithm which prioritizes the instances and thefeatures by their negative impacts, such that themodel can yield large improvements with mini-mal human feedback. We validate IAL on vari-ous time-series datasets from multiple domains(healthcare, real-estate, and computer vision) onwhich it significantly outperforms baselines withconventional attention mechanisms, or withoutcost-effective reranking, with substantially lessretraining and human-model interaction cost.

1. IntroductionDeep neural networks are arguably the most prevalent toolsfor predictive modeling tasks nowadays, thanks to theirability to learn complex functions with multiple layers ofnon-linear transformations. However, the complex nature

1Korea Advanced Institute of Science and Technology (KAIST),Daejeon, South Korea 2Yonsei University College of Medicine,Seoul, South Korea 3AITRICS, Seoul, South Korea. Correspon-dence to: Jay Heo <[email protected]>, Sung Ju Hwang <[email protected]>.

Proceedings of the 37 th International Conference on MachineLearning, Vienna, Austria, PMLR 108, 2020. Copyright 2020 bythe author(s).

of the model, at the same time, makes it difficult to interpretwhat they have learned, which has led to the recent surge ofinterest in interpretable models that are capable of providinginterpretations of the model and the prediction in human-understandable forms (Gilpin et al., 2018).

Although recent works propose diverse solutions to inter-pretability (Choi et al., 2016; Ahmad et al., 2018; Lageet al., 2018), including attention mechanisms, activationvisualization, and optimization for human-interpretabilityunder human-in-the-loop, we face yet another challenge:not all machine-generated interpretations are correct or hu-man understandable. This is mainly due to two reasons:1) correctness and reliability of a learning model heavilydepends on the quantity and quality of the training data. 2)neural networks tend to learn non-robust features that helpwith predictions but are not human-perceptible (Ilyas et al.,2019). Such unreliability of the interpretations is highlyproblematic for safety-critical applications such as clinicalrisk predictions (Ahmad et al., 2018; Sankar et al., 2019) orautonomous driving (Chi & Mu, 2017).

The main limitation of the existing models is that theymostly only consider passive roles for human supervisors,where they simply take the provided interpretations as is.Yet, a more effective way to use the interpretations is touse them as channels for human-model communications,such that the models learn by continuously interacting withthe human supervisors, where they iteratively correct themodel-generated interpretations. From a cognitive scienceperspective, human learning is done by internal reflection(back-propagation) and external explanation (human feed-back) during social interactions (Clark et al., 2015).

Based on this motivation, we propose an interactive learningframework, where the model learns by iteratively interactingwith the human supervisors who manipulate the model byadjusting the provided interpretations, which is depicted inFigure 1. The specific interpretation mechanism we considerin this work is the attention mechanism (Bahdanau et al.,2014). While active learning asks for supervision at theinstance level, in our interactive learning model, it asks forsupervision at the attention level. However, this leads tomultiple challenges regarding efficiency, which hinders theirapplications to practical scenarios:

• Model retraining cost and overfitting: To reflect hu-

arX

iv:2

006.

0541

9v1

[cs

.LG

] 9

Jun

202

0

Page 2: Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Cost-effective Interactive Attention Learning with Neural Attention Process

Challenge 1. Retraining Cost,Overfitting

Attentions

# of Instance

DimensionTime step

Annotation MaskHuman Supervisor

Challenge 2. Human Labeling Cost

Main Network

Time-series Data

𝒚"

𝒇𝜽Neural

Processes

𝒈𝝓

Attention Generator Instance-wise

Instance-level Reranking

1234

1234

Feature-wise

Feature-level Reranking

0

annotate

New Observations

-1 : “I don’t know”0 : ”Not attend”1 : “Attend”

000

00

1 1

1

111 1-1

-1-1

Influence / Uncertainty / Counterfactual

(A) Neural Attention Process (B) Cost-Effective Re-ranking (C) Human AnnotationFigure 1. Our Interactive Attention Learning (IAL) framework. IAL is an interactive learning framework which iteratively learns byinteracting with the human supervisor, via the learned attentions. It allows efficient model update using (A) Neural Attention Processwhich does not require retraining, and cost-effective interaction via (B) Cost-effective reranking of the instances and features.

man feedback, the model needs to be retrained, whichis costly. Moreover, retraining the model with scarcehuman feedback may result in the model overfitting.

• Expensive human supervision cost: Obtaining hu-man feedback on datasets with large numbers of train-ing instances and features is extremely costly. Further,obtaining feedback on already correct interpretationsis wasteful.

To tackle these practical challenges, we propose a novelinteractive learning framework, which we refer to as Inter-active Attention Learning (IAL), that allows both efficientmodel retraining and sample-efficient learning that mini-mizes human supervision cost. IAL consists of two maincomponents: 1) Neural Attention Processes (NAP) and 2)Cost-Effective instance and feature Reranking (CER).Basically, our model minimizes retraining cost via NAPwhich allows the model to correct its attention-generatingbehaviour in a sample-efficient manner by incorporatingnew labeled instances without retraining. NAP also pre-vents overfitting, which is inevitable with scarce humanfeedbacks when using a conventional attention mechanism.Secondly, to address the expensive human labeling cost,CER reranks the instances, features, and timesteps (fortime-series data) by their negative impacts. This enablesthe model to minimize human interaction cost, such that thehuman supervisors only correct the interpretations that arelikely to be incorrect and influential to the prediction. Theimportance of each sample and feature is measured eitherby the uncertainty, influence function (Cook & Weisberg,1980), or counterfactual estimation.

We validate our IAL framework on a variety of real worldtasks with time-series data, including cerebral infarction riskprediction from electronic health records (EHR), New YorkCity real-estate price forecast, and squat-posture predictiontask. The experimental results show that our model outper-forms baseline interactive learning schemes with significantmargins, with considerably smaller interaction cost in termsof both model retraining and human annotation cost. Ourcontributions are as follows:

• We propose a novel interactive learning framework

which iteratively updates the model by interacting withthe human supervisor via the generated attentions.

• To minimize the retraining cost, we propose a novelprobabilistic attention mechanism which sample-efficiently incorporates new attention-level supervi-sions on-the-fly without retraining and overfitting.

• To minimize human supervision cost, we propose anefficient instance and feature reranking algorithm,that prioritizes them based on their negative impactson the prediction, measured either by uncertainty, in-fluence function, or counterfactual estimation.

• We validate our model on five real-world datasetswith binary, multi-label classification, and regressiontasks, and show that our model obtains significant im-provements over baselines with substantially less re-training and human feedback cost.

2. Related workInterpretable machine learning The literature on inter-pretable machine learning is vast, but we only discuss afew. A popular approach to obtain interpretable model isto build a simple proxy model that mimics the (local) be-haviours of a complex model, using either simplified lin-ear models (Ribeiro et al., 2016) or decision trees (Sato& Tsukimoto, 2001; Salzberg, 1994). Another approach,specific for neural networks, is analyzing their learned rep-resentations (Sharif Razavian et al., 2014; Yosinski et al.,2014) at each unit via visualization. Bau et al. (2017) fur-ther consider interpretability of representations in light oftheir correspondence to semantic concepts, and utilize itfor controlling the behaviours of generative adversarial net-works (Bau et al., 2019). In this work, we propose a novelinteractive learning framework that leverages the model’sinterpretation to iteratively correct the model’s behaviour,while minimizing the interaction cost.

Attention Mechanism Attention mechanism (Bahdanauet al., 2014) is an effective approach to adaptively select asubset of features in an input-dependent manner, such that

Page 3: Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Cost-effective Interactive Attention Learning with Neural Attention Process

the model dynamically focuses on more relevant featuresfor prediction. This mechanism works by input-adaptivelygenerating coefficients for input features to allocate moreweights to more relevant features for prediction. Attentionmechanisms have achieved success with various applica-tions, including image translation (Xu et al., 2015), naturallanguage understanding (Bahdanau et al., 2015; Vaswaniet al., 2017), and visual question answering (Das et al.,2017). However, in the interactive learning setting, con-ventional attention mechanisms are either not trainable, orrequire retraining of the attention generator on the newlydelivered attention-level annotations, which may lead toperformance degeneration due to catastrophic forgetting. Inthis work, we incorporate benefits from the nonparametricand amortized inference of Neural Process (NPs) (Garneloet al., 2018) into an attention mechanism such that it gener-alizes well with scarce human labels in a semi-supervisedmanner and can incorporate new labeled instances withoutretraining via an approximation of stochastic process.

Active learning While there are vast literature on annota-tion methodology and active learning (Tong, 2001; Sener &Savarese, 2017), we here discuss a few relevant pre-existingworks for learning from rationales, which is a popular an-notation technique in natural language processing (Zaidan& Eisner, 2008) and vision (Donahue & Grauman, 2011),where a human highlights the important region of input.However, while these works directly zero out or modifyinput features, the attention generator in IAL provides itsinterpretation in the form of the attention, and the human su-pervisor corrects them. Furthermore, in conventional activelearning settings, annotators’ roles are relatively passive, asthey simply provide labels to each given instance such thatthey can’t see the effect of one’s annotation. However, theannotators in IAL actively interpret the generated attentions,directly modify the learning manifold of the model by mask-ing them, and can immediately see the effect of the newlyadded annotation.

3. Interactive Attention LearningSuppose we have a pre-trained neural network FΘ with aparameter Θ trained on a dataset Dtrain = (x(1:T )

i ,yi)Ni=1.x(1:T )

i = [x(1)

i , . . . ,x(T )

i ] is a time-series instance withx(t)

i ∈ RD, and yi ∈ RL is the corresponding label. Wedenote each labeled instance as ui = (x(1:T )

i ,yi). Θ istrained to minimize the empirical risk, the expectation ofindividual loss L(Θ,ui) over all training instances; we usemean-squared error for regressions or the categorical cross-entropy for classification problems. We further assume thatΘ consists of two sub-parameters (θ,φ), where θ corre-sponds to the parameter of the main neural network fθ andφ corresponds to the parameter of the attention-generatingnetwork gφ. gφ generates an attention α(1:T )

i for x(1:T )

i ,

Algorithm 1 Interactive Attention Learning FrameworkInput: Dtrain = x(1:T )

i ,yiNi=1, Θ = θ,φ, rounds S.Output: Θ.1: Pretrain Θ(0) = argminΘ L(Θ,Dtrain) + Ω(Θ).2: for s = 1, ..., S do3: D(s)

selection, α(1:T )

k Kk=1 = CER(Θ(s−1)).. Cost-Effective Re-ranking (CER)

4: m(1:T )

k Kk=1 = Evaluate(D(s)selection, α

(1:T )

k Kk=1). Get attention masks for α

5: φ(s) = NAP(D(s)selection, m

(1:T )

k Kk=1,φ(s−1))

. Learn human feedback with quick forwardpass using Neural Attention Process (NAP).

6: if s = 1 then7: Retrain Θ(1) = argminΘ L(Θ,Dtrain) + Ω(Θ) with an

adapted network containing NAP.8: end if9: end for

where each α(t)

i is separated into an attention for time-axisβ(1:T )

i and an attention for feature-axis γ(1:T )

i (see (6) fordetailed definition). The attentions are applied to the Dfeatures along T time-steps, and let the model focus on aspecific features of the representations of inputs relevant tothe prediction. Hence, the attention provides an interpreta-tion of the model’s decision.

Our goal in this paper is to correct the behaviour of theattention-generating network gφ with human supervision.This may be done by incrementally retraining gφ over multi-ple rounds, where for each round human supervisors inspectthe attentions generated by gφ and update φ. We assumethat a human supervisor provides an attention mask m(1:T )

i

for each sample x(1:T )

i as ground-truth label, after manuallyexamining the attention α(1:T )

i produced by gφ. An atten-tion mask for a certain axis is defined to be a ternary value−1, 0, 1, where −1 indicates "I don’t know", 0 indicates"Not attend", and 1 indicates "Attend". Note that a naïveretraining of gφ leads to the costly retraining of fθ via gra-dient back-propagation. Instead, we choose to fix θ andupdate φ only to minimize the cost of retraining. We referto this general framework that learns by interacting withthe human supervisor via learned attention, as InteractiveAttention Learning framework (IAL).

Yet, as discussed in the introduction, there are still remain-ing challenges that need to be tackled. First, the retrainingof gφ will still incur a non-negligible cost and may alsoresult in overfitting when human feedback is scarce. Totackle this, we propose a novel attention generator that canreadily incorporate human annotations without retraining.Another challenge is reducing the human interaction cost.Ideally, a human annotator may have a look on the entireattentions generated by gφ. This involves examining allinstances (ui, . . . ,uN ), and within each instance, all fea-tures over all time-steps (u

(1)i,1 , . . . ,u

(T )i,D). This is not fea-

sible and wasteful since many attention values are already

Page 4: Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Cost-effective Interactive Attention Learning with Neural Attention Process

𝒓𝒄

𝒎𝒄$𝟏

𝒎𝒄$𝟐

𝒎𝒄$𝒌

𝒓𝒄$𝟏

𝒓𝒄$𝟐

𝒓𝒄$𝒌

𝒛

𝜶Attention

Context Points

𝒈𝝓

𝒗𝒄 𝒚𝒄 𝒎𝒄

𝒛𝒍

New Observations

𝒗𝒄*𝟏 𝒚𝒄*𝟏 𝒎𝒄*𝟏

𝜶Attention

𝒈𝝓 𝒛𝒍

New Observations

𝒗𝒄(𝟐 𝒚𝒄(𝟐 𝒎𝒄(𝟐

Context Points

𝒗𝒄 𝒚𝒄 𝒎𝒄𝒗𝒄(𝟏 𝒚𝒄(𝟏 𝒎𝒄(𝟏

(a) Neural Attention Process (NAP) (b) First Round (s=1) (c) Further Rounds (s=2,3,..)Figure 2. (a): NAP naturally reflects the information from the annotation summarization z via amortization. (b) For new observations(annotation mask mc+1), NAP accepts them as input and generates the mean and variance parameter for z. (c) NAP doesn’t requireretraining for further new observations, in that NAP automatically adapt to them at the cost of a forward pass through a network gφ.

correct. To tackle this problem, we further propose a cost-effective reranking method which prioritizes the instancesand features by their impacts on the model’s prediction, tomaximize performance gains with minimal human effort.

Algorithm 1 describes the detailed algorithm for our IALframework that leverages the proposed attention mechanismand re-ranking method. In the next two subsections, wedescribe the two components that minimize both the modelretraining cost and human-model interaction cost.

3.1. Neural Attention Process

In this section, we describe Neural Attention Process (NAP),an novel attention generator based on NPs (Garnelo et al.,2018). NAP can effectively update the model without re-training by amortization using sparse human annotations.

Before describing our approach, we briefly explain howattention is applied for time-series prediction, using RE-TAIN (Choi et al., 2016) as our base model. Let v(1:T ) =Wembx

(1:T ) be a linear embedding of an input. We restrictv(1:T ) to have the same dimensionality (D) as x(1:T ), so thatwe can directly compute the contribution of a certain featureto a prediction1. The model computes attention coefficientsfor both time-steps and input-features as,

o(1:T ) = RNNβ(v(1:T )), (1)h(1:T ) = RNNγ(v(1:T )), (2)

e(t) = w>β o(t) + bβ for t = 1, . . . , T, (3)

q(t) = Wγh(t) + bγ for t = 1, . . . , T, (4)β(1:T ) = Softmax(e(1), . . . , e(T )), (5)γ(t) = tanh(q(t)) for t = 1, . . . , T. (6)

Here, β(1:T ) are attention weights applied for time-stepsand γ(1:T ) are attention weights for the input features. Wemay also consider the stochastic attention as in (Xu et al.,2015). Given α(1:T ) = β(1:T ),γ(1:T ), the model makes

1Please refer to the supplementary material to see how to com-pute the contribution of input features to predictions based onattentions and embedding v(1:T ). For now, treat each dimension ofv(1:T ) to be directly linked to the corresponding feature in x(1:T ).

predictions as y = h(∑Tt=1 β

(t) · (γ(t) v(t))) where isthe element-wise multiplication and h is an output layer.

Now we describe NAP, especially how it amortizes the pro-cedure of updating the model given human annotations. Letm(1:T )

k Kk=1 be a set of attention masks given by human an-notators for a subset Dselection = (x(1:T )

k ,yk)Kk=1 ⊆ Dtrainwith K N . Instead of exhaustively retraining gφ, NAPlearns to summarize Dselection to a latent vector, and givethe summarization as an additional input to the attentiongenerating network. This approach, when trained properly,can automatically adapt to new annotations without hav-ing to retrain the parameters. From below, we describe thecomponents of NAP in more detail.

Embedding & summarizing the annotations We firstfeed the input embedding v(1:T ) to LSTM (Hochreiter& Schmidhuber, 1997) (RNNβ ,RNNγ) to generate time-series representation l(1:T ) = [o(1:T ),h(1:T )]. Given attentionmasks m(1:T )

k Kk=1, we build an intermediate representationr(1:T )

k Kk=1 via another LSTM. Then, for each time step,we build a summarized representation r(t) by a permutation-invariant operation (for instance, average),

r(t) = r(t)

1 ⊕ · · · ⊕ r(t)

K . (7)

Having r(1:T ), we define a distribution for the summaryvariable z as Gaussian:

z(t) ∼ N (µ(r(t)),σ2(r(t))), (8)µ(r(t)) = Wµr(t) + bµ, (9)σ(r(t)) = softplus(Wσ r(t) + bσ). (10)

Generating attentions & Training NAP Now we gener-ate the attention by a similar procedure to (6), but insteadof feeding only l(1:T ) = (o(1:T ),h(1:T )), we feed both l(1:T )

and the annotation summarization vector z(1:T ) by concate-nation. This allows the network to naturally reflect theinformation obtained from z(1:T ) without having to retrainthe whole attention network parameter φ. The original NPis meta-trained using many training examples. Likewise,NAP requires a meta-training for adapting the attention

Page 5: Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Cost-effective Interactive Attention Learning with Neural Attention Process

generating network gφ to take z(1:T ) as an additional in-put (Figure 2, (b)). We found that this adaptation requiressignificantly fewer training examples than the typical NPtraining, possibly because the network is pretrained usingDtrain in advance. For such adaptation training, given aset of annotated examples, we randomly subsample anno-tations for each training step to comprise a random taskto meta-train the model. The subsampling prevents NAPfrom completely being over-fitted to the entire annotationset, leading to effective generalization to newly deliveredannotations across rounds. We also regularize z(1:T ) bypositing a standard Gaussian prior distribution as in Garneloet al. (2018). We train the parameters of NAP via stochasticgradient variational inference.

3.2. Cost-Effective instance and feature Reranking

As we discussed earlier, letting human annotators inspectattentions for all instances and features is inefficient evenfor a small dataset. We may reduce this cost by randomlysubsampling from all attention values, but it may result inselecting instances or features that are already correct orhave little impact to the model’s prediction. Thus, we wantto prioritize the attentions by their negative impact on themodel’s prediction, such that each feedback given by the hu-man supervisor results in large performance improvements.In this section, we propose a general framework, depictedin Figure 3, to select important instances and features. Forinstance-level selection, we use the influence score and un-certainty score. For feature-level, we use the influence score,uncertainty score, and counterfactual score.

3.2.1. INSTANCE-LEVEL RERANKING

Influence score We use the influence function (Koh &Liang, 2017) to approximate the impact of individual train-ing points on the model prediction. The idea behind thisis simple; given a validation point uval, how would thevalidation loss change if a certain training instance u isexcluded from training procedure? Formally, let Θ be theminimizer of empirical risk for the original training set,1N

∑Ni=1 L(Θ,ui), and Θ−u be simply the one computed

from empirical risk without u, 1N−1

∑ui 6=u L(Θ,ui). The

effect of removing u is then measured as L(Θ−u,uval)−

L(Θu,uval). Since exactly computing this involves N

retraining procedures and quite expensive, Koh & Liang(2017) propose to use the influence function I(u,uval) toapproximate it as follows:

L(Θ−u,uval)− L(Θ,uval) ≈ − 1

NI(u,uval), (11)

I(u,uval)def= −∇ΘL(uval, Θ)H−1

Θ∇ΘL(u, Θ), (12)

where HΘ = 1N

∑Ni=1∇2

ΘL(Θ,ui) is the Hessian. Tosummarize, the influence function I(u,uval) approximates

𝑷𝑲

… … …

TrainTrainValid Valid

Estimate 𝑰(𝒖&) / 𝐕𝐚𝐫(𝒖&)

Instance-level Reranking

Train

Feature-level Reranking

Select Re-rank & select

Estimate 𝑰(𝒖+ ,-

(.) ) / 𝐕𝐚𝒓(𝒖+,-(.) )/ 𝜓(𝒖+,-

(.) )

Feature… …

Feature

𝑭

Re-rank & select

Figure 3. Cost-Effective Re-ranking Procedure (CER).

the change in the validation loss (up to a constant) withouthaving to retrain the model.

During training, we are given a set of validation instancesDvalid = uval

j Mj=1. Then, we first select P instances thathave the highest validation loss L(Θ,uval

j ) to compriseD′valid = uval

p Pp=1. The intuition behind is that we wantto select the training instances having large impact on thevalidation instances that are mis-predicted by the currentmodel. In the supplementary file, we empirically show thatthis indeed improves the performance. Having D′valid, theinfluence score of a training instance ui is computed asI(ui) =

∑Pp=1 I(ui,u

valp ).

Uncertainty score While influence scores provide directmeasures of the negative impact of an instance, it is expen-sive because of the Hessian computation. An alternative,and less expensive approach to measure the negative im-pacts is using the uncertainty. We assume that instanceshaving high-predictive uncertainties are potential candidateto be corrected. This is a common approach in active learn-ing or Bayesian optimization literature, where the pointswith high-uncertainties are explored. Instance-level predic-tive uncertainty can simply be obtained by Monte-Carlo(MC) sampling (Gal & Ghahramani, 2016). We denote theinstance-level uncertainty score as Var(ui).

3.2.2. FEATURE-LEVEL RERANKING

Influence score We can also estimate the feature-levelinfluence score by a similar idea; if certain feature valueis modified, how would the validation loss change? Letu = (x(1:T ),y) be a training instance, and suppose we wantto compute the influence of u

(t)i,d, which is the d-th input

feature for timestep t, x(t)

d ∈ R. Define a perturbed data

point uδdef= (x(1:T ) + δet,d,y) where et,d is an one-hot

vector having d-th feature of t-th time step as one. LetΘuδ,−u be the empirical risk minimizer with u replaced byuδ . Then, as before, we have

L(Θuδ,−u,uval)− L(Θ,uval)

≈ − 1

N(I(uδ,u

val)− I(u,uval)). (13)

Based on this approximation, we sampled δ from mean± 2·std of features, and computed the average influencescore over multiple perturbations to rank features. As for

Page 6: Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Cost-effective Interactive Attention Learning with Neural Attention Process

Algorithm 2 Cost-Effective Re-rankingInput: Dtrain = uiNi=1, Dvalid = uval

j Mj=1, P , K, F , Θ(s−1).Output: D(s)

selection = ukKk=1, α(1:T )

k Kk=1.

1: Evaluate the loss for Dvalid.2: Sort uval

j Mj=1 in the descending order ofL(Θ(s−1),uvalj ) and

select top-P valid points D′valid.3: . Instance-level re-ranking4: for i = 1, ..., N do5: Compute the influence I(ui) or uncertainty score Var(ui).6: Select the top K-training points Dselection w.r.t the score.7: end for8: . Feature-level re-ranking9: for k = 1, . . . ,K do

10: for (t, d) = (1, 1), . . . , (T,D) do11: Compute influence I(u(t)

k,d) or uncertainty Var(u(t)

k,d) orcounterfactual ψ(u(t)

k,d) score.12: Select top-F features.13: end for14: end for

Patient Label : 0

Year : 2012 2009 2010 2011 2012

Patient ID : 19 Previous NextSave

0

0.2

0.4

0.6

0.8

1Attention

Female 53 23.708 136.007 86.003 287.012 313.013 14.398(Female/Male) (Years) (kg/m) (mmHg) (mmHg) (mg/dL) (mg/dL) (mg/dL)

Disease : Cardiovascular1 Previous Next

1. Sex 2. Age 3. BMI 4. SystolicBP

5. DiastolicBP

6. FastingGlucose

7. TotalCholesterol

8. Hemo-globin

𝑦# 𝑦/%('))

0.32

Feature importance:

7.Total Cholesterol

1. Sex 2. Age 3. BMI 4. SBP

5. DBP 6. FastingGlucose

7. Chole-sterol

8. Hemo-globin

Execute

Counterfactual Estimation

Figure 4. Attention annotation interface (risk prediction for Car-diovascular Disease (CVD)) with counterfactual estimation tool.

the instance-level influence score, we add up the influencescores for all selected validation samples. We denote I(u(t)

i,d)

the influence score obtained by perturbing u(t)i,d.

Uncertainty score NAP induces stochasticity to the atten-tions applied to the individual features, and this naturallyleads to feature-level uncertainty scores. As for the instance-level uncertainty score, we computed variances of attentionsapplied for each feature by MC sampling. We denote thefeature-level uncertainty score of u

(t)i,d as Var(u

(t)i,d).

Conterfactual score The last score, which we call ascounterfactual score, is the most direct measure of the nega-tive impact of a feature. It answers the following question:how would the prediction change if we ignore a certainfeature by manually turning off the corresponding attentionvalue? This does not require retraining since we can simplyset its attention value to zero, yet still effective because ourgoal is to rank the features w.r.t. their importance in atten-tion feedback. Recall that given an attention (β(1:T ),γ(1:T ))generated from gφ, a prediction is given as

yi = h

( T∑t=1

β(t)

i γ(t)

i v(t)

i

), (14)

where v(1:T )

i is the linear embedding of x(1:T ). The effect ofperturbing u

(t)i,d can be then computed as follows:

yi,−(t,d) = h

(∑t′ 6=t

β(t′)i γ(t′)

i v(t)

i + β(t)

i γ(t)

i,−d v(t)

i

)ψ(u(t)

i,d) = yi − yi,−(t,d), (15)

where γ(t)

i,−d is the attention where γ(t)i,d = 0. We empirically

found that the counterfactual score is the most effectivemeasure for feature-level reranking (See Table 2).

3.3. Human Annotation

Finally, given a subset selected using CER whose instancesand features also sorted by their negative impacts, we vi-sualize and present the attentions to human annotators, us-ing an online interactive user interface. We provide anexample of this interface in Figure 4 for the clinical riskprediction task. On the interface, the annotators set the at-tention mask for each feature to one of the following values:mk = −1 : I don’t know, 0 : Not attend, 1 : Attend. Theinterface visually emphasizes the features with high atten-tions using either a bar plot (for tabular data) or an attentionmap (for image data) depending on the given task. Then, theannotators examine attention weights to check whether theyare incorrectly allocated, and correct them when necessary.

4. Experiments4.1. Datasets and Baselines

1) Medical Check-ups These datasets are subsets of theelectronic health records (EHR) database of a major hospital,which consists of medical check-ups from 2009 to 2012 (4timesteps) for patients over the age of 15 in out-patientunits. We extracted 245, 000 patient records from the totalof 1.5 million records, each of which contains 34 variablesincluding general information (e.g., sex and height), vitalsigns (e.g., hemoglobin level), and risk-inducing behaviors(e.g., alcohol consumption). The task is to predict the onsetof the following disease in the next year: 1) Heart Failure,2) Cerebral Infarction, 3) Cardiovascular Disease (CVD).

2) Fitness - Squat Pose Correction This dataset contains4, 000 video frames of human subject performing squats,where the task is to predict whether the person is performingthe squat with the correct posture or with one of ten differenttypes of incorrect postures (e.g., 0: Correct posture, 1: Exag-gerated knees-forward movement, 2: Sitting on the thighs).Thus this is a multi-label classification task. We extract 14pairs of key points from joints (e.g., left shoulder or rightankle) over all frames, to clearly visualize which body jointsan attention generator attends to for each instance.

Page 7: Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Cost-effective Interactive Attention Learning with Neural Attention Process

EHR Fitness Real EstateHeart Failure Cerebral Infarction CVD Squat Forecasting

One-timeTraining

RETAIN 0.6069 ± 0.01 0.6394 ± 0.02 0.6018 ± 0.02 0.8425 ± 0.03 0.2136 ± 0.01Random-RETAIN 0.5952 ± 0.02 0.6256 ± 0.02 0.5885 ± 0.01 0.8221 ± 0.05 0.2140 ± 0.01

IF-RETAIN 0.6134 ± 0.03 0.6422 ± 0.02 0.5882 ± 0.02 0.8363 ± 0.03 0.2049 ± 0.01Random

Re-rankingRandom-UA 0.6231 ± 0.03 0.6491 ± 0.01 0.6112 ± 0.02 0.8521 ± 0.02 0.2222 ± 0.02

Random-NAP 0.6414 ± 0.01 0.6674 ± 0.02 0.6284 ± 0.01 0.8525 ± 0.01 0.2061 ± 0.01IAL

(Cost-effective)AILA 0.6363 ± 0.03 0.6602 ± 0.03 0.6193 ± 0.02 0.8425 ± 0.01 0.2119 ± 0.01

IAL-NAP 0.6612 ± 0.02 0.6892 ± 0.03 0.6371 ± 0.02 0.8689 ± 0.01 0.1835 ± 0.01Table 1. The binary & multi-class classification performance on the three electronic health records datasets and one fitness dataset. Thereported numbers are mean-AUROC for EHR and mean-Accuracy for squat. In the real estate forecasting task, the number indicatesmean-percentage error, meaning a lower error indicates better performance.

IAL-NAP Variants EHR Fitness Real EstateInstance-level Feature-level Heart Failure Cerebral Infarction CVD Squat Forecasting

Influence Function Uncertainty 0.6563 ± 0.01 0.6821 ± 0.02 0.6308 ± 0.02 0.8712 ± 0.01 0.1921 ± 0.01Influence Function Influence Function 0.6514 ± 0.02 0.6825 ± 0.01 0.6329 ± 0.03 0.8632 ± 0.01 0.1865 ± 0.02Influence Function Counterfactual 0.6592 ± 0.02 0.6921 ± 0.03 0.6379 ± 0.02 0.8682 ± 0.01 0.1863 ± 0.02

Uncertainty Counterfactual 0.6612 ± 0.01 0.6892 ± 0.03 0.6371 ± 0.02 0.8689 ± 0.02 0.1835 ± 0.02

Table 2. Results of Ablation study with proposed IAL-NAP combinations for instance- and feature-level reranking on all tasks.

3) Real Estate Sales Transactions This datasets is a sub-set of public rolling sales transaction database (Zhu &Sobolevsky, 2018) from New York City Department of Fi-nance that is publicly available, which consists of 70, 700house records with 27, 000 sales transaction records over 10years from 2010 to 2019 (10 time-steps). The subset used forexperiments includes 3, 100 housing transactions, each ofwhich includes 47 variables that describes the property (e.g.number of rooms), neighborhood (e.g. minimum distance toa supermarket), and macro-economy indicators (e.g., mort-gage rate). The task is to make an one-year forecast for theprice of a given residential property.

Baselines and our models1) RETAIN: This is the attentional recurrent neural networkmodel (RETAIN) proposed in (Choi et al., 2016).2) Random-RETAIN: RETAIN, which is newly trainedfrom a training set without K randomly selected samples.3) IF-RETAIN: RETAIN that is newly trained from thetraining set without the top K-negative points, which areobtained using the influence function (Koh & Liang, 2017).4) Random-UA: This is the Uncertainty-Aware attentionalnetwork (UA) (Heo et al., 2018) which is trained using IALwith random instance and feature selection.5) Random-NAP: Our IAL framework with Neural Atten-tion Process model (NAP), which is trained using randominstance and feature selection.6) Cost-effective AILA: This is a modified version of theinteractive attention learning model proposed by (Choi et al.,2019) which retrains the attention generator by using a bi-nary cross entropy loss function between the attention vectorαk and the attention annotation mk. We train the modelwith CER to verify the effectiveness of the NAP.7) IAL-NAP Our IAL framework with Neural AttentionProcess (NAP) and cost-effective instance and feature

Reranking (CER), which uses uncertainty for instance-wisereranking and counterfactual score for feature reranking.

Experimental setup For all datasets, we generatetrain/valid/test splits with the ratio of 70%:10%:20%. ForRandom-UA and AILA model, we use `2-regularization‖φ(s) − φ(s−1)‖22 to prevent overfitting. Please see sup-plementary file for more details of the datasets, networkconfigurations, and hyperparameters. We will also publiclyrelease the codes and all datasets used in the experiments.

4.2. Experimental results

We first examine the prediction performance of the base-lines and our models. Table 1 shows the results, where theperformance is measured with Area Under the ROC curve(AUROC) on the risk prediction tasks, accuracy on squatposture task with multi-labels, and mean percentage erroron real estate price forecasts. Note that IF-RETAIN, whichuses influence functions to remove instances with negativeinfluence scores, performs relatively better on most tasksthan other RETAIN baselines, but fails to improve on CVDand squat posture task. We observe that Random-UA, whichis retrained with human attention-level supervision on ran-domly selected samples, performs worse than Random-NAPon all tasks. This is due to overfitting to few supervised la-bels, while NAP does not suffer from overfitting. IAL-NAPsignificantly outperforms Random-NAP on all tasks, whichshows that the effect of attention annotation cannot havemuch effect on the model when the instances are randomlyselected. AILA with cost-effective reranking also performsworse than IAL-NAP, due to severe overfitting even withregularizations to prevent it. We further perform an abla-tion study of cost-effective reranking with different scoringmeasures in table 2. The results show that for instance-levelscoring, influence and uncertainty scores work similarly,

Page 8: Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Cost-effective Interactive Attention Learning with Neural Attention Process

1 2 3 4Round (s)

0

10

20

30

40

50

60Tim

e (se

c)Random-UARandom-NAP

AILAIAL-NAP

1 2 3 4Round (s)

0102030405060708090

Time (

sec)

Random-UARandom-NAP

AILAIAL-NAP

1 2 3 4Round (s)

0153045607590

105120

Time (

sec)

Random-UARandom-NAP

AILAIAL-NAP

1 2 3 4Round (s)

0

10

20

30

40

50

60

Time (

sec)

Random-UARandom-NAP

AILAIAL-NAP

1 2 3 4Round (s)

120

150

180

210

240

270

300

Time (

sec)

Random-UARandom-NAP

AILAIAL-NAP

1 2 3 4Round (s)

50

100

150

200

250

Time (

sec)

Random-NAPIAL-NAP

1 2 3 4Round (s)

50

100

150

200

250

Time (

sec)

Random-NAPIAL-NAP

1 2 3 4Round (s)

50

100

150

200

250

Time (

sec)

Random-NAPIAL-NAP

1 2 3 4Round (s)

30

60

90

120

150

180

Time (

sec)

Random-NAPIAL-NAP

1 2 3 4Round (s)

0

100

200

300

400

500

Time (

sec)

Random-NAPIAL-NAP

(a) Heart Failure (b) Cerebral Infarction (c) CVD (d) Squat (e) Real EstateFigure 5. (top) Retraining Time to retrain examples of human annotation on all task for Random-UA, AILA, Random-NAP, and IAL-NAP.(bottom) mean Response Time (mean-RT) of human labeling on three risk prediction task, one squat posture classification task, and onerealestate forecasting task (IAL-NAP with features ranked by uncertainty vs Random-NAP with features ranked randomly).

Age Smoking SysBP HDL LDL2009 31 Yes 139 54 972010 32 Yes 134 55 97Current State 33 yrs Yes 141 mmHg 55 mg/dL 102 mg/dL

Age Smoking(Amount)

SystolicBP

HDL LDL

Feature

0

1

2

3

4

5

Atte

ntion

False PositiveFalse NegativeTrue Positive

Age Smoking(Amount)

SystolicBP

HDL LDL

Feature

0

1

2

3

4

5

Atte

ntion

False PositiveFalse NegativeTrue Positive

Age Smoking(Amount)

SystolicBP

HDL LDL

Feature

0

1

2

3

4

5At

tent

ionFalse PositiveFalse NegativeTrue Positive

(a) Pretrained (b) s=1 (c) s=2Figure 6. Visualization of attention for a selected patient on Cardiovascular Disease(CVD) prediction task. Contribution indicates the extent to which each individualfeature affects the onset of CVD in 1 year. Age - Age, Smoking - Whether currentlysmokes a cigarette, SysBP - Systolic blood pressure, HDL - High-density lipoproteinscholesterol, LDL - Low-density lipoprotein cholesterol. Bars correspond to attentions.

(a) Heart Failure (b) Cerebral Infarction

(c) CVD (d) Squat

Figure 7. Change of accuracy with 100 annotationsacross four rounds (S) between IAL-NAP (blue) vsRandom-NAP (red).

while the counterfactual score was the most effective forfeature-wise reranking. However, considering the compu-tation cost, the combination of uncertainty-counterfactualis the most cost-effective solution since it avoids expensivecomputation of the Hessians.

Effect of Neural Attention Process Line plots in Fig-ure 5 (top) shows averaged time to retrain examples over therounds of interactions with Random-UA, AILA, Random-NAP, and IAL-NAP on the five tasks. IAL-NAP andRandom-NAP shows shorter retraining time, while Random-UA and AILA which fine-tune the attention-generating net-work take a longer time to retrain. This shows anotherbenefit of our neural attention process, which is its ability toperform amortized inference. A more responsive system canalso improve the quality of the interaction, in the interactivelearning setting.

Effect of Cost-Effective Re-ranking We further measurethe average response time of the annotators with and with-out cost-effective reranking. Figure 5 (bottom) shows thatannotators spend less time with annotation if variables areprioritized by their negative impacts measured using un-certainty (blue bars) compared to presenting them in theoriginal order (grey bars), on all tasks. Figure 7 shows

the change in model accuracy over training rounds withand without cost-effective reranking, where the negativeimpacts are measured by the influence score. On the riskprediction and squat posture tasks, the accuracy of IAL-NAPincreases over the 4 rounds of interaction, while Random-NAP achieves only marginal increases. Especially, on theheart failure task (a), the line plot shows that IAL-NAP usesa smaller number of annotated examples (100 examples)than Random-NAP (400 examples) to improve the modelwith comparable accuracy (auc: 0.6414), which shows thatIAL-NAP improves the model with fewer examples.

Qualitative analysis We further analyze the contributionof each feature for a CVD patient (label=1) whose recordsshowed significant changes in attention with the help ofphysicians in Figure 6. The table (top in Figure 6) showsthe patient’s medical records at the previous (2009, 2010)and the current time-step (2011), yearly registered records.The three graphs shows the values of the allocated atten-tions across three rounds. Our model, IAL-NAP failed topredict the label at pretrained round (a), but makes a correctprediction at s=2 (c). We visualized five variables that haveclinically meaningful changes. Across the change of atten-tions from (a) to (c), the physicians consider that attentionson age, HDL, and LDL in (a) are false positive (red bars)

Page 9: Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Cost-effective Interactive Attention Learning with Neural Attention Process

and smoking as false negative (blue bars), except SysBP astrue positive (grey bars). Noting that the patient’s age (30) isyounger than the median age (50 years-old) of female CVDpatient (Garcia et al., 2016), initial IAL-NAP (a) allocatedtoo much weights on age, which led to an overconfidentattention model and in turn resulted in the incorrect predic-tion. However, our model gradually allocated less weightson age over rounds, as it started to learn what to attend tofrom interactive attention learning. Note that attention onsmoking highly increased at s=2 (c), which is also clini-cally guided by a physician for the reason that CVD riskincreases by 25% for women who smoke cigarettes (Huxley& Woodward, 2011). Previous incorrect attentions on HDLand LDL (a) decrease over rounds, since the HDL level (55mg/dL) is in the normal range (40-60) and the level of LDL(102 mg/dL) is still lower than borderline high (130-159).

5. ConclusionWe proposed an interactive learning framework which itera-tively learns by interacting with the human supervisors viathe generated attentions. The framework utilizes a novelstochastic attention mechanism based on neural processthat can correct the model’s interpretation from scarce hu-man feedback without retraining or overfitting. Further, ituses cost-effective reranking of the instances and featuresby their negative impacts to maximize the effect of eachhuman-machine interaction. We validated our model onfive real-world tasks from the healthcare, real estate, and fit-ness domains, on which our model significantly outperformsbaselines with smaller retraining and human annotation cost.Qualitative analysis of our model shows that it generatesmore human-interpretable attentions that is crucial for itsreliability on safety-critical tasks.

AcknowledgementsThis work was supported by Institute for Information &communications Technology Planning & Evaluation (IITP)grant funded by the Korea government (MSIT) (No.2017-0-01779, A machine learning and statistical inference frame-work for explainable artificial intelligence).

ReferencesAhmad, M. A., Eckert, C., and Teredesai, A. Interpretable

machine learning in healthcare. In Proceedings of the2018 ACM International Conference on Bioinformatics,Computational Biology, and Health Informatics, pp. 559–560. ACM, 2018.

Bahdanau, D., Cho, K., and Bengio, Y. Neural machinetranslation by jointly learning to align and translate. arXivpreprint arXiv:1409.0473, 2014.

Bahdanau, D., Cho, K., and Bengio, Y. Neural machine

translation by jointly learning to align and translate. ICLR,2015.

Bau, D., Zhou, B., Khosla, A., Oliva, A., and Torralba,A. Network dissection: Quantifying interpretability ofdeep visual representations. In Proceedings of the IEEEConference on Computer Vision and Pattern Recognition,pp. 6541–6549, 2017.

Bau, D., Zhu, J.-Y., Strobelt, H., Zhou, B., Tenenbaum,J. B., Freeman, W. T., and Torralba, A. Visualizing andunderstanding generative adversarial networks. arXivpreprint arXiv:1901.09887, 2019.

Chi, L. and Mu, Y. Deep steering: Learning end-to-end driv-ing model from spatial and temporal visual cues. arXivpreprint arXiv:1708.03798, 2017.

Choi, E., Bahadori, M. T., Sun, J., Kulas, J., Schuetz, A., andStewart, W. Retain: An interpretable predictive modelfor healthcare using reverse time attention mechanism. InAdvances in Neural Information Processing Systems, pp.3504–3512, 2016.

Choi, M., Park, C., Yang, S., Kim, Y., Choo, J., and Hong,S. R. Aila: Attentive interactive labeling assistant for doc-ument classification through attention-based deep neuralnetworks. In Proceedings of the 2019 CHI Conferenceon Human Factors in Computing Systems, pp. 230. ACM,2019.

Clark, Ian, and Dumas, G. Toward a neural basis for peer-interaction: what makes peer-learning tick? Frontiers inpsychology, 2015.

Cook, R. D. and Weisberg, S. Characterizations of an em-pirical influence function for detecting influential casesin regression. Technometrics, 22(4):495–508, 1980.

Das, A., Agrawal, H., Zitnick, L., Parikh, D., and Batra,D. Human attention in visual question answering: Dohumans and deep networks look at the same regions?Computer Vision and Image Understanding, 163:90–100,2017.

Donahue, J. and Grauman, K. Annotator rationales forvisual recognition. In 2011 International Conference onComputer Vision, pp. 1395–1402. IEEE, 2011.

Gal, Y. and Ghahramani, Z. Dropout as a bayesian approxi-mation: Representing model uncertainty in deep learning.In ICML, 2016.

Garcia, M., Mulvagh, S. L., Bairey Merz, C. N., Buring,J. E., and Manson, J. E. Cardiovascular disease in women:clinical perspectives. Circulation research, 118(8):1273–1293, 2016.

Page 10: Cost-effective Interactive Attention Learning with …work which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated

Cost-effective Interactive Attention Learning with Neural Attention Process

Garnelo, M., Schwarz, J., Rosenbaum, D., Viola, F.,Rezende, D. J., Eslami, S. M. A., and Teh, Y. W. Neu-ral processes. CoRR, abs/1807.01622, 2018. URLhttp://arxiv.org/abs/1807.01622.

Gilpin, L. H., Bau, D., Yuan, B. Z., Bajwa, A., Specter, M.,and Kagal, L. Explaining explanations: An overview ofinterpretability of machine learning. In 2018 IEEE 5thInternational Conference on Data Science and AdvancedAnalytics (DSAA), pp. 80–89. IEEE, 2018.

Heo, J., Lee, H. B., Kim, S., Lee, J., Kim, K. J., Yang, E.,and Hwang, S. J. Uncertainty-aware attention for reliableinterpretation and prediction. In Advances in NeuralInformation Processing Systems, pp. 909–918, 2018.

Hochreiter, S. and Schmidhuber, J. Long short term memory.Neural Computation, 9:1735–1780, 1997.

Huxley, R. R. and Woodward, M. Cigarette smoking asa risk factor for coronary heart disease in women com-pared with men: a systematic review and meta-analysisof prospective cohort studies. The Lancet, 378(9799):1297–1305, 2011.

Ilyas, A., Santurkar, S., Tsipras, D., Engstrom, L., Tran, B.,and Madry, A. Adversarial examples are not bugs, theyare features. arXiv preprint arXiv:1905.02175, 2019.

Koh, P. W. and Liang, P. Understanding black-box predic-tions via influence functions. In Proceedings of the 34thInternational Conference on Machine Learning-Volume70, pp. 1885–1894. JMLR. org, 2017.

Lage, I., Ross, A., Gershman, S. J., Kim, B., and Doshi-Velez, F. Human-in-the-loop interpretability prior. InAdvances in Neural Information Processing Systems, pp.10159–10168, 2018.

Ribeiro, M. T., Singh, S., and Guestrin, C. "why should itrust you?": Explaining the predictions of any classifier.In Proceedings of the 22Nd ACM SIGKDD InternationalConference on Knowledge Discovery and Data Mining,KDD ’16, pp. 1135–1144, New York, NY, USA, 2016.ACM. ISBN 978-1-4503-4232-2. doi: 10.1145/2939672.2939778. URL http://doi.acm.org/10.1145/2939672.2939778.

Salzberg, S. L. C4. 5: Programs for machine learning byj. ross quinlan. morgan kaufmann publishers, inc., 1993.Machine Learning, 16(3):235–240, 1994.

Sankar, V., Kumar, D., Clausi, D. A., Taylor, G. W.,and Wong, A. Sisc: End-to-end interpretable dis-covery radiomics-driven lung cancer prediction viastacked interpretable sequencing cells. arXiv preprintarXiv:1901.04641, 2019.

Sato, M. and Tsukimoto, H. Rule extraction from neuralnetworks via decision tree induction. In IJCNN’01. Inter-national Joint Conference on Neural Networks. Proceed-ings (Cat. No. 01CH37222), volume 3, pp. 1870–1875.IEEE, 2001.

Sener, O. and Savarese, S. Active learning for convolutionalneural networks: A core-set approach. arXiv preprintarXiv:1708.00489, 2017.

Sharif Razavian, A., Azizpour, H., Sullivan, J., and Carlsson,S. Cnn features off-the-shelf: an astounding baseline forrecognition. In Proceedings of the IEEE conference oncomputer vision and pattern recognition workshops, pp.806–813, 2014.

Tong, S. Active learning: theory and applications, volume 1.Stanford University USA, 2001.

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones,L., Gomez, A. N., Kaiser, Ł., and Polosukhin, I. Atten-tion is all you need. In Advances in neural informationprocessing systems, pp. 5998–6008, 2017.

Xu, K., Ba, J. L., Kiros, R., Cho, K., Courville, A., Salakhut-dinov, R., Zemel, R. S., and Bengio, Y. Show, attend andtell: Neural image caption generation with visual atten-tion. In ICML, 2015.

Yosinski, J., Clune, J., Bengio, Y., and Lipson, H. Howtransferable are features in deep neural networks? InAdvances in neural information processing systems, pp.3320–3328, 2014.

Zaidan, O. and Eisner, J. Modeling annotators: A generativeapproach to learning from annotator rationales. In Pro-ceedings of the 2008 conference on Empirical methods innatural language processing, pp. 31–40, 2008.

Zhu, E. and Sobolevsky, S. House price modeling withdigital census. arXiv preprint arXiv:1809.03834, 2018.URL https://www1.nyc.gov/site/finance/taxes/property-rolling-sales-data.page.