Top Banner
Text Mining An inter-disciplinary research area focusing on the process of deriving knowledge from texts Exploit techniques in linguistics & NLP, statistics, machine learning and information retrieval to achieve its goal: from texts to knowledge
56

Text Mining

Jan 06, 2016

Download

Documents

Keegan

Text Mining. An inter-disciplinary research area focusing on the process of deriving knowledge from texts Exploit techniques in linguistics & NLP, statistics, machine learning and information retrieval to achieve its goal: from texts to knowledge. Typical tasks in text mining. - PowerPoint PPT Presentation
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Text Mining

Text Mining

An inter-disciplinary research area focusing on the process of deriving knowledge from texts

Exploit techniques in linguistics & NLP, statistics, machine learning and information retrieval to achieve its goal: from texts to knowledge

Page 2: Text Mining

Typical tasks in text mining

Text classification and clustering Concept extraction Named entity extraction Semantic relation learning Text summarization Sentiment analysis (facebook/twitter posts

analysis) …

Page 3: Text Mining

Techniques for Text Mining

Information retrieval + Natural language processing Word frequency distribution, morphological analysis Parts-of-Speech tagging and annotation Parsing, semantic analysis

Statistical methods Machine learning/data mining methods

Supervised text classification Unsupervised text clustering Association/linkage analysis Visualization techniques

Page 4: Text Mining

Machine Learning Techniques forAutomatic Ontology Extraction from Domain

Texts

Janardhana R. Punuru Jianhua Chen

Computer Science Dept.

Louisiana State University, USA

Page 5: Text Mining

Presentation Outline

Introduction Concept extraction Taxonomical relation learning Non-taxonomical relation learning Conclusions and Future Works

Page 6: Text Mining

Introduction Ontology

An ontology OL of a domain D is a specification of a conceptualisation of D, or simply, a data model describing D. An OL typically consists of: A list of concepts important for domain D A list of attributes describing the concepts A list of taxonomical (hierarchical) relationships

among these concepts A list of (non-hierarchical) semantical

relationships among these concepts

Page 7: Text Mining

Sample (partial) Ontology – Electronic Voting Domain

Concepts: person, voter, worker, poll watcher, location, county, precinct, vote, ballot, machine, voting machine, manufacturer, etc.

Attributes: name of person, model of machine, etc. Taxonomical relations:

Voter is a person; precinct is a location; voting machine is a machine, etc.

Non-hierarchical relations: Voter cast ballot; voter trust machine; county adopt

machine; equipment miscount ballot, etc.

Page 8: Text Mining

Sample (partial) Ontology – Electronic Voting Domain

Page 9: Text Mining

Applications of Ontologies

Knowledge representation and knowledge management systems

Intelligent query-answering systems Information retrieval and extraction Semantic Web

Web pages annotated with ontologies User queries for Web pages analysed at knowledge

level and answered by inferencing on ontological knowledge

Page 10: Text Mining

Task: automatic ontology extraction from domain texts

Ontology extraction

textsontology

Page 11: Text Mining

Challenges in Text Processing

Unstructured texts Ambiguity in English text

Multiple senses of a word Multiple parts of speech – e.g., “like” can occur in 8 PoS:

Verb: “Fruit flies like banana” Noun: “We may not see its like again” Adjective: “People of like tastes agree” Adverb: “The rate is more like 12 percent” Preposition: “Time flies like an arrow” etc

Lack of closed domain of lexical categories Noisy texts Requirement of very large training text sets Lack of standards in text processing

Page 12: Text Mining

Challenges in Knowledge Acquisition from Texts

Lack of standards in knowledge representation Lack of fully automatic techniques for KA Lack of techniques for coverage of whole texts Existing techniques typically consider word

frequencies, co-occurrence statistics, syntactic patterns, and ignore other useful information from the texts

Full-fledged natural language understanding is still computationally infeasible for large text collections

Page 13: Text Mining

Our Approach

Page 14: Text Mining

Our Approach

Page 15: Text Mining

Concept Extraction: Existing Methods

Frequency-based methods Text-to-Onto [Maedche & Volz 2001]

Use syntactic patterns and extract concepts matching the patterns [Paice, Jones 1993]

Use WordNet [Gelfand et. Al. 2004] start from a base word list,

for each w in the list, add the hypernyms and hyponyms in WordNet to the list

Page 16: Text Mining

Concept Extraction: Our Approach

Parts of Speech tagging and NP chunking Morphological processing – word stemming,

converting words to root form stopword removal Focus on top % freq. NP Focus on NP with fewer number of WordNet

senses

Page 17: Text Mining

Concept Extraction: WordNet Sense Count Approach

Page 18: Text Mining

Background: WordNet

General lexical knowledge base Contains ~ 150,000 words (noun, verb, adj, adv) A word can have multiple senses: “plant” as a noun has 4

senses Each concept (under each sense and PoS) is represented

by a set of synonyms (a syn-set). Semantic relations such as hypernym/antonym/meronym

of a syn-set are represented WordNet - Princeton University Cognitive Science Labo

ratory

Page 19: Text Mining

Background: Electronic Voting Domain

15 documents from New York Times (www.nytimes.com)

Contains more than 10,000 words Pre-processing produced 768 distinct noun phrases

(concepts) 329 relevant to electronic voting 439 irrelevant

Page 20: Text Mining

Background: Text Processing

Many local election officials and voting machine companies are fighting paper trails, in partbecause they will create more work and will raise difficult questions if the paper and electronictallies do not match.

● POS Tagging: Many/JJ local/JJ election/NN officials/NNS and/CC voting/NN machine/NN companies/NNS are/VBP fighting/VBG paper/NN trails,/NN in/IN part/NN because/IN they/PRP will/MD create/VB more/JJR work/NN and/CC will/MD raise/VB difficult/JJ questions/NNS if/IN the/DT paper/NN and/CC electronic/JJ tallies/NNS do/VBP not/RB match./JJ

● NP Chuking: [ Many/JJ local/JJ election/NN officials/NNS ] and/CC [ voting/NN machine/NN companies/NNS ] are/VBP fighting/VBG [ paper/NN trails,/NN ] in/IN [ part/NN ] because/IN [ they/PRP ] will/MD create/VB [ more/JJR work/NN ] and/CC will/MD raise/VB [ difficult/JJ questions/NNS ] if/IN [ the/DT paper/NN ] and/CC [ electronic/JJ tallies/NNS ] do/VBP not/RB [ match./JJ]

● Stopword Elimination: local/JJ election/NN officials/NNS, voting/NN machine/NN companies/NNS , paper/NN trails,/NN, part/NN, work/NN, difficult/JJ questions/NNS, paper/NN, electronic/JJ tallies/NNS, match./JJ

● Morphological Analysis: local election official, voting machine company, paper trail, part, work, difficult question, paper, electronic tally

Page 21: Text Mining

WNSCA + {PE, POP}

Take top n% of NP, and select only those with less than 4 senses in WordNet ==> obtain T, a set of noun phrases

Make a base list L of words from T PE: add to T, any noun phrase np from NP, if the head-

word (ending word) in np is in L POP: add to T, any noun phrase np from NP, if some

word in np is in L

Page 22: Text Mining

Evaluation: Precision and Recall

S T

| |

| |

S T

S

| |

| |

S T

T

Precision: n

Recall:

Page 23: Text Mining

Evaluations on the E-voting Domain

0

10

20

30

40

50

60

70

80

90

100

Top10%

Top20%

Top50%

Top75%

frequency threshold

pre

cisi

on Raw Freq

WNSCA

W +PE

W + POP

Page 24: Text Mining

Evaluations on the E-voting Domain

0

10

20

30

40

50

60

70

80

90

Top10%

Top25%

Top50%

Top75%

frequency threshold

reca

ll

Raw Freq

WNSCA

W +PE

W + POP

Page 25: Text Mining

TF*IDF Measure

TF*IDF: Term Frequency Inverted Document Frequency

|D|: total number of documents

|Di|: total number of documents containing term ti

TF*IDF(tij): TF*IDF measure for term ti in document dj

fij: frequency of term ti in document dj

TF IDF t f LogD

Dij iji

* ( )| |

| |*

Page 26: Text Mining

Comparison with the tf.idf method

Retrieved R & Rel Precision Recall F-measure0

50

100

150

200

250

300

350

tf.idf

WNSCA

W+PE

W+POP

Page 27: Text Mining

Evaluations on the TNM Domain

TNM Corpus: 270 texts in the TIPSTER Vol. 1 data from NIST: 3 years (87, 88, 89) news articles from Wall Street Journal, in the category of “Tender offers, Mergers and Acquisitions”

30 MB in size 183, 348 concepts extracted - only used the top 10%

frequent ones in the experiments - manually label the 18,334 concepts: only 3,388 concepts are relevant

Use the top 1% frequent concepts as the initial cut

Page 28: Text Mining

Evaluations on the TNM Domain

0

10

20

30

40

50

60

70

80

90

Pre. Recall f-measure

tf*idf

WNSCA

W+10%PE

W+10%POP

Page 29: Text Mining

Taxonomy Extraction: Existing Methods

A taxonomy: an “is-A” hierarchy on concepts Existing approaches:

Hierarchical clustering: Text-To-Onto but this needs users to manually label the internal nodes Use lexico-syntactic patterns: [Hearst 1992, Iwanska

1999] “musical instruments, such as piano and violin … “ Use seed concepts and semantic variants: [Morin &

Jacqumin 2003] “An apple is a fruit” “Apple juice is fruit juice”

Page 30: Text Mining

Taxonomy Extraction: Our Method

3 techniques for taxonomy extraction Compound term heuristic: “voting machine” is a

machine WordNet-based method – needs word sense

disambiguation (WSD) Supervised learning (Naive-Bayes) for semantic

class labeling (SCL) of concepts

Page 31: Text Mining
Page 32: Text Mining

Semantic Class Labeling of Concepts

Given: semantic classes T ={T1, ..., T

k } and

concepts C = { C1, ..., C

n}

Find: a labeling L: C --> T, namely, L(c) identifies the semantic class of concept c for each c in C.

For example, C = {voter, poll worker, voting machine} and T = {person, location, artifacts}

Page 33: Text Mining

SCL

Page 34: Text Mining

Naïve Bayes Learning for SCL

Four attributes are used to describe any concept 1. The last 2 characters of the concept

2. The head word of the concept

3. The pronoun following the concept

4. The preposition proceeding the concept

Page 35: Text Mining

Naïve Bayes Learning for SCL

Naïve Bayes Classifier:

Given an instance x = <a1, ..., an>, and

a set of classes Y = {y1, ..., yk}

NB(x) =

arg m ax P r( ) P r( | )y Y

j

j

n

y a y

1

Page 36: Text Mining

Evaluations

On E-voting domain:

622 instances, 6-fold cross-validation: 93.6% prediction accuracy

Larger experiment: from WordNet 2326 in the person category 447 in the artifacts category 196 in the location category 223 in the action category

2624 instances from the Reuters data, 6-fold cross-val.

produced 91.0% accuracy

Reuters data: 21578 Reuters news wire articles in 1987

Page 37: Text Mining

Attribute Analysis for SCL

Page 38: Text Mining

Non-taxonomical relation learning

We focus on learning non-hierarchical relations of form <Ci, R, Cj>

Here R is a non-hierarchical relation, and Ci, C

j are

concepts Example relations: < voter, cast, ballot>

<official, tell, voter>

<machine, record, ballot>

Page 39: Text Mining

Related Works

Non-hierarchical relation learning is relatively less tackled

Several works on this problem make restrictive assumptions: Define a fixed set of concepts, then look for relations

among these concepts Define a fixed set of non-hierarchical relations, then

look for concept pairs satisfying these relations Syntactical structure of the form (subject, verb, object)

is often used

Page 40: Text Mining

Ciaramita et al(2005): Use a pre-defined set of relations Extract concept pairs satisfying such a relation Use chi-square test to verify the statistical significance Experimented with the Molecular Biology domain texts

Schutz and Buitelaar (2004): Also use a pre-defined set of relations Build triples from concept pairs and relations Experimented with the football domain texts

Page 41: Text Mining

Kavalec et al(2004) No pre-defined set of relations Use the following AE measure to estimate the

strength of the triple:

Experimented with the tourism domain texts We have also implemented the AE measure for the

purpose of performance comparisons

AE C C VP C C V

P C V P C V(( ) | )

(( ) | )

( | ) ( | )( )1 2

1 2

1 21

Page 42: Text Mining

Our Method

The the framework of our method

Page 43: Text Mining

Extracting concepts and concept pairs

Domain concepts C are extracted using WNSCA + PE/POP

Concept pairs are obtained in two ways: RCL: Consider pairs (Ci, Cj), both from C, and

occurring together in at least one setence

SVO: Consider pairs (Ci, Cj), both from C, and occurring as subject and object in a sentence

Both use log-likelihood ratio to choose good pairs

Page 44: Text Mining

Verb extraction using VF*ICF Measure

Focus on verbs specific to the domain

Filter out overly general ones such as “do”, “is”

|C|: total number of concepts

VF(V): number of counts of V in all domain texts

CF(V): number of concepts in the same sentence as V

VF ICF V Log VF V LogC

CF V* ( ) ( ( ( ))

| |

( )( ) 1 2

Page 45: Text Mining

Sample top verbs from the electronic voting domain

Verb V VF*ICF(V)

produce 25.010 check 24.674 ensure 23.971 purge 23.863 create 23.160 include 23.160 say 23.151 restore 23.088 certify 23.047 pass 23.047

Page 46: Text Mining

Relation label assignment by Log-likelihood ratio measure

Candidate triples: (C1, V, C2) (C1, C2) is a candidate concept pair (by log-likelihood measure) V is a candidate verb (by VF*ICF measure) The triple occurs in a sentence

Question: Is the co-occurrence of V and the pair (C1, C2) accidental? Consider the following two hypotheses:

H P V C C P V C C

H P V C C P V C C

1 1 2 1 2

2 1 2 1 2

: ( | ( )) ( | ( ))

: ( | ( )) ( | ( ))

Page 47: Text Mining

S(C1, C2): set of sentences containing both C1, C2

S(V): set of sentences containing V

n S C C n S V n S C C S VC V CV | ( , ) | | ( ) | | ( , ) ( ) |1 2 1 2

N S V S C Cij k

C

i

n

j k

| ( ) ( , ) |,

| |

11

Page 48: Text Mining

Log-likelihood ratio:

For concept pair (C1, C2), select V with highest value for

L o g L o gL HL H

( )( )

12

L H b n n p b n n N n p

L H b n n p b n n N n p

cv c v cv c

cv c v cv c

( ) ( ; , ) ( ; , )

( ) ( ; , ) ( ; , )

1

2 1 2

b k n pn

kp pk n k( ; , ) ( ) ( )

1

2 Log

pnNV p

nnC V

C1 p

n nN nV C V

C2

Page 49: Text Mining

Experiments on the E-voting Domain

Recap: E-voting domain 15 articles from New York Times More than 10,000 distinct English words 164 relevant concepts were used in the experiments

For VF*ICF validation: First removed stop words Then apply VF*ICF measure to sort the verbs Take the top 20% of the sorted list as relevant verbs Achieved 57% precision with the top 20%

Page 50: Text Mining

Experiments -Continued

Criteria for evaluating a triple (C1, V, C2)

C1 and C2 are related non-hierarchically

V is a semantic label for either C1 C2 or

C2 C1

V is a semantic label for C1 C2 but not for C2 C1

Page 51: Text Mining

Experiments -Continued

yiuy787878uyuiuuuiuiuiiii Table II Example concept pairs

Concept pairs (C1, C2)

(election, official) (company, voting machine) (ballot, voter) (manufacturer, voting machine) (polling place, worker) (polling place, precinct) (poll, security)

Page 52: Text Mining

Experiments –RCL method

Table III RCL method example triples

 

CConcept C1 Label V Concept C2

machine produce paper ballot cast voter paper produce voting polling place Show up voter polling place turn voter election insist official ballot include paper manufacturer install voting machine

Page 53: Text Mining

Experiments –SVO method

Table IV SVO method example triples

 

Concept C1 Label V Concept C2

machine produce paper voter cast ballot voter record vote official tell voter voter trust machine worker direct voter county adopt machine company provide machine machine record ballot

Page 54: Text Mining

Comparisons

Table V Accuracy comparisons

Method (C1, C2) (C1, V, C2) (C1 V C2)

AE 89.00% 6.00% 4.00%

RCL 81.58% 30.36% 9.82%

SVO 89.47% 68.42% 68.42%

Page 55: Text Mining

Conclusions and Future Work

Presented techniques for automatic ontology extraction from texts

Combination of knowledge-base (WordNet), machine learning, information retrieval, syntactic patterns and heuristics

For concept extraction, WNSCA gives good precision and WNSCA + POP gives good recall

For taxonomy extraction, SCL and compound word heuristics are quite useful. The naïve Bayes classifier works well for SCL

For non-taxonomy extraction, SVO method has good accuracy, but Require using syntactical parsing Coverage (recall) not good

Page 56: Text Mining

Conclusions and Future Work

Both WNSCA and SVO are unsupervised method whereas SCL is a supervised one - what about un-supervised SCL?

The quality of extracted concepts heavily influences subsequent ontology extraction tasks

Better word sense disambiguation method would help to produce better taxonomy extraction results using WordNet

Consideration of other syntactic/semantic information may be needed to further improve non-taxonomical relation extraction Prepositional phrases Use WordNet Incorporate other knowledge

More experiments with larger text collections