-
Definitions
- for the first time mentioned in Feldman et al. [FD95]
(Hotho, Nürnberger, and Paaß 2005)
- Text mining refers generally to the process of extracting interesting information and knowledge from unstructured text
(Hotho, Nürnberger, and Paaß 2005)
- the application of algorithms and methods from the fields machine learning and statistics to texts with the goal of finding useful patterns. For this purpose it is necessary to pre-process the texts accordingly. Many authors use information extraction methods, natural language processing or some simple pre- processing steps in order to extract data from texts. To the extracted data then data mining algorithms can be applied (see [NM02, Gai03]).
(Hotho, Nürnberger, and Paaß 2005)
- Text Mining [1] is the discovery by computer of new,
previously unknown information, by automatically
extracting information from different written resources.
(Gupta and Lehal 2009)
- refers generally to the process of extracting interesting and non-trivial information and knowledge from unstructured text.
(Gupta and Lehal 2009)
- can work with unstructured or semi-structured data sets such as emails, full-text documents and HTML files etc.
(Gupta and Lehal 2009)
-
a.k.a
- knowledge discovery from text (KDT)
(Hotho, Nürnberger, and Paaß 2005)
- Intelligent Text Analysis, Text Data Mining or Knowledge-Discovery in Text (KDT)
(Gupta and Lehal 2009)
- Text Analytics
(Ppts APomares)
- Text Data Mining
(Ppts APomares)
-
Related Areas
- information retrieval, machine learning, statistics, computational linguistics and especially data mining.
(Hotho, Nürnberger, and Paaß 2005)
-
Data Mining
(Gupta and Lehal 2009)
- tries to find interesting patterns from large databases.
(Gupta and Lehal 2009)
-
Web mining
- is to explore interesting information and potential patterns from the contents of web page, the information of accessing the web page linkages and resources of e-commerce by using techniques of data mining, which can help people extract knowledge, improve web sites design, and develop e- commerce better.
(Gupta and Lehal 2009)
- Graph Mining
- Computational Linguistics
(Gupta and Lehal 2009)
-
Information Retrieval
- Information retrieval is the finding of documents which contain answers to questions and not the finding of answers itself [Hea99]
(Hotho, Nürnberger, and Paaß 2005)
-
Natural Language Processing NLP
- The general goal of NLP is to achieve a better understanding of natural language by use of computers [Kod99].
(Hotho, Nürnberger, and Paaß 2005)
-
Information Extraction
- The goal of information extraction methods is the ex- traction of specific information from text documents. These are stored in data base-like patterns (see [Wil97])
(Hotho, Nürnberger, and Paaß 2005)
- IE addresses the problem of transforming a corpus of textual documents into a more structured database, the database constructed by an IE module can be provided to the KDD module for further mining of knowledge
(Gupta and Lehal 2009)
-
Databases
- are necessary in order to analyze large quantities of data efficiently.
(Hotho, Nürnberger, and Paaß 2005)
-
Machine Learning
- is an area of artificial intelligence concerned with the development of techniques which allow computers to ”learn” by the analysis of data sets.
(Hotho, Nürnberger, and Paaß 2005)
-
Statistics
- Statistics has its grounds in mathematics and deals with the science and practice for the analysis of empirical data. It is based on statistical theory which is a branch of ap- plied mathematics
(Hotho, Nürnberger, and Paaß 2005)
-
”Knowledge Discovery in Databases (KDD)
- is the non-trivial process of identifying valid, novel, potentially useful, and ultimately understandable patterns in data
(Hotho, Nürnberger, and Paaß 2005)
-
Sub Areas
-
Text stream mining
-
Text Stream
- Created by very large scale interacions of individuals or structured creations of particular kinds of content by dedicated organizations. i.e. news-wire services (Reuters, AP)
(Aggarwal, 2012)
- Provide unprecedented challenges to data mining algorithms from an efficiency perspective
(Aggarwal, 2012)
- Ubiquitous in recent years because wide variety of applications in social networks, news collection. In general continuous creation of massive streams
(Aggarwal, 2012)
-
Applications
- Social networks
(Aggarwal, 2012)
- Users continuously communicate with one another with the use of text messages
- Interesting due to text messages are reflective of user interests, an the same applies to chat and email networks
- News aggregator services i.e. Google News
(Aggarwal, 2012)
- Recieves news articles continously over time
- Web crawlers
(Aggarwal, 2012)
- Collect large volume of documents from networks in small time frame
- Combines search results from major search engines like Google, Yahoo! and Bing
-
Opportunites
- Methods for online summarizations need to be designed
(Aggarwal, 2012)
-
Motivations
- There are estimates that 85% of business information lives in the form of text
(Hotho, Nürnberger, and Paaß 2005)
- As most information (over 80%) is stored as text, text mining is believed to have a high commercial potential value.
(Gupta and Lehal 2009), (Tan 1999)
- Humans have the ability to distinguish and apply linguistic patterns to text and humans can easily overcome obstacles that computers cannot easily handle such as slang, spelling variations and contextual meaning. However, although our language capabilities allow us to comprehend unstructured data, we lack the computer’s ability to process text in large volumes or at high speeds.
(Gupta and Lehal 2009)
- As the most natural form of storing information is text, text mining is believed to have a commercial potential higher than that of data mining
(Tan 1999)
-
Methodologies
-
CRoss Industry Standard Process for Data Mining (Crisp DM)
(Hotho, Nürnberger, and Paaß 2005)
- (1) business understanding2, (2) data understanding, (3) data preparation, (4) modelling, (5) evaluation, (6) deployment
-
Phases
-
Preprocessing
- For mining large document collections it is necessary to pre-process the text documents and store the information in a data structure, which is more appropriate for further processing than a plain text file
(Hotho, Nürnberger, and Paaß 2005)
-
Filtering
(Hotho, Nürnberger, and Paaß 2005)
- remove words from the dictionary and thus from the documents.
-
Lemmatization
(Hotho, Nürnberger, and Paaß 2005)
- Try to map verb forms to the infinite tense and nouns to the singular form.
-
Stemming
(Hotho, Nürnberger, and Paaß 2005)
- Try to build the basic forms of words, i.e. strip the plural ’s’ from nouns, the ’ing’ from verbs, or other affixes.
-
Linguistic
(Hotho, Nürnberger, and Paaß 2005)
-
Part-of-speech tagging (POS)
- determines the part of speech tag, e.g. noun, verb, adjective, etc. for each term.
-
Text chunking
- aims at grouping adjacent words in a sentence.
-
Word Sense Disambiguation (WSD)
- Tries to resolve the ambiguity in the meaning of single words or phrases.
-
Parsing
- produces a full parse tree of a sentence.
-
Index Term Selection
(Hotho, Nürnberger, and Paaß 2005)
- In this case, only the selected keywords are used to describe the documents.
-
The most commonly used criteria is the entropy
- The entropy gives a measure how well a word is suited to separate documents by keyword search.
- The entropy can be seen as a measure of the importance of a word in the given domain context.
-
Dimensionality reduction
-
Principal Components Analysis (PCA)
- a.k.a Kahrhunen-Lo`eve procedure, eigenvector analysis, and empirical orthogonal functions depending on the context in which one is being used
(Berry et al. 2008)
- recently it has been used primarily in statistical data analysis and image processing
(Berry et al. 2008)
- For text and data mining that focuses on covariance matrix analysis (COV)
(Berry et al. 2008)
-
Latent Semantic Indexing (LSI)
- Given a database with M documents and N distinguishing attributes for relevancy
ranking, let A denote the corresponding M-by-N document-attribute matrix model
with entries a(i, j) that represent the importance of the ith term in the jth document. The fundamental idea in LSI is to reduce the dimension of the IR problem to k, where k M,N, by projecting the problem into the space spanned by the rows of the closest rank-k matrix to A in the Frobenius norm [DDF+90]. Projection
(Berry et al. 2008)
- Singular Value Decomposition (SVD)
(Ppts APomares)
-
Storage
-
The Vector Space Model
-
Each element of the vector usually represents a word (or a group of words) of the document collection, i.e. the size of the vector is defined by the number of words (or groups of words) of the complete document collection
(Hotho, Nürnberger, and Paaß 2005)
-
Bag of words
- A document is represented as a set of words, together with their associated frequency in the document.
(Aggarwal, 2012)
- This representation is essentially independent of the sequence of words in the
collection
(Aggarwal, 2012)
- simplicity for classification purposes
(Aggarwal, 2012)
-
Techniques
-
Classification
(Hotho, Nürnberger, and Paaß 2005)
- Classification aims at assigning pre-defined classes to text documents
-
a.k.a Categorization
- Categorization involves identifying the main themes of a document by placing the document into a pre-defined set of topics.
(Gupta and Lehal 2009)
-
Na ̈ıve Bayes Classifier
- We may assign the class with highest posterior probability to our document.
- Combining this “na ̈ıve” independence assumption with the Bayes formula
-
Nearest Neighbor Classifier
- Instead of building explicit models for the different classes we may select documents from the training set which are “similar” to the target document. The class of the target document subsequently may be inferred from the class labels of these similar documents.
-
Decision Trees
- Decision trees are classifiers which consist of a set of rules which are applied in a sequential way and finally yield a decision.
-
Support Vector Machines
- The SVM algorithm determines a hyperplane which is located between the positive and negative examples of the training set
- The most important property of SVMs is that learning is nearly independent of the dimensionality of the feature space. It rarely requires feature selection as it inherently selects data points (the support vectors) required for a good classification.
-
Clustering
(Hotho, Nürnberger, and Paaß 2005)
- Clustering method can be used in order to find groups of documents with similar content.
- Clustering [7] is a technique used to group similar documents, but it differs from categorization in that documents are clustered on the fly instead of through the use of predefined topics.
(Gupta and Lehal 2009)
-
K-means
- the documents are assigned to the nearest of the k centroids (also called cluster prototype). Step four calculates a new centroids on the basis of the new allocations. We repeat the two steps in a loop (step five) until the cluster centroids do not change any more
-
Self organizing maps (SOM)
- The neurons in the input layer correspond to the input dimensions, here the words of the document vector. The output layer (map) contains as many neurons as clusters needed.
-
Fuzzy Clustering
- While most classical clustering algorithms assign each datum to exactly one cluster, thus forming a crisp partition of the given data, fuzzy clustering al- lows for degrees of membership, to which a datum belongs to different clusters
-
Information Extraction
(Gupta and Lehal 2009)
- The main task is to extract parts of text and assign specific attributes to it.
-
Hidden Markov Models
- Hidden Markov models require the conditional independence of features of different words given the labels.
-
Visualization
- Visual text mining, or information visualization [3], puts large textual sources in a visual hierarchy or map and provides browsing capabilities, in addition to simple searching.
(Gupta and Lehal 2009)
-
Topic Tracking
- choose keywords and notifies them when news relating to those topics becomes available.
(Gupta and Lehal 2009)
-
Summarization
- Text summarization is immensely helpful for trying to figure out whether or not a lengthy document meets the user’s needs and is worth reading for further information.
(Gupta and Lehal 2009)
-
Concept Linkage
- Concept linkage tools [3] connect related documents by identifying their commonly-shared concepts and help users find information that they perhaps wouldn’t have found using traditional searching methods.
(Gupta and Lehal 2009)
-
Question Answering
- Another application area of natural language processing is natural language queries, or question answering (Q&A), which deals with how to find the best answer to a given question.
(Gupta and Lehal 2009)
-
Association Rule Mining
- Association rule mining (ARM) [33] is a technique used to discover relationships among a large set of variables in a data set.
(Gupta and Lehal 2009)
- ARM discovers what items customers typically purchase together.
(Gupta and Lehal 2009)
-
Application Areas
(Hotho, Nürnberger, and Paaß 2005)
-
Patent Analysis
- The users like to have these stories tagged with categories and the names of important persons, organizations and places
-
Text Classification for News Agencies
- Automatic clustering for news assigning topics
-
Bioinformatics
- Bio-entity recognition aims to identify and classify technical terms in the domain of molecular biology that correspond to instances of concepts that are of interest to biologists.
-
Knowledge and Human Resource management
- The need to organize and modify their strategies according to demands and to the opportunities that the market present requires that companies collect information about themselves, the market and their competitors, and to manage enormous amount of data, and analyzing them to make plans.
(Gupta and Lehal 2009)
-
Customer Relationship Management (CRM)
- automatically rerouting specific requests to the appropriate service or at supplying immediate answers to the most frequently asked questions.
(Gupta and Lehal 2009)
-
Market Analysis
- analyze competitors and/or monitor customers' opinions to identify new potential customers, as well as to determine the companies’ image through the analysis of press reviews and other relevant sources.
(Gupta and Lehal 2009)
-
Technology watch
- analyses the characteristics of existing technologies, as well as identifying emerging technologies
(Gupta and Lehal 2009)
-
Open Problems and future directions
- It remains a challenge to see how semantic analysis can be made much more efficient and scalable for very large text corpora.
(Tan 1999)
- It is essential to develop text refining algorithms, that process multilingual text documents and produce language-independent intermediate forms
(Tan 1999)
- It is interesting to explore how one can take advantage of domain information to improve parsing efficiency and derive a more compact intermediate form.
(Tan 1999)
- Current text mining products and applications are still tools designed for trained knowledge specialists. Future text mining tools, as part of the knowledge management systems, should be readily usable by technical users as well as management executives.
(Tan 1999)
-
Real Applications
(Ppts APomares)
-
opinioncrawl.com
- Sentiment Analysis
- allows visitors to assess Web sentiment on a topic - a person, an event, a company or a product
-
socialmention.com
- social media search and analysis platform that aggregates user generated content from across the universe into a single stream of information.
-
nlp.stanford.edu:8080/sentiment/rntnDemo.html
- Sentiment Analysis