Natural Language Processing: State of The Art, Current Trends and Challenges
Natural Language Processing: State of The Art, Current Trends and Challenges
Natural Language Processing: State of The Art, Current Trends and Challenges
Challenges
Diksha Khurana1, Aditya Koli1, Kiran Khatter1,2 and Sukhdev Singh1,2
1
Manav Rachna University, India
2
Accendere Knowledge Management Services Pvt. Ltd., India
Abstract
Current paradigm of information technology uses human-computer interaction in the form of
natural language—the language we use for day-to-day communication. Natural Language
Processing (NLP) has recently gained much attention for representing and analysing human
language computationally. It has spread its applications in various fields such as machine
translation, email spam detection, information extraction, summarization, medical, and
question answering etc. The paper distinguishes various phases by discussing different levels
of NLP and components of Natural Language Generation (NLG) followed by presenting the
history and evolution of NLP, state of the art presenting the various applications of NLP and
current trends and challenges.
Keywords: Natural Language Processing, Human Computer Interaction, Natural Language
Generation
1. Introduction
Natural Language Processing (NLP) is a tract of Artificial Intelligence and Linguistics,
devoted to make computers understand the statements or words written in human languages.
Natural language processing came into existence to ease the user’s work and to satisfy the
wish to communicate with the computer in natural language. Since all the users may not be
well-versed in machine specific language, NLP caters those users who do not have enough
time to learn new languages or get perfection in it.
A language can be defined as a set of rules or a set of symbols. Symbols are combined and
used for conveying information or broadcasting the information. Symbols are tyrannized by
the rules. Natural Language Processing basically can be classified into two parts i.e. Natural
Language Understanding and Natural Language Generation which evolves the task to
understand and generate the text (Figure 1).
Noah Chomsky, one of the first linguists of the twelfth century who started syntactic theories,
marked a unique position in the field of theoretical linguistics because he revolutionized the
area of syntax1 which can be broadly categorized into two levels: Higher Level which
includes speech recognition and Lower Level which corresponds to natural language. Few of
the researched tasks of NLP are Automatic Summarization, Co-Reference Resolution,
Discourse Analysis, Machine Translation, Morphological Segmentation, Named Entity
Recognition, Optical Character Recognition, Part Of Speech Tagging etc. Some of these tasks
have direct real world applications such as Machine translation, Named entity recognition,
Optical character recognition etc. Automatic summarization produces an understandable
summary of a set of text and provides summaries or detailed information of text, of a known
type. Co-reference resolution refers to a sentence or a large set of text that determines which
words refer to the same object. Discourse analysis refers to the task of identifying the
discourse structure of connected text. Machine translation refers to the automatic translation
of text from one human language to another. Morphological segmentation refers to separating
a word into individual morphemes and identify the class of the morphemes. Named Entity
Recognition (NER) describes a stream of text, determines which items in the text relate to
proper names. Optical Character Recognition (OCR) gives an image representing printed
text, which helps in determining the corresponding or related text. Part of speech tagging
describes a sentence, determines the part of speech for each word. Though NLP tasks are
obviously very closely interweaved but they are used frequently,for convenience .Some of the
tasks such as automatic summarisation, co-reference analysis etc. act as subtasks that are used
in solving larger tasks.
Linguistic is the science which involves meaning of language, language context and various
forms of language. The various important terminologies of Natural Language Processing are:-
1. Phonology
Phonology is the part of linguistics which refers to the systematic arrangement of sound. The
term phonology comes from Ancient Greek and the term phono- means voice or sound, and
the suffix –logy refers to word or speech. In 1993 Nikolai Trubetzkoy stated that Phonology
that is “the study of sound pertaining to the system of language,whereas Lass in 1998 wrote
that phonology refers broadly with the sounds of language, concerned with the lathe sub
discipline of linguistics, it could be better explained as, "phonology proper is concerned with
the function, behaviour and organization of sounds as linguistic items. It includes semantic
use of sound to encode meaning of any human language 6.
2. Morphology
The different part of the word represent the smallest units of meaning known as Morphemes.
Morphology which comprises of nature of words, are initiated by morphemes. An example of
Morpheme could be, the word precancellation which can be morphologically scrutinized into
three separate morphemes: the prefix pre, the root cancella, and the suffix -tion. The
interpretation of morpheme stays same across all the words and just to understand the
meaning humans can break any unknown word into morphemes. For example, adding the
suffix –ed to a verb, conveys that the action of the verb took place in the past. The words that
cannot be divided and have meaning by themselves are called Lexical morphemes (e.g.: table,
chair).The words (e.g. -ed, -ing, -est, -ly, -ful) that are combined with the lexical morpheme
are known as Grammatical morphemes (eg. Worked, Consulting, Smallest, Likely, Use).
Those grammatical morphemes that occur in combination called bound morphemes (eg. -ed, -
ing). Grammatical morphemes can be further divided into bound morphemes and derivational
morphemes.
3. Lexical
In Lexical, humans as well as NLP systems can interpret the meaning of individual words.
Several types of processing bestow to word-level understanding – the first of these being a
part-of-speech tag to each word. In this processing, words that can act as more than one part-
of-speech which are assigned to the most probable part-of speech tag based on the context in
which they occur. At the lexical level, semantic representations can be replaced by the words
that have one meaning. In NLP system, the nature of the representation varies according to
the semantic theory deployed.
4. Syntactic
This level emphasis to examine the words in a sentence so as to uncover the grammatical
structure of the sentence. Both grammar and parser are required in this level. The output of
this level of processing is the representation of the sentence that communicate the structural
dependency relationships between the words.There are various grammars that can be
hindered. Not all NLP applications require a full parse of sentences. They even face a lot of
challenges in parsing of prepositional phrase attachment and conjunction audit no longer
impede that plea for which phrasal and clausal dependencies are adequate 7.The syntax
conveys meaning in most languages because order and dependency contribute to connotation.
For example, the two sentences: ‘The cat chased the mouse’ and ‘The mouse chased the cat’
differ only in terms of syntax, yet convey quite different meaning.
5. Semantic
In semantic, most people think that meaning is determined, however this is not it is all the
levels that bestow to meaning. Semantic processing determines the possible meanings of a
sentence by pivoting on the interactions among word-level meanings in the sentence. This
level of processing can incorporate the semantic disambiguation of words with multiple
senses; in a cognate way to how syntactic disambiguation of words that can work as multiple
parts-of-speech is handy at the syntactic level. For example, amongst other meanings, ‘file’ as
a noun can mean either a binder for gathering papers, or a tool to form one’s fingernails, or a
line of individuals in a queue 7. The semantic level survey words for their dictionary report,
but also for the report they derive from the context of the sentence. Semantics context that
most words have more than one report but that we can spot the appropriate one by looking at
the rest of the sentence8 .
6. Discourse
While syntax and semantics travail with sentence-length units, the discourse level of NLP
travail with units of text longer than a sentence i.e, it does not interpret multi sentence texts as
just sequence sentences, a piece of which can be reported singly. Rather, discourse focuses on
the properties of the text as a whole that convey meaning by making connections between
component sentences7. The two of the most common levels are Anaphora Resolution -It is the
replacing of words such as pronouns, which are semantically stranded, with the pertinent
entity to which they refer. Discourse/Text Structure Recognition - Discourse/text structure
recognition sway the functions of sentences in the text, which in turn adds to the meaningful
representation of the text.
7. Pragmatic:
Pragmatic is concerned with the firm use of language in situations and utilizes nub over and
above the nub of the text for understanding the goal and to explain how extra meaning is read
into texts without literally being encoded in them. This requisite much world knowledge,
including the understanding of intentions, plans, and goals. For example, the following two
sentences need aspiration of the anaphoric term ‘they’, but this aspiration requires pragmatic
or world knowledge7.
3. Natural Language Generation
Natural Language Generation (NLG) is the process of producing phrases, sentences and
paragraphs that are meaningful from an internal representation. It is a part of Natural
Language Processing and happens in four phases: identifying the goals, planning on how
goals maybe achieved by evaluating the situation and available communicative sources and
realizing the plans as a text as shown in Figure 3. It is opposite to Understanding.
Speaker and Generator – To generate the text we need to have a speaker or an application
and a generator or a program that renders the application’s intentions into fluent phrase
relevant to the situation.
Application or Speaker – This is only for maintaining the model of the situation. Here the
speaker just initiates the process doesn’t take part in the language generation. It stores the
history, structures the content that is potentially relevant and deploys a representation of what
is actually known. All these form the situation, while selecting subset of propositions that
speaker has. The only requirement is the speaker that has to make a sense of the situation9 .
4. History of NLP
In late 1940s the term wasn’t even in existence, but the work regarding machine translation
(MT) had started. Research in this period was not completely localized. Russian and English
were the dominant languages for MT, but others, like Chinese were used for MT
(Booth ,1967)10. MT/NLP research almost died in 1966 according to ALPAC report, which
concluded that MT is going now here. But later on some MT production systems were
providing output to their customers11. By this time, they started working on the use of
computers for literary and linguistic studies had also started.
As early as 1960 signature work influenced by AI began, with the BASEBALL Q-A
systems12. LUNAR13 and Winograd SHRDLU were natural successors of these systems but
they were seen as stepped up sophistication, in terms of their linguistic and their task
processing capabilities. There was a widespread belief that progress could only be made on
the two sides, one is ARPA Speech Understanding Research (SUR) project and other in some
major system developments projects building database front ends. The front-end projects 14
were intended to go beyond LUNAR in interfacing the large databases.
In early 1980s computational grammar theory became a very active area of research linked
with logics for meaning and knowledge’s ability to deal with the user’s beliefs and intentions
and with functions like emphasis and themes.
By the end of the decade the powerful general purpose sentence processors like SRI’s Core
Language Engine15 and Discourse Representation Theory16 offered a means of tackling more
extended discourse within the grammatico-logical framework. This period was one of the
growing community. Practical resources, grammars, and tools and parsers became available
e.g the Alvey Natural Language Tools 17.The (D)ARPA speech recognition and message
understanding information extraction conferences were not only for the tasks they addressed
but for the emphasis on heavy evaluation, thus starting a trend that became a major feature in
the 1990s18,19. They work on user modelling 20 that was one strand in research paper and on
discourse structure serving21. At the same time, as McKeown22 showed, rhetorical schemas
which could be used for producing both linguistically coherent and communicatively
effective text. Some researches in NLP marked important topics for future like word sense
disambiguation23and probabilistic networks, statistically coloured NLP, the work on the
lexicon, also pointed in this direction.
Statistical language processing was a major thing in 90s 24, because this not only involves data
analysts. Information extraction and automatic summarising 25 was also a point of focus.
Recent researches are mainly focused on unsupervised and semi-supervised learning
algorithms.
5. Related Work
Many researchers worked on NLP, building tools and systems which makes NLP what it is
today. Tools like Sentiment Analyser, Parts of Speech (POS)Taggers, Chunking, Named
Entity Recognitions (NER), Emotion detection, Semantic Role Labelling made NLP a good
topic for research.
Sentiment analyser26 works by extracting sentiments about given topic. Sentiment analysis
consists of a topic specific feature term extraction, sentiment extraction, and association by
relationship analysis. Sentiment Analysis utilizes two linguistic resources for the analysis:
The sentiment lexicon and the sentiment pattern database. It analyses the documents for
positive and negative words and try to give ratings on scale -5 to +5.
Parts of speech taggers for the languages like European languages, research is being done on
making parts of speech taggers for other languages like Arabic, Sanskrit 27, Hindi28 etc. It can
efficiently tag and classify words as nouns, adjectives, verbs etc. The most procedures for
part of speech can work efficiently on European languages, but it won’t on Asian languages
or middle eastern languages. Sanskrit part of speech tagger is specifically uses treebank
technique. Arabic uses Support Vector Machine (SVM)29 approach to automatically tokenize,
parts of speech tag and annotate base phrases in Arabic text.
Usage of Named Entity Recognition in places such as Internet is a problem as people don’t
use traditional or standard English. This degrades the performance of standard natural
language processing tools substantially. By annotating the phrases or tweets and building
tools trained on unlabelled, in domain and out domain data 33. It improves the performance as
compared to standard natural language processing tools.
Emotion Detection34 is similar to sentiment analysis, but it works on social media platforms
on mixing of two languages I.e English + Any other Indian Language. It categorizes
statements into six groups based on emotions. During this process, they were able to identify
the language of ambiguous words which were common in Hindi and English and tag lexical
category or parts of speech in mixed script by identifying the base language of the speaker.
Sematic Role Labelling – SRL works by giving a semantic role to a sentence. For example in
the PropBank35 formalism, one assigns roles to words that are arguments of a verb in the
sentence. The precise arguments depend on verb frame and if there exists multiple verbs in a
sentence, it might have multiple tags. State-of-the-art SRL systems comprise of several
stages: creating a parse tree, identifying which parse tree nodes represent the arguments of a
given verb, and finally classifying these nodes to compute the corresponding SRL tags.
Event discovery in social media feeds 36, using a graphical model to analyse any social media
feeds to determine whether it contains name of a person or name of a venue, place, time etc.
The model operates on noisy feeds of data to extract records of events by aggregating
multiple information across multiple messages, despite the noise of irrelevant noisy messages
and very irregular message language, this model was able to extract records with high
accuracy. However, there is some scope for improvement using broader array of features on
factors.
6. Applications of NLP
Natural Language Processing can be applied into various areas like Machine Translation,
Email Spam detection, Information Extraction, Summarization, Question Answering etc.
Discovery of knowledge is becoming important areas of research over the recent years.
Knowledge discovery research use a variety of techniques in order to extract useful
information from source documents like
Parts of Speech (POS) tagging, Chunking or Shadow Parsing, Stop-words I.e Keywords that
are used and must be removed before processing documents, Stemming I.e Mapping words to
some base for, it has two methods, dictionary based stemming and Porter style stemming 55.
Former one has higher accuracy but higher cost of implementation while latter has lower
implementation cost and is usually insufficient for IR. Compound or Statistical Phrases
Compounds and statistical phrases index multi token units instead of single tokens. Word
Sense Disambiguation refers to Word sense disambiguation is the task of understanding the
correct sense of a word in context. When used for information retrieval, terms are replaced by
their senses in the document vector.
Its extracted information can be applied on a variety of purpose, for example to prepare a
summary, to build databases, identify keywords, classifying text items according to some pre-
defined categories etc. For example CONSTRUE, it was developed for Reuters, that is used
in classifying news stories57. It has been suggested that many IE systems can successfully
extract terms from documents, acquiring relations between the terms is still a difficulty.
PROMETHEE is a system that extracts lexico-syntactic patterns relative to a specific
conceptual relation58. IE systems should work at many levels, from word recognition to
discourse analysis at the level of the complete document. An application of the Blank Slate
Language Processor (BSLP)59 approach for the analysis of a real life natural language corpus
that consists of responses to open-ended questionnaires in the field of advertising.
There’s a system called MITA (Metlife’s Intelligent Text Analyzer) 60 that extracts
information from life insurance applications. Ahonen et al. 61 suggested a mainstream
framework for text mining that uses pragmatic and discourse level analyses of text.
6.5 Summarization
Overload of information is the real thing in this digital age, and already our reach and access
to knowledge and information exceeds our capacity to understand it. This trend is not slowing
down, so an ability to summarize the data while keeping the meaning intact is highly
required. This is important not just allowing us the ability to recognize the understand the
important information for a large set of data, it is used to understand the deeper emotional
meanings; For example, a company determine the general sentiment on social media and use
it on their latest product offering. This application is useful as a valuable marketing asset.
The types of text summarization depends on the basis of the number of documents and the
two important categories are single document summarization and multi document
summarization62,63. Summaries can also be of two types: generic or query-focused 64,65,66,67.
Summarization task can be either supervised or unsupervised 68,63,69. Training data is required
in a supervised system for selecting relevant material from the documents. Large amount of
annotated data is needed for learning techniques. Few techniques are as follows–
- Bayesian Sentence based Topic Model (BSTM) uses both term-sentences and term
document associations for summarizing multiple documents70.
- Factorization with Given Bases (FGB) is a language model where sentence bases are
the given bases and it utilizes document-term and sentence term matrices. This
approach groups and summarizes the documents simultaneously71.
- Topic Aspect-Oriented Summarization (TAOS) is based on topic factors. These topic
factors are various features that describe topics such as capital words are used to
represent entity. Various topics can have various aspects and various preferences of
features are used to represent various aspects72.
6.7 Medicine
NLP is applied in medicine field as well. The Linguistic String Project-Medical Language
Processor is one the large scale projects of NLP in the field of medicine 74,75,76,77,78. The LSP-
MLP helps enabling physicians to extract and summarize information of any signs or
symptoms, drug dosage and response data with aim of identifying possible side effects of any
medicine while highlighting or flagging data items 74. The National Library of Medicine is
developing The Specialist System79,80,81,82,83. It is expected to function as Information
Extraction tool for Biomedical Knowledge Bases, particularly Medline abstracts. The lexicon
was created using MeSH (Medical Subject Headings), Dorland’s Illustrated Medical
Dictionary and general English Dictionaries. The Centre d’Informatique Hospitaliere of the
Hopital Cantonal de Geneve is working on an electronic archiving environment with NLP
features84,85. In first phase, patient records were archived . At later stage the LSP-MLP has
been adapted for French86,87,88,89 , and finally , a proper NLP system called RECIT 90,91,92,93 has
been developed using a method called Proximity Processing 94. It’s task was to implement a
robust and multilingual system able to analyze/comprehend medical sentences, and to
preserve a knowledge of free text into a language independent knowledge representation 95,96.
The Columbia university of New York has developed an NLP system called MEDLEE
(MEDical Language Extraction and Encoding System) that identifies clinical information in
narrative reports and transforms the textual information into structured representation 97.
7. Approaches
Rationalist approach or symbolic approach assume that crucial part of the knowledge in the
human mind is not derived by the sense but is firm in advance, probably by genetic in
heritance. Noam Chomsky was the strongest advocate of this approach. It was trusted that
machine can be made to function like human brain by giving some fundamental knowledge
and reasoning mechanism linguistics knowledge is directly encoded in rule or other forms of
representation. This helps automatic process of natural languages. [98] Statistical and machine
learning entail evolution of algorithms that allow a program to infer patterns. An iterative
process is used to characterize a given algorithm’s underlying algorithm that are optimised by
a numerical measure that characterize numerical parameters and learning phase. Machine-
learning models can be predominantly categorized as either generative or discriminative.
Generative methods can generate synthetic data because of which they create rich models of
probability distributions. Discriminative methods are more functional and have right
estimating posterior probabilities and are based on observations.
Srihari99 explains the different generative models as one with a resemblance that is used to
spot an unknown speaker’s language and would bid the deep knowledge of numerous
language to perform the match. Whereas discriminative methods rely on a less knowledge-
intensive approach and using distinction between language. Whereas generative models, can
become troublesome when many features are used and discriminative models allow use of
more features100. Few of the examples of discriminative methods are Logistic regression and
conditional random fields (CRFs), generative methods are Naive Bayes classifiers and hidden
Markov models (HMMs).
Hidden Markov Models are extensively used for speech recognition, where the output
sequence is matched to the sequence of individual phonemes. Frederick Jelinek, a statistical-
NLP advocate who first instigated HMMs at IBM’s Speech Recognition Group, reportedly
joked, every time a linguist leaves my group, the speech recognizer’s performance improves.
[101]
HMM is not restricted to this application it has several others such as bioinformatics
problems, for example, multiple sequence alignment102. Sonnhammer mentioned that Pfam
hold multiple alignments and hidden Markov model based profiles (HMM-profiles) of entire
protein domains. The cue of domain boundaries, family members and alignment is done
semi-automatically found on expert knowledge, sequence similarity, other protein family
databases and the capability of HMM-profiles to correctly identify and align the members 103.
7.2 Naive Bayes Classifiers
The choice of area is wide ranging covering usual items like word segmentation and
translation but also unusual areas like segmentation for infant learning and identifying
documents for opinions and facts. In addition, exclusive article was selected for its use of
Bayesian methods to aid the research in designing algorithms for their investigation.
8. NLP in Talk
This section discusses the recent developments in the NLP projects implemented by various
companies and these are as follows:
8.1 ACE Powered GDPR Robot Launched by RAVN Systems 104
RAVN Systems, an leading expert in Artificial Intelligence (AI), Search and Knowledge
Management Solutions, announced the launch of a RAVN ("Applied Cognitive Engine") i.e
powered software Robot to help and facilitate the GDPR ("General Data Protection
Regulation") compliance.
The Robot uses AI techniques to automatically analyse documents and other types of data in
any business system which is subject to GDPR rules. It allows users to quickly and easily
search, retrieve, flag, classify and report on data mediated to be supersensitive under GDPR.
Users also have the ability to identify personal data from documents, view feeds on the latest
personal data that requires attention and provide reports on the data suggested to be deleted or
secured. RAVN's GDPR Robot is also able to hasten requests for information (Data Subject
Access Requests - "DSAR") in a simple and efficient way, removing the need for a physical
approach to these requests which tends to be very labour thorough. Peter Wallqvist, CSO at
RAVN Systems commented, "GDPR compliance is of universal paramountcy as it will
exploit to any organisation that control and process data concerning EU citizens.
LINK:http://markets.financialcontent.com/stocks/news/read/33888795/
RAVN_Systems_Launch_the_ACE_Powered_GDPR_Robot
This provides a different platform than other brands that launch chatbots like Facebook
Messenger and Skype. They believed that Facebook has too much access of private
information of a person, which could get them into trouble with privacy laws of U.S.
financial institutions work under. Like any Facebook Page admin can access full transcripts
of the bot’s conversations. If that would be the case then the admins could easily view the
personal banking information of customers with is not correct
LINK: https://www.macobserver.com/analysis/capital-one-natural-language-chatbot-eno/
8.3 Future of BI in Natural Language Processing106
Several companies in BI spaces are trying to get with the trend and trying hard to ensure that
data becomes more friendly and is easily accessible. But still there is a long way for this.BI
will also make it easier to access as GUI is not needed. Because now a days the queries are
made by text or voice command on smartphones.one of the most common example is Google
might tell you today what will be the tomorrows weather. But soon enough, we will be able
to ask our personal data (chatbot) about customer sentiment today, and how do we feel about
their brand next week; all while walking down the street. Today, NLP tends to be based on
turning natural language into machine language. But with time the technology matures –
especially the AI component –the computer will get better at “understanding” the query and
start to deliver answers rather than search results.
Initially, the data chatbot will probably ask the question as how have revenues changed over
the last three-quarters?’ and then return pages of data for you to analyse. But once it learns
the semantic relations and inferences of the question, it will be able to automatically perform
the filtering and formulation necessary to provide an intelligible answer, rather than simply
showing you data.
Link: http://www.smartdatacollective.com/eran-levy/489410/here-s-why-natural-language-processing-future-bi
Unique concepts in each abstract are extracted using Meta Map and their pairwise
cooccurrence are determined. Then the information is used to construct a network graph of
concept co-occurrence that is further analysed to identify content for the new conceptual
model. 142 abstracts are analysed. Medication adherence is the most studied drug therapy
problem and co-occurred with concepts related to patient-centred interventions targeting self-
management. The enhanced model consists of 65 concepts clustered into 14 constructs. The
framework requires additional refinement and evaluation to determine its relevance and
applicability across a broad audience including underserved settings.
Link: https://www.ncbi.nlm.nih.gov/pubmed/28269895?dopt=Abstract
Simultaneously, the user will hear the translated version of the speech on the second earpiece.
Moreover, it is not necessary that conversation would be taking place between two people
only the users can join in and discuss as a group. As if now the user may experience a few
second lag interpolated the speech and translation, which Waverly Labs pursue to reduce.
The Pilot earpiece will be available from September, but can be pre-ordered now for $249.
The earpieces can also be used for streaming music, answering voice calls and getting audio
notifications.
Link:https://www.indiegogo.com/projects/meet-the-pilot-smart-earpiece-language-translator-
headphones-travel#/
REFRENCES
[1] Chomsky, Noam. Aspects of the Theory of Syntax. MASSACHUSETTS INST OF TECH
CAMBRIDGE RESEARCH LAB OF ELECTRONICS, 1964.
[2] Rospocher, Marco, et al. "Building event-centric knowledge graphs from news." Web
Semantics: Science, Services and Agents on the World Wide Web 37 (2016): 132-151.
[4] Emele, Martin C., and Michael Dorna. "Ambiguity preserving machine translation using
packed representations." Proceedings of the 36th Annual Meeting of the Association for
Computational Linguistics and 17th International Conference on Computational Linguistics-
Volume 1. Association for Computational Linguistics, 1998.
[5] Knight, Kevin, and Irene Langkilde. "Preserving ambiguities in generation via automata
intersection." AAAI/IAAI. 2000.
[6] Nation, Kate, Margaret J. Snowling, and Paula Clarke. "Dissecting the relationship
between language skills and learning to read: Semantic and phonological contributions to
new vocabulary learning in children with poor reading comprehension." Advances in Speech
Language Pathology9.2 (2007): 131-139.
[8] Feldman, Susan. "NLP meets the Jabberwocky: Natural language processing in
information retrieval." ONLINE-WESTON THEN WILTON- 23 (1999): 62-73.
[9] "Natural Language Processing." Natural Language Processing RSS. N.p., n.d. Web. 25
Mar. 2017
[10] Hutchins, William John. Machine translation: past, present, future. Chichester: Ellis
Horwood, 1986.
[11] Hutchins, W. John, ed. Early years in machine translation: memoirs and biographies of
pioneers. Vol. 97. John Benjamins Publishing, 2000.
[12] Green Jr, Bert F., et al. "Baseball: an automatic question-answerer." Papers presented at
the May 9-11, 1961, western joint IRE-AIEE-ACM computer conference. ACM, 1961.
[14] Hendrix, Gary G., et al. "Developing a natural language interface to complex data."
ACM Transactions on Database Systems (TODS) 3.2 (1978): 105-147.
[15] Alshawi, Hiyan, ed. The core language engine. MIT press, 1992.
[16] Kamp, Hans, and Uwe Reyle. "From discourse to logic: Introduction to modeltheoretic
semantics of natural language, formal logic and discourse representation." Studies in
Linguistics and Philosophy. Kluwer (1993).
[17] Lea , W.A Trends in speech recognition , Englewoods Cliffs , NJ: Prentice Hall , 1980.
[18] Young, Steve J., and Lin Lawrence Chase. "Speech recognition evaluation: a review of
the US CSR and LVCSR programmes." Computer Speech & Language 12.4 (1998): 263-279.
[19] Sundheim, Beth M., and Nancy A. Chinchor. "Survey of the message understanding
conferences." Proceedings of the workshop on Human Language Technology. Association for
Computational Linguistics, 1993.
[20] Wahlster, Wolfgang, and Alfred Kobsa. "User models in dialog systems." User models
in dialog systems. Springer, Berlin, Heidelberg, 1989. 4-34.
[21] McKeown, K.R. Text generation , Cambridge: Cambridge University Press , 1985.
[22] Small S.L., Cortell G.W., and Tanenhaus , M.K. Lexical Ambiguity Resolutions , San
Mateo , CA : Morgan Kauffman, 1988.
[23] Manning, Christopher D., and Hinrich Schütze. Foundations of statistical natural
language processing. MIT press, 1999.
[24] Mani, Inderjeet, and Mark T. Maybury. Advances in automatic text summarization. MIT
press, 1999.
[25] Yi, Jeonghee, et al. "Sentiment analyzer: Extracting sentiments about a given topic using
natural language processing techniques." Data Mining, 2003. ICDM 2003. Third IEEE
International Conference on. IEEE, 2003.
[26] Yi, Jeonghee, et al. "Sentiment analyzer: Extracting sentiments about a given topic using
natural language processing techniques." Data Mining, 2003. ICDM 2003. Third IEEE
International Conference on. IEEE, 2003.
[27] Tapaswi, Namrata, and Suresh Jain. "Treebank based deep grammar acquisition and
Part-Of-Speech Tagging for Sanskrit sentences." Software Engineering (CONSEG), 2012 CSI
Sixth International Conference on. IEEE, 2012.
[28] Ranjan, Pradipta, and Harish V. Sudeshna Sarkar Anupam Basu. "Part of speech tagging
and local word grouping techniques for natural language parsing in Hindi." Proceedings of
the 1st International Conference on Natural Language Processing (ICON 2003). 2003.
[29] Diab, Mona, Kadri Hacioglu, and Daniel Jurafsky. "Automatic tagging of Arabic text:
From raw text to base phrase chunks." Proceedings of HLT-NAACL 2004: Short papers.
Association for Computational Linguistics, 2004.
[30] Sha, Fei, and Fernando Pereira. "Shallow parsing with conditional random fields."
Proceedings of the 2003 Conference of the North American Chapter of the Association for
Computational Linguistics on Human Language Technology-Volume 1. Association for
Computational Linguistics, 2003.
[31] McDonald, Ryan, Koby Crammer, and Fernando Pereira. "Flexible text segmentation
with structured multilabel classification." Proceedings of the conference on Human Language
Technology and Empirical Methods in Natural Language Processing. Association for
Computational Linguistics, 2005.
[32] Sun, Xu, et al. "Modeling latent-dynamic in shallow parsing: a latent conditional model
with improved inference." Proceedings of the 22nd International Conference on
Computational Linguistics-Volume 1. Association for Computational Linguistics, 2008.
[33] Ritter, Alan, Sam Clark, and Oren Etzioni. "Named entity recognition in tweets: an
experimental study." Proceedings of the conference on empirical methods in natural
language processing. Association for Computational Linguistics, 2011..
[34] Sharma, S., Srinivas, PYKL, & Balabantaray, RC (2016). Emotion Detection using
Online Machine Learning Method and TLBO on Mixed Script. In Proceedings of Language
Resources and Evaluation Conference 2016 (pp. 47-51).
[35] Palmer, Martha, Daniel Gildea, and Paul Kingsbury. "The proposition bank: An
annotated corpus of semantic roles." Computational linguistics 31.1 (2005): 71-106.
[36] Benson, Edward, Aria Haghighi, and Regina Barzilay. "Event discovery in social media
feeds." Proceedings of the 49th Annual Meeting of the Association for Computational
Linguistics: Human Language Technologies-Volume 1. Association for Computational
Linguistics, 2011.
[37] Tillmann, Christoph, et al. "Accelerated DP based search for statistical translation." Fifth
European Conference on Speech Communication and Technology. 1997.
[38] Bangalore, Srinivas, Owen Rambow, and Steve Whittaker. "Evaluation metrics for
generation." Proceedings of the first international conference on Natural language
generation-Volume 14. Association for Computational Linguistics, 2000.
[39] Nießen, Sonja, et al. "An Evaluation Tool for Machine Translation: Fast Evaluation for
MT Research." LREC. 2000.
[40] Papineni, Kishore, et al. "BLEU: a method for automatic evaluation of machine
translation." Proceedings of the 40th annual meeting on association for computational
linguistics. Association for Computational Linguistics, 2002.
[41] Doddington, George. "Automatic evaluation of machine translation quality using n-gram
co-occurrence statistics." Proceedings of the second international conference on Human
Language Technology Research. Morgan Kaufmann Publishers Inc., 2002.
[42] Papineni, Kishore, et al. "BLEU: a method for automatic evaluation of machine
translation." Proceedings of the 40th annual meeting on association for computational
linguistics. Association for Computational Linguistics, 2002.
[43] Doddington, George. "Automatic evaluation of machine translation quality using n-gram
co-occurrence statistics." Proceedings of the second international conference on Human
Language Technology Research. Morgan Kaufmann Publishers Inc., 2002.
[44] Hayes, Phillip J. "Intelligent high-volume text processing using shallow, domain-specific
techniques." Text-based intelligent systems: Current research and practice in information
extraction and retrieval (1992): 227-242.
[45] Cohen, William W. "Learning rules that classify e-mail." AAAI spring symposium on
machine learning in information access. Vol. 18. 1996.
[46] Sahami, Mehran, et al. "A Bayesian approach to filtering junk e-mail." Learning for Text
Categorization: Papers from the 1998 workshop. Vol. 62. 1998.
[47] Androutsopoulos, Ion, et al. "Learning to filter spam e-mail: A comparison of a naive
bayesian and a memory-based approach." arXiv preprint cs/0009009 (2000).
[48] Rennie, Jason. "ifile: An application of machine learning to e-mail filtering." Proc. KDD
2000 Workshop on Text Mining, Boston, MA. 2000.
[49] Drucker, Harris, Donghui Wu, and Vladimir N. Vapnik. "Support vector machines for
spam categorization." IEEE Transactions on Neural networks 10.5 (1999): 1048-1054.
[50] Carreras, Xavier, and Lluis Marquez. "Boosting trees for anti-spam email filtering."
arXiv preprint cs/0109015 (2001).
[51] Berger, Adam L., Vincent J. Della Pietra, and Stephen A. Della Pietra. "A maximum
entropy approach to natural language processing." Computational linguistics 22.1 (1996): 39-
71.
[52] Sakkis, Georgios, et al. "Stacking classifiers for anti-spam filtering of e-mail." arXiv
preprint cs/0106040 (2001).
[53] Lewis, David D. "Naive (Bayes) at forty: The independence assumption in information
retrieval." European conference on machine learning. Springer, Berlin, Heidelberg, 1998.
[54] McCallum, Andrew, and Kamal Nigam. "A comparison of event models for naive bayes
text classification." AAAI-98 workshop on learning for text categorization. Vol. 752. No. 1.
1998.
[55] McCallum, Andrew, and Kamal Nigam. "A comparison of event models for naive bayes
text classification." AAAI-98 workshop on learning for text categorization. Vol. 752. No. 1.
1998.
[56] Porter, Martin F. "An algorithm for suffix stripping." Program14.3 (1980): 130-137.
[57] Hayes, Phillip J. "Intelligent high-volume text processing using shallow, domain-specific
techniques." Text-based intelligent systems: Current research and practice in information
extraction and retrieval (1992): 227-242.
[58] Morin, Emmanuel. "Automatic acquisition of semantic relations between terms from
technical corpora." Proc. of the Fifth International Congress on Terminology and Knowledge
Engineering-TKE’99. 1999.
[61] Ahonen, Helena, et al. "Applying data mining techniques for descriptive phrase
extraction in digital document collections." Research and Technology Advances in Digital
Libraries, 1998. ADL 98. Proceedings. IEEE International Forum on. IEEE, 1998.
[62] Zajic, David M., Bonnie J. Dorr, and Jimmy Lin. "Single-document and multi-document
summarization techniques for email threads using sentence compression." Information
Processing & Management 44.4 (2008): 1600-1610.
[63] Fattah, Mohamed Abdel, and Fuji Ren. "GA, MR, FFNN, PNN and GMM based models
for automatic text summarization." Computer Speech & Language 23.1 (2009): 126-144.
[64] Gong, Yihong, and Xin Liu. "Generic text summarization using relevance measure and
latent semantic analysis." Proceedings of the 24th annual international ACM SIGIR
conference on Research and development in information retrieval. ACM, 2001.
[65] Dunlavy, Daniel M., et al. "QCS: A system for querying, clustering and summarizing
documents." Information processing & management 43.6 (2007): 1588-1605.
[66] Wan, Xiaojun. "Using only cross-document relationships for both generic and topic-
focused multi-document summarizations." Information Retrieval 11.1 (2008): 25-49.
[68] Mani, Inderjeet, and Mark T. Maybury. Advances in automatic text summarization. MIT
press, 1999.
[69] Riedhammer, Korbinian, Benoit Favre, and Dilek Hakkani-Tür. "Long story short–
global unsupervised models for keyphrase based meeting summarization." Speech
Communication 52.10 (2010): 801-815.
[72] Fang, Hanyin, et al. "Topic aspect-oriented summarization via group selection."
Neurocomputing 149 (2015): 1613-1619.
[73] Sager, Naomi, et al. "Medical language processing: applications to patient data
representation and automatic encoding." Methods of information in medicine 34.1-2 (1995):
140-146.
[74] Chi, E. C., et al. "A database of computer-structured narrative: methods of computing
complex relations." Proceedings of the Annual Symposium on Computer Application in
Medical Care. American Medical Informatics Association, 1985.
[75] Grishman, Ralph, et al. "The linguistic string parser." Proceedings of the June 4-8, 1973,
national computer conference and exposition. ACM, 1973.
[76] Hirschman, Lynette, Ralph Grishman, and Naomi Sager. "From text to structured
information: automatic processing of medical reports." Proceedings of the June 7-10, 1976,
national computer conference and exposition. ACM, 1976.
[78] Lyman, M., et al. "Computer-structured narrative in ambulatory care: its use in
longitudinal review of clinical data." Proceedings of the Annual Symposium on Computer
Application in Medical Care. American Medical Informatics Association, 1985.
[79] McCray, Alexa T., and Stuart J. Nelson. "The representation of meaning in the UMLS."
Methods of information in medicine34.01/02 (1995): 193-201.
[80] McGray, Alexa T., et al. "The role of lexical knowledge in biomedical text
understanding." Proceedings of the Annual Symposium on Computer Application in Medical
Care. American Medical Informatics Association, 1987.
[81] McCray, Alexa T. "Natural language processing for intelligent information retrieval."
Engineering in Medicine and Biology Society, 1991. Vol. 13: 1991., Proceedings of the
Annual International Conference of the IEEE. IEEE, 1991.
[82] McCray, Alexa T. "Extending a natural language parser with UMLS knowledge."
Proceedings of the Annual Symposium on Computer Application in Medical Care. American
Medical Informatics Association, 1991.
[83] McCray, Alexa T., Suresh Srinivasan, and Allen C. Browne. "Lexical methods for
managing variation in biomedical terminologies." Proceedings of the Annual Symposium on
Computer Application in Medical Care. American Medical Informatics Association, 1994.
[84] McCray, A. T., and A. Razi. "The UMLS Knowledge Source server." Medinfo.
MEDINFO 8 (1995): 144-147.
[85] Scherrer, J-R., et al. "Medical office automation integrated into the distributed
architecture of a hospital information system." Methods of information in medicine 33.02
(1994): 174-179.
[86] Baud, R. H., A-M. Rassinoux, and J-R. Scherrer. "Natural language processing and
semantical representation of medical texts." Methods of information in medicine 31.02
(1992): 117-125.
[87] Lyman, M., et al. "Medical Language Processing for Knowledge Representation and
Retrievals." Proceedings. Symposium on Computer Applications in Medical Care. American
Medical Informatics Association, 1989.
[88] Nhàn, N. T., et al. "A Medical Language Processor for Two Indo-European Languages."
Proceedings. Symposium on Computer Applications in Medical Care. American Medical
Informatics Association, 1989.
[89] Sager, Naomi, et al. "Adapting a medical language processor from English to French."
Medinfo 89 (1989): 795-799.
[90] Borst, F., et al. "Analyse automatique de comptes rendus d'hospitalisation." Degoulet P,
Stephan JC, Venot A, Yvon PJ, rédacteurs. Informatique et Santé, Informatique et Gestion
des Unités de Soins, Comptes Rendus du Colloque AIM-IF, Paris. 1989.
[91] Baud, Robert H., A-M. Rassinoux, and J-R. Scherrer. "Knowledge representation of
discharge summaries." AIME 91. Springer, Berlin, Heidelberg, 1991. 173-182.
[92] Baud, R. H., Laurence Alpay, and Christian Lovis. "Let’s Meet the Users with Natural
Language Understanding." Knowledge and Decisions in Health Telematics: The Next Decade
12 (1994): 103.
[93] Rassinoux, A. M., R. H. Baud, and J. R. Scherrer. "Conceptual graphs model extension
for knowledge representation of medical texts." MEDINFO 92 (1992): 1368-1374.
[95] Rassinoux, Anne-Marie, et al. "Natural language processing of medical texts within the
HELIOS environment." Computer methods and programs in biomedicine 45 (1994): S79-96.
[96] Rassinoux, A-M., et al. "Analysis of medical jargon: The RECIT system." Conference
on Artificial Intelligence in Medicine in Europe. Springer, Berlin, Heidelberg, 1995..
[97] Friedman, Carol, James J. Cimino, and Stephen B. Johnson. "A conceptual model for
clinical radiology reports." Proceedings of the Annual Symposium on Computer Application
in Medical Care. American Medical Informatics Association, 1993.
[98] "Natural Language Processing." Natural Language Processing RSS. N.p., n.d. Web. 23
Mar. 2017.
[99] [Srihari S. Machine Learning: Generative and Discriminative Models. 2010. http://
www.cedar.buffalo.edu/wsrihari/CSE574/Discriminative-Generative.pdf (accessed 31 May
2011).]
[100] [Elkan C. Log-Linear Models and Conditional Random Fields. 2008. http://cseweb.
ucsd.edu/welkan/250B/cikmtutorial.pdf (accessed 28 Jun 2011). 62. Hearst MA, Dumais ST,
Osman E, et al. Support vector machines]
[101] [Jurafsky D, Martin JH. Speech and Language Processing. 2nd edn. Englewood Cliffs,
NJ: Prentice-Hall, 2008.]
[102] [Sonnhammer ELL, Eddy SR, Birney E, et al. Pfam: Multiple sequence alignments and
HMM-profiles of protein domains. Nucleic Acids Res 1998;26:320]
[103] [Sonnhammer, E. L., Eddy, S. R., Birney, E., Bateman, A., & Durbin, R. (1998). Pfam:
multiple sequence alignments and HMM-profiles of protein domains. Nucleic acids
research, 26(1), 320-322]
[104] Systems, RAVN. "RAVN Systems Launch the ACE Powered GDPR Robot - Artificial
Intelligence to Expedite GDPR Compliance." Stock Market. PR Newswire, n.d. Web. 19
Mar. 2017.
[105] "Here's Why Natural Language Processing is the Future of BI." SmartData Collective.
N.p., n.d. Web. 19 Mar. 2017
[106] "Using Natural Language Processing and Network Analysis to Develop a Conceptual
Framework for Medication Therapy Management Research." AMIA ... Annual Symposium
proceedings. AMIA Symposium. U.S. National Library of Medicine, n.d. Web. 19 Mar. 2017
[107] Ogallo, W., & Kanter, A. S. (2017, February 10). Using Natural Language Processing
and Network Analysis to Develop a Conceptual Framework for Medication Therapy
Management Research. Retrieved April 10, 2017, from
https://www.ncbi.nlm.nih.gov/pubmed/28269895?dopt=Abstract
[108] Ochoa, A. (2016, May 25). Meet the Pilot: Smart Earpiece Language Translator.
Retrieved April 10, 2017, from https://www.indiegogo.com/projects/meet-the-pilot-smart-
earpiece-language-translator-headphones-travel