This article outlines the concept and python implementation of Named Entity Recognition using StanfordNERTagger. The technical challenges such as installation issues, version conflict issues, operating system issues that are very common to this analysis are out of scope for this article.
Table of contents:
1. Named Entity Recognition defined
2. Business Use cases
3. Installation Pre-requisites
4. Python Code for implementation
5. Additional Reading: CRF model, Multiple models available in the package
6. Disclaimer
There is a need for NER across multiple domains. Below are a few sample business use cases for your reference.
1.Download Stanford NER from http://nlp.stanford.edu/software/stanford-ner-2015-04-20.zip
Output
StanfordNER algorithm leverages a general implementation of linear chain Conditional Random Fields (CRFs) sequence models. CRFs seem very similar to Hidden Markov Model but are very different.
Below are some key points to note about the CRFs in general.
![]() |
NER NLP using Python |
Table of contents:
1. Named Entity Recognition defined
2. Business Use cases
3. Installation Pre-requisites
4. Python Code for implementation
5. Additional Reading: CRF model, Multiple models available in the package
6. Disclaimer
1. Named Entity Recognition Defined
The process of detecting and classifying proper names mentioned in a text can be defined as Named Entity Recognition (NER). In simple words, it locates person name, organization and location etc. in the content. This is generally the first step in most of the Information Extraction (IE) tasks of Natural Language Processing.
2. Business Use Cases
There is a need for NER across multiple domains. Below are a few sample business use cases for your reference.
- Investment research: To identify the various announcements of the companies, people’s reaction towards them and its impact on the stock prices, one needs to identify people and organisation names in the text
- Chat-bots in multiple domains: To identify places and dates for booking hotel rooms, air tickets etc.
- Insurance domain: Identify and mask people’s names in the feedback forms before analyzing. This is needed for being regulatory compliant(example: HIPAA)
3. Installation Prerequisites
1.Download Stanford NER from http://nlp.stanford.edu/software/stanford-ner-2015-04-20.zip
2.Unzip the zipped folder and save in a drive.
3.Copy the “stanford-ner.jar” from the folder and save it just outside the folder as shown in the image
4.Download the caseless models from https://stanfordnlp.github.io/CoreNLP/history.html by clicking on “caseless” as given below. The models in the first link work as well. However, the caseless models help in identifying named entities even when they are not capitalised as required by formal grammar rules.
5.Save the folder in the same location as the Stanford NER folder for ease of access
![]() |
Stanford NER Installation - Step1 |
NER Installation - Step2 |
4. Python Code for implementation:
#Import all the required libraries.import osfrom nltk.tag import StanfordNERTaggerimport pandas as pd#Set environmental variables programmatically.#Set the classpath to the path where the jar file is locatedos.environ['CLASSPATH'] = "<path to the file>/stanford-ner-2015-04-20/stanford-ner.jar"
#Set the Stanford models to the path where the models are storedos.environ['STANFORD_MODELS'] = '<path to the file>/stanford-corenlp-caseless-2015-04-20-models/edu/stanford/nlp/models/ner'
#Set the java jdk pathjava_path = "C:/Program Files/Java/jdk1.8.0_161/bin/java.exe"os.environ['JAVAHOME'] = java_path#Set the path to the model that you would like to usestanford_classifier = '<path to the file>/stanford-corenlp-caseless-2015-04-20-models/edu/stanford/nlp/models/ner/english.all.3class.caseless.distsim.crf.ser.gz'#Build NER tagger objectst = StanfordNERTagger(stanford_classifier)#A sample text for NER taggingtext = 'srinivas ramanujan went to the united kingdom. There he studied at cambridge university.'#Tag the sentence and print outputtagged = st.tag(str(text).split())print(tagged)
Output
[(u'srinivas', u'PERSON'),
(u'ramanujan', u'PERSON'),
(u'went', u'O'),
(u'to', u'O'),
(u'the', u'O'),
(u'united', u'LOCATION'),
(u'kingdom.', u'LOCATION'),
(u'There', u'O'),
(u'he', u'O'),
(u'studied', u'O'),
(u'at', u'O'),
(u'cambridge', u'ORGANIZATION'),
(u'university', u'ORGANIZATION')]
5. Additional Reading
StanfordNER algorithm leverages a general implementation of linear chain Conditional Random Fields (CRFs) sequence models. CRFs seem very similar to Hidden Markov Model but are very different.
Below are some key points to note about the CRFs in general.
- It is a discriminative model unlike the HMM model and thus models the conditional probability
- It does not assume independence of features unlike the HMM model. This means that the current word, previous word, next word are all considered for model as features
- Relative to HMM or Max ent Markov Models, CRFs are the slowest
6. Disclaimer
This article explains the implementation of StanfordNER algorithm for research purposes and does not promote it for commercial gains. For any questions on commercial aspects of implementing this algorithm, please contact Stanford University