Named Entity Recognition using Transformers and Spacy in Python

Learn how you can perform named entity recognition using HuggingFace Transformers and spaCy libraries in Python.
  · 11 min read · Updated may 2022 · Machine Learning · Natural Language Processing

Welcome! Meet our Python Code Assistant, your new coding buddy. Why wait? Start exploring now!

Open In Colab

Named Entity Recognition (NER) is a typical natural language processing (NLP) task that automatically identifies and recognizes predefined entities in a given text. Entities like person names, organizations, dates and times, and locations are valuable information to extract from unstructured and unlabeled raw text.

At the end of this tutorial, you will be able to perform named entity recognition on any given English text with HuggingFace Transformers and SpaCy in Python; here's an example of the resulting NER:

Named Entity Recognition

SpaCy is an open-source library in Python for advanced NLP. It is built on the latest research and designed to be used in real-world products. We'll be using two NER models on SpaCy, namely the regular en_core_web_sm and the transformer en_core_web_trf. We'll also use spaCy's NER amazing visualizer.

To get started, let's install the required libraries for this tutorial. First, installing transformers:

$ pip install --upgrade transformers sentencepiece

Next, we need to install spacy and spacy-transformers. To do that, I've grabbed the latest .whl file from the spacy-models releases for installation:

$ pip install https://github.com/explosion/spacy-models/releases/download/en_core_web_trf-3.2.0/en_core_web_trf-3.2.0-py3-none-any.whl

Of course, if you're reading this tutorial in the future, make sure to get the latest release from this page if you encounter any problems regarding the above command.

Next, we have to download the spaCy's en_core_web_sm regular model:

$ python -m spacy download en_core_web_sm

en_core_web_sm is an English model pipeline optimized for CPU. It is small, with only 13MB in size, and under the MIT license. For larger models, you can use en_core_web_md for medium-sized and en_core_web_lg for the large one.

Once done with the installation, let's get started with the code:

import spacy
from transformers import *

For this tutorial, we'll be performing NER on this text that I've grabbed from Wikipedia:

# sample text from Wikipedia
text = """
Albert Einstein was a German-born theoretical physicist, widely acknowledged to be one of the greatest and most influential physicists of all time. 
Einstein is best known for developing the theory of relativity, but he also made important contributions to the development of the theory of quantum mechanics.
Einstein was born in the German Empire, but moved to Switzerland in 1895, forsaking his German citizenship (as a subject of the Kingdom of Württemberg) the following year. 
In 1897, at the age of 17, he enrolled in the mathematics and physics teaching diploma program at the Swiss Federal polytechnic school in Zürich, graduating in 1900
"""

NER with Transformers

We'll be using the HuggingFace Transformers' pipeline API for loading the models:

# load BERT model fine-tuned for Named Entity Recognition (NER)
ner = pipeline("ner", model="dslim/bert-base-NER")

We're using a BERT model (bert-base-encased) that was fine-tuned on the CoNLL-2003 Named Entity Recognition dataset. You can use dslim/bert-large-NER for a larger version of this one.

Let's extract the entities for our text using this model:

# perform inference on the transformer model
doc_ner = ner(text)
# print the output
doc_ner

Output:

[{'end': 7,
  'entity': 'B-PER',
  'index': 1,
  'score': 0.99949145,
  'start': 1,
  'word': 'Albert'},
 {'end': 16,
  'entity': 'I-PER',
  'index': 2,
  'score': 0.998417,
  'start': 8,
  'word': 'Einstein'},
 {'end': 29,
  'entity': 'B-MISC',
  'index': 5,
  'score': 0.99211043,
  'start': 23,
  'word': 'German'},
 {'end': 158,
  'entity': 'B-PER',
  'index': 28,
  'score': 0.99736506,
  'start': 150,
  'word': 'Einstein'},
 {'end': 318,
  'entity': 'B-PER',
  'index': 55,
  'score': 0.9977113,
  'start': 310,
  'word': 'Einstein'},
 {'end': 341,
  'entity': 'B-LOC',
  'index': 60,
  'score': 0.50242233,
  'start': 335,
  'word': 'German'},
 {'end': 348,
  'entity': 'I-LOC',
  'index': 61,
  'score': 0.95330054,
  'start': 342,
  'word': 'Empire'},
 {'end': 374,
  'entity': 'B-LOC',
  'index': 66,
  'score': 0.99978524,
  'start': 363,
  'word': 'Switzerland'},
 {'end': 404,
  'entity': 'B-MISC',
  'index': 74,
  'score': 0.9995827,
  'start': 398,
  'word': 'German'},
 {'end': 460,
  'entity': 'B-LOC',
  'index': 84,
  'score': 0.9994709,
  'start': 449,
  'word': 'Württemberg'},
 {'end': 590,
  'entity': 'B-MISC',
  'index': 111,
  'score': 0.9888771,
  'start': 585,
  'word': 'Swiss'},
 {'end': 627,
  'entity': 'B-LOC',
  'index': 119,
  'score': 0.9977405,
  'start': 621,
  'word': 'Zürich'}]

As you can see, the output is a list of dictionaries that has the start and end positions of the entity in the text, the prediction score, the word itself, the index, and the entity name.

The named entities of this datasets are:

  • O: Outside of a named entity.
  • B-MIS: Beginning of a miscellaneous entity right after another miscellaneous entity.
  • I-MIS: Miscellaneous entity.
  • B-PER: Beginning of a person’s name right after another person’s name.
  • I-PER: Person’s name.
  • B-ORG: The beginning of an organization right after another organization.
  • I-ORG: Organization.
  • B-LOC: Beginning of a location right after another location.
  • I-LOC: Location.

Next, let's make a function that uses spaCy to visualize this Python dictionary:

def get_entities_html(text, ner_result, title=None):
  """Visualize NER with the help of SpaCy"""
  ents = []
  for ent in ner_result:
    e = {}
    # add the start and end positions of the entity
    e["start"] = ent["start"]
    e["end"] = ent["end"]
    # add the score if you want in the label
    # e["label"] = f"{ent["entity"]}-{ent['score']:.2f}"
    e["label"] = ent["entity"]
    if ents and -1 <= ent["start"] - ents[-1]["end"] <= 1 and ents[-1]["label"] == e["label"]:
      # if the current entity is shared with previous entity
      # simply extend the entity end position instead of adding a new one
      ents[-1]["end"] = e["end"]
      continue
    ents.append(e)
  # construct data required for displacy.render() method
  render_data = [
    {
      "text": text,
      "ents": ents,
      "title": title,
    }
  ]
  spacy.displacy.render(render_data, style="ent", manual=True, jupyter=True)

The above function uses the spacy.displacy.render() function to render a named entity extracted text. We use manual=True indicating that's a manual visualization and not a spaCy document. We also set jupyter to True as we're currently on a Juypter notebook or Colab.

The whole purpose of the for loop is to construct a list of dictionaries with the start and end positions, and the entity's label. We also check to see if there are some same entities nearby, so we combine them.

Let's call it:

# get HTML representation of NER of our text
get_entities_html(text, doc_ner)
AlbertB-PER EinsteinI-PER was a GermanB-MISC-born theoretical physicist, widely acknowledged to be one of the greatest and most influential physicists of all time. EinsteinB-PER is best known for developing the theory of relativity, but he also made important contributions to the development of the theory of quantum mechanics.EinsteinB-PER was born in the GermanB-LOC EmpireI-LOC, but moved to SwitzerlandB-LOC in 1895, forsaking his GermanB-MISC citizenship (as a subject of the Kingdom of WürttembergB-LOC) the following year. In 1897, at the age of 17, he enrolled in the mathematics and physics teaching diploma program at the SwissB-MISC Federal polytechnic school in ZürichB-LOC, graduating in 1900

Next, let's load another relatively larger and better model that is based on roberta-large:

# load roberta-large model
ner2 = pipeline("ner", model="xlm-roberta-large-finetuned-conll03-english")

Performing inference:

# perform inference on this model
doc_ner2 = ner2(text)

Visualizing:

# get HTML representation of NER of our text
get_entities_html(text, doc_ner2)
Albert EinsteinI-PER was a GermanI-MISC-born theoretical physicist, widely acknowledged to be one of the greatest and most influential physicists of all time. EinsteinI-PER is best known for developing the theory of relativity, but he also made important contributions to the development of the theory of quantum mechanics.EinsteinI-PER was born in the German EmpireI-LOC, but moved to SwitzerlandI-LOC in 1895, forsaking his GermanI-MISC citizenship (as a subject of the Kingdom of WürttembergI-LOC) the following year. In 1897, at the age of 17, he enrolled in the mathematics and physics teaching diploma program at the SwissI-MISC FederalI-ORG polytechnic school in ZürichI-LOC, graduating in 1900

As you can see, now it's improved, naming Albert Einstein as a single entity and also the Kingdom of Wurttemberg.

There are a lot of other models that were fine-tuned on the same dataset. Here's yet another one:

# load yet another roberta-large model
ner3 = pipeline("ner", model="Jean-Baptiste/roberta-large-ner-english")
# perform inference on this model
doc_ner3 = ner3(text)
# get HTML representation of NER of our text
get_entities_html(text, doc_ner3)
Albert EinsteinPER was a German-bornMISC theoretical physicist, widely acknowledged to be one of the greatest and most influential physicists of all time. EinsteinPER is best known for developing the theory of relativity, but he also made important contributions to the development of the theory of quantum mechanics.EinsteinPER was born in the German EmpireLOC, but moved to SwitzerlandLOC in 1895, forsaking his GermanMISC citizenship (as a subject of the Kingdom of WürttembergLOC) the following year. In 1897, at the age of 17, he enrolled in the mathematics and physics teaching diploma program at the SwissMISC FederalORG polytechnic school in ZürichLOC, graduating in 1900

This model, however, only has PER, MISC, LOC, and ORG entities. SpaCy automatically colors the familiar entities.

NER with SpaCy

To perform NER using SpaCy, we must first load the model using spacy.load() function:

# load the English CPU-optimized pipeline
nlp = spacy.load("en_core_web_sm")

We're loading the model we've downloaded. Make sure you download the model you want to use before loading it here. Next, let's generate our document:

# predict the entities
doc = nlp(text)

And then visualizing it:

# display the doc with jupyter mode
spacy.displacy.render(doc, style="ent", jupyter=True)
Albert EinsteinPERSON was a GermanNORP-born theoretical physicist, widely acknowledged to be one of the greatest and most influential physicists of all time. EinsteinPERSON is best known for developing the theory of relativity, but he also made important contributions to the development of the theory of quantum mechanicsORG.EinsteinPERSON was born in the German EmpireGPE, but moved to SwitzerlandGPE in 1895DATE, forsaking his GermanNORP citizenship (as a subject of the Kingdom of WürttembergGPE) the following yearDATE. In 1897DATE, at the age of 17DATE, he enrolled in the mathematics and physics teaching diploma program at the SwissNORP Federal polytechnic school in ZürichGPE, graduating in 1900DATE

This one looks much better, and there are a lot more entities (18) than the previous ones, namely CARDINAL, DATE, EVENT, FAC, GPE, LANGUAGE, LAW, LOC, MONEY, NORP, ORDINAL, ORG, PERCENT, PERSON, PRODUCT, QUANTITY, TIME, WORK_OF_ART.

However, quantum mechanics was mistakenly labeled as an organization, so let's use the Transformer model that spaCy is offering:

# load the English transformer pipeline (roberta-base) using spaCy
nlp_trf = spacy.load('en_core_web_trf')

Let's perform inference and visualize the text:

# perform inference on the model
doc_trf = nlp_trf(text)
# display the doc with jupyter mode
spacy.displacy.render(doc_trf, style="ent", jupyter=True)
Albert EinsteinPERSON was a GermanNORP-born theoretical physicist, widely acknowledged to be one of the greatest and most influential physicists of all time. EinsteinPERSON is best known for developing the theory of relativity, but he also made important contributions to the development of the theory of quantum mechanics.EinsteinPERSON was born in the German EmpireGPE, but moved to SwitzerlandGPE in 1895DATE, forsaking his GermanNORP citizenship (as a subject of the Kingdom of WürttembergGPE) the following yearDATE. In 1897DATE, at the age of 17DATE, he enrolled in the mathematics and physics teaching diploma program at the Swiss FederalORG polytechnic school in ZürichGPE, graduating in 1900DATE

This time Swiss Federal was labeled as an organization, even though it wasn't complete (it should be Swiss Federal polytechnic school), and quantum mechanics is no longer an organization.

The en_core_web_trf model performs much better than the previous ones. Check this table that shows each English model offered by spaCy with their size and metrics evaluation of each:

Model Name Model Size Precision Recall F-Score
en_core_web_sm 13MB 0.85 0.84 0.84
en_core_web_md 43MB 0.85 0.84 0.85
en_core_web_lg 741MB 0.86 0.85 0.85
en_core_web_trf 438MB 0.90 0.90 0.90

 

Conclusion

Make sure you try other types of texts and see for yourself if your text confirms the above table! You can check this page on spaCy to see the details of each model.

For other languages, spaCy strives to make these models available for every language globally. You can check this page to see the available models for each language.

Here are some related NLP tutorials that you may find useful:

Open In Colab

Happy learning ♥

Why juggle between languages when you can convert? Check out our Code Converter. Try it out today!

View Full Code Explain The Code for Me
Sharing is caring!



Read Also



Comment panel

    Got a coding query or need some guidance before you comment? Check out this Python Code Assistant for expert advice and handy tips. It's like having a coding tutor right in your fingertips!