Natural Language Processing with spaCy
Azadeh Mobasher
Principal Data Scientist
spaCy first tokenizes the text to produce a Doc objectDoc is processed in several different steps of processing pipeline
import spacy
nlp = spacy.load("en_core_web_sm")
doc = nlp(example_text)
spaCy NER pipeline:print([ent.text for ent in doc.ents])
sentencizer: spaCy pipeline component for sentence segmentation.text = " ".join(["This is a test sentence."]*10000)en_core_sm_nlp = spacy.load("en_core_web_sm") start_time = time.time() doc = en_core_sm_nlp(text)print(f"Finished processing with en_core_web_sm model in {round((time.time() - start_time)/60.0 , 5)} minutes")
>>> Finished processing with en_core_web_sm model in 0.09332 minutes
sentencizer pipe:blank_nlp = spacy.blank("en")blank_nlp.add_pipe("sentencizer")start_time = time.time() doc = blank_nlp(text) print(f"Finished processing with blank model in {round((time.time() - start_time)/60.0 , 5)} minutes")
>>> Finished processing with blank model in 0.00091 minutes
nlp.analyze_pipes() analyzes a spaCy pipeline to determine:
pretty to True will print a table instead of only returning the structured data.import spacy
nlp = spacy.load("en_core_web_sm")
analysis = nlp.analyze_pipes(pretty=True)
Natural Language Processing with spaCy