nlp-ner / app.py
mgoin's picture
Update app.py
5ab71e8
raw
history blame
4.42 kB
from deepsparse import Pipeline
import time
import gradio as gr
markdownn = '''
# Named Entity Recognition Pipeline with DeepSparse
Named Entity Recognition is the task of extracting and locating named entities in a sentence. The entities include, people's names, location, organizations, etc.
![Named Entity Recognition Pipeline with DeepSparse](https://huggingface.co/spaces/neuralmagic/nlp-ner/resolve/main/named.png)
## What is DeepSparse?
DeepSparse is sparsity-aware inference runtime offering GPU-class performance on CPUs and APIs to integrate ML into your application. DeepSparse provides sparsified pipelines for computer vision and NLP.
The pipelines are similar to Hugging Face pipelines but are faster because they have been pruned and quantized. SparseML Named Entity Recognition Pipelines integrate with Hugging Face’s Transformers library to enable the sparsification of a large set of transformers models.
### Inference
Here is sample code for a token classification pipeline:
```python
from deepsparse import Pipeline
pipeline = Pipeline.create(task="ner", model_path="zoo:nlp/token_classification/distilbert-none/pytorch/huggingface/conll2003/pruned80_quant-none-vnni")
inference = pipeline(text)
print(inference)
```
## Use case example
The Named Entity Recognition Pipeline can process text before storing the information in a database.
For example, you may want to process text and store the entities in different columns depending on the entity type.
'''
task = "ner"
dense_qa_pipeline = Pipeline.create(
task=task,
model_path="zoo:nlp/token_classification/bert-base/pytorch/huggingface/conll2003/base-none",
)
sparse_qa_pipeline = Pipeline.create(
task=task,
model_path="zoo:nlp/token_classification/bert-base/pytorch/huggingface/conll2003/12layer_pruned80_quant-none-vnni",
)
def map_ner(inference):
entities = []
for item in dict(inference)['predictions'][0]:
dictionary = dict(item)
entity = dictionary['entity']
if entity == "LABEL_0":
value = "O"
elif entity == "LABEL_1":
value = "B-PER"
elif entity == "LABEL_2":
value = "I-PER"
elif entity == "LABEL_3":
value = "-ORG"
elif entity == "LABEL_4":
value = "I-ORG"
elif entity == "LABEL_5":
value = "B-LOC"
elif entity == "LABEL_6":
value = "I-LOC"
elif entity == "LABEL_7":
value = "B-MISC"
else:
value = "I-MISC"
dictionary['entity'] = value
entities.append(dictionary)
return entities
def run_pipeline(text):
dense_start = time.perf_counter()
dense_output = dense_qa_pipeline(text)
dense_entities = map_ner(dense_output)
dense_output = {"text": text, "entities": dense_entities}
dense_end = time.perf_counter()
dense_duration = (dense_end - dense_start) * 1000.0
sparse_start = time.perf_counter()
sparse_output = sparse_qa_pipeline(text)
sparse_entities = map_ner(sparse_output)
sparse_output = {"text": text, "entities": sparse_entities}
sparse_result = dict(sparse_output)
sparse_end = time.perf_counter()
sparse_duration = (sparse_end - sparse_start) * 1000.0
return sparse_output, sparse_duration, dense_output, dense_duration
with gr.Blocks() as demo:
with gr.Row():
with gr.Column():
gr.Markdown(markdownn)
with gr.Column():
gr.Markdown("""
### Named Entity Recognition Demo
Using [token_classification/bert-base](https://sparsezoo.neuralmagic.com/models/nlp%2Ftoken_classification%2Fdistilbert-none%2Fpytorch%2Fhuggingface%2Fconll2003%2Fpruned80_quant-none-vnni)
""")
text = gr.Text(label="Text")
btn = gr.Button("Submit")
gr.Examples( [["We are flying from Texas to California"] ],inputs=[text],)
dense_answers = gr.HighlightedText(label="Dense model answer")
dense_duration = gr.Number(label="Dense Latency (ms):")
sparse_answers = gr.HighlightedText(label="Sparse model answers")
sparse_duration = gr.Number(label="Sparse Latency (ms):")
btn.click(
run_pipeline,
inputs=[text],
outputs=[sparse_answers,sparse_duration,dense_answers,dense_duration],
)
if __name__ == "__main__":
demo.launch()