{% extends "base.html.j2" %} {% from "services/_breadcrumbs.html.j2" import breadcrumbs with context %} {% import "materialize/wtf.html.j2" as wtf %} {% block main_attribs %} class="service-scheme" data-service="spacy-nlp-pipeline"{% endblock main_attribs %} {% block page_content %}

{{ title }}

 

 

layersTokenization

Your text is split up into sentences and words, so called tokens, which can then be analyzed.

layersLemmatization

All inflected forms of a word are grouped together so that it can be analyzed as a single item.

layersPart-of-speech Tagging

In accordance with its definition and context, each word is marked up as corresponding to a particular part of speech.

layersNamed-Entity Recognition

Named entities are located and classified into specific categories like persons or locations.

Submit a job

{{ form.hidden_tag() }}
{{ wtf.render_field(form.title, material_icon='title') }}
{{ wtf.render_field(form.description, material_icon='description') }}
{{ wtf.render_field(form.txt, accept='text/plain', placeholder='Choose a plain text file') }}
{{ wtf.render_field(form.model, material_icon='language') }}
{{ wtf.render_field(form.version, material_icon='apps') }}
Preprocessing

{{ form.encoding_detection.label.text }}

If the input files are not created with the nopaque OCR service or you do not know if your text files are UTF-8 encoded, check this switch. We will try to automatically determine the right encoding for your texts to process them.

{{ wtf.render_field(form.submit, material_icon='send') }}
{% endblock page_content %} {% block modals %} {{ super() }} {% endblock modals %} {% block scripts %} {{ super() }} {% endblock scripts %}