mirror of
https://gitlab.ub.uni-bielefeld.de/sfb1288inf/nlp.git
synced 2024-12-25 20:14:17 +00:00
Fix
This commit is contained in:
parent
3fc6ebff4c
commit
887e814020
@ -52,7 +52,7 @@ with open(args.i, encoding=encoding) as input_file:
|
||||
text = input_file.read()
|
||||
# spaCys NLP is limited to strings with maximum 1 million characters at
|
||||
# once. So we split it into suitable chunks.
|
||||
text_chunks = textwrap.wrap(text, 1000, break_long_words=False)
|
||||
text_chunks = textwrap.wrap(text, 1000000, break_long_words=False)
|
||||
# the text variable potentially occupies a lot of system memory and is no
|
||||
# longer needed...
|
||||
del text
|
||||
@ -66,7 +66,6 @@ nlp = spacy.load(SPACY_MODELS[args.language])
|
||||
# See: http://cwb.sourceforge.net/files/CWB_Encoding_Tutorial/node3.html
|
||||
output_file_original_filename = args.o
|
||||
output_file_stand_off_filename = args.o.replace('.vrt', '.stand-off.vrt')
|
||||
output_file_tokens_filename = args.o.replace('.vrt', '.tokens.txt')
|
||||
xml_head = '''<?xml version="1.0" encoding="UTF-8"?>\n\
|
||||
<corpus>\n\
|
||||
<text>\n\
|
||||
@ -78,12 +77,10 @@ xml_head = '''<?xml version="1.0" encoding="UTF-8"?>\n\
|
||||
spacy_model=SPACY_MODELS[args.language])
|
||||
|
||||
with open(output_file_original_filename, 'w+') as output_file_original, \
|
||||
open(output_file_stand_off_filename, 'w+') as output_file_stand_off, \
|
||||
open(output_file_tokens_filename, 'w+') as output_file_tokens:
|
||||
open(output_file_stand_off_filename, 'w+') as output_file_stand_off:
|
||||
|
||||
output_file_original.write(xml_head)
|
||||
output_file_stand_off.write(xml_head)
|
||||
output_file_tokens.write(xml_head)
|
||||
text_offset = 0
|
||||
for text_chunk in text_chunks:
|
||||
doc = nlp(text_chunk)
|
||||
@ -110,10 +107,8 @@ with open(output_file_original_filename, 'w+') as output_file_original, \
|
||||
+ '\t{}'.format(token.pos_)
|
||||
+ '\t{}'.format(token.tag_)
|
||||
+ '\t{}\n'.format(token.ent_type_ or 'NULL'))
|
||||
output_file_tokens.write('{}\n'.format(escape(token.text)))
|
||||
output_file_original.write('</s>\n')
|
||||
output_file_stand_off.write('</s>\n')
|
||||
text_offset = token_end + 1
|
||||
output_file_original.write('</metadata>\n</text>\n</corpus>')
|
||||
output_file_stand_off.write('</metadata>\n</text>\n</corpus>')
|
||||
output_file_tokens.write('</metadata>\n</text>\n</corpus>')
|
||||
|
Loading…
Reference in New Issue
Block a user