-
Notifications
You must be signed in to change notification settings - Fork 71
/
content_processor.py
98 lines (87 loc) · 4.19 KB
/
content_processor.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
from apply_bpe import BPE
from mosestokenizer import MosesSentenceSplitter, MosesPunctuationNormalizer, MosesTokenizer, MosesDetokenizer
import sentencepiece
class ContentProcessor():
def __init__(self, srclang,
targetlang, sourcebpe=None, targetbpe=None,sourcespm=None,targetspm=None):
self.bpe_source = None
self.bpe_target = None
self.sp_processor_source = None
self.sp_processor_target = None
self.sentences=[]
# load BPE model for pre-processing
if sourcebpe:
# print("load BPE codes from " + sourcebpe, flush=True)
BPEcodes = open(sourcebpe, 'r', encoding="utf-8")
self.bpe_source = BPE(BPEcodes)
if targetbpe:
# print("load BPE codes from " + targetbpe, flush=True)
BPEcodes = open(targetbpe, 'r', encoding="utf-8")
self.bpe_target = BPE(BPEcodes)
# load SentencePiece model for pre-processing
if sourcespm:
# print("load sentence piece model from " + sourcespm, flush=True)
self.sp_processor_source = sentencepiece.SentencePieceProcessor()
self.sp_processor_source.Load(sourcespm)
if targetspm:
# print("load sentence piece model from " + targetspm, flush=True)
self.sp_processor_target = sentencepiece.SentencePieceProcessor()
self.sp_processor_target.Load(targetspm)
# pre- and post-processing tools
self.tokenizer = None
self.detokenizer = None
# TODO: should we have support for other sentence splitters?
# print("start pre- and post-processing tools")
self.sentence_splitter = MosesSentenceSplitter(srclang)
self.normalizer = MosesPunctuationNormalizer(srclang)
if self.bpe_source:
self.tokenizer = MosesTokenizer(srclang)
self.detokenizer = MosesDetokenizer(targetlang)
def preprocess(self, srctxt):
normalized_text = '\n'.join(self.normalizer(line) for line in srctxt.split('\n')) # normalizer do not accept '\n'
sentSource = self.sentence_splitter([normalized_text])
self.sentences=[]
for s in sentSource:
if self.tokenizer:
# print('raw sentence: ' + s, flush=True)
tokenized = ' '.join(self.tokenizer(s))
# print('tokenized sentence: ' + tokenized, flush=True)
segmented = self.bpe_source.process_line(tokenized)
elif self.sp_processor_source:
print('raw sentence: ' + s, flush=True)
segmented = ' '.join(self.sp_processor_source.EncodeAsPieces(s))
# print(segmented, flush=True)
else:
raise RuntimeError("No tokenization / segmentation method defines, can't preprocess")
self.sentences.append(segmented)
return self.sentences
def postprocess(self, recievedsentences):
sentTranslated = []
for index, s in enumerate(recievedsentences):
received = s.strip().split(' ||| ')
# print(received, flush=True)
# undo segmentation
if self.bpe_source:
translated = received[0].replace('@@ ','')
elif self.sp_processor_target:
translated = self.sp_processor_target.DecodePieces(received[0].split(' '))
else:
translated = received[0].replace(' ','').replace('▁',' ').strip()
alignment = ''
if len(received) == 2:
alignment = received[1]
links = alignment.split(' ')
fixedLinks = []
outputLength = len(received[0].split(' '))
for link in links:
ids = link.split('-')
if ids[0] != '-1' and int(ids[0])<len(self.sentences[index]):
if int(ids[1])<outputLength:
fixedLinks.append('-'.join(ids))
alignment = ' '.join(fixedLinks)
if self.detokenizer:
detokenized = self.detokenizer(translated.split())
else:
detokenized = translated
sentTranslated.append(detokenized)
return sentTranslated