forked from TurkuNLP/Finnish-dep-parser
-
Notifications
You must be signed in to change notification settings - Fork 1
/
limit_sentence_len.py
146 lines (123 loc) · 5.88 KB
/
limit_sentence_len.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
from visualize import read_conll
import sys
import codecs
try:
import argparse
except ImportError:
import compat.argparse as argparse
out8=codecs.getwriter("utf-8")(sys.stdout)
ID,FORM,LEMMA,PLEMMA,POS,PPOS,FEAT,PFEAT,HEAD,PHEAD,DEPREL,PDEPREL=range(12)
def cutCandidates(sent):
#Cut-point candidates for sent (list of CoNLL09 columns)
candidates=[] #list of: (score, position of cut (new sent starts here))
for lIdx in range(args.leeway,len(sent)-args.leeway): #Don't want to cut too short
l=sent[lIdx]
if l[FORM] in (u";",): #semicolon is a pretty decent candidate
candidates.append((0,lIdx+1))
elif l[FORM]==u")" and sent[lIdx-1][FORM].isdigit() and sent[lIdx-2][FORM]==u",": #lists like ", 2 ) ..."
candidates.append((5, lIdx-1))
elif l[FORM] in u":\u2012\u2013\u2014\u2015\u2053~----": #colon and dashes are a worse candidate still. The \u chars are different dashes
candidates.append((10, lIdx+1))
elif l[FORM] in (u",",): #and comma is the worst candidate
candidates.append((20, lIdx+1))
return candidates
def cutPointsNearPlaces(nearPlaces,cutCandidates):
cutPoints=[]
for p in nearPlaces:
cands=[(ccPenalty,abs(cc-p),cc) for (ccPenalty,cc) in cutCandidates if cc>=p-args.leeway and cc<=p+args.leeway] #list all candidates near cut points
cands.sort()
if cands:
cutPoints.append(cands[0][2]) #This is the best place to cut around this point
else:
cutPoints.append(p) #Meh, we're screwed, just cut and never look back
return cutPoints
def cutSent(sent):
nearPlaces=list(range(args.chunk_size,len(sent),args.chunk_size)) #The places nearby which a cut should ideally happen
assert len(nearPlaces)>0, len(sent)
candidates=cutCandidates(sent)
cutPoints=cutPointsNearPlaces(nearPlaces,candidates)
if len(sent)-cutPoints[-1]<args.leeway:
cutPoints.pop(-1) #We don't want to have too short final segment
subsents=[]
prev=0
for x in cutPoints+[len(sent)]:
#create a new sub-sentence
subsents.append(sent[prev:x])
#...renumber tokens
newSubSent=subsents[-1]
for idx in range(len(newSubSent)):
newSubSent[idx][ID]=unicode(idx+1)
prev=x
return subsents
def print_sent(sent,comments,add_newline=True):
if comments:
print >> out8, u"\n".join(comments)
print >> out8, u"\n".join(u"\t".join(cols) for cols in sent)
if add_newline:
print >> out8
def renumber(sent,offset):
#Renumber all tokens and head references by a given offset (can be negative)
for cols in sent:
for col in (ID,HEAD,PHEAD): #renumber these columns, unless they are a "_"
if cols[col]!=u"_":
cols[col]=unicode(int(cols[col])+offset)
def get_root(sent):
"""
On the off-chance there'd be several roots, let's dep them together and return the first one
"""
r=[] #list of 0-based indices into sent pointing to root rows
for idx,cols in enumerate(sent):
if cols[HEAD]==u"0":
r.append(idx)
if len(r)>1: #Ooops - this really happens?
for i in range(1,len(r)):
gov,d_row=sent[r[i-1]][ID],sent[r[i]] #gov is the string with ID
d_row[HEAD]=gov
d_row[PHEAD]=gov
d_row[DEPREL]=u"dep"
d_row[PDEPREL]=u"dep"
return int(sent[r[0]][ID])-1 #the row index of the root
def cutAndPrintSentence(sent,comments):
subsents=cutSent(sent)
#The first sub-sentence will take on the comment of the whole sentence
print_sent(subsents[0],comments)
#and every other will have a comment marking it as a continuation of the prev. one
for x in subsents[1:]:
assert len(x)>0
print_sent(x,[u"#---sentence---splitter---JOIN-TO-PREVIOUS-SENTENCE---"])
if __name__=="__main__":
parser = argparse.ArgumentParser(description='Split/merge long sentences. Use --reverse to merge.')
parser.add_argument('--reverse', default=False, action="store_true", help='Reverse the splitting.')
parser.add_argument('-N', '--max-len', type=int, default=120, help='Pass sentences shorter or equal to this number of tokens through, split the rest. This will also be the absolute maximum chunk size ever fed into the parser. Default %(default)d.')
parser.add_argument('-C', '--chunk-size', type=int, default=80, help='Split into chunks of approximately this size. Default %(default)d.')
parser.add_argument('input', nargs='?', help='Input. Nothing or "-" for stdin.')
args = parser.parse_args()
args.leeway=args.chunk_size//3 #TODO - better value maybe?
if args.reverse:
last_len=None
last_root=None
for sent,comments in read_conll(args.input,0):
if len(comments)==1 and comments[0]==u"#---sentence---splitter---JOIN-TO-PREVIOUS-SENTENCE---":
part_root=get_root(sent) #root of this chunk
renumber(sent,last_len)
sent[part_root][HEAD]=unicode(last_root+1) #...dep to the previous one
sent[part_root][PHEAD]=unicode(last_root+1)
sent[part_root][DEPREL]=u"dep"
sent[part_root][PDEPREL]=u"dep"
last_root=int(sent[part_root][ID])-1 #...and remember the renumbered root for the possible next one
print_sent(sent,[],False)
last_len+=len(sent)
else:
if last_len is not None:
print >> out8
last_root=get_root(sent)
print_sent(sent,comments,False)
last_len=len(sent)
else:
print >> out8
else:
for sent,comments in read_conll(args.input,0):
if len(sent)<=args.max_len: #this one's good
print_sent(sent,comments)
else:
cutAndPrintSentence(sent,comments)