-
Notifications
You must be signed in to change notification settings - Fork 0
/
train.go
233 lines (187 loc) · 5.33 KB
/
train.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
package bpe
import (
"bufio"
"context"
"io"
"strings"
"unicode"
"unicode/utf8"
"github.com/pkg/errors"
)
const (
defaultMaxNumberOfTokens = 50000
defaultMaxTokenLength = 32
maxScanBufferSize = 64 * 1024
BeginOfWord = "<w>"
EndOfWord = "</w>"
BeginOfSentence = "<s>"
EndOfSentence = "</s>"
UnknownToken = "<u>"
)
// Train returns BPE instance with vocabulary learned from source.
func Train(ctx context.Context, source io.Reader, opts ...TrainOption) (*BPE, error) {
options := defaultTrainOptions()
options.Apply(opts...)
tft, err := calculateTokensFrequency(ctx, source, options)
if err != nil {
return nil, err
}
model := newModelFromTokensFrequencyTable(tft, options.MaxNumberOfTokens)
return model, nil
}
func defaultTrainOptions() *trainOptions {
return &trainOptions{
MaxNumberOfTokens: defaultMaxNumberOfTokens,
MaxTokenLength: defaultMaxTokenLength,
ScanBufferSize: maxScanBufferSize,
}
}
type trainOptions struct {
MaxNumberOfTokens int
MaxTokenLength int
ScanBufferSize int
WordsOnly bool
}
func (o *trainOptions) Apply(opts ...TrainOption) {
for _, opt := range opts {
opt(o)
}
}
type TrainOption func(opts *trainOptions)
func WithMaxNumberOfTokens(n int) TrainOption {
return func(opts *trainOptions) {
opts.MaxNumberOfTokens = n
}
}
func WithMaxTokenLength(length int) TrainOption {
return func(opts *trainOptions) {
opts.MaxTokenLength = length
}
}
func WithScanBufferSize(size int) TrainOption {
return func(opts *trainOptions) {
opts.ScanBufferSize = size
}
}
func WithWordsOnly() TrainOption {
return func(opts *trainOptions) {
opts.WordsOnly = true
}
}
type tokensFrequencyTable map[string]int
func calculateTokensFrequency(ctx context.Context, r io.Reader, options *trainOptions) (tokensFrequencyTable, error) {
tokensFrequency := make(tokensFrequencyTable, options.MaxNumberOfTokens) // Approximate size. Avoid extra allocations.
scanner := bufio.NewScanner(r)
scanner.Split(scanSentences)
scanner.Buffer(make([]byte, 0, options.ScanBufferSize), options.ScanBufferSize)
// TODO read in separate threads.
for scanner.Scan() {
select {
case <-ctx.Done():
return nil, ctx.Err()
default:
sentence := scanner.Text()
tokenize(tokensFrequency, sentence, options.MaxTokenLength, options.WordsOnly)
}
}
if err := scanner.Err(); err != nil && err != io.EOF {
return nil, errors.Wrap(err, "file scan")
}
return tokensFrequency, nil
}
const abbreviationLength = 4
// isEndOfSentence checks some heuristics to understand whether it's the end of sentence or not.
func isEndOfSentence(lastSymbol rune, prev, next []byte) bool {
if len(prev) == 0 {
return false
}
switch lastSymbol {
case '\r', '\n', '!', '?':
return true
case '.':
// Sad but dot isn't explicit marker of the end of sentence.
// It can be used for name or other abbreviation.
// Check last runes.
// If they're \w{abbreviationLength} – looks like a Mrs., Dr. or other abbreviation.
// It previous and next rune are numbers – it's a float.
// Otherwise it looks like an end of sentence.
prevRune, width := utf8.DecodeLastRune(prev)
if unicode.IsLetter(prevRune) {
nextAfterCurrent := prevRune // Is needed to check letter capitalization after getting first space.
margin := width
// Let's find the space.
for i := 0; i < abbreviationLength; i++ {
if len(prev)-margin <= 0 {
return false
}
currentRune, currentWidth := utf8.DecodeLastRune(prev[:len(prev)-margin])
// We've found the space. Let's check that the next character after space is a capitalized letter.
if unicode.IsSpace(currentRune) {
return !unicode.IsUpper(nextAfterCurrent)
}
// Is token is inside some group?
if strings.ContainsAny(string(currentRune), `[({"'`) {
return false
}
// If it's not letter it's not an abbreviation.
if !unicode.IsLetter(currentRune) {
return true
}
nextAfterCurrent = currentRune
margin += currentWidth
}
// If the last n characters was letters it's probably is the end of string.
return true
}
if unicode.IsDigit(prevRune) {
if len(next) == 0 {
return true
}
nextRune, _ := utf8.DecodeRune(next)
// Looks like it's a float number.
if unicode.IsDigit(nextRune) {
return false
}
}
return true
}
return false
}
// Preserve Unicode symbols.
func tokenize(tft tokensFrequencyTable, sentence string, maxTokenLength int, wordsOnly bool) {
words := strings.Fields(sentence)
for _, word := range words {
if wordsOnly && !isWord(word) {
continue
}
wordTokens := strings.Split(word, "")
// Add special tokens.
wordTokens[0] = BeginOfWord + wordTokens[0]
wordTokens[len(wordTokens)-1] = wordTokens[len(wordTokens)-1] + EndOfWord
tokenizeWord(tft, wordTokens, maxTokenLength)
}
}
func tokenizeWord(tft tokensFrequencyTable, word []string, maxTokenLength int) {
for i, firstToken := range word {
tft[firstToken]++
b := strings.Builder{}
b.WriteString(firstToken)
for i2, token := range word[i+1:] {
// Current index plus first token.
if i2+1 >= maxTokenLength {
break
}
b.WriteString(token)
tft[b.String()]++
}
}
}
// isWord checks that given string contains only letters or dashes.
func isWord(word string) bool {
for _, r := range word {
if !unicode.IsLetter(r) && r != '-' {
return false
}
}
return true
}