/
tokenizer.go
60 lines (52 loc) · 943 Bytes
/
tokenizer.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
package internal
import "strings"
type Tokenizer struct {
value string
separator uint8
index int
buf strings.Builder
}
func NewTokenizer(value string, sep uint8) Tokenizer {
return Tokenizer{
value: value,
separator: sep,
index: -1,
}
}
func (t *Tokenizer) HasMoreTokens() bool {
return len(t.value) != t.index
}
func (t *Tokenizer) NextToken() string {
if len(t.value) == t.index {
return ""
}
end := t.index + 1
quoted, escaped := false, false
buf := t.buf
buf.Reset()
for end != len(t.value) {
c := t.value[end]
if c == '"' {
if !escaped {
quoted = !quoted
}
buf.WriteByte(c)
escaped = false
} else {
if escaped || quoted {
buf.WriteByte(c)
escaped = false
} else if c == '\\' {
buf.WriteByte(c)
escaped = true
} else if c == t.separator {
break
} else {
buf.WriteByte(c)
}
}
end++
}
t.index = end
return buf.String()
}