1
0
mirror of https://github.com/danog/liquid.git synced 2024-11-30 10:08:58 +01:00
liquid/parser/scanner.go

64 lines
1.8 KiB
Go
Raw Normal View History

2017-07-07 11:41:37 +02:00
package parser
2017-06-25 17:23:20 +02:00
import (
2017-07-21 04:32:06 +02:00
"fmt"
2017-06-25 17:23:20 +02:00
"regexp"
2017-06-30 14:42:11 +02:00
"strings"
2017-06-25 17:23:20 +02:00
)
2017-07-09 17:18:35 +02:00
// Scan breaks a string into a sequence of Tokens.
2017-07-21 04:32:06 +02:00
func Scan(data string, loc SourceLoc, delims []byte) (tokens []Token) {
// Configure the token matcher to respect the delimeters passed to it
if len(delims) != 3 {
delims = []byte{'{', '}', '%'}
}
objectLeft := string(delims[0]) + string(delims[0])
objectRight := string(delims[1]) + string(delims[1])
tagLeft := string(delims[0]) + string(delims[2])
tagRight := string(delims[2]) + string(delims[1])
var tokenMatcher = regexp.MustCompile(fmt.Sprintf(`%v-?\s*(.+?)\s*-?%v|%v-?\s*(\w+)(?:\s+((?:[^%%]|%%[^}])+?))?\s*-?%v`, objectLeft, objectRight, tagLeft, tagRight))
2017-06-26 18:41:41 +02:00
// TODO error on unterminated {{ and {%
// TODO probably an error when a tag contains a {{ or {%, at least outside of a string
2017-07-16 20:15:11 +02:00
p, pe := 0, len(data)
2017-07-09 17:18:35 +02:00
for _, m := range tokenMatcher.FindAllStringSubmatchIndex(data, -1) {
2017-06-25 17:23:20 +02:00
ts, te := m[0], m[1]
if p < ts {
2017-07-16 20:15:11 +02:00
tokens = append(tokens, Token{Type: TextTokenType, SourceLoc: loc, Source: data[p:ts]})
loc.LineNo += strings.Count(data[p:ts], "\n")
2017-06-25 17:23:20 +02:00
}
2017-06-30 14:42:11 +02:00
source := data[ts:te]
2017-06-25 17:23:20 +02:00
switch data[ts+1] {
2017-07-21 04:32:06 +02:00
case delims[0]:
2017-07-16 20:15:11 +02:00
tok := Token{
Type: ObjTokenType,
2017-07-16 20:15:11 +02:00
SourceLoc: loc,
Source: source,
Args: data[m[2]:m[3]],
2017-07-16 20:15:11 +02:00
TrimLeft: source[2] == '-',
TrimRight: source[len(source)-3] == '-',
}
tokens = append(tokens, tok)
2017-07-21 04:32:06 +02:00
case delims[2]:
2017-07-16 20:15:11 +02:00
tok := Token{
Type: TagTokenType,
2017-07-16 20:15:11 +02:00
SourceLoc: loc,
Source: source,
Name: data[m[4]:m[5]],
2017-07-16 20:15:11 +02:00
TrimLeft: source[2] == '-',
TrimRight: source[len(source)-3] == '-',
2017-06-29 18:20:16 +02:00
}
2017-06-25 17:23:20 +02:00
if m[6] > 0 {
2017-07-16 20:15:11 +02:00
tok.Args = data[m[6]:m[7]]
2017-06-25 17:23:20 +02:00
}
2017-07-16 20:15:11 +02:00
tokens = append(tokens, tok)
2017-06-25 17:23:20 +02:00
}
2017-07-16 20:15:11 +02:00
loc.LineNo += strings.Count(source, "\n")
2017-06-25 23:26:14 +02:00
p = te
2017-06-25 17:23:20 +02:00
}
if p < pe {
2017-07-16 20:15:11 +02:00
tokens = append(tokens, Token{Type: TextTokenType, SourceLoc: loc, Source: data[p:]})
2017-06-25 17:23:20 +02:00
}
2017-07-16 20:15:11 +02:00
return tokens
2017-06-25 17:23:20 +02:00
}