mirror of https://github.com/gogits/gogs.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
859 lines
21 KiB
859 lines
21 KiB
// Copyright 2013 The Go Authors. All rights reserved. |
|
// Use of this source code is governed by a BSD-style |
|
// license that can be found in the LICENSE file. |
|
|
|
package language |
|
|
|
import ( |
|
"bytes" |
|
"errors" |
|
"fmt" |
|
"sort" |
|
"strconv" |
|
"strings" |
|
|
|
"golang.org/x/text/internal/tag" |
|
) |
|
|
|
// isAlpha returns true if the byte is not a digit. |
|
// b must be an ASCII letter or digit. |
|
func isAlpha(b byte) bool { |
|
return b > '9' |
|
} |
|
|
|
// isAlphaNum returns true if the string contains only ASCII letters or digits. |
|
func isAlphaNum(s []byte) bool { |
|
for _, c := range s { |
|
if !('a' <= c && c <= 'z' || 'A' <= c && c <= 'Z' || '0' <= c && c <= '9') { |
|
return false |
|
} |
|
} |
|
return true |
|
} |
|
|
|
// errSyntax is returned by any of the parsing functions when the |
|
// input is not well-formed, according to BCP 47. |
|
// TODO: return the position at which the syntax error occurred? |
|
var errSyntax = errors.New("language: tag is not well-formed") |
|
|
|
// ValueError is returned by any of the parsing functions when the |
|
// input is well-formed but the respective subtag is not recognized |
|
// as a valid value. |
|
type ValueError struct { |
|
v [8]byte |
|
} |
|
|
|
func mkErrInvalid(s []byte) error { |
|
var e ValueError |
|
copy(e.v[:], s) |
|
return e |
|
} |
|
|
|
func (e ValueError) tag() []byte { |
|
n := bytes.IndexByte(e.v[:], 0) |
|
if n == -1 { |
|
n = 8 |
|
} |
|
return e.v[:n] |
|
} |
|
|
|
// Error implements the error interface. |
|
func (e ValueError) Error() string { |
|
return fmt.Sprintf("language: subtag %q is well-formed but unknown", e.tag()) |
|
} |
|
|
|
// Subtag returns the subtag for which the error occurred. |
|
func (e ValueError) Subtag() string { |
|
return string(e.tag()) |
|
} |
|
|
|
// scanner is used to scan BCP 47 tokens, which are separated by _ or -. |
|
type scanner struct { |
|
b []byte |
|
bytes [max99thPercentileSize]byte |
|
token []byte |
|
start int // start position of the current token |
|
end int // end position of the current token |
|
next int // next point for scan |
|
err error |
|
done bool |
|
} |
|
|
|
func makeScannerString(s string) scanner { |
|
scan := scanner{} |
|
if len(s) <= len(scan.bytes) { |
|
scan.b = scan.bytes[:copy(scan.bytes[:], s)] |
|
} else { |
|
scan.b = []byte(s) |
|
} |
|
scan.init() |
|
return scan |
|
} |
|
|
|
// makeScanner returns a scanner using b as the input buffer. |
|
// b is not copied and may be modified by the scanner routines. |
|
func makeScanner(b []byte) scanner { |
|
scan := scanner{b: b} |
|
scan.init() |
|
return scan |
|
} |
|
|
|
func (s *scanner) init() { |
|
for i, c := range s.b { |
|
if c == '_' { |
|
s.b[i] = '-' |
|
} |
|
} |
|
s.scan() |
|
} |
|
|
|
// restToLower converts the string between start and end to lower case. |
|
func (s *scanner) toLower(start, end int) { |
|
for i := start; i < end; i++ { |
|
c := s.b[i] |
|
if 'A' <= c && c <= 'Z' { |
|
s.b[i] += 'a' - 'A' |
|
} |
|
} |
|
} |
|
|
|
func (s *scanner) setError(e error) { |
|
if s.err == nil || (e == errSyntax && s.err != errSyntax) { |
|
s.err = e |
|
} |
|
} |
|
|
|
// resizeRange shrinks or grows the array at position oldStart such that |
|
// a new string of size newSize can fit between oldStart and oldEnd. |
|
// Sets the scan point to after the resized range. |
|
func (s *scanner) resizeRange(oldStart, oldEnd, newSize int) { |
|
s.start = oldStart |
|
if end := oldStart + newSize; end != oldEnd { |
|
diff := end - oldEnd |
|
if end < cap(s.b) { |
|
b := make([]byte, len(s.b)+diff) |
|
copy(b, s.b[:oldStart]) |
|
copy(b[end:], s.b[oldEnd:]) |
|
s.b = b |
|
} else { |
|
s.b = append(s.b[end:], s.b[oldEnd:]...) |
|
} |
|
s.next = end + (s.next - s.end) |
|
s.end = end |
|
} |
|
} |
|
|
|
// replace replaces the current token with repl. |
|
func (s *scanner) replace(repl string) { |
|
s.resizeRange(s.start, s.end, len(repl)) |
|
copy(s.b[s.start:], repl) |
|
} |
|
|
|
// gobble removes the current token from the input. |
|
// Caller must call scan after calling gobble. |
|
func (s *scanner) gobble(e error) { |
|
s.setError(e) |
|
if s.start == 0 { |
|
s.b = s.b[:+copy(s.b, s.b[s.next:])] |
|
s.end = 0 |
|
} else { |
|
s.b = s.b[:s.start-1+copy(s.b[s.start-1:], s.b[s.end:])] |
|
s.end = s.start - 1 |
|
} |
|
s.next = s.start |
|
} |
|
|
|
// deleteRange removes the given range from s.b before the current token. |
|
func (s *scanner) deleteRange(start, end int) { |
|
s.setError(errSyntax) |
|
s.b = s.b[:start+copy(s.b[start:], s.b[end:])] |
|
diff := end - start |
|
s.next -= diff |
|
s.start -= diff |
|
s.end -= diff |
|
} |
|
|
|
// scan parses the next token of a BCP 47 string. Tokens that are larger |
|
// than 8 characters or include non-alphanumeric characters result in an error |
|
// and are gobbled and removed from the output. |
|
// It returns the end position of the last token consumed. |
|
func (s *scanner) scan() (end int) { |
|
end = s.end |
|
s.token = nil |
|
for s.start = s.next; s.next < len(s.b); { |
|
i := bytes.IndexByte(s.b[s.next:], '-') |
|
if i == -1 { |
|
s.end = len(s.b) |
|
s.next = len(s.b) |
|
i = s.end - s.start |
|
} else { |
|
s.end = s.next + i |
|
s.next = s.end + 1 |
|
} |
|
token := s.b[s.start:s.end] |
|
if i < 1 || i > 8 || !isAlphaNum(token) { |
|
s.gobble(errSyntax) |
|
continue |
|
} |
|
s.token = token |
|
return end |
|
} |
|
if n := len(s.b); n > 0 && s.b[n-1] == '-' { |
|
s.setError(errSyntax) |
|
s.b = s.b[:len(s.b)-1] |
|
} |
|
s.done = true |
|
return end |
|
} |
|
|
|
// acceptMinSize parses multiple tokens of the given size or greater. |
|
// It returns the end position of the last token consumed. |
|
func (s *scanner) acceptMinSize(min int) (end int) { |
|
end = s.end |
|
s.scan() |
|
for ; len(s.token) >= min; s.scan() { |
|
end = s.end |
|
} |
|
return end |
|
} |
|
|
|
// Parse parses the given BCP 47 string and returns a valid Tag. If parsing |
|
// failed it returns an error and any part of the tag that could be parsed. |
|
// If parsing succeeded but an unknown value was found, it returns |
|
// ValueError. The Tag returned in this case is just stripped of the unknown |
|
// value. All other values are preserved. It accepts tags in the BCP 47 format |
|
// and extensions to this standard defined in |
|
// http://www.unicode.org/reports/tr35/#Unicode_Language_and_Locale_Identifiers. |
|
// The resulting tag is canonicalized using the default canonicalization type. |
|
func Parse(s string) (t Tag, err error) { |
|
return Default.Parse(s) |
|
} |
|
|
|
// Parse parses the given BCP 47 string and returns a valid Tag. If parsing |
|
// failed it returns an error and any part of the tag that could be parsed. |
|
// If parsing succeeded but an unknown value was found, it returns |
|
// ValueError. The Tag returned in this case is just stripped of the unknown |
|
// value. All other values are preserved. It accepts tags in the BCP 47 format |
|
// and extensions to this standard defined in |
|
// http://www.unicode.org/reports/tr35/#Unicode_Language_and_Locale_Identifiers. |
|
// The resulting tag is canonicalized using the the canonicalization type c. |
|
func (c CanonType) Parse(s string) (t Tag, err error) { |
|
// TODO: consider supporting old-style locale key-value pairs. |
|
if s == "" { |
|
return und, errSyntax |
|
} |
|
if len(s) <= maxAltTaglen { |
|
b := [maxAltTaglen]byte{} |
|
for i, c := range s { |
|
// Generating invalid UTF-8 is okay as it won't match. |
|
if 'A' <= c && c <= 'Z' { |
|
c += 'a' - 'A' |
|
} else if c == '_' { |
|
c = '-' |
|
} |
|
b[i] = byte(c) |
|
} |
|
if t, ok := grandfathered(b); ok { |
|
return t, nil |
|
} |
|
} |
|
scan := makeScannerString(s) |
|
t, err = parse(&scan, s) |
|
t, changed := t.canonicalize(c) |
|
if changed { |
|
t.remakeString() |
|
} |
|
return t, err |
|
} |
|
|
|
func parse(scan *scanner, s string) (t Tag, err error) { |
|
t = und |
|
var end int |
|
if n := len(scan.token); n <= 1 { |
|
scan.toLower(0, len(scan.b)) |
|
if n == 0 || scan.token[0] != 'x' { |
|
return t, errSyntax |
|
} |
|
end = parseExtensions(scan) |
|
} else if n >= 4 { |
|
return und, errSyntax |
|
} else { // the usual case |
|
t, end = parseTag(scan) |
|
if n := len(scan.token); n == 1 { |
|
t.pExt = uint16(end) |
|
end = parseExtensions(scan) |
|
} else if end < len(scan.b) { |
|
scan.setError(errSyntax) |
|
scan.b = scan.b[:end] |
|
} |
|
} |
|
if int(t.pVariant) < len(scan.b) { |
|
if end < len(s) { |
|
s = s[:end] |
|
} |
|
if len(s) > 0 && tag.Compare(s, scan.b) == 0 { |
|
t.str = s |
|
} else { |
|
t.str = string(scan.b) |
|
} |
|
} else { |
|
t.pVariant, t.pExt = 0, 0 |
|
} |
|
return t, scan.err |
|
} |
|
|
|
// parseTag parses language, script, region and variants. |
|
// It returns a Tag and the end position in the input that was parsed. |
|
func parseTag(scan *scanner) (t Tag, end int) { |
|
var e error |
|
// TODO: set an error if an unknown lang, script or region is encountered. |
|
t.lang, e = getLangID(scan.token) |
|
scan.setError(e) |
|
scan.replace(t.lang.String()) |
|
langStart := scan.start |
|
end = scan.scan() |
|
for len(scan.token) == 3 && isAlpha(scan.token[0]) { |
|
// From http://tools.ietf.org/html/bcp47, <lang>-<extlang> tags are equivalent |
|
// to a tag of the form <extlang>. |
|
lang, e := getLangID(scan.token) |
|
if lang != 0 { |
|
t.lang = lang |
|
copy(scan.b[langStart:], lang.String()) |
|
scan.b[langStart+3] = '-' |
|
scan.start = langStart + 4 |
|
} |
|
scan.gobble(e) |
|
end = scan.scan() |
|
} |
|
if len(scan.token) == 4 && isAlpha(scan.token[0]) { |
|
t.script, e = getScriptID(script, scan.token) |
|
if t.script == 0 { |
|
scan.gobble(e) |
|
} |
|
end = scan.scan() |
|
} |
|
if n := len(scan.token); n >= 2 && n <= 3 { |
|
t.region, e = getRegionID(scan.token) |
|
if t.region == 0 { |
|
scan.gobble(e) |
|
} else { |
|
scan.replace(t.region.String()) |
|
} |
|
end = scan.scan() |
|
} |
|
scan.toLower(scan.start, len(scan.b)) |
|
t.pVariant = byte(end) |
|
end = parseVariants(scan, end, t) |
|
t.pExt = uint16(end) |
|
return t, end |
|
} |
|
|
|
var separator = []byte{'-'} |
|
|
|
// parseVariants scans tokens as long as each token is a valid variant string. |
|
// Duplicate variants are removed. |
|
func parseVariants(scan *scanner, end int, t Tag) int { |
|
start := scan.start |
|
varIDBuf := [4]uint8{} |
|
variantBuf := [4][]byte{} |
|
varID := varIDBuf[:0] |
|
variant := variantBuf[:0] |
|
last := -1 |
|
needSort := false |
|
for ; len(scan.token) >= 4; scan.scan() { |
|
// TODO: measure the impact of needing this conversion and redesign |
|
// the data structure if there is an issue. |
|
v, ok := variantIndex[string(scan.token)] |
|
if !ok { |
|
// unknown variant |
|
// TODO: allow user-defined variants? |
|
scan.gobble(mkErrInvalid(scan.token)) |
|
continue |
|
} |
|
varID = append(varID, v) |
|
variant = append(variant, scan.token) |
|
if !needSort { |
|
if last < int(v) { |
|
last = int(v) |
|
} else { |
|
needSort = true |
|
// There is no legal combinations of more than 7 variants |
|
// (and this is by no means a useful sequence). |
|
const maxVariants = 8 |
|
if len(varID) > maxVariants { |
|
break |
|
} |
|
} |
|
} |
|
end = scan.end |
|
} |
|
if needSort { |
|
sort.Sort(variantsSort{varID, variant}) |
|
k, l := 0, -1 |
|
for i, v := range varID { |
|
w := int(v) |
|
if l == w { |
|
// Remove duplicates. |
|
continue |
|
} |
|
varID[k] = varID[i] |
|
variant[k] = variant[i] |
|
k++ |
|
l = w |
|
} |
|
if str := bytes.Join(variant[:k], separator); len(str) == 0 { |
|
end = start - 1 |
|
} else { |
|
scan.resizeRange(start, end, len(str)) |
|
copy(scan.b[scan.start:], str) |
|
end = scan.end |
|
} |
|
} |
|
return end |
|
} |
|
|
|
type variantsSort struct { |
|
i []uint8 |
|
v [][]byte |
|
} |
|
|
|
func (s variantsSort) Len() int { |
|
return len(s.i) |
|
} |
|
|
|
func (s variantsSort) Swap(i, j int) { |
|
s.i[i], s.i[j] = s.i[j], s.i[i] |
|
s.v[i], s.v[j] = s.v[j], s.v[i] |
|
} |
|
|
|
func (s variantsSort) Less(i, j int) bool { |
|
return s.i[i] < s.i[j] |
|
} |
|
|
|
type bytesSort [][]byte |
|
|
|
func (b bytesSort) Len() int { |
|
return len(b) |
|
} |
|
|
|
func (b bytesSort) Swap(i, j int) { |
|
b[i], b[j] = b[j], b[i] |
|
} |
|
|
|
func (b bytesSort) Less(i, j int) bool { |
|
return bytes.Compare(b[i], b[j]) == -1 |
|
} |
|
|
|
// parseExtensions parses and normalizes the extensions in the buffer. |
|
// It returns the last position of scan.b that is part of any extension. |
|
// It also trims scan.b to remove excess parts accordingly. |
|
func parseExtensions(scan *scanner) int { |
|
start := scan.start |
|
exts := [][]byte{} |
|
private := []byte{} |
|
end := scan.end |
|
for len(scan.token) == 1 { |
|
extStart := scan.start |
|
ext := scan.token[0] |
|
end = parseExtension(scan) |
|
extension := scan.b[extStart:end] |
|
if len(extension) < 3 || (ext != 'x' && len(extension) < 4) { |
|
scan.setError(errSyntax) |
|
end = extStart |
|
continue |
|
} else if start == extStart && (ext == 'x' || scan.start == len(scan.b)) { |
|
scan.b = scan.b[:end] |
|
return end |
|
} else if ext == 'x' { |
|
private = extension |
|
break |
|
} |
|
exts = append(exts, extension) |
|
} |
|
sort.Sort(bytesSort(exts)) |
|
if len(private) > 0 { |
|
exts = append(exts, private) |
|
} |
|
scan.b = scan.b[:start] |
|
if len(exts) > 0 { |
|
scan.b = append(scan.b, bytes.Join(exts, separator)...) |
|
} else if start > 0 { |
|
// Strip trailing '-'. |
|
scan.b = scan.b[:start-1] |
|
} |
|
return end |
|
} |
|
|
|
// parseExtension parses a single extension and returns the position of |
|
// the extension end. |
|
func parseExtension(scan *scanner) int { |
|
start, end := scan.start, scan.end |
|
switch scan.token[0] { |
|
case 'u': |
|
attrStart := end |
|
scan.scan() |
|
for last := []byte{}; len(scan.token) > 2; scan.scan() { |
|
if bytes.Compare(scan.token, last) != -1 { |
|
// Attributes are unsorted. Start over from scratch. |
|
p := attrStart + 1 |
|
scan.next = p |
|
attrs := [][]byte{} |
|
for scan.scan(); len(scan.token) > 2; scan.scan() { |
|
attrs = append(attrs, scan.token) |
|
end = scan.end |
|
} |
|
sort.Sort(bytesSort(attrs)) |
|
copy(scan.b[p:], bytes.Join(attrs, separator)) |
|
break |
|
} |
|
last = scan.token |
|
end = scan.end |
|
} |
|
var last, key []byte |
|
for attrEnd := end; len(scan.token) == 2; last = key { |
|
key = scan.token |
|
keyEnd := scan.end |
|
end = scan.acceptMinSize(3) |
|
// TODO: check key value validity |
|
if keyEnd == end || bytes.Compare(key, last) != 1 { |
|
// We have an invalid key or the keys are not sorted. |
|
// Start scanning keys from scratch and reorder. |
|
p := attrEnd + 1 |
|
scan.next = p |
|
keys := [][]byte{} |
|
for scan.scan(); len(scan.token) == 2; { |
|
keyStart, keyEnd := scan.start, scan.end |
|
end = scan.acceptMinSize(3) |
|
if keyEnd != end { |
|
keys = append(keys, scan.b[keyStart:end]) |
|
} else { |
|
scan.setError(errSyntax) |
|
end = keyStart |
|
} |
|
} |
|
sort.Sort(bytesSort(keys)) |
|
reordered := bytes.Join(keys, separator) |
|
if e := p + len(reordered); e < end { |
|
scan.deleteRange(e, end) |
|
end = e |
|
} |
|
copy(scan.b[p:], bytes.Join(keys, separator)) |
|
break |
|
} |
|
} |
|
case 't': |
|
scan.scan() |
|
if n := len(scan.token); n >= 2 && n <= 3 && isAlpha(scan.token[1]) { |
|
_, end = parseTag(scan) |
|
scan.toLower(start, end) |
|
} |
|
for len(scan.token) == 2 && !isAlpha(scan.token[1]) { |
|
end = scan.acceptMinSize(3) |
|
} |
|
case 'x': |
|
end = scan.acceptMinSize(1) |
|
default: |
|
end = scan.acceptMinSize(2) |
|
} |
|
return end |
|
} |
|
|
|
// Compose creates a Tag from individual parts, which may be of type Tag, Base, |
|
// Script, Region, Variant, []Variant, Extension, []Extension or error. If a |
|
// Base, Script or Region or slice of type Variant or Extension is passed more |
|
// than once, the latter will overwrite the former. Variants and Extensions are |
|
// accumulated, but if two extensions of the same type are passed, the latter |
|
// will replace the former. A Tag overwrites all former values and typically |
|
// only makes sense as the first argument. The resulting tag is returned after |
|
// canonicalizing using the Default CanonType. If one or more errors are |
|
// encountered, one of the errors is returned. |
|
func Compose(part ...interface{}) (t Tag, err error) { |
|
return Default.Compose(part...) |
|
} |
|
|
|
// Compose creates a Tag from individual parts, which may be of type Tag, Base, |
|
// Script, Region, Variant, []Variant, Extension, []Extension or error. If a |
|
// Base, Script or Region or slice of type Variant or Extension is passed more |
|
// than once, the latter will overwrite the former. Variants and Extensions are |
|
// accumulated, but if two extensions of the same type are passed, the latter |
|
// will replace the former. A Tag overwrites all former values and typically |
|
// only makes sense as the first argument. The resulting tag is returned after |
|
// canonicalizing using CanonType c. If one or more errors are encountered, |
|
// one of the errors is returned. |
|
func (c CanonType) Compose(part ...interface{}) (t Tag, err error) { |
|
var b builder |
|
if err = b.update(part...); err != nil { |
|
return und, err |
|
} |
|
t, _ = b.tag.canonicalize(c) |
|
|
|
if len(b.ext) > 0 || len(b.variant) > 0 { |
|
sort.Sort(sortVariant(b.variant)) |
|
sort.Strings(b.ext) |
|
if b.private != "" { |
|
b.ext = append(b.ext, b.private) |
|
} |
|
n := maxCoreSize + tokenLen(b.variant...) + tokenLen(b.ext...) |
|
buf := make([]byte, n) |
|
p := t.genCoreBytes(buf) |
|
t.pVariant = byte(p) |
|
p += appendTokens(buf[p:], b.variant...) |
|
t.pExt = uint16(p) |
|
p += appendTokens(buf[p:], b.ext...) |
|
t.str = string(buf[:p]) |
|
} else if b.private != "" { |
|
t.str = b.private |
|
t.remakeString() |
|
} |
|
return |
|
} |
|
|
|
type builder struct { |
|
tag Tag |
|
|
|
private string // the x extension |
|
ext []string |
|
variant []string |
|
|
|
err error |
|
} |
|
|
|
func (b *builder) addExt(e string) { |
|
if e == "" { |
|
} else if e[0] == 'x' { |
|
b.private = e |
|
} else { |
|
b.ext = append(b.ext, e) |
|
} |
|
} |
|
|
|
var errInvalidArgument = errors.New("invalid Extension or Variant") |
|
|
|
func (b *builder) update(part ...interface{}) (err error) { |
|
replace := func(l *[]string, s string, eq func(a, b string) bool) bool { |
|
if s == "" { |
|
b.err = errInvalidArgument |
|
return true |
|
} |
|
for i, v := range *l { |
|
if eq(v, s) { |
|
(*l)[i] = s |
|
return true |
|
} |
|
} |
|
return false |
|
} |
|
for _, x := range part { |
|
switch v := x.(type) { |
|
case Tag: |
|
b.tag.lang = v.lang |
|
b.tag.region = v.region |
|
b.tag.script = v.script |
|
if v.str != "" { |
|
b.variant = nil |
|
for x, s := "", v.str[v.pVariant:v.pExt]; s != ""; { |
|
x, s = nextToken(s) |
|
b.variant = append(b.variant, x) |
|
} |
|
b.ext, b.private = nil, "" |
|
for i, e := int(v.pExt), ""; i < len(v.str); { |
|
i, e = getExtension(v.str, i) |
|
b.addExt(e) |
|
} |
|
} |
|
case Base: |
|
b.tag.lang = v.langID |
|
case Script: |
|
b.tag.script = v.scriptID |
|
case Region: |
|
b.tag.region = v.regionID |
|
case Variant: |
|
if !replace(&b.variant, v.variant, func(a, b string) bool { return a == b }) { |
|
b.variant = append(b.variant, v.variant) |
|
} |
|
case Extension: |
|
if !replace(&b.ext, v.s, func(a, b string) bool { return a[0] == b[0] }) { |
|
b.addExt(v.s) |
|
} |
|
case []Variant: |
|
b.variant = nil |
|
for _, x := range v { |
|
b.update(x) |
|
} |
|
case []Extension: |
|
b.ext, b.private = nil, "" |
|
for _, e := range v { |
|
b.update(e) |
|
} |
|
// TODO: support parsing of raw strings based on morphology or just extensions? |
|
case error: |
|
err = v |
|
} |
|
} |
|
return |
|
} |
|
|
|
func tokenLen(token ...string) (n int) { |
|
for _, t := range token { |
|
n += len(t) + 1 |
|
} |
|
return |
|
} |
|
|
|
func appendTokens(b []byte, token ...string) int { |
|
p := 0 |
|
for _, t := range token { |
|
b[p] = '-' |
|
copy(b[p+1:], t) |
|
p += 1 + len(t) |
|
} |
|
return p |
|
} |
|
|
|
type sortVariant []string |
|
|
|
func (s sortVariant) Len() int { |
|
return len(s) |
|
} |
|
|
|
func (s sortVariant) Swap(i, j int) { |
|
s[j], s[i] = s[i], s[j] |
|
} |
|
|
|
func (s sortVariant) Less(i, j int) bool { |
|
return variantIndex[s[i]] < variantIndex[s[j]] |
|
} |
|
|
|
func findExt(list []string, x byte) int { |
|
for i, e := range list { |
|
if e[0] == x { |
|
return i |
|
} |
|
} |
|
return -1 |
|
} |
|
|
|
// getExtension returns the name, body and end position of the extension. |
|
func getExtension(s string, p int) (end int, ext string) { |
|
if s[p] == '-' { |
|
p++ |
|
} |
|
if s[p] == 'x' { |
|
return len(s), s[p:] |
|
} |
|
end = nextExtension(s, p) |
|
return end, s[p:end] |
|
} |
|
|
|
// nextExtension finds the next extension within the string, searching |
|
// for the -<char>- pattern from position p. |
|
// In the fast majority of cases, language tags will have at most |
|
// one extension and extensions tend to be small. |
|
func nextExtension(s string, p int) int { |
|
for n := len(s) - 3; p < n; { |
|
if s[p] == '-' { |
|
if s[p+2] == '-' { |
|
return p |
|
} |
|
p += 3 |
|
} else { |
|
p++ |
|
} |
|
} |
|
return len(s) |
|
} |
|
|
|
var errInvalidWeight = errors.New("ParseAcceptLanguage: invalid weight") |
|
|
|
// ParseAcceptLanguage parses the contents of a Accept-Language header as |
|
// defined in http://www.ietf.org/rfc/rfc2616.txt and returns a list of Tags and |
|
// a list of corresponding quality weights. It is more permissive than RFC 2616 |
|
// and may return non-nil slices even if the input is not valid. |
|
// The Tags will be sorted by highest weight first and then by first occurrence. |
|
// Tags with a weight of zero will be dropped. An error will be returned if the |
|
// input could not be parsed. |
|
func ParseAcceptLanguage(s string) (tag []Tag, q []float32, err error) { |
|
var entry string |
|
for s != "" { |
|
if entry, s = split(s, ','); entry == "" { |
|
continue |
|
} |
|
|
|
entry, weight := split(entry, ';') |
|
|
|
// Scan the language. |
|
t, err := Parse(entry) |
|
if err != nil { |
|
id, ok := acceptFallback[entry] |
|
if !ok { |
|
return nil, nil, err |
|
} |
|
t = Tag{lang: id} |
|
} |
|
|
|
// Scan the optional weight. |
|
w := 1.0 |
|
if weight != "" { |
|
weight = consume(weight, 'q') |
|
weight = consume(weight, '=') |
|
// consume returns the empty string when a token could not be |
|
// consumed, resulting in an error for ParseFloat. |
|
if w, err = strconv.ParseFloat(weight, 32); err != nil { |
|
return nil, nil, errInvalidWeight |
|
} |
|
// Drop tags with a quality weight of 0. |
|
if w <= 0 { |
|
continue |
|
} |
|
} |
|
|
|
tag = append(tag, t) |
|
q = append(q, float32(w)) |
|
} |
|
sortStable(&tagSort{tag, q}) |
|
return tag, q, nil |
|
} |
|
|
|
// consume removes a leading token c from s and returns the result or the empty |
|
// string if there is no such token. |
|
func consume(s string, c byte) string { |
|
if s == "" || s[0] != c { |
|
return "" |
|
} |
|
return strings.TrimSpace(s[1:]) |
|
} |
|
|
|
func split(s string, c byte) (head, tail string) { |
|
if i := strings.IndexByte(s, c); i >= 0 { |
|
return strings.TrimSpace(s[:i]), strings.TrimSpace(s[i+1:]) |
|
} |
|
return strings.TrimSpace(s), "" |
|
} |
|
|
|
// Add hack mapping to deal with a small number of cases that that occur |
|
// in Accept-Language (with reasonable frequency). |
|
var acceptFallback = map[string]langID{ |
|
"english": _en, |
|
"deutsch": _de, |
|
"italian": _it, |
|
"french": _fr, |
|
"*": _mul, // defined in the spec to match all languages. |
|
} |
|
|
|
type tagSort struct { |
|
tag []Tag |
|
q []float32 |
|
} |
|
|
|
func (s *tagSort) Len() int { |
|
return len(s.q) |
|
} |
|
|
|
func (s *tagSort) Less(i, j int) bool { |
|
return s.q[i] > s.q[j] |
|
} |
|
|
|
func (s *tagSort) Swap(i, j int) { |
|
s.tag[i], s.tag[j] = s.tag[j], s.tag[i] |
|
s.q[i], s.q[j] = s.q[j], s.q[i] |
|
}
|
|
|