123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860 |
- // Copyright 2013 The Go Authors. All rights reserved.
- // Use of this source code is governed by a BSD-style
- // license that can be found in the LICENSE file.
- package language
- import (
- "bytes"
- "errors"
- "fmt"
- "sort"
- "strconv"
- "strings"
- "golang.org/x/text/internal/tag"
- )
- // isAlpha returns true if the byte is not a digit.
- // b must be an ASCII letter or digit.
- func isAlpha(b byte) bool {
- return b > '9'
- }
- // isAlphaNum returns true if the string contains only ASCII letters or digits.
- func isAlphaNum(s []byte) bool {
- for _, c := range s {
- if !('a' <= c && c <= 'z' || 'A' <= c && c <= 'Z' || '0' <= c && c <= '9') {
- return false
- }
- }
- return true
- }
- // errSyntax is returned by any of the parsing functions when the
- // input is not well-formed, according to BCP 47.
- // TODO: return the position at which the syntax error occurred?
- var errSyntax = errors.New("language: tag is not well-formed")
- // ValueError is returned by any of the parsing functions when the
- // input is well-formed but the respective subtag is not recognized
- // as a valid value.
- type ValueError struct {
- v [8]byte
- }
- func mkErrInvalid(s []byte) error {
- var e ValueError
- copy(e.v[:], s)
- return e
- }
- func (e ValueError) tag() []byte {
- n := bytes.IndexByte(e.v[:], 0)
- if n == -1 {
- n = 8
- }
- return e.v[:n]
- }
- // Error implements the error interface.
- func (e ValueError) Error() string {
- return fmt.Sprintf("language: subtag %q is well-formed but unknown", e.tag())
- }
- // Subtag returns the subtag for which the error occurred.
- func (e ValueError) Subtag() string {
- return string(e.tag())
- }
- // scanner is used to scan BCP 47 tokens, which are separated by _ or -.
- type scanner struct {
- b []byte
- bytes [max99thPercentileSize]byte
- token []byte
- start int // start position of the current token
- end int // end position of the current token
- next int // next point for scan
- err error
- done bool
- }
- func makeScannerString(s string) scanner {
- scan := scanner{}
- if len(s) <= len(scan.bytes) {
- scan.b = scan.bytes[:copy(scan.bytes[:], s)]
- } else {
- scan.b = []byte(s)
- }
- scan.init()
- return scan
- }
- // makeScanner returns a scanner using b as the input buffer.
- // b is not copied and may be modified by the scanner routines.
- func makeScanner(b []byte) scanner {
- scan := scanner{b: b}
- scan.init()
- return scan
- }
- func (s *scanner) init() {
- for i, c := range s.b {
- if c == '_' {
- s.b[i] = '-'
- }
- }
- s.scan()
- }
- // restToLower converts the string between start and end to lower case.
- func (s *scanner) toLower(start, end int) {
- for i := start; i < end; i++ {
- c := s.b[i]
- if 'A' <= c && c <= 'Z' {
- s.b[i] += 'a' - 'A'
- }
- }
- }
- func (s *scanner) setError(e error) {
- if s.err == nil || (e == errSyntax && s.err != errSyntax) {
- s.err = e
- }
- }
- // resizeRange shrinks or grows the array at position oldStart such that
- // a new string of size newSize can fit between oldStart and oldEnd.
- // Sets the scan point to after the resized range.
- func (s *scanner) resizeRange(oldStart, oldEnd, newSize int) {
- s.start = oldStart
- if end := oldStart + newSize; end != oldEnd {
- diff := end - oldEnd
- if end < cap(s.b) {
- b := make([]byte, len(s.b)+diff)
- copy(b, s.b[:oldStart])
- copy(b[end:], s.b[oldEnd:])
- s.b = b
- } else {
- s.b = append(s.b[end:], s.b[oldEnd:]...)
- }
- s.next = end + (s.next - s.end)
- s.end = end
- }
- }
- // replace replaces the current token with repl.
- func (s *scanner) replace(repl string) {
- s.resizeRange(s.start, s.end, len(repl))
- copy(s.b[s.start:], repl)
- }
- // gobble removes the current token from the input.
- // Caller must call scan after calling gobble.
- func (s *scanner) gobble(e error) {
- s.setError(e)
- if s.start == 0 {
- s.b = s.b[:+copy(s.b, s.b[s.next:])]
- s.end = 0
- } else {
- s.b = s.b[:s.start-1+copy(s.b[s.start-1:], s.b[s.end:])]
- s.end = s.start - 1
- }
- s.next = s.start
- }
- // deleteRange removes the given range from s.b before the current token.
- func (s *scanner) deleteRange(start, end int) {
- s.setError(errSyntax)
- s.b = s.b[:start+copy(s.b[start:], s.b[end:])]
- diff := end - start
- s.next -= diff
- s.start -= diff
- s.end -= diff
- }
- // scan parses the next token of a BCP 47 string. Tokens that are larger
- // than 8 characters or include non-alphanumeric characters result in an error
- // and are gobbled and removed from the output.
- // It returns the end position of the last token consumed.
- func (s *scanner) scan() (end int) {
- end = s.end
- s.token = nil
- for s.start = s.next; s.next < len(s.b); {
- i := bytes.IndexByte(s.b[s.next:], '-')
- if i == -1 {
- s.end = len(s.b)
- s.next = len(s.b)
- i = s.end - s.start
- } else {
- s.end = s.next + i
- s.next = s.end + 1
- }
- token := s.b[s.start:s.end]
- if i < 1 || i > 8 || !isAlphaNum(token) {
- s.gobble(errSyntax)
- continue
- }
- s.token = token
- return end
- }
- if n := len(s.b); n > 0 && s.b[n-1] == '-' {
- s.setError(errSyntax)
- s.b = s.b[:len(s.b)-1]
- }
- s.done = true
- return end
- }
- // acceptMinSize parses multiple tokens of the given size or greater.
- // It returns the end position of the last token consumed.
- func (s *scanner) acceptMinSize(min int) (end int) {
- end = s.end
- s.scan()
- for ; len(s.token) >= min; s.scan() {
- end = s.end
- }
- return end
- }
- // Parse parses the given BCP 47 string and returns a valid Tag. If parsing
- // failed it returns an error and any part of the tag that could be parsed.
- // If parsing succeeded but an unknown value was found, it returns
- // ValueError. The Tag returned in this case is just stripped of the unknown
- // value. All other values are preserved. It accepts tags in the BCP 47 format
- // and extensions to this standard defined in
- // http://www.unicode.org/reports/tr35/#Unicode_Language_and_Locale_Identifiers.
- // The resulting tag is canonicalized using the default canonicalization type.
- func Parse(s string) (t Tag, err error) {
- return Default.Parse(s)
- }
- // Parse parses the given BCP 47 string and returns a valid Tag. If parsing
- // failed it returns an error and any part of the tag that could be parsed.
- // If parsing succeeded but an unknown value was found, it returns
- // ValueError. The Tag returned in this case is just stripped of the unknown
- // value. All other values are preserved. It accepts tags in the BCP 47 format
- // and extensions to this standard defined in
- // http://www.unicode.org/reports/tr35/#Unicode_Language_and_Locale_Identifiers.
- // The resulting tag is canonicalized using the the canonicalization type c.
- func (c CanonType) Parse(s string) (t Tag, err error) {
- // TODO: consider supporting old-style locale key-value pairs.
- if s == "" {
- return und, errSyntax
- }
- if len(s) <= maxAltTaglen {
- b := [maxAltTaglen]byte{}
- for i, c := range s {
- // Generating invalid UTF-8 is okay as it won't match.
- if 'A' <= c && c <= 'Z' {
- c += 'a' - 'A'
- } else if c == '_' {
- c = '-'
- }
- b[i] = byte(c)
- }
- if t, ok := grandfathered(b); ok {
- return t, nil
- }
- }
- scan := makeScannerString(s)
- t, err = parse(&scan, s)
- t, changed := t.canonicalize(c)
- if changed {
- t.remakeString()
- }
- return t, err
- }
- func parse(scan *scanner, s string) (t Tag, err error) {
- t = und
- var end int
- if n := len(scan.token); n <= 1 {
- scan.toLower(0, len(scan.b))
- if n == 0 || scan.token[0] != 'x' {
- return t, errSyntax
- }
- end = parseExtensions(scan)
- } else if n >= 4 {
- return und, errSyntax
- } else { // the usual case
- t, end = parseTag(scan)
- if n := len(scan.token); n == 1 {
- t.pExt = uint16(end)
- end = parseExtensions(scan)
- } else if end < len(scan.b) {
- scan.setError(errSyntax)
- scan.b = scan.b[:end]
- }
- }
- if int(t.pVariant) < len(scan.b) {
- if end < len(s) {
- s = s[:end]
- }
- if len(s) > 0 && tag.Compare(s, scan.b) == 0 {
- t.str = s
- } else {
- t.str = string(scan.b)
- }
- } else {
- t.pVariant, t.pExt = 0, 0
- }
- return t, scan.err
- }
- // parseTag parses language, script, region and variants.
- // It returns a Tag and the end position in the input that was parsed.
- func parseTag(scan *scanner) (t Tag, end int) {
- var e error
- // TODO: set an error if an unknown lang, script or region is encountered.
- t.lang, e = getLangID(scan.token)
- scan.setError(e)
- scan.replace(t.lang.String())
- langStart := scan.start
- end = scan.scan()
- for len(scan.token) == 3 && isAlpha(scan.token[0]) {
- // From http://tools.ietf.org/html/bcp47, <lang>-<extlang> tags are equivalent
- // to a tag of the form <extlang>.
- lang, e := getLangID(scan.token)
- if lang != 0 {
- t.lang = lang
- copy(scan.b[langStart:], lang.String())
- scan.b[langStart+3] = '-'
- scan.start = langStart + 4
- }
- scan.gobble(e)
- end = scan.scan()
- }
- if len(scan.token) == 4 && isAlpha(scan.token[0]) {
- t.script, e = getScriptID(script, scan.token)
- if t.script == 0 {
- scan.gobble(e)
- }
- end = scan.scan()
- }
- if n := len(scan.token); n >= 2 && n <= 3 {
- t.region, e = getRegionID(scan.token)
- if t.region == 0 {
- scan.gobble(e)
- } else {
- scan.replace(t.region.String())
- }
- end = scan.scan()
- }
- scan.toLower(scan.start, len(scan.b))
- t.pVariant = byte(end)
- end = parseVariants(scan, end, t)
- t.pExt = uint16(end)
- return t, end
- }
- var separator = []byte{'-'}
- // parseVariants scans tokens as long as each token is a valid variant string.
- // Duplicate variants are removed.
- func parseVariants(scan *scanner, end int, t Tag) int {
- start := scan.start
- varIDBuf := [4]uint8{}
- variantBuf := [4][]byte{}
- varID := varIDBuf[:0]
- variant := variantBuf[:0]
- last := -1
- needSort := false
- for ; len(scan.token) >= 4; scan.scan() {
- // TODO: measure the impact of needing this conversion and redesign
- // the data structure if there is an issue.
- v, ok := variantIndex[string(scan.token)]
- if !ok {
- // unknown variant
- // TODO: allow user-defined variants?
- scan.gobble(mkErrInvalid(scan.token))
- continue
- }
- varID = append(varID, v)
- variant = append(variant, scan.token)
- if !needSort {
- if last < int(v) {
- last = int(v)
- } else {
- needSort = true
- // There is no legal combinations of more than 7 variants
- // (and this is by no means a useful sequence).
- const maxVariants = 8
- if len(varID) > maxVariants {
- break
- }
- }
- }
- end = scan.end
- }
- if needSort {
- sort.Sort(variantsSort{varID, variant})
- k, l := 0, -1
- for i, v := range varID {
- w := int(v)
- if l == w {
- // Remove duplicates.
- continue
- }
- varID[k] = varID[i]
- variant[k] = variant[i]
- k++
- l = w
- }
- if str := bytes.Join(variant[:k], separator); len(str) == 0 {
- end = start - 1
- } else {
- scan.resizeRange(start, end, len(str))
- copy(scan.b[scan.start:], str)
- end = scan.end
- }
- }
- return end
- }
- type variantsSort struct {
- i []uint8
- v [][]byte
- }
- func (s variantsSort) Len() int {
- return len(s.i)
- }
- func (s variantsSort) Swap(i, j int) {
- s.i[i], s.i[j] = s.i[j], s.i[i]
- s.v[i], s.v[j] = s.v[j], s.v[i]
- }
- func (s variantsSort) Less(i, j int) bool {
- return s.i[i] < s.i[j]
- }
- type bytesSort [][]byte
- func (b bytesSort) Len() int {
- return len(b)
- }
- func (b bytesSort) Swap(i, j int) {
- b[i], b[j] = b[j], b[i]
- }
- func (b bytesSort) Less(i, j int) bool {
- return bytes.Compare(b[i], b[j]) == -1
- }
- // parseExtensions parses and normalizes the extensions in the buffer.
- // It returns the last position of scan.b that is part of any extension.
- // It also trims scan.b to remove excess parts accordingly.
- func parseExtensions(scan *scanner) int {
- start := scan.start
- exts := [][]byte{}
- private := []byte{}
- end := scan.end
- for len(scan.token) == 1 {
- extStart := scan.start
- ext := scan.token[0]
- end = parseExtension(scan)
- extension := scan.b[extStart:end]
- if len(extension) < 3 || (ext != 'x' && len(extension) < 4) {
- scan.setError(errSyntax)
- end = extStart
- continue
- } else if start == extStart && (ext == 'x' || scan.start == len(scan.b)) {
- scan.b = scan.b[:end]
- return end
- } else if ext == 'x' {
- private = extension
- break
- }
- exts = append(exts, extension)
- }
- sort.Sort(bytesSort(exts))
- if len(private) > 0 {
- exts = append(exts, private)
- }
- scan.b = scan.b[:start]
- if len(exts) > 0 {
- scan.b = append(scan.b, bytes.Join(exts, separator)...)
- } else if start > 0 {
- // Strip trailing '-'.
- scan.b = scan.b[:start-1]
- }
- return end
- }
- // parseExtension parses a single extension and returns the position of
- // the extension end.
- func parseExtension(scan *scanner) int {
- start, end := scan.start, scan.end
- switch scan.token[0] {
- case 'u':
- attrStart := end
- scan.scan()
- for last := []byte{}; len(scan.token) > 2; scan.scan() {
- if bytes.Compare(scan.token, last) != -1 {
- // Attributes are unsorted. Start over from scratch.
- p := attrStart + 1
- scan.next = p
- attrs := [][]byte{}
- for scan.scan(); len(scan.token) > 2; scan.scan() {
- attrs = append(attrs, scan.token)
- end = scan.end
- }
- sort.Sort(bytesSort(attrs))
- copy(scan.b[p:], bytes.Join(attrs, separator))
- break
- }
- last = scan.token
- end = scan.end
- }
- var last, key []byte
- for attrEnd := end; len(scan.token) == 2; last = key {
- key = scan.token
- keyEnd := scan.end
- end = scan.acceptMinSize(3)
- // TODO: check key value validity
- if keyEnd == end || bytes.Compare(key, last) != 1 {
- // We have an invalid key or the keys are not sorted.
- // Start scanning keys from scratch and reorder.
- p := attrEnd + 1
- scan.next = p
- keys := [][]byte{}
- for scan.scan(); len(scan.token) == 2; {
- keyStart, keyEnd := scan.start, scan.end
- end = scan.acceptMinSize(3)
- if keyEnd != end {
- keys = append(keys, scan.b[keyStart:end])
- } else {
- scan.setError(errSyntax)
- end = keyStart
- }
- }
- sort.Sort(bytesSort(keys))
- reordered := bytes.Join(keys, separator)
- if e := p + len(reordered); e < end {
- scan.deleteRange(e, end)
- end = e
- }
- copy(scan.b[p:], bytes.Join(keys, separator))
- break
- }
- }
- case 't':
- scan.scan()
- if n := len(scan.token); n >= 2 && n <= 3 && isAlpha(scan.token[1]) {
- _, end = parseTag(scan)
- scan.toLower(start, end)
- }
- for len(scan.token) == 2 && !isAlpha(scan.token[1]) {
- end = scan.acceptMinSize(3)
- }
- case 'x':
- end = scan.acceptMinSize(1)
- default:
- end = scan.acceptMinSize(2)
- }
- return end
- }
- // Compose creates a Tag from individual parts, which may be of type Tag, Base,
- // Script, Region, Variant, []Variant, Extension, []Extension or error. If a
- // Base, Script or Region or slice of type Variant or Extension is passed more
- // than once, the latter will overwrite the former. Variants and Extensions are
- // accumulated, but if two extensions of the same type are passed, the latter
- // will replace the former. A Tag overwrites all former values and typically
- // only makes sense as the first argument. The resulting tag is returned after
- // canonicalizing using the Default CanonType. If one or more errors are
- // encountered, one of the errors is returned.
- func Compose(part ...interface{}) (t Tag, err error) {
- return Default.Compose(part...)
- }
- // Compose creates a Tag from individual parts, which may be of type Tag, Base,
- // Script, Region, Variant, []Variant, Extension, []Extension or error. If a
- // Base, Script or Region or slice of type Variant or Extension is passed more
- // than once, the latter will overwrite the former. Variants and Extensions are
- // accumulated, but if two extensions of the same type are passed, the latter
- // will replace the former. A Tag overwrites all former values and typically
- // only makes sense as the first argument. The resulting tag is returned after
- // canonicalizing using CanonType c. If one or more errors are encountered,
- // one of the errors is returned.
- func (c CanonType) Compose(part ...interface{}) (t Tag, err error) {
- var b builder
- if err = b.update(part...); err != nil {
- return und, err
- }
- t, _ = b.tag.canonicalize(c)
- if len(b.ext) > 0 || len(b.variant) > 0 {
- sort.Sort(sortVariant(b.variant))
- sort.Strings(b.ext)
- if b.private != "" {
- b.ext = append(b.ext, b.private)
- }
- n := maxCoreSize + tokenLen(b.variant...) + tokenLen(b.ext...)
- buf := make([]byte, n)
- p := t.genCoreBytes(buf)
- t.pVariant = byte(p)
- p += appendTokens(buf[p:], b.variant...)
- t.pExt = uint16(p)
- p += appendTokens(buf[p:], b.ext...)
- t.str = string(buf[:p])
- } else if b.private != "" {
- t.str = b.private
- t.remakeString()
- }
- return
- }
- type builder struct {
- tag Tag
- private string // the x extension
- ext []string
- variant []string
- err error
- }
- func (b *builder) addExt(e string) {
- if e == "" {
- } else if e[0] == 'x' {
- b.private = e
- } else {
- b.ext = append(b.ext, e)
- }
- }
- var errInvalidArgument = errors.New("invalid Extension or Variant")
- func (b *builder) update(part ...interface{}) (err error) {
- replace := func(l *[]string, s string, eq func(a, b string) bool) bool {
- if s == "" {
- b.err = errInvalidArgument
- return true
- }
- for i, v := range *l {
- if eq(v, s) {
- (*l)[i] = s
- return true
- }
- }
- return false
- }
- for _, x := range part {
- switch v := x.(type) {
- case Tag:
- b.tag.lang = v.lang
- b.tag.region = v.region
- b.tag.script = v.script
- if v.str != "" {
- b.variant = nil
- for x, s := "", v.str[v.pVariant:v.pExt]; s != ""; {
- x, s = nextToken(s)
- b.variant = append(b.variant, x)
- }
- b.ext, b.private = nil, ""
- for i, e := int(v.pExt), ""; i < len(v.str); {
- i, e = getExtension(v.str, i)
- b.addExt(e)
- }
- }
- case Base:
- b.tag.lang = v.langID
- case Script:
- b.tag.script = v.scriptID
- case Region:
- b.tag.region = v.regionID
- case Variant:
- if !replace(&b.variant, v.variant, func(a, b string) bool { return a == b }) {
- b.variant = append(b.variant, v.variant)
- }
- case Extension:
- if !replace(&b.ext, v.s, func(a, b string) bool { return a[0] == b[0] }) {
- b.addExt(v.s)
- }
- case []Variant:
- b.variant = nil
- for _, x := range v {
- b.update(x)
- }
- case []Extension:
- b.ext, b.private = nil, ""
- for _, e := range v {
- b.update(e)
- }
- // TODO: support parsing of raw strings based on morphology or just extensions?
- case error:
- err = v
- }
- }
- return
- }
- func tokenLen(token ...string) (n int) {
- for _, t := range token {
- n += len(t) + 1
- }
- return
- }
- func appendTokens(b []byte, token ...string) int {
- p := 0
- for _, t := range token {
- b[p] = '-'
- copy(b[p+1:], t)
- p += 1 + len(t)
- }
- return p
- }
- type sortVariant []string
- func (s sortVariant) Len() int {
- return len(s)
- }
- func (s sortVariant) Swap(i, j int) {
- s[j], s[i] = s[i], s[j]
- }
- func (s sortVariant) Less(i, j int) bool {
- return variantIndex[s[i]] < variantIndex[s[j]]
- }
- func findExt(list []string, x byte) int {
- for i, e := range list {
- if e[0] == x {
- return i
- }
- }
- return -1
- }
- // getExtension returns the name, body and end position of the extension.
- func getExtension(s string, p int) (end int, ext string) {
- if s[p] == '-' {
- p++
- }
- if s[p] == 'x' {
- return len(s), s[p:]
- }
- end = nextExtension(s, p)
- return end, s[p:end]
- }
- // nextExtension finds the next extension within the string, searching
- // for the -<char>- pattern from position p.
- // In the fast majority of cases, language tags will have at most
- // one extension and extensions tend to be small.
- func nextExtension(s string, p int) int {
- for n := len(s) - 3; p < n; {
- if s[p] == '-' {
- if s[p+2] == '-' {
- return p
- }
- p += 3
- } else {
- p++
- }
- }
- return len(s)
- }
- var errInvalidWeight = errors.New("ParseAcceptLanguage: invalid weight")
- // ParseAcceptLanguage parses the contents of a Accept-Language header as
- // defined in http://www.ietf.org/rfc/rfc2616.txt and returns a list of Tags and
- // a list of corresponding quality weights. It is more permissive than RFC 2616
- // and may return non-nil slices even if the input is not valid.
- // The Tags will be sorted by highest weight first and then by first occurrence.
- // Tags with a weight of zero will be dropped. An error will be returned if the
- // input could not be parsed.
- func ParseAcceptLanguage(s string) (tag []Tag, q []float32, err error) {
- var entry string
- for s != "" {
- if entry, s = split(s, ','); entry == "" {
- continue
- }
- entry, weight := split(entry, ';')
- // Scan the language.
- t, err := Parse(entry)
- if err != nil {
- id, ok := acceptFallback[entry]
- if !ok {
- return nil, nil, err
- }
- t = Tag{lang: id}
- }
- // Scan the optional weight.
- w := 1.0
- if weight != "" {
- weight = consume(weight, 'q')
- weight = consume(weight, '=')
- // consume returns the empty string when a token could not be
- // consumed, resulting in an error for ParseFloat.
- if w, err = strconv.ParseFloat(weight, 32); err != nil {
- return nil, nil, errInvalidWeight
- }
- // Drop tags with a quality weight of 0.
- if w <= 0 {
- continue
- }
- }
- tag = append(tag, t)
- q = append(q, float32(w))
- }
- sortStable(&tagSort{tag, q})
- return tag, q, nil
- }
- // consume removes a leading token c from s and returns the result or the empty
- // string if there is no such token.
- func consume(s string, c byte) string {
- if s == "" || s[0] != c {
- return ""
- }
- return strings.TrimSpace(s[1:])
- }
- func split(s string, c byte) (head, tail string) {
- if i := strings.IndexByte(s, c); i >= 0 {
- return strings.TrimSpace(s[:i]), strings.TrimSpace(s[i+1:])
- }
- return strings.TrimSpace(s), ""
- }
- // Add hack mapping to deal with a small number of cases that that occur
- // in Accept-Language (with reasonable frequency).
- var acceptFallback = map[string]langID{
- "english": _en,
- "deutsch": _de,
- "italian": _it,
- "french": _fr,
- "*": _mul, // defined in the spec to match all languages.
- }
- type tagSort struct {
- tag []Tag
- q []float32
- }
- func (s *tagSort) Len() int {
- return len(s.q)
- }
- func (s *tagSort) Less(i, j int) bool {
- return s.q[i] > s.q[j]
- }
- func (s *tagSort) Swap(i, j int) {
- s.tag[i], s.tag[j] = s.tag[j], s.tag[i]
- s.q[i], s.q[j] = s.q[j], s.q[i]
- }
|