mirror of
https://github.com/go-gitea/gitea.git
synced 2024-12-23 23:18:10 +08:00
388 lines
8.9 KiB
Go
388 lines
8.9 KiB
Go
// Copyright 2015 Unknwon
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License"): you may
|
|
// not use this file except in compliance with the License. You may obtain
|
|
// a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
// WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
// License for the specific language governing permissions and limitations
|
|
// under the License.
|
|
|
|
package ini
|
|
|
|
import (
|
|
"bufio"
|
|
"bytes"
|
|
"fmt"
|
|
"io"
|
|
"strconv"
|
|
"strings"
|
|
"unicode"
|
|
)
|
|
|
|
type tokenType int
|
|
|
|
const (
|
|
_TOKEN_INVALID tokenType = iota
|
|
_TOKEN_COMMENT
|
|
_TOKEN_SECTION
|
|
_TOKEN_KEY
|
|
)
|
|
|
|
type parser struct {
|
|
buf *bufio.Reader
|
|
isEOF bool
|
|
count int
|
|
comment *bytes.Buffer
|
|
}
|
|
|
|
func newParser(r io.Reader) *parser {
|
|
return &parser{
|
|
buf: bufio.NewReader(r),
|
|
count: 1,
|
|
comment: &bytes.Buffer{},
|
|
}
|
|
}
|
|
|
|
// BOM handles header of UTF-8, UTF-16 LE and UTF-16 BE's BOM format.
|
|
// http://en.wikipedia.org/wiki/Byte_order_mark#Representations_of_byte_order_marks_by_encoding
|
|
func (p *parser) BOM() error {
|
|
mask, err := p.buf.Peek(2)
|
|
if err != nil && err != io.EOF {
|
|
return err
|
|
} else if len(mask) < 2 {
|
|
return nil
|
|
}
|
|
|
|
switch {
|
|
case mask[0] == 254 && mask[1] == 255:
|
|
fallthrough
|
|
case mask[0] == 255 && mask[1] == 254:
|
|
p.buf.Read(mask)
|
|
case mask[0] == 239 && mask[1] == 187:
|
|
mask, err := p.buf.Peek(3)
|
|
if err != nil && err != io.EOF {
|
|
return err
|
|
} else if len(mask) < 3 {
|
|
return nil
|
|
}
|
|
if mask[2] == 191 {
|
|
p.buf.Read(mask)
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (p *parser) readUntil(delim byte) ([]byte, error) {
|
|
data, err := p.buf.ReadBytes(delim)
|
|
if err != nil {
|
|
if err == io.EOF {
|
|
p.isEOF = true
|
|
} else {
|
|
return nil, err
|
|
}
|
|
}
|
|
return data, nil
|
|
}
|
|
|
|
func cleanComment(in []byte) ([]byte, bool) {
|
|
i := bytes.IndexAny(in, "#;")
|
|
if i == -1 {
|
|
return nil, false
|
|
}
|
|
return in[i:], true
|
|
}
|
|
|
|
func readKeyName(in []byte) (string, int, error) {
|
|
line := string(in)
|
|
|
|
// Check if key name surrounded by quotes.
|
|
var keyQuote string
|
|
if line[0] == '"' {
|
|
if len(line) > 6 && string(line[0:3]) == `"""` {
|
|
keyQuote = `"""`
|
|
} else {
|
|
keyQuote = `"`
|
|
}
|
|
} else if line[0] == '`' {
|
|
keyQuote = "`"
|
|
}
|
|
|
|
// Get out key name
|
|
endIdx := -1
|
|
if len(keyQuote) > 0 {
|
|
startIdx := len(keyQuote)
|
|
// FIXME: fail case -> """"""name"""=value
|
|
pos := strings.Index(line[startIdx:], keyQuote)
|
|
if pos == -1 {
|
|
return "", -1, fmt.Errorf("missing closing key quote: %s", line)
|
|
}
|
|
pos += startIdx
|
|
|
|
// Find key-value delimiter
|
|
i := strings.IndexAny(line[pos+startIdx:], "=:")
|
|
if i < 0 {
|
|
return "", -1, ErrDelimiterNotFound{line}
|
|
}
|
|
endIdx = pos + i
|
|
return strings.TrimSpace(line[startIdx:pos]), endIdx + startIdx + 1, nil
|
|
}
|
|
|
|
endIdx = strings.IndexAny(line, "=:")
|
|
if endIdx < 0 {
|
|
return "", -1, ErrDelimiterNotFound{line}
|
|
}
|
|
return strings.TrimSpace(line[0:endIdx]), endIdx + 1, nil
|
|
}
|
|
|
|
func (p *parser) readMultilines(line, val, valQuote string) (string, error) {
|
|
for {
|
|
data, err := p.readUntil('\n')
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
next := string(data)
|
|
|
|
pos := strings.LastIndex(next, valQuote)
|
|
if pos > -1 {
|
|
val += next[:pos]
|
|
|
|
comment, has := cleanComment([]byte(next[pos:]))
|
|
if has {
|
|
p.comment.Write(bytes.TrimSpace(comment))
|
|
}
|
|
break
|
|
}
|
|
val += next
|
|
if p.isEOF {
|
|
return "", fmt.Errorf("missing closing key quote from '%s' to '%s'", line, next)
|
|
}
|
|
}
|
|
return val, nil
|
|
}
|
|
|
|
func (p *parser) readContinuationLines(val string) (string, error) {
|
|
for {
|
|
data, err := p.readUntil('\n')
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
next := strings.TrimSpace(string(data))
|
|
|
|
if len(next) == 0 {
|
|
break
|
|
}
|
|
val += next
|
|
if val[len(val)-1] != '\\' {
|
|
break
|
|
}
|
|
val = val[:len(val)-1]
|
|
}
|
|
return val, nil
|
|
}
|
|
|
|
// hasSurroundedQuote check if and only if the first and last characters
|
|
// are quotes \" or \'.
|
|
// It returns false if any other parts also contain same kind of quotes.
|
|
func hasSurroundedQuote(in string, quote byte) bool {
|
|
return len(in) >= 2 && in[0] == quote && in[len(in)-1] == quote &&
|
|
strings.IndexByte(in[1:], quote) == len(in)-2
|
|
}
|
|
|
|
func (p *parser) readValue(in []byte,
|
|
ignoreContinuation, ignoreInlineComment, unescapeValueDoubleQuotes, unescapeValueCommentSymbols bool) (string, error) {
|
|
|
|
line := strings.TrimLeftFunc(string(in), unicode.IsSpace)
|
|
if len(line) == 0 {
|
|
return "", nil
|
|
}
|
|
|
|
var valQuote string
|
|
if len(line) > 3 && string(line[0:3]) == `"""` {
|
|
valQuote = `"""`
|
|
} else if line[0] == '`' {
|
|
valQuote = "`"
|
|
} else if unescapeValueDoubleQuotes && line[0] == '"' {
|
|
valQuote = `"`
|
|
}
|
|
|
|
if len(valQuote) > 0 {
|
|
startIdx := len(valQuote)
|
|
pos := strings.LastIndex(line[startIdx:], valQuote)
|
|
// Check for multi-line value
|
|
if pos == -1 {
|
|
return p.readMultilines(line, line[startIdx:], valQuote)
|
|
}
|
|
|
|
if unescapeValueDoubleQuotes && valQuote == `"` {
|
|
return strings.Replace(line[startIdx:pos+startIdx], `\"`, `"`, -1), nil
|
|
}
|
|
return line[startIdx : pos+startIdx], nil
|
|
}
|
|
|
|
// Won't be able to reach here if value only contains whitespace
|
|
line = strings.TrimSpace(line)
|
|
|
|
// Check continuation lines when desired
|
|
if !ignoreContinuation && line[len(line)-1] == '\\' {
|
|
return p.readContinuationLines(line[:len(line)-1])
|
|
}
|
|
|
|
// Check if ignore inline comment
|
|
if !ignoreInlineComment {
|
|
i := strings.IndexAny(line, "#;")
|
|
if i > -1 {
|
|
p.comment.WriteString(line[i:])
|
|
line = strings.TrimSpace(line[:i])
|
|
}
|
|
}
|
|
|
|
// Trim single and double quotes
|
|
if hasSurroundedQuote(line, '\'') ||
|
|
hasSurroundedQuote(line, '"') {
|
|
line = line[1 : len(line)-1]
|
|
} else if len(valQuote) == 0 && unescapeValueCommentSymbols {
|
|
if strings.Contains(line, `\;`) {
|
|
line = strings.Replace(line, `\;`, ";", -1)
|
|
}
|
|
if strings.Contains(line, `\#`) {
|
|
line = strings.Replace(line, `\#`, "#", -1)
|
|
}
|
|
}
|
|
return line, nil
|
|
}
|
|
|
|
// parse parses data through an io.Reader.
|
|
func (f *File) parse(reader io.Reader) (err error) {
|
|
p := newParser(reader)
|
|
if err = p.BOM(); err != nil {
|
|
return fmt.Errorf("BOM: %v", err)
|
|
}
|
|
|
|
// Ignore error because default section name is never empty string.
|
|
name := DEFAULT_SECTION
|
|
if f.options.Insensitive {
|
|
name = strings.ToLower(DEFAULT_SECTION)
|
|
}
|
|
section, _ := f.NewSection(name)
|
|
|
|
var line []byte
|
|
var inUnparseableSection bool
|
|
for !p.isEOF {
|
|
line, err = p.readUntil('\n')
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
line = bytes.TrimLeftFunc(line, unicode.IsSpace)
|
|
if len(line) == 0 {
|
|
continue
|
|
}
|
|
|
|
// Comments
|
|
if line[0] == '#' || line[0] == ';' {
|
|
// Note: we do not care ending line break,
|
|
// it is needed for adding second line,
|
|
// so just clean it once at the end when set to value.
|
|
p.comment.Write(line)
|
|
continue
|
|
}
|
|
|
|
// Section
|
|
if line[0] == '[' {
|
|
// Read to the next ']' (TODO: support quoted strings)
|
|
// TODO(unknwon): use LastIndexByte when stop supporting Go1.4
|
|
closeIdx := bytes.LastIndex(line, []byte("]"))
|
|
if closeIdx == -1 {
|
|
return fmt.Errorf("unclosed section: %s", line)
|
|
}
|
|
|
|
name := string(line[1:closeIdx])
|
|
section, err = f.NewSection(name)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
comment, has := cleanComment(line[closeIdx+1:])
|
|
if has {
|
|
p.comment.Write(comment)
|
|
}
|
|
|
|
section.Comment = strings.TrimSpace(p.comment.String())
|
|
|
|
// Reset aotu-counter and comments
|
|
p.comment.Reset()
|
|
p.count = 1
|
|
|
|
inUnparseableSection = false
|
|
for i := range f.options.UnparseableSections {
|
|
if f.options.UnparseableSections[i] == name ||
|
|
(f.options.Insensitive && strings.ToLower(f.options.UnparseableSections[i]) == strings.ToLower(name)) {
|
|
inUnparseableSection = true
|
|
continue
|
|
}
|
|
}
|
|
continue
|
|
}
|
|
|
|
if inUnparseableSection {
|
|
section.isRawSection = true
|
|
section.rawBody += string(line)
|
|
continue
|
|
}
|
|
|
|
kname, offset, err := readKeyName(line)
|
|
if err != nil {
|
|
// Treat as boolean key when desired, and whole line is key name.
|
|
if IsErrDelimiterNotFound(err) && f.options.AllowBooleanKeys {
|
|
kname, err := p.readValue(line,
|
|
f.options.IgnoreContinuation,
|
|
f.options.IgnoreInlineComment,
|
|
f.options.UnescapeValueDoubleQuotes,
|
|
f.options.UnescapeValueCommentSymbols)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
key, err := section.NewBooleanKey(kname)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
key.Comment = strings.TrimSpace(p.comment.String())
|
|
p.comment.Reset()
|
|
continue
|
|
}
|
|
return err
|
|
}
|
|
|
|
// Auto increment.
|
|
isAutoIncr := false
|
|
if kname == "-" {
|
|
isAutoIncr = true
|
|
kname = "#" + strconv.Itoa(p.count)
|
|
p.count++
|
|
}
|
|
|
|
value, err := p.readValue(line[offset:],
|
|
f.options.IgnoreContinuation,
|
|
f.options.IgnoreInlineComment,
|
|
f.options.UnescapeValueDoubleQuotes,
|
|
f.options.UnescapeValueCommentSymbols)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
key, err := section.NewKey(kname, value)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
key.isAutoIncrement = isAutoIncr
|
|
key.Comment = strings.TrimSpace(p.comment.String())
|
|
p.comment.Reset()
|
|
}
|
|
return nil
|
|
}
|