mirror of
https://github.com/syncthing/syncthing.git
synced 2024-11-17 19:08:46 -07:00
2c89f04be7
This uses persistent Matcher objects that can reload their content and provide a hash string that can be used to check if it's changed. The cache is local to each Matcher object instead of kept globally.
286 lines
6.3 KiB
Go
286 lines
6.3 KiB
Go
// Copyright (C) 2014 The Syncthing Authors.
|
|
//
|
|
// This program is free software: you can redistribute it and/or modify it
|
|
// under the terms of the GNU General Public License as published by the Free
|
|
// Software Foundation, either version 3 of the License, or (at your option)
|
|
// any later version.
|
|
//
|
|
// This program is distributed in the hope that it will be useful, but WITHOUT
|
|
// ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
// FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
// more details.
|
|
//
|
|
// You should have received a copy of the GNU General Public License along
|
|
// with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package ignore
|
|
|
|
import (
|
|
"bufio"
|
|
"bytes"
|
|
"crypto/md5"
|
|
"fmt"
|
|
"io"
|
|
"os"
|
|
"path/filepath"
|
|
"regexp"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/syncthing/syncthing/internal/fnmatch"
|
|
)
|
|
|
|
type Pattern struct {
|
|
match *regexp.Regexp
|
|
include bool
|
|
}
|
|
|
|
func (p Pattern) String() string {
|
|
if p.include {
|
|
return p.match.String()
|
|
} else {
|
|
return "(?exclude)" + p.match.String()
|
|
}
|
|
}
|
|
|
|
type Matcher struct {
|
|
patterns []Pattern
|
|
withCache bool
|
|
matches *cache
|
|
curHash string
|
|
stop chan struct{}
|
|
mut sync.Mutex
|
|
}
|
|
|
|
func New(withCache bool) *Matcher {
|
|
m := &Matcher{
|
|
withCache: withCache,
|
|
stop: make(chan struct{}),
|
|
}
|
|
if withCache {
|
|
go m.clean(2 * time.Hour)
|
|
}
|
|
return m
|
|
}
|
|
|
|
func (m *Matcher) Load(file string) error {
|
|
// No locking, Parse() does the locking
|
|
|
|
fd, err := os.Open(file)
|
|
if err != nil {
|
|
// We do a parse with empty patterns to clear out the hash, cache etc.
|
|
m.Parse(&bytes.Buffer{}, file)
|
|
return err
|
|
}
|
|
defer fd.Close()
|
|
|
|
return m.Parse(fd, file)
|
|
}
|
|
|
|
func (m *Matcher) Parse(r io.Reader, file string) error {
|
|
m.mut.Lock()
|
|
defer m.mut.Unlock()
|
|
|
|
seen := map[string]bool{file: true}
|
|
patterns, err := parseIgnoreFile(r, file, seen)
|
|
// Error is saved and returned at the end. We process the patterns
|
|
// (possibly blank) anyway.
|
|
|
|
newHash := hashPatterns(patterns)
|
|
if newHash == m.curHash {
|
|
// We've already loaded exactly these patterns.
|
|
return err
|
|
}
|
|
|
|
m.curHash = newHash
|
|
m.patterns = patterns
|
|
if m.withCache {
|
|
m.matches = newCache(patterns)
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
func (m *Matcher) Match(file string) (result bool) {
|
|
m.mut.Lock()
|
|
defer m.mut.Unlock()
|
|
|
|
if len(m.patterns) == 0 {
|
|
return false
|
|
}
|
|
|
|
if m.matches != nil {
|
|
// Check the cache for a known result.
|
|
res, ok := m.matches.get(file)
|
|
if ok {
|
|
return res
|
|
}
|
|
|
|
// Update the cache with the result at return time
|
|
defer func() {
|
|
m.matches.set(file, result)
|
|
}()
|
|
}
|
|
|
|
// Check all the patterns for a match.
|
|
for _, pattern := range m.patterns {
|
|
if pattern.match.MatchString(file) {
|
|
return pattern.include
|
|
}
|
|
}
|
|
|
|
// Default to false.
|
|
return false
|
|
}
|
|
|
|
// Patterns return a list of the loaded regexp patterns, as strings
|
|
func (m *Matcher) Patterns() []string {
|
|
m.mut.Lock()
|
|
defer m.mut.Unlock()
|
|
|
|
patterns := make([]string, len(m.patterns))
|
|
for i, pat := range m.patterns {
|
|
patterns[i] = pat.String()
|
|
}
|
|
return patterns
|
|
}
|
|
|
|
func (m *Matcher) Hash() string {
|
|
m.mut.Lock()
|
|
defer m.mut.Unlock()
|
|
return m.curHash
|
|
}
|
|
|
|
func (m *Matcher) Stop() {
|
|
close(m.stop)
|
|
}
|
|
|
|
func (m *Matcher) clean(d time.Duration) {
|
|
t := time.NewTimer(d / 2)
|
|
for {
|
|
select {
|
|
case <-m.stop:
|
|
return
|
|
case <-t.C:
|
|
m.mut.Lock()
|
|
if m.matches != nil {
|
|
m.matches.clean(d)
|
|
}
|
|
t.Reset(d / 2)
|
|
m.mut.Unlock()
|
|
}
|
|
}
|
|
}
|
|
|
|
func hashPatterns(patterns []Pattern) string {
|
|
h := md5.New()
|
|
for _, pat := range patterns {
|
|
h.Write([]byte(pat.String()))
|
|
h.Write([]byte("\n"))
|
|
}
|
|
return fmt.Sprintf("%x", h.Sum(nil))
|
|
}
|
|
|
|
func loadIgnoreFile(file string, seen map[string]bool) ([]Pattern, error) {
|
|
if seen[file] {
|
|
return nil, fmt.Errorf("Multiple include of ignore file %q", file)
|
|
}
|
|
seen[file] = true
|
|
|
|
fd, err := os.Open(file)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer fd.Close()
|
|
|
|
return parseIgnoreFile(fd, file, seen)
|
|
}
|
|
|
|
func parseIgnoreFile(fd io.Reader, currentFile string, seen map[string]bool) ([]Pattern, error) {
|
|
var patterns []Pattern
|
|
|
|
addPattern := func(line string) error {
|
|
include := true
|
|
if strings.HasPrefix(line, "!") {
|
|
line = line[1:]
|
|
include = false
|
|
}
|
|
|
|
if strings.HasPrefix(line, "/") {
|
|
// Pattern is rooted in the current dir only
|
|
exp, err := fnmatch.Convert(line[1:], fnmatch.FNM_PATHNAME)
|
|
if err != nil {
|
|
return fmt.Errorf("Invalid pattern %q in ignore file", line)
|
|
}
|
|
patterns = append(patterns, Pattern{exp, include})
|
|
} else if strings.HasPrefix(line, "**/") {
|
|
// Add the pattern as is, and without **/ so it matches in current dir
|
|
exp, err := fnmatch.Convert(line, fnmatch.FNM_PATHNAME)
|
|
if err != nil {
|
|
return fmt.Errorf("Invalid pattern %q in ignore file", line)
|
|
}
|
|
patterns = append(patterns, Pattern{exp, include})
|
|
|
|
exp, err = fnmatch.Convert(line[3:], fnmatch.FNM_PATHNAME)
|
|
if err != nil {
|
|
return fmt.Errorf("Invalid pattern %q in ignore file", line)
|
|
}
|
|
patterns = append(patterns, Pattern{exp, include})
|
|
} else if strings.HasPrefix(line, "#include ") {
|
|
includeFile := filepath.Join(filepath.Dir(currentFile), line[len("#include "):])
|
|
includes, err := loadIgnoreFile(includeFile, seen)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
patterns = append(patterns, includes...)
|
|
} else {
|
|
// Path name or pattern, add it so it matches files both in
|
|
// current directory and subdirs.
|
|
exp, err := fnmatch.Convert(line, fnmatch.FNM_PATHNAME)
|
|
if err != nil {
|
|
return fmt.Errorf("Invalid pattern %q in ignore file", line)
|
|
}
|
|
patterns = append(patterns, Pattern{exp, include})
|
|
|
|
exp, err = fnmatch.Convert("**/"+line, fnmatch.FNM_PATHNAME)
|
|
if err != nil {
|
|
return fmt.Errorf("Invalid pattern %q in ignore file", line)
|
|
}
|
|
patterns = append(patterns, Pattern{exp, include})
|
|
}
|
|
return nil
|
|
}
|
|
|
|
scanner := bufio.NewScanner(fd)
|
|
var err error
|
|
for scanner.Scan() {
|
|
line := strings.TrimSpace(scanner.Text())
|
|
switch {
|
|
case line == "":
|
|
continue
|
|
case strings.HasPrefix(line, "//"):
|
|
continue
|
|
case strings.HasPrefix(line, "#"):
|
|
err = addPattern(line)
|
|
case strings.HasSuffix(line, "/**"):
|
|
err = addPattern(line)
|
|
case strings.HasSuffix(line, "/"):
|
|
err = addPattern(line)
|
|
if err == nil {
|
|
err = addPattern(line + "**")
|
|
}
|
|
default:
|
|
err = addPattern(line)
|
|
if err == nil {
|
|
err = addPattern(line + "/**")
|
|
}
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
return patterns, nil
|
|
}
|