Auto-rediscover new glob patterns

This commit is contained in:
2026-03-23 20:39:12 +01:00
parent d3160c7dd4
commit b89caa594c
3 changed files with 89 additions and 28 deletions

View File

@@ -11,30 +11,27 @@ import (
"path/filepath"
"strings"
"syscall"
"time"
pb "git.ipng.ch/ipng/nginx-logtail/proto/logtailpb"
"google.golang.org/grpc"
)
func main() {
listen := flag.String("listen", ":9090", "gRPC listen address")
logPaths := flag.String("logs", "", "comma-separated log file paths/globs to tail")
logsFile := flag.String("logs-file", "", "file containing one log path/glob per line")
source := flag.String("source", hostname(), "name for this collector (default: hostname)")
v4prefix := flag.Int("v4prefix", 24, "IPv4 prefix length for client bucketing")
v6prefix := flag.Int("v6prefix", 48, "IPv6 prefix length for client bucketing")
listen := flag.String("listen", ":9090", "gRPC listen address")
logPaths := flag.String("logs", "", "comma-separated log file paths/globs to tail")
logsFile := flag.String("logs-file", "", "file containing one log path/glob per line")
source := flag.String("source", hostname(), "name for this collector (default: hostname)")
v4prefix := flag.Int("v4prefix", 24, "IPv4 prefix length for client bucketing")
v6prefix := flag.Int("v6prefix", 48, "IPv6 prefix length for client bucketing")
scanInterval := flag.Duration("scan-interval", 10*time.Second, "how often to rescan glob patterns for new/removed files")
flag.Parse()
patterns := collectPatterns(*logPaths, *logsFile)
if len(patterns) == 0 {
log.Fatal("collector: no log paths specified; use --logs or --logs-file")
}
paths := expandGlobs(patterns)
if len(paths) == 0 {
log.Fatal("collector: no log files matched the specified patterns")
}
log.Printf("collector: tailing %d file(s)", len(paths))
log.Printf("collector: watching %d pattern(s), rescan every %s", len(patterns), *scanInterval)
ctx, stop := signal.NotifyContext(context.Background(), syscall.SIGINT, syscall.SIGTERM)
defer stop()
@@ -45,7 +42,7 @@ func main() {
store := NewStore(*source)
go store.Run(ch)
tailer := NewMultiTailer(paths, *v4prefix, *v6prefix, ch)
tailer := NewMultiTailer(patterns, *scanInterval, *v4prefix, *v6prefix, ch)
go tailer.Run(ctx)
lis, err := net.Listen("tcp", *listen)

View File

@@ -28,20 +28,25 @@ type reopenMsg struct {
// fsnotify.Watcher (one inotify instance). This scales to hundreds of files
// without hitting the kernel limit on inotify instances per user.
type MultiTailer struct {
paths []string
v4bits int
v6bits int
ch chan<- LogRecord
patterns []string
scanInterval time.Duration
v4bits int
v6bits int
ch chan<- LogRecord
}
func NewMultiTailer(paths []string, v4bits, v6bits int, ch chan<- LogRecord) *MultiTailer {
return &MultiTailer{paths: paths, v4bits: v4bits, v6bits: v6bits, ch: ch}
func NewMultiTailer(patterns []string, scanInterval time.Duration, v4bits, v6bits int, ch chan<- LogRecord) *MultiTailer {
return &MultiTailer{patterns: patterns, scanInterval: scanInterval, v4bits: v4bits, v6bits: v6bits, ch: ch}
}
// Run tails all configured files until ctx is cancelled.
// All files share one fsnotify.Watcher. Log rotation is handled per-file:
// on RENAME/REMOVE the old fd is drained then a retry goroutine re-opens
// the original path and hands it back via a channel.
//
// A periodic rescan re-expands the glob patterns so that files created after
// startup are picked up automatically and files that have disappeared (and
// are no longer matched by any pattern) are retired.
func (mt *MultiTailer) Run(ctx context.Context) {
watcher, err := fsnotify.NewWatcher()
if err != nil {
@@ -49,21 +54,33 @@ func (mt *MultiTailer) Run(ctx context.Context) {
}
defer watcher.Close()
files := make(map[string]*fileState, len(mt.paths))
reopenCh := make(chan reopenMsg, len(mt.paths))
files := make(map[string]*fileState)
retrying := make(map[string]struct{}) // paths currently in a retryOpen goroutine
reopenCh := make(chan reopenMsg, 32)
// Open all files and seek to EOF.
for _, path := range mt.paths {
startRetry := func(path string) {
if _, already := retrying[path]; already {
return
}
retrying[path] = struct{}{}
go retryOpen(ctx, path, watcher, reopenCh)
}
// Initial scan.
for _, path := range expandGlobs(mt.patterns) {
fs, err := openAndSeekEOF(path, watcher)
if err != nil {
log.Printf("tailer: %s not found, will retry: %v", path, err)
go retryOpen(ctx, path, watcher, reopenCh)
startRetry(path)
continue
}
files[path] = fs
log.Printf("tailer: watching %s", path)
}
ticker := time.NewTicker(mt.scanInterval)
defer ticker.Stop()
for {
select {
case <-ctx.Done():
@@ -76,12 +93,16 @@ func (mt *MultiTailer) Run(ctx context.Context) {
if !ok {
return
}
delete(retrying, msg.path)
files[msg.path] = &fileState{f: msg.f, reader: bufio.NewReader(msg.f)}
if err := watcher.Add(msg.path); err != nil {
log.Printf("tailer: watcher re-add failed for %s: %v", msg.path, err)
}
log.Printf("tailer: re-opened %s after rotation", msg.path)
case <-ticker.C:
mt.rescan(ctx, watcher, files, retrying, reopenCh, startRetry)
case event, ok := <-watcher.Events:
if !ok {
return
@@ -99,7 +120,7 @@ func (mt *MultiTailer) Run(ctx context.Context) {
fs.f.Close()
delete(files, event.Name)
_ = watcher.Remove(event.Name)
go retryOpen(ctx, event.Name, watcher, reopenCh)
startRetry(event.Name)
}
case err, ok := <-watcher.Errors:
@@ -111,6 +132,49 @@ func (mt *MultiTailer) Run(ctx context.Context) {
}
}
// rescan re-expands the glob patterns and reconciles against the current file
// set: new matches are opened (or queued for retry), and files no longer
// matched by any pattern are drained, closed, and retired.
func (mt *MultiTailer) rescan(
ctx context.Context,
watcher *fsnotify.Watcher,
files map[string]*fileState,
retrying map[string]struct{},
_ chan reopenMsg,
startRetry func(string),
) {
current := make(map[string]struct{})
for _, path := range expandGlobs(mt.patterns) {
current[path] = struct{}{}
if _, inFiles := files[path]; inFiles {
continue
}
if _, isRetrying := retrying[path]; isRetrying {
continue
}
// Newly matched file — try to open it right away.
fs, err := openAndSeekEOF(path, watcher)
if err != nil {
startRetry(path)
continue
}
files[path] = fs
log.Printf("tailer: discovered %s", path)
}
// Retire files that no longer match any pattern and are not being rotated
// (rotation is handled by the RENAME/REMOVE event path, not here).
for path, fs := range files {
if _, matched := current[path]; !matched {
mt.readLines(fs.reader)
fs.f.Close()
_ = watcher.Remove(path)
delete(files, path)
log.Printf("tailer: retired %s (no longer matched by any pattern)", path)
}
}
}
// openAndSeekEOF opens path, seeks to EOF, and registers it with watcher.
func openAndSeekEOF(path string, watcher *fsnotify.Watcher) (*fileState, error) {
f, err := os.Open(path)

View File

@@ -28,7 +28,7 @@ func TestMultiTailerReadsLines(t *testing.T) {
defer f.Close()
ch := make(chan LogRecord, 100)
mt := NewMultiTailer([]string{path}, 24, 48, ch)
mt := NewMultiTailer([]string{path}, time.Hour, 24, 48, ch)
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
@@ -67,7 +67,7 @@ func TestMultiTailerMultipleFiles(t *testing.T) {
}
ch := make(chan LogRecord, 200)
mt := NewMultiTailer(paths, 24, 48, ch)
mt := NewMultiTailer(paths, time.Hour, 24, 48, ch)
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
go mt.Run(ctx)
@@ -95,7 +95,7 @@ func TestMultiTailerLogRotation(t *testing.T) {
}
ch := make(chan LogRecord, 100)
mt := NewMultiTailer([]string{path}, 24, 48, ch)
mt := NewMultiTailer([]string{path}, time.Hour, 24, 48, ch)
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
go mt.Run(ctx)