
The whitespace linter was added in #673. Enable it and fix found issues. Add auto-fixing to the whitespace linter.
319 lines
7.9 KiB
Go
319 lines
7.9 KiB
Go
package lint
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"runtime/debug"
|
|
"sort"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/golangci/golangci-lint/pkg/lint/lintersdb"
|
|
|
|
"github.com/golangci/golangci-lint/pkg/fsutils"
|
|
|
|
"github.com/golangci/golangci-lint/pkg/config"
|
|
"github.com/golangci/golangci-lint/pkg/goutil"
|
|
"github.com/golangci/golangci-lint/pkg/lint/astcache"
|
|
"github.com/golangci/golangci-lint/pkg/lint/linter"
|
|
"github.com/golangci/golangci-lint/pkg/logutils"
|
|
"github.com/golangci/golangci-lint/pkg/packages"
|
|
"github.com/golangci/golangci-lint/pkg/result"
|
|
"github.com/golangci/golangci-lint/pkg/result/processors"
|
|
"github.com/golangci/golangci-lint/pkg/timeutils"
|
|
)
|
|
|
|
type Runner struct {
|
|
Processors []processors.Processor
|
|
Log logutils.Log
|
|
}
|
|
|
|
func NewRunner(astCache *astcache.Cache, cfg *config.Config, log logutils.Log, goenv *goutil.Env,
|
|
lineCache *fsutils.LineCache, dbManager *lintersdb.Manager) (*Runner, error) {
|
|
icfg := cfg.Issues
|
|
excludePatterns := icfg.ExcludePatterns
|
|
if icfg.UseDefaultExcludes {
|
|
excludePatterns = append(excludePatterns, config.GetDefaultExcludePatternsStrings()...)
|
|
}
|
|
|
|
var excludeTotalPattern string
|
|
if len(excludePatterns) != 0 {
|
|
excludeTotalPattern = fmt.Sprintf("(%s)", strings.Join(excludePatterns, "|"))
|
|
}
|
|
|
|
skipFilesProcessor, err := processors.NewSkipFiles(cfg.Run.SkipFiles)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
skipDirs := cfg.Run.SkipDirs
|
|
if cfg.Run.UseDefaultSkipDirs {
|
|
skipDirs = append(skipDirs, packages.StdExcludeDirRegexps...)
|
|
}
|
|
skipDirsProcessor, err := processors.NewSkipDirs(skipDirs, log.Child("skip dirs"), cfg.Run.Args)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
var excludeRules []processors.ExcludeRule
|
|
for _, r := range icfg.ExcludeRules {
|
|
excludeRules = append(excludeRules, processors.ExcludeRule{
|
|
Text: r.Text,
|
|
Source: r.Source,
|
|
Path: r.Path,
|
|
Linters: r.Linters,
|
|
})
|
|
}
|
|
|
|
return &Runner{
|
|
Processors: []processors.Processor{
|
|
processors.NewCgo(goenv),
|
|
processors.NewFilenameUnadjuster(astCache, log.Child("filename_unadjuster")), // must go after Cgo
|
|
processors.NewPathPrettifier(), // must be before diff, nolint and exclude autogenerated processor at least
|
|
skipFilesProcessor,
|
|
skipDirsProcessor, // must be after path prettifier
|
|
|
|
processors.NewAutogeneratedExclude(astCache),
|
|
processors.NewIdentifierMarker(), // must be befor exclude
|
|
processors.NewExclude(excludeTotalPattern),
|
|
processors.NewExcludeRules(excludeRules, lineCache, log.Child("exclude_rules")),
|
|
processors.NewNolint(astCache, log.Child("nolint"), dbManager),
|
|
|
|
processors.NewUniqByLine(cfg),
|
|
processors.NewDiff(icfg.Diff, icfg.DiffFromRevision, icfg.DiffPatchFilePath),
|
|
processors.NewMaxPerFileFromLinter(cfg),
|
|
processors.NewMaxSameIssues(icfg.MaxSameIssues, log.Child("max_same_issues"), cfg),
|
|
processors.NewMaxFromLinter(icfg.MaxIssuesPerLinter, log.Child("max_from_linter"), cfg),
|
|
processors.NewSourceCode(lineCache, log.Child("source_code")),
|
|
processors.NewPathShortener(),
|
|
},
|
|
Log: log,
|
|
}, nil
|
|
}
|
|
|
|
type lintRes struct {
|
|
linter *linter.Config
|
|
err error
|
|
issues []result.Issue
|
|
}
|
|
|
|
func (r *Runner) runLinterSafe(ctx context.Context, lintCtx *linter.Context,
|
|
lc *linter.Config) (ret []result.Issue, err error) {
|
|
defer func() {
|
|
if panicData := recover(); panicData != nil {
|
|
err = fmt.Errorf("panic occurred: %s", panicData)
|
|
r.Log.Warnf("Panic stack trace: %s", debug.Stack())
|
|
}
|
|
}()
|
|
|
|
specificLintCtx := *lintCtx
|
|
specificLintCtx.Log = r.Log.Child(lc.Name())
|
|
issues, err := lc.Linter.Run(ctx, &specificLintCtx)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
for _, i := range issues {
|
|
i.FromLinter = lc.Name()
|
|
}
|
|
|
|
return issues, nil
|
|
}
|
|
|
|
func (r Runner) runWorker(ctx context.Context, lintCtx *linter.Context,
|
|
tasksCh <-chan *linter.Config, lintResultsCh chan<- lintRes, name string) {
|
|
sw := timeutils.NewStopwatch(name, r.Log)
|
|
defer sw.Print()
|
|
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
return
|
|
case lc, ok := <-tasksCh:
|
|
if !ok {
|
|
return
|
|
}
|
|
if ctx.Err() != nil {
|
|
// XXX: if check it in only int a select
|
|
// it's possible to not enter to this case until tasksCh is empty.
|
|
return
|
|
}
|
|
var issues []result.Issue
|
|
var err error
|
|
sw.TrackStage(lc.Name(), func() {
|
|
issues, err = r.runLinterSafe(ctx, lintCtx, lc)
|
|
})
|
|
lintResultsCh <- lintRes{
|
|
linter: lc,
|
|
err: err,
|
|
issues: issues,
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
func (r Runner) logWorkersStat(workersFinishTimes []time.Time) {
|
|
lastFinishTime := workersFinishTimes[0]
|
|
for _, t := range workersFinishTimes {
|
|
if t.After(lastFinishTime) {
|
|
lastFinishTime = t
|
|
}
|
|
}
|
|
|
|
logStrings := []string{}
|
|
for i, t := range workersFinishTimes {
|
|
if t.Equal(lastFinishTime) {
|
|
continue
|
|
}
|
|
|
|
logStrings = append(logStrings, fmt.Sprintf("#%d: %s", i+1, lastFinishTime.Sub(t)))
|
|
}
|
|
|
|
r.Log.Infof("Workers idle times: %s", strings.Join(logStrings, ", "))
|
|
}
|
|
|
|
func getSortedLintersConfigs(linters []*linter.Config) []*linter.Config {
|
|
ret := make([]*linter.Config, len(linters))
|
|
copy(ret, linters)
|
|
|
|
sort.Slice(ret, func(i, j int) bool {
|
|
return ret[i].GetSpeed() < ret[j].GetSpeed()
|
|
})
|
|
|
|
return ret
|
|
}
|
|
|
|
func (r *Runner) runWorkers(ctx context.Context, lintCtx *linter.Context, linters []*linter.Config) <-chan lintRes {
|
|
tasksCh := make(chan *linter.Config, len(linters))
|
|
lintResultsCh := make(chan lintRes, len(linters))
|
|
var wg sync.WaitGroup
|
|
|
|
workersFinishTimes := make([]time.Time, lintCtx.Cfg.Run.Concurrency)
|
|
|
|
for i := 0; i < lintCtx.Cfg.Run.Concurrency; i++ {
|
|
wg.Add(1)
|
|
go func(i int) {
|
|
defer wg.Done()
|
|
name := fmt.Sprintf("worker.%d", i+1)
|
|
r.runWorker(ctx, lintCtx, tasksCh, lintResultsCh, name)
|
|
workersFinishTimes[i] = time.Now()
|
|
}(i)
|
|
}
|
|
|
|
lcs := getSortedLintersConfigs(linters)
|
|
for _, lc := range lcs {
|
|
tasksCh <- lc
|
|
}
|
|
close(tasksCh)
|
|
|
|
go func() {
|
|
wg.Wait()
|
|
close(lintResultsCh)
|
|
|
|
r.logWorkersStat(workersFinishTimes)
|
|
}()
|
|
|
|
return lintResultsCh
|
|
}
|
|
|
|
func (r Runner) processLintResults(inCh <-chan lintRes) <-chan lintRes {
|
|
outCh := make(chan lintRes, 64)
|
|
|
|
go func() {
|
|
sw := timeutils.NewStopwatch("processing", r.Log)
|
|
|
|
var issuesBefore, issuesAfter int
|
|
defer close(outCh)
|
|
|
|
for res := range inCh {
|
|
if res.err != nil {
|
|
r.Log.Warnf("Can't run linter %s: %s", res.linter.Name(), res.err)
|
|
continue
|
|
}
|
|
|
|
if len(res.issues) != 0 {
|
|
issuesBefore += len(res.issues)
|
|
res.issues = r.processIssues(res.issues, sw)
|
|
issuesAfter += len(res.issues)
|
|
outCh <- res
|
|
}
|
|
}
|
|
|
|
// finalize processors: logging, clearing, no heavy work here
|
|
|
|
for _, p := range r.Processors {
|
|
p := p
|
|
sw.TrackStage(p.Name(), func() {
|
|
p.Finish()
|
|
})
|
|
}
|
|
|
|
if issuesBefore != issuesAfter {
|
|
r.Log.Infof("Issues before processing: %d, after processing: %d", issuesBefore, issuesAfter)
|
|
}
|
|
sw.PrintStages()
|
|
}()
|
|
|
|
return outCh
|
|
}
|
|
|
|
func collectIssues(resCh <-chan lintRes) <-chan result.Issue {
|
|
retIssues := make(chan result.Issue, 1024)
|
|
go func() {
|
|
defer close(retIssues)
|
|
|
|
for res := range resCh {
|
|
if len(res.issues) == 0 {
|
|
continue
|
|
}
|
|
|
|
for _, i := range res.issues {
|
|
retIssues <- i
|
|
}
|
|
}
|
|
}()
|
|
|
|
return retIssues
|
|
}
|
|
|
|
func (r Runner) Run(ctx context.Context, linters []*linter.Config, lintCtx *linter.Context) <-chan result.Issue {
|
|
lintResultsCh := r.runWorkers(ctx, lintCtx, linters)
|
|
processedLintResultsCh := r.processLintResults(lintResultsCh)
|
|
if ctx.Err() != nil {
|
|
// XXX: always process issues, even if timeout occurred
|
|
finishedLintersN := 0
|
|
for range processedLintResultsCh {
|
|
finishedLintersN++
|
|
}
|
|
|
|
r.Log.Errorf("%d/%d linters finished: deadline exceeded",
|
|
finishedLintersN, len(linters))
|
|
}
|
|
|
|
return collectIssues(processedLintResultsCh)
|
|
}
|
|
|
|
func (r *Runner) processIssues(issues []result.Issue, sw *timeutils.Stopwatch) []result.Issue {
|
|
for _, p := range r.Processors {
|
|
var newIssues []result.Issue
|
|
var err error
|
|
p := p
|
|
sw.TrackStage(p.Name(), func() {
|
|
newIssues, err = p.Process(issues)
|
|
})
|
|
|
|
if err != nil {
|
|
r.Log.Warnf("Can't process result by %s processor: %s", p.Name(), err)
|
|
} else {
|
|
issues = newIssues
|
|
}
|
|
|
|
if issues == nil {
|
|
issues = []result.Issue{}
|
|
}
|
|
}
|
|
|
|
return issues
|
|
}
|