| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900 |
- package detect
- import (
- "bufio"
- "context"
- "fmt"
- "os"
- "strings"
- "sync"
- "sync/atomic"
- "time"
- "github.com/zricethezav/gitleaks/v8/config"
- "github.com/zricethezav/gitleaks/v8/detect/codec"
- "github.com/zricethezav/gitleaks/v8/logging"
- "github.com/zricethezav/gitleaks/v8/regexp"
- "github.com/zricethezav/gitleaks/v8/report"
- "github.com/zricethezav/gitleaks/v8/sources"
- ahocorasick "github.com/BobuSumisu/aho-corasick"
- "github.com/fatih/semgroup"
- "github.com/rs/zerolog"
- "github.com/spf13/viper"
- "golang.org/x/exp/maps"
- )
- const (
- gitleaksAllowSignature = "gitleaks:allow"
- // SlowWarningThreshold is the amount of time to wait before logging that a file is slow.
- // This is useful for identifying problematic files and tuning the allowlist.
- SlowWarningThreshold = 5 * time.Second
- )
- var (
- newLineRegexp = regexp.MustCompile("\n")
- )
- // Detector is the main detector struct
- type Detector struct {
- // Config is the configuration for the detector
- Config config.Config
- // Redact is a flag to redact findings. This is exported
- // so users using gitleaks as a library can set this flag
- // without calling `detector.Start(cmd *cobra.Command)`
- Redact uint
- // verbose is a flag to print findings
- Verbose bool
- // MaxDecodeDepths limits how many recursive decoding passes are allowed
- MaxDecodeDepth int
- // MaxArchiveDepth limits how deep the sources will explore nested archives
- MaxArchiveDepth int
- // files larger than this will be skipped
- MaxTargetMegaBytes int
- // followSymlinks is a flag to enable scanning symlink files
- FollowSymlinks bool
- // NoColor is a flag to disable color output
- NoColor bool
- // IgnoreGitleaksAllow is a flag to ignore gitleaks:allow comments.
- IgnoreGitleaksAllow bool
- // commitMutex is to prevent concurrent access to the
- // commit map when adding commits
- commitMutex *sync.Mutex
- // commitMap is used to keep track of commits that have been scanned.
- // This is only used for logging purposes and git scans.
- commitMap map[string]bool
- // findingMutex is to prevent concurrent access to the
- // findings slice when adding findings.
- findingMutex *sync.Mutex
- // findings is a slice of report.Findings. This is the result
- // of the detector's scan which can then be used to generate a
- // report.
- findings []report.Finding
- // prefilter is a ahocorasick struct used for doing efficient string
- // matching given a set of words (keywords from the rules in the config)
- prefilter ahocorasick.Trie
- // a list of known findings that should be ignored
- baseline []report.Finding
- // path to baseline
- baselinePath string
- // gitleaksIgnore
- gitleaksIgnore map[string]struct{}
- // Sema (https://github.com/fatih/semgroup) controls the concurrency
- Sema *semgroup.Group
- // report-related settings.
- ReportPath string
- Reporter report.Reporter
- TotalBytes atomic.Uint64
- }
- // Fragment is an alias for sources.Fragment for backwards compatibility
- //
- // Deprecated: This will be replaced with sources.Fragment in v9
- type Fragment sources.Fragment
- // NewDetector creates a new detector with the given config
- func NewDetector(cfg config.Config) *Detector {
- return NewDetectorContext(context.Background(), cfg)
- }
- // NewDetectorContext is the same as NewDetector but supports passing in a
- // context to use for timeouts
- func NewDetectorContext(ctx context.Context, cfg config.Config) *Detector {
- return &Detector{
- commitMap: make(map[string]bool),
- gitleaksIgnore: make(map[string]struct{}),
- findingMutex: &sync.Mutex{},
- commitMutex: &sync.Mutex{},
- findings: make([]report.Finding, 0),
- Config: cfg,
- prefilter: *ahocorasick.NewTrieBuilder().AddStrings(maps.Keys(cfg.Keywords)).Build(),
- Sema: semgroup.NewGroup(ctx, 40),
- }
- }
- // NewDetectorDefaultConfig creates a new detector with the default config
- func NewDetectorDefaultConfig() (*Detector, error) {
- viper.SetConfigType("toml")
- err := viper.ReadConfig(strings.NewReader(config.DefaultConfig))
- if err != nil {
- return nil, err
- }
- var vc config.ViperConfig
- err = viper.Unmarshal(&vc)
- if err != nil {
- return nil, err
- }
- cfg, err := vc.Translate()
- if err != nil {
- return nil, err
- }
- return NewDetector(cfg), nil
- }
- func (d *Detector) AddGitleaksIgnore(gitleaksIgnorePath string) error {
- logging.Debug().Str("path", gitleaksIgnorePath).Msgf("found .gitleaksignore file")
- file, err := os.Open(gitleaksIgnorePath)
- if err != nil {
- return err
- }
- defer func() {
- // https://github.com/securego/gosec/issues/512
- if err := file.Close(); err != nil {
- logging.Warn().Err(err).Msgf("Error closing .gitleaksignore file")
- }
- }()
- scanner := bufio.NewScanner(file)
- replacer := strings.NewReplacer("\\", "/")
- for scanner.Scan() {
- line := strings.TrimSpace(scanner.Text())
- // Skip lines that start with a comment
- if line == "" || strings.HasPrefix(line, "#") {
- continue
- }
- // Normalize the path.
- // TODO: Make this a breaking change in v9.
- s := strings.Split(line, ":")
- switch len(s) {
- case 3:
- // Global fingerprint.
- // `file:rule-id:start-line`
- s[0] = replacer.Replace(s[0])
- case 4:
- // Commit fingerprint.
- // `commit:file:rule-id:start-line`
- s[1] = replacer.Replace(s[1])
- default:
- logging.Warn().Str("fingerprint", line).Msg("Invalid .gitleaksignore entry")
- }
- d.gitleaksIgnore[strings.Join(s, ":")] = struct{}{}
- }
- return nil
- }
- // DetectBytes scans the given bytes and returns a list of findings
- func (d *Detector) DetectBytes(content []byte) []report.Finding {
- return d.DetectString(string(content))
- }
- // DetectString scans the given string and returns a list of findings
- func (d *Detector) DetectString(content string) []report.Finding {
- return d.Detect(Fragment{
- Raw: content,
- })
- }
- // DetectSource scans the given source and returns a list of findings
- func (d *Detector) DetectSource(ctx context.Context, source sources.Source) ([]report.Finding, error) {
- err := source.Fragments(ctx, func(fragment sources.Fragment, err error) error {
- logContext := logging.With()
- if len(fragment.FilePath) > 0 {
- logContext = logContext.Str("path", fragment.FilePath)
- }
- if len(fragment.CommitSHA) > 6 {
- logContext = logContext.Str("commit", fragment.CommitSHA[:7])
- d.addCommit(fragment.CommitSHA)
- } else if len(fragment.CommitSHA) > 0 {
- logContext = logContext.Str("commit", fragment.CommitSHA)
- d.addCommit(fragment.CommitSHA)
- logger := logContext.Logger()
- logger.Warn().Msg("commit SHAs should be >= 7 characters long")
- }
- logger := logContext.Logger()
- if err != nil {
- // Log the error and move on to the next fragment
- logger.Error().Err(err).Send()
- return nil
- }
- // both the fragment's content and path should be empty for it to be
- // considered empty at this point because of path based matches
- if len(fragment.Raw) == 0 && len(fragment.FilePath) == 0 {
- logger.Trace().Msg("skipping empty fragment")
- return nil
- }
- var timer *time.Timer
- // Only start the timer in debug mode
- if logger.GetLevel() <= zerolog.DebugLevel {
- timer = time.AfterFunc(SlowWarningThreshold, func() {
- logger.Debug().Msgf("Taking longer than %s to inspect fragment", SlowWarningThreshold.String())
- })
- }
- for _, finding := range d.DetectContext(ctx, Fragment(fragment)) {
- d.AddFinding(finding)
- }
- // Stop the timer if it was created
- if timer != nil {
- timer.Stop()
- }
- return nil
- })
- if _, isGit := source.(*sources.Git); isGit {
- logging.Info().Msgf("%d commits scanned.", len(d.commitMap))
- logging.Debug().Msg("Note: this number might be smaller than expected due to commits with no additions")
- }
- return d.Findings(), err
- }
- // Detect scans the given fragment and returns a list of findings
- func (d *Detector) Detect(fragment Fragment) []report.Finding {
- return d.DetectContext(context.Background(), fragment)
- }
- // DetectContext is the same as Detect but supports passing in a
- // context to use for timeouts
- func (d *Detector) DetectContext(ctx context.Context, fragment Fragment) []report.Finding {
- if fragment.Bytes == nil {
- d.TotalBytes.Add(uint64(len(fragment.Raw)))
- }
- d.TotalBytes.Add(uint64(len(fragment.Bytes)))
- var (
- findings []report.Finding
- logger = func() zerolog.Logger {
- l := logging.With().Str("path", fragment.FilePath)
- if fragment.CommitSHA != "" {
- l = l.Str("commit", fragment.CommitSHA)
- }
- return l.Logger()
- }()
- )
- // check if filepath is allowed
- if fragment.FilePath != "" {
- // is the path our config or baseline file?
- if fragment.FilePath == d.Config.Path || (d.baselinePath != "" && fragment.FilePath == d.baselinePath) {
- logging.Trace().Msg("skipping file: matches config or baseline path")
- return findings
- }
- }
- // check if commit or filepath is allowed.
- if isAllowed, event := checkCommitOrPathAllowed(logger, fragment, d.Config.Allowlists); isAllowed {
- event.Msg("skipping file: global allowlist")
- return findings
- }
- // setup variables to handle different decoding passes
- currentRaw := fragment.Raw
- encodedSegments := []*codec.EncodedSegment{}
- currentDecodeDepth := 0
- decoder := codec.NewDecoder()
- ScanLoop:
- for {
- select {
- case <-ctx.Done():
- break ScanLoop
- default:
- // build keyword map for prefiltering rules
- keywords := make(map[string]bool)
- normalizedRaw := strings.ToLower(currentRaw)
- matches := d.prefilter.MatchString(normalizedRaw)
- for _, m := range matches {
- keywords[normalizedRaw[m.Pos():int(m.Pos())+len(m.Match())]] = true
- }
- for _, rule := range d.Config.Rules {
- select {
- case <-ctx.Done():
- break ScanLoop
- default:
- if len(rule.Keywords) == 0 {
- // if no keywords are associated with the rule always scan the
- // fragment using the rule
- findings = append(findings, d.detectRule(fragment, currentRaw, rule, encodedSegments)...)
- continue
- }
- // check if keywords are in the fragment
- for _, k := range rule.Keywords {
- if _, ok := keywords[strings.ToLower(k)]; ok {
- findings = append(findings, d.detectRule(fragment, currentRaw, rule, encodedSegments)...)
- break
- }
- }
- }
- }
- // increment the depth by 1 as we start our decoding pass
- currentDecodeDepth++
- // stop the loop if we've hit our max decoding depth
- if currentDecodeDepth > d.MaxDecodeDepth {
- break ScanLoop
- }
- // decode the currentRaw for the next pass
- currentRaw, encodedSegments = decoder.Decode(currentRaw, encodedSegments)
- // stop the loop when there's nothing else to decode
- if len(encodedSegments) == 0 {
- break ScanLoop
- }
- }
- }
- return filter(findings, d.Redact)
- }
- // detectRule scans the given fragment for the given rule and returns a list of findings
- func (d *Detector) detectRule(fragment Fragment, currentRaw string, r config.Rule, encodedSegments []*codec.EncodedSegment) []report.Finding {
- var (
- findings []report.Finding
- logger = func() zerolog.Logger {
- l := logging.With().Str("rule-id", r.RuleID).Str("path", fragment.FilePath)
- if fragment.CommitSHA != "" {
- l = l.Str("commit", fragment.CommitSHA)
- }
- return l.Logger()
- }()
- )
- if r.SkipReport && !fragment.InheritedFromFinding {
- return findings
- }
- // check if commit or file is allowed for this rule.
- if isAllowed, event := checkCommitOrPathAllowed(logger, fragment, r.Allowlists); isAllowed {
- event.Msg("skipping file: rule allowlist")
- return findings
- }
- if r.Path != nil {
- if r.Regex == nil && len(encodedSegments) == 0 {
- // Path _only_ rule
- if r.Path.MatchString(fragment.FilePath) || (fragment.WindowsFilePath != "" && r.Path.MatchString(fragment.WindowsFilePath)) {
- finding := report.Finding{
- Commit: fragment.CommitSHA,
- RuleID: r.RuleID,
- Description: r.Description,
- File: fragment.FilePath,
- SymlinkFile: fragment.SymlinkFile,
- Match: "file detected: " + fragment.FilePath,
- Tags: r.Tags,
- }
- if fragment.CommitInfo != nil {
- finding.Author = fragment.CommitInfo.AuthorName
- finding.Date = fragment.CommitInfo.Date
- finding.Email = fragment.CommitInfo.AuthorEmail
- finding.Link = createScmLink(fragment.CommitInfo.Remote, finding)
- finding.Message = fragment.CommitInfo.Message
- }
- return append(findings, finding)
- }
- } else {
- // if path is set _and_ a regex is set, then we need to check both
- // so if the path does not match, then we should return early and not
- // consider the regex
- if !(r.Path.MatchString(fragment.FilePath) || (fragment.WindowsFilePath != "" && r.Path.MatchString(fragment.WindowsFilePath))) {
- return findings
- }
- }
- }
- // if path only rule, skip content checks
- if r.Regex == nil {
- return findings
- }
- // if flag configure and raw data size bigger then the flag
- if d.MaxTargetMegaBytes > 0 {
- rawLength := len(currentRaw) / 1_000_000
- if rawLength > d.MaxTargetMegaBytes {
- logger.Debug().
- Int("size", rawLength).
- Int("max-size", d.MaxTargetMegaBytes).
- Msg("skipping fragment: size")
- return findings
- }
- }
- matches := r.Regex.FindAllStringIndex(currentRaw, -1)
- if len(matches) == 0 {
- return findings
- }
- // TODO profile this, probably should replace with something more efficient
- newlineIndices := newLineRegexp.FindAllStringIndex(fragment.Raw, -1)
- // use currentRaw instead of fragment.Raw since this represents the current
- // decoding pass on the text
- for _, matchIndex := range r.Regex.FindAllStringIndex(currentRaw, -1) {
- // Extract secret from match
- secret := strings.Trim(currentRaw[matchIndex[0]:matchIndex[1]], "\n")
- // For any meta data from decoding
- var metaTags []string
- currentLine := ""
- // Check if the decoded portions of the segment overlap with the match
- // to see if its potentially a new match
- if len(encodedSegments) > 0 {
- segments := codec.SegmentsWithDecodedOverlap(encodedSegments, matchIndex[0], matchIndex[1])
- if len(segments) == 0 {
- // This item has already been added to a finding
- continue
- }
- matchIndex = codec.AdjustMatchIndex(segments, matchIndex)
- metaTags = append(metaTags, codec.Tags(segments)...)
- currentLine = codec.CurrentLine(segments, currentRaw)
- } else {
- // Fixes: https://github.com/gitleaks/gitleaks/issues/1352
- // removes the incorrectly following line that was detected by regex expression '\n'
- matchIndex[1] = matchIndex[0] + len(secret)
- }
- // determine location of match. Note that the location
- // in the finding will be the line/column numbers of the _match_
- // not the _secret_, which will be different if the secretGroup
- // value is set for this rule
- loc := location(newlineIndices, fragment.Raw, matchIndex)
- if matchIndex[1] > loc.endLineIndex {
- loc.endLineIndex = matchIndex[1]
- }
- finding := report.Finding{
- Commit: fragment.CommitSHA,
- RuleID: r.RuleID,
- Description: r.Description,
- StartLine: fragment.StartLine + loc.startLine,
- EndLine: fragment.StartLine + loc.endLine,
- StartColumn: loc.startColumn,
- EndColumn: loc.endColumn,
- Line: fragment.Raw[loc.startLineIndex:loc.endLineIndex],
- Match: secret,
- Secret: secret,
- File: fragment.FilePath,
- SymlinkFile: fragment.SymlinkFile,
- Tags: append(r.Tags, metaTags...),
- }
- if fragment.CommitInfo != nil {
- finding.Author = fragment.CommitInfo.AuthorName
- finding.Date = fragment.CommitInfo.Date
- finding.Email = fragment.CommitInfo.AuthorEmail
- finding.Link = createScmLink(fragment.CommitInfo.Remote, finding)
- finding.Message = fragment.CommitInfo.Message
- }
- if !d.IgnoreGitleaksAllow && strings.Contains(finding.Line, gitleaksAllowSignature) {
- logger.Trace().
- Str("finding", finding.Secret).
- Msg("skipping finding: 'gitleaks:allow' signature")
- continue
- }
- if currentLine == "" {
- currentLine = finding.Line
- }
- // Set the value of |secret|, if the pattern contains at least one capture group.
- // (The first element is the full match, hence we check >= 2.)
- groups := r.Regex.FindStringSubmatch(finding.Secret)
- if len(groups) >= 2 {
- if r.SecretGroup > 0 {
- if len(groups) <= r.SecretGroup {
- // Config validation should prevent this
- continue
- }
- finding.Secret = groups[r.SecretGroup]
- } else {
- // If |secretGroup| is not set, we will use the first suitable capture group.
- for _, s := range groups[1:] {
- if len(s) > 0 {
- finding.Secret = s
- break
- }
- }
- }
- }
- // check entropy
- entropy := shannonEntropy(finding.Secret)
- finding.Entropy = float32(entropy)
- if r.Entropy != 0.0 {
- // entropy is too low, skip this finding
- if entropy <= r.Entropy {
- logger.Trace().
- Str("finding", finding.Secret).
- Float32("entropy", finding.Entropy).
- Msg("skipping finding: low entropy")
- continue
- }
- }
- // check if the result matches any of the global allowlists.
- if isAllowed, event := checkFindingAllowed(logger, finding, fragment, currentLine, d.Config.Allowlists); isAllowed {
- event.Msg("skipping finding: global allowlist")
- continue
- }
- // check if the result matches any of the rule allowlists.
- if isAllowed, event := checkFindingAllowed(logger, finding, fragment, currentLine, r.Allowlists); isAllowed {
- event.Msg("skipping finding: rule allowlist")
- continue
- }
- findings = append(findings, finding)
- }
- // Handle required rules (multi-part rules)
- if fragment.InheritedFromFinding || len(r.RequiredRules) == 0 {
- return findings
- }
- // Process required rules and create findings with auxiliary findings
- return d.processRequiredRules(fragment, currentRaw, r, encodedSegments, findings, logger)
- }
- // processRequiredRules handles the logic for multi-part rules with auxiliary findings
- func (d *Detector) processRequiredRules(fragment Fragment, currentRaw string, r config.Rule, encodedSegments []*codec.EncodedSegment, primaryFindings []report.Finding, logger zerolog.Logger) []report.Finding {
- if len(primaryFindings) == 0 {
- logger.Debug().Msg("no primary findings to process for required rules")
- return primaryFindings
- }
- // Pre-collect all required rule findings once
- allRequiredFindings := make(map[string][]report.Finding)
- for _, requiredRule := range r.RequiredRules {
- rule, ok := d.Config.Rules[requiredRule.RuleID]
- if !ok {
- logger.Error().Str("rule-id", requiredRule.RuleID).Msg("required rule not found in config")
- continue
- }
- // Mark fragment as inherited to prevent infinite recursion
- inheritedFragment := fragment
- inheritedFragment.InheritedFromFinding = true
- // Call detectRule once for each required rule
- requiredFindings := d.detectRule(inheritedFragment, currentRaw, rule, encodedSegments)
- allRequiredFindings[requiredRule.RuleID] = requiredFindings
- logger.Debug().
- Str("rule-id", requiredRule.RuleID).
- Int("findings", len(requiredFindings)).
- Msg("collected required rule findings")
- }
- var finalFindings []report.Finding
- // Now process each primary finding against the pre-collected required findings
- for _, primaryFinding := range primaryFindings {
- var requiredFindings []*report.RequiredFinding
- for _, requiredRule := range r.RequiredRules {
- foundRequiredFindings, exists := allRequiredFindings[requiredRule.RuleID]
- if !exists {
- continue // Rule wasn't found earlier, skip
- }
- // Filter findings that are within proximity of the primary finding
- for _, requiredFinding := range foundRequiredFindings {
- if d.withinProximity(primaryFinding, requiredFinding, requiredRule) {
- req := &report.RequiredFinding{
- RuleID: requiredFinding.RuleID,
- StartLine: requiredFinding.StartLine,
- EndLine: requiredFinding.EndLine,
- StartColumn: requiredFinding.StartColumn,
- EndColumn: requiredFinding.EndColumn,
- Line: requiredFinding.Line,
- Match: requiredFinding.Match,
- Secret: requiredFinding.Secret,
- }
- requiredFindings = append(requiredFindings, req)
- }
- }
- }
- // Check if we have at least one auxiliary finding for each required rule
- if len(requiredFindings) > 0 && d.hasAllRequiredRules(requiredFindings, r.RequiredRules) {
- // Create a finding with auxiliary findings
- newFinding := primaryFinding // Copy the primary finding
- newFinding.AddRequiredFindings(requiredFindings)
- finalFindings = append(finalFindings, newFinding)
- logger.Debug().
- Str("primary-rule", r.RuleID).
- Int("primary-line", primaryFinding.StartLine).
- Int("auxiliary-count", len(requiredFindings)).
- Msg("multi-part rule satisfied")
- }
- }
- return finalFindings
- }
- // hasAllRequiredRules checks if we have at least one auxiliary finding for each required rule
- func (d *Detector) hasAllRequiredRules(auxiliaryFindings []*report.RequiredFinding, requiredRules []*config.Required) bool {
- foundRules := make(map[string]bool)
- // AuxiliaryFinding
- for _, aux := range auxiliaryFindings {
- foundRules[aux.RuleID] = true
- }
- for _, required := range requiredRules {
- if !foundRules[required.RuleID] {
- return false
- }
- }
- return true
- }
- func (d *Detector) withinProximity(primary, required report.Finding, requiredRule *config.Required) bool {
- // fmt.Println(requiredRule.WithinLines)
- // If neither within_lines nor within_columns is set, findings just need to be in the same fragment
- if requiredRule.WithinLines == nil && requiredRule.WithinColumns == nil {
- return true
- }
- // Check line proximity (vertical distance)
- if requiredRule.WithinLines != nil {
- lineDiff := abs(primary.StartLine - required.StartLine)
- if lineDiff > *requiredRule.WithinLines {
- return false
- }
- }
- // Check column proximity (horizontal distance)
- if requiredRule.WithinColumns != nil {
- // Use the start column of each finding for proximity calculation
- colDiff := abs(primary.StartColumn - required.StartColumn)
- if colDiff > *requiredRule.WithinColumns {
- return false
- }
- }
- return true
- }
- // abs returns the absolute value of an integer
- func abs(x int) int {
- if x < 0 {
- return -x
- }
- return x
- }
- // AddFinding synchronously adds a finding to the findings slice
- func (d *Detector) AddFinding(finding report.Finding) {
- globalFingerprint := fmt.Sprintf("%s:%s:%d", finding.File, finding.RuleID, finding.StartLine)
- if finding.Commit != "" {
- finding.Fingerprint = fmt.Sprintf("%s:%s:%s:%d", finding.Commit, finding.File, finding.RuleID, finding.StartLine)
- } else {
- finding.Fingerprint = globalFingerprint
- }
- // check if we should ignore this finding
- logger := logging.With().Str("finding", finding.Secret).Logger()
- if _, ok := d.gitleaksIgnore[globalFingerprint]; ok {
- logger.Debug().
- Str("fingerprint", globalFingerprint).
- Msg("skipping finding: global fingerprint")
- return
- } else if finding.Commit != "" {
- // Awkward nested if because I'm not sure how to chain these two conditions.
- if _, ok := d.gitleaksIgnore[finding.Fingerprint]; ok {
- logger.Debug().
- Str("fingerprint", finding.Fingerprint).
- Msgf("skipping finding: fingerprint")
- return
- }
- }
- if d.baseline != nil && !IsNew(finding, d.Redact, d.baseline) {
- logger.Debug().
- Str("fingerprint", finding.Fingerprint).
- Msgf("skipping finding: baseline")
- return
- }
- d.findingMutex.Lock()
- d.findings = append(d.findings, finding)
- if d.Verbose {
- printFinding(finding, d.NoColor)
- }
- d.findingMutex.Unlock()
- }
- // Findings returns the findings added to the detector
- func (d *Detector) Findings() []report.Finding {
- return d.findings
- }
- // AddCommit synchronously adds a commit to the commit slice
- func (d *Detector) addCommit(commit string) {
- d.commitMutex.Lock()
- d.commitMap[commit] = true
- d.commitMutex.Unlock()
- }
- // checkCommitOrPathAllowed evaluates |fragment| against all provided |allowlists|.
- //
- // If the match condition is "OR", only commit and path are checked.
- // Otherwise, if regexes or stopwords are defined this will fail.
- func checkCommitOrPathAllowed(
- logger zerolog.Logger,
- fragment Fragment,
- allowlists []*config.Allowlist,
- ) (bool, *zerolog.Event) {
- if fragment.FilePath == "" && fragment.CommitSHA == "" {
- return false, nil
- }
- for _, a := range allowlists {
- var (
- isAllowed bool
- allowlistChecks []bool
- commitAllowed, _ = a.CommitAllowed(fragment.CommitSHA)
- pathAllowed = a.PathAllowed(fragment.FilePath) || (fragment.WindowsFilePath != "" && a.PathAllowed(fragment.WindowsFilePath))
- )
- // If the condition is "AND" we need to check all conditions.
- if a.MatchCondition == config.AllowlistMatchAnd {
- if len(a.Commits) > 0 {
- allowlistChecks = append(allowlistChecks, commitAllowed)
- }
- if len(a.Paths) > 0 {
- allowlistChecks = append(allowlistChecks, pathAllowed)
- }
- // These will be checked later.
- if len(a.Regexes) > 0 {
- continue
- }
- if len(a.StopWords) > 0 {
- continue
- }
- isAllowed = allTrue(allowlistChecks)
- } else {
- isAllowed = commitAllowed || pathAllowed
- }
- if isAllowed {
- event := logger.Trace().Str("condition", a.MatchCondition.String())
- if commitAllowed {
- event.Bool("allowed-commit", commitAllowed)
- }
- if pathAllowed {
- event.Bool("allowed-path", pathAllowed)
- }
- return true, event
- }
- }
- return false, nil
- }
- // checkFindingAllowed evaluates |finding| against all provided |allowlists|.
- //
- // If the match condition is "OR", only regex and stopwords are run. (Commit and path should be handled separately).
- // Otherwise, all conditions are checked.
- //
- // TODO: The method signature is awkward. I can't think of a better way to log helpful info.
- func checkFindingAllowed(
- logger zerolog.Logger,
- finding report.Finding,
- fragment Fragment,
- currentLine string,
- allowlists []*config.Allowlist,
- ) (bool, *zerolog.Event) {
- for _, a := range allowlists {
- allowlistTarget := finding.Secret
- switch a.RegexTarget {
- case "match":
- allowlistTarget = finding.Match
- case "line":
- allowlistTarget = currentLine
- }
- var (
- checks []bool
- isAllowed bool
- commitAllowed bool
- commit string
- pathAllowed bool
- regexAllowed = a.RegexAllowed(allowlistTarget)
- containsStopword, word = a.ContainsStopWord(finding.Secret)
- )
- // If the condition is "AND" we need to check all conditions.
- if a.MatchCondition == config.AllowlistMatchAnd {
- // Determine applicable checks.
- if len(a.Commits) > 0 {
- commitAllowed, commit = a.CommitAllowed(fragment.CommitSHA)
- checks = append(checks, commitAllowed)
- }
- if len(a.Paths) > 0 {
- pathAllowed = a.PathAllowed(fragment.FilePath) || (fragment.WindowsFilePath != "" && a.PathAllowed(fragment.WindowsFilePath))
- checks = append(checks, pathAllowed)
- }
- if len(a.Regexes) > 0 {
- checks = append(checks, regexAllowed)
- }
- if len(a.StopWords) > 0 {
- checks = append(checks, containsStopword)
- }
- isAllowed = allTrue(checks)
- } else {
- isAllowed = regexAllowed || containsStopword
- }
- if isAllowed {
- event := logger.Trace().
- Str("finding", finding.Secret).
- Str("condition", a.MatchCondition.String())
- if commitAllowed {
- event.Str("allowed-commit", commit)
- }
- if pathAllowed {
- event.Bool("allowed-path", pathAllowed)
- }
- if regexAllowed {
- event.Bool("allowed-regex", regexAllowed)
- }
- if containsStopword {
- event.Str("allowed-stopword", word)
- }
- return true, event
- }
- }
- return false, nil
- }
- func allTrue(bools []bool) bool {
- for _, check := range bools {
- if !check {
- return false
- }
- }
- return true
- }
|