This is the changelog for Nosey Parker. All notable changes to the project will be documented in this file.
The format is based on Keep a Changelog, and this project aspires to use Semantic Versioning.
-
Directories that appear to be Nosey Parker datastore directories are now skipped from scanning (#224).
-
The
/proc
,/sys
, and/dev
paths (special filesystems on Linux) are now ignored by default (#225). This suppresses many innocuous errors that would previously be seen when scanning the root filesystem of a Linux system. -
Lockfiles from a few languages (e.g.,
Cargo.lock
,Pipfile.lock
,go.sum
) are now ignored by default. -
Rules have been modified:
Age Recipient (X25519 public key)
andThingsBoard Access Token
have expanded category metdata.Credentials in ODBC Connection String
detects more occurrences (#227).
-
When using the
--copy-blobs
option, the default output format is nowparquet
(when theparquet
feature is enabled, which it is unless you build with--no-default-features
) (#229).
-
New rules have been added:
-
The
rules check
command now has an optional--pedantic
mode that verifies some additional non-material properties. -
The
scan
command now has a new--copy-blobs-format=FORMAT
option that controls the format used when the--copy-blobs
option is used (#229). A newparquet
format is available and is the default when theparquet
feature is enabled (which it is unless you build with--no-default-features
).
v0.20.0 (2024-10-04)
The most significant feature addition to this release is a new "extensible enumerator" mechanism, which makes it possible to scan content from arbitrary sources with Nosey Parker without having to write it to the filesystem.
This release also includes several changes that speed up and slim down the scanning process. A 10-30% reduction in wall clock time and a 10-50% reduction in memory use are typical, but in some unusual cases, wall clock and memory use are reduced 10-20x.
Happy secret hunting!
-
An experimental "extensible enumerator mechanism" has been added to the
scan
command (#220). This allows Nosey Parker to scan inputs produced by any program that can emit JSON objects to stdout, without having to first write the inputs to the filesystem. It is invoked with the new--enumerator=FILE
option, whereFILE
is a JSON Lines file. Each line of the enumerator file should be a JSON object with one of the following forms:{ "content_base64": "base64-encoded bytestring to scan", "provenance": <arbitrary object> } { "content": "utf8 string to scan", "provenance": <arbitrary object> }
Shell process substitution can make streaming invocation ergonomic, e.g.,
scan --enumerator=<(my-enumerator-program)
.
-
Inputs are now enumerated incrementally as scanning proceeds rather than done in an initial batch step (#216). This reduces peak memory use and wall clock time 10-20%, particularly in environments with slow I/O. A consequence of this change is that the total amount of data to scan is not known until it has actually been scanned, and so the scanning progress bar no longer shows a completion percentage.
-
When cloning Git repositories while scanning, the progress bar for now includes the current repository URL (#212).
-
When scanning, automatically cloned Git repositories are now recorded with the path given on the command line instead of the canonicalized path (#212). This makes datastores slightly more portable across different environments, such as within a Docker container and on the host machine, as relative paths can now be recorded.
-
The deprecated
--rules=PATH
alias for--rules-path=PATH
has been removed from thescan
andrules
commands. -
The built-in support for enumerating and interacting with GitHub is now a compile time-selectable feature that is enabled by default (#213). This makes it possible to build a slimmer release for environments where GitHub functionality is unused.
-
A new rule has been added:
- Bitbucket App Password (#219 from @gemesa)
-
The default number of parallel scanner jobs is now higher on many systems (#222). This value is determined in part by the amount of system RAM; due to several memory use improvements, the required minim RAM per job has been reduced, allowing for more parallelism.
-
The
Google OAuth Credentials
rule has been revised to avoid runtime errors about an empty capture group. -
The
AWS Secret Access Key
rule has been revised to avoid runtimeRegex failed to match
errors. -
The code that determines first-commit provenance information for blobs from Git repositories has been reworked to improve memory use (#222). In typical cases of scanning Git repositories, this reduces both peak memory use and wall clock time by 20-50%. In certain pathological cases, such as Homebrew or nixpkgs, the new implementation uses up to 20x less peak memory and up to 5x less wall clock time.
-
When determining blob provenance informatino from Git repositories, blobs that first appear multiple times within a single commit will now be reported with all names they appear with (#222). Previously, one of the pathnames would be arbitrarily selected.
v0.19.0 (2024-07-30)
-
The
scan
andgithub repos list
commands offer a new--github-repo-type={all,source,fork}
option to select a subset of repositories (#204). -
A category mechanism is now provided for rules (#208). Each rule can have zero or more freeform text categories assigned to it. The existing rules have been updated with category information with the following meanings:
secret
: the rule detects things that are in fact secretsidentifier
: the rule detects things that are not secrets but could be used to enumerate additional resources (e.g., S3 bucket names)hashed
: the rule detects hashed payloads (e.g., bcrypt hashes)test
: the rule detects test deployment-specific payloads (e.g., stripe test keys)api
: the rule detects payloads used for API accessgeneric
: the rule is a "generic" one rather than one that detects a specific type of payload (e.g., username/password pairs)fuzzy
: the rule pattern requires matching of non-payload surrounding context
The category information is included in output in the
rules list
command.
-
The
scan
andgithub repos list
commands now only consider non-forked repositories by default (#204). This behavior can be reverted to the previous behavior using the--github-repo-type=all
option. -
The Alpine-based Docker image has been updated to use the
alpine:latest
base image instead ofalpine:3.18
(#201). -
The "Blynk Organization" rules have been refined (#208). The two "Blynk Organization Client ID" and two "Blynk Organization Client Secret" variations have been subsumed by two new
Blynk Organization Client Credential
rules. These new rules combine the client ID and client secret into single findings instead of reporting them as two separate findings as previous. -
Several rules have been renamed (#208):
AWS S3 Bucket (subdomain style)
->AWS S3 Bucket
AWS S3 Bucket (path style)
->AWS S3 Bucket
Blynk Organization Access Token (URL first)
->Blynk Organization Access Token
.Blynk Organization Access Token (URL last)
->Blynk Organization Access Token
.Generic Password (double quoted)
->Generic Password
Generic Password (single quoted)
->Generic Password
Generic Username and Password (quoted)
->Generic Username and Password
Generic Username and Password (unquoted)
->Generic Username and Password
Google Cloud Storage Bucket (path style)
->Google Cloud Storage Bucket
Google Cloud Storage Bucket (subdomain style)
->Google Cloud Storage Bucket
Google OAuth Client Secret (prefixed)
->Google OAuth Client Secret
New Relic License Key (non-suffixed)
->New Relic License Key
particle.io Access Token (URL first)
->particle.io Access Token
particle.io Access Token (URL last)
->particle.io Access Token
Note that although several rules share the same name now, they all still have distinct IDs.
-
The default set of patterns for the existing gitignore-style path-based exclusion mechanism (
scan --ignore=GITIGNORE_FILE
) has been expanded (#209). The new patterns cover test files from things like vendored Python, Node.js, and Go packages. -
The gitignore-style path-based exclusion patterns (
scan --ignore=GITIGNORE_FILE
) now also apply to content found within Git history, and not just paths on the filesystem (#209). When a blob is found in Git history with at least 1 associated pathname, if all of the associated pathnames match the ignore rules, the blob is not scanned. -
The Rust version required to build has been bumped from 1.76 to 1.77. This is necessary to support C-string literals in the
rusqlite
crate.
v0.18.1 (2024-07-12)
-
Nosey Parker no longer crashes upon startup when running in environments with less than 4 GiB of RAM (#202).
-
The
Base64-PEM-Encoded Private Key
rule has been refined to reduce false positives and avoid a rare performance pitfall.
v0.18.0 (2024-06-27)
-
The README now includes several animated GIFs that demonstrate simple example use cases (#154).
-
The
report
command now offers a new--finding-status=STATUS
filtering option (#162). This option causes findings with an assigned status that does not matchSTATUS
to be suppressed from the report. -
The
report
command now offers a new--min-score=SCORE
filtering option (#184). This option causes findings that have a mean score less thanSCORE
to be suppressed from the report. This option is set by default with a value of 0.05. -
A new
datastore export
command has been added (#166). This command exports the essential content from a Nosey Parker datastore as a .tgz file that can be extracted wherever it is needed. -
New experimental
annotations export
andannotations import
commands have been added (#171). These commands allow annotations (finding comments, match comments, and match statuses) to be converted between JSON and datastore representations. -
New rules have been added:
-
Prebuilt releases now included separate debug symbols (.dSYM or .dwp files) (#191). Having the debug symbols available makes stack traces more useful in the rare event of a crash. The Alpine-based Docker image does not include these debug symbols, as its point of existing is to provide a small distribution.
-
The
summarize
command now includes additional columns for the assigned finding status (#196).
-
The vendored copy of Boost included in the internal
vectorscan-sys
crate has been removed in favor of using the system-provided Boost (#150 from @seqre). This change is only relevant to building Nosey Parker from source. -
The vendored copy of the Vectorscan regular expression library included in the internal
vectorscan-sys
crate has been removed (#151 from @seqre). Instead, a copy of the Vectorscan 5.4.11 source tarball is included in this repository, and is extracted and patched during the build phase. -
SARIF reporting format is now listed as experimental.
-
In the
scan
andrules
command, the command-line option to load additional rules and rulesets from files has been renamed from--rules
to--rules-path
. The old--rules
option is still supported as an alias, but this is deprecated and will be removed in the v0.19 release. -
The
rules list
command now includes additional fields when using JSON format (#161). -
The
vectorscan
andvectorscan-sys
crates have been split off into a separate project with crates published on crates.io (#168). -
The
scan
command is now more conservative in its default degree of parallelism (#174). Previously the default value was determined only by the number of available vCPUs. Now the default value is additionally limited to ensure at least 4 GiB of system RAM per job. -
The
scan
command now records its results incrementally to the datastore instead of in one enormous transaction (#189). Now, results are recorded in transactions about every second. This helps avoid complete loss of scan results in the rare event of a crash.
-
A rare crash when parsing malformed Git commit timestamps has been fixed by updating the
gix-date
dependency (#185). -
Upon
noseyparker
startup, if resource limits cannot be adjusted, instead of crashing, a warning is printed and the process attempts to continue (#170). -
The prepackaged releases and binaries produced by the default settings of
cargo build
should now be more portable across microarchitectures (#175). Previously, the builds would be tied to the microarchitecture of the build system; this would sometimes result in binaries that were not portable across machines, particularly on x86_64. -
The
--ignore-certs
command-line option is now a global option and can be specified anywhere on the command line.
v0.17.0 (2024-03-05)
-
A new
--ignore-certs
command-line option has been added to thescan
andgithub
commands. This option causes TLS certificate validation to be skipped (#125; thank you @seqre). -
The
scan
andgithub
commands now support the--all-organizations
flag. When supplied along with a custom GitHub API URL, Nosey Parker will scan the provided GitHub instance for all organizations to be further enumerated for additional repositories (#126; thank you @seqre). -
New rules have been added (thank you @gemesa):
- Adafruit IO Key (#114)
- Blynk Device Access Token (#117)
- Blynk Organization Access Token (URL first) (#117)
- Blynk Organization Access Token (URL last) (#117)
- Blynk Organization Client ID (URL first) (#117)
- Blynk Organization Client ID (URL last) (#117)
- Blynk Organization Client Secret (URL first) (#117)
- Blynk Organization Client Secret (URL last) (#117)
- Docker Hub Personal Access Token (#108)
- Doppler CLI Token (#111)
- Doppler Personal Token (#111)
- Doppler Service Token (#111)
- Doppler Service Account Token (#111)
- Doppler SCIM Token (#111)
- Doppler Audit Token (#111)
- Dropbox Access Token (#106)
- particle.io Access Token (URL first) (#113)
- particle.io Access Token (URL last) (#113)
- ThingsBoard Access Token (#112)
- ThingsBoard Provision Device Key (#112)
- ThingsBoard Provision Device Secret (#112)
- TrueNAS API Key (WebSocket) (#110)
- TrueNAS API Key (REST API) (#110)
- WireGuard Private Key (#104)
- WireGuard Preshared Key (#104)
-
A new
generate
command has been added, which generates various assets that are included in prebuilt releases:
-
Several rules have been fixed that in certain circumstances would fail to match and produce a runtime error message:
- Google API Key
- ODBC Connection String
- Sauce Token
-
The
netrc Credentials
rule has been modified to avoid a runtime message about an empty capture group. -
The
JSON Web Token (base64url-encoded)
rule has been improved to reduce false positives. Thank you @saullocarvalho for the bug report. -
The prebuilt releases now include shell completion scripts for bash, fish, elvish, powershell, and zsh, instead of 5 copies of the zsh completions (#132; thank you @Marcool04).
-
The minimum supported Rust version has been changed from 1.70 to 1.76.
-
The data model and datastore have been significantly overhauled:
-
The rules used during scanning are now explicitly recorded in the datastore. Each rule is additionally accompanied by a content-based identifier that uniquely identifies the rule based on its pattern.
-
Each match is now associated with the rule that produced it, rather than just the rule's name (which can change as rules are modified).
-
Each match is now assigned a unique content-based identifier.
-
Findings (i.e., groups of matches with the same capture groups, produced by the same rule) are now represented explicitly in the datastore. Each finding is assigned a unique content-based identifier.
-
Now, each time a rule matches, a single match object is produced. Each match in the datastore is now associated with an array of capture groups. Previously, a rule whose pattern had multiple capture groups would produce one match object for each group, with each one being associated with a single capture group.
-
Provenance metadata for blobs is recorded in a much simpler way than before. The new representation explicitly records file and git-based provenance, but also adds explicit support for extensible provenance. This change will make it possible in the future to have Nosey Parker scan and usefully report blobs produced by custom input data enumerators (e.g., a Python script that lists files from the Common Crawl WARC files).
-
Scores are now associated with matches instead of findings.
-
Comments can now be associated with both matches and findings, instead of just findings.
-
-
The JSON and JSONL report formats have changed. These will stabilize in a future release (#101).
-
The
matching_input
field for matches has been removed and replaced with a newgroups
field, which contains an array of base64-encoded bytestrings. -
Each match now includes additional
rule_text_id
,rule_structural_id
, andstructural_id
fields. -
The
provenance
field of each match is now slightly different.
-
-
Schema migration of older Nosey Parker datastores is no longer performed. Previously, this would automatically and silently be done when opening a datastore from an older version. Explicit support for datastore migration may be added back in a future release.
-
The
shell-completions
command has been moved from the top level to a subcommand ofgenerate
.
v0.16.0 (2023-12-06)
-
The
scan
command now supports a new--copy-blobs={all,matching,none}
parameter. When specified asmatching
, a copy of each encountered blob that has matches will be saved to the datastore'sblobs
directory. When specified asall
, a copy of each encountered blob will be saved. The default value isnone
. This mechanism exists to aid in ad-hoc downstream investigation. Copied blobs are not used elsewhere in Nosey Parker at this point. -
A new advanced global command-line parameter has been exposed:
--sqlite-cache-size=SIZE
to control thepragma cache_size
value used in sqlite database connections
-
The datastore now contains two additional tables for to represent freeform comments and accept/reject status associated with findings. These additional tables are not currently populated in the open-source version of Nosey Parker. The
report
command now emits finding status and comment if populated. Note: the datastore format is not settled and is subject to change. -
A new "ruleset" mechanism has been added. A ruleset is a named collection of rules that can be selected as a group. The new
--ruleset=NAME
parameter toscan
can be used to enable alternative rulesets. Three built-in rulesets are provided (default
,np.assets
andnp.hashes
); the special ruleset nameall
enables all known rules. See the built-in rulesets atcrates/noseyparker/data/default/builtin/rulesets
for an example for writing your own. -
The default collection of rules has been pruned down to further emphasize signal-to-noise. Only rules that detect secret things are included in the default collection. Rules that detect other things, such as cloud assets, application IDs, or public keys, are not included in this set. Instead, those are in the
np.assets
ruleset, which is not enabled by default. No rules have been removed from Nosey Parker; rather, the defaults have been adjusted to support the most common use case (secrets detection). -
Additional checks have been added to the
rules check
command:- Each regex rule must have at least one capture group
- Each ruleset must have a globally-unique ID
- A ruleset's included rules must resolve to actual rules
- A ruleset should not include duplicate rules
-
A new
rules list
command is available, which lists available rules and rulesets. This command can emit its output in human-oriented format or in JSON format. -
New rules have been added:
- Dependency-Track API Key (Thank you @tpat13!)
- Password Hash (sha256crypt)
- Password Hash (sha512crypt)
- Password Hash (Cisco IOS PBKDF2 with SHA256)
- React App Username
- React App Password
-
A new global
--quiet
/-q
option has been added, which suppresses non-error feedback messages and disables progress bars (#97).
-
Command-line parameters that can meaningfully accept negative numbers can now be specified without having to use
--PARAMETER=NEGATIVE_VALUE
syntax; a space can now separate the paraemter and the value. -
Fixed three rules that were missing capture groups:
- Age Recipient (X25519 public key)
- Age Identity (X22519 secret key)
- crates.io API Key
Due to nuanced details of how scanning is performed, rules without capture groups will never produce reported matches. An additional check was added to the
rules check
command and a couple assertions were added that should help prevent this type of error in the future. -
Fixed several rules:
- Amazon MWS Auth Token: the capture group was smaller than it should have been
- Microsoft Teams Webhook: changed 3 capture groups to 1; full URL is now included
- Slack Webhook: full URL is now included
-
The LICENSE, README.md, and CHANGELOG.md files are now included in prebuilt binary releases.
-
ANSI formatting sequences are now no longer included by default by the
report
command when the output is redirected to a file using the-o
/--outfile
parameter (#55). -
The
scan
command should no longer emit warnings likeFailed to decode entry in tree
. These warnings were due to a bug in the Git object parsing code in thegix
dependency, which was fixed upstream.
-
The
rules check
command invocation now behaves differently. It now no longer requires input paths to be specified. It will check the built-in rules for problems, and if additional paths are specified, will check those rules as well. This change was made so that thescan
,rules check
, andrules list
invocations have consistent interfaces. -
The default path-based ignore rules in Nosey Parker now ignore
packed-refs
files from Git repositories. -
Several rules have been changed:
- The
Slack
rule (idnp.slack.1
) has been removed, as it was redundant withSlack Token
. Slack Token
has been split intoSlack Bot Token
,Slack Legacy Bot Token
,Slack User Token
, andSlack App Token
.CodeClimate
was enhanced to detect additional cases and was renamed toCodeClimate Reporter ID
.md5crypt Hash
(idnp.md5.1
) has been renamed toPassword Hash (md5crypt)
and re-identified asnp.pwhash.1
.bcrypt Hash
(idnp.bcrypt.1
) has been renamed toPassword Hash (bcrypt)
and re-identified asnp.pwhash.2
.
- The
-
Log messages are written to stderr instead of stdout (#97).
v0.15.0 (2023-10-12)
-
A default value (
datastore.np
) is now set for commands that take a datastore parameter (#74). This makes simplernoseyparker
command-line invocations possible. -
A new
shell-completions
command has been added, which generates shell-specific completion scripts for zsh, bash, fish, powershell, and elvish (#76). These generated completion scripts make discovery of Nosey Parker's command-line API simpler. Thank you @Coruscant11! -
The
report
command supports a new--max-matches=N
parameter to control the maximum number of matches that will be output for any single finding (#75). A negative number means "no limit". -
The
scan
command now supports a new--git-history={full,none}
parameter to control whether encountered Git history will be scanned. This defaults tofull
, but specifying a value ofnone
will cause Git history to be ignored. -
New rules have been added:
- Mapbox Temporary Access Token
- Salesforce Access Token
-
A new
disable_tracing
Cargo feature has been added, which disablestrace
-level logging and tracing messages. This feature is also aliased by a newrelease
feature, which is enabled in prebuilt releases. -
The
NP_LOG
environment variable is inspected at runtime to allow find-grain control over Nosey Parker's diagnostic output. The syntax of this variable are defined by thetracing-subscriber
Rust crate.
-
All the output formats for the
report
command now respect the new--max-matches=N
parameter. Previously, the output formats other thanhuman
would run without limit (i.e., as though--max-matches=-1
had been specified). -
The release process is now codified in a shell script:
scripts/create-release.zsh
. This emits a release tree atrelease
in the top-level of the repository, which includes the prebuilt binary as well as shell completions (#80). -
The
report
command has improved performance when using JSON output format. Previously, the entire JSON output document needed to be accumulated in memory and then written in one step at the end. Now, the JSON output document is written in a streaming fashion, one finding at a time. -
mimalloc
is now used as the global allocator (#81). This reduces peak resident memory when scanning large inputs with a high degree of parallelism.
- Fixed a bug in the
report
command when--format=sarif
is used which caused some metadata to be unintentionally omitted from the output.
v0.14.0 (2023-08-17)
A prebuilt multiplatform Docker image for this release is available for x86_64 and ARM64 architectures:
docker pull ghcr.io/praetorian-inc/noseyparker:v0.14.0
-
Running
noseyparker --version
now emits many compile-time details about the build, which can be useful for troubleshooting (#48). -
The
github
andscan
commands now support accessing GitHub Enterprise Server instances using the new--github-api-url URL
parameter (#53—thank you @AdnaneKhan!). -
New rules have been added:
- Amazon Resource Name
- AWS S3 Bucket (subdomain style)
- AWS S3 Bucket (path style)
- Google Cloud Storage Bucket (subdomain style)
- Google Cloud Storage Bucket (path style)
- HuggingFace User Access Token (#54—thank you @AdnaneKhan!)
-
Rules are now required to have a globally-unique identifier (#62)
-
Two new advanced global command-line parameters have been exposed:
--rlimit-nofile LIMIT
to control the maximum number of open file descriptors--enable-backtraces BOOL
to control whether backtraces are printed upon panic
-
The snippet length for matches found by the
scan
command can now be controlled with the new--snippet-length BYTES
parameter. -
The Git repository cloning behavior in the
scan
command can now be controlled with the new--git-clone-mode {mirror,bare}
parameter. -
The
scan
command now collects additional metadata about blobs. This metadata includes size in bytes and guessed mime type based on filename extension. Optionally, if the non-defaultlibmagic
Cargo feature is enabled, the mime type and charset are guessed by passing the content of the blob throughlibmagic
(the guts of thefile
command-line program).By default, all this additional metadata is recorded into the datastore for each blob in which matches are found. This can be more precisely controlled using the new
--blob-metadata={all,matching,none}
parameter.This newly-collected metadata is included in output of the
report
command. -
The
scan
command now collects additional metadata about blobs found within Git repositories. Specifically, for each blob found in Git repository history, the set of commits where it was introduced and the accompanying pathname for the blob is collected (#16). This is enabled by default, but can be controlled using the new--git-blob-provenance={first-seen,minimal}
parameter.This newly-collected metadata is included in output of the
report
command.
-
The datastore schema has been changed in an incompatible way such that migrating existing datastores to the new version is not possible. This was necessary to support the significantly increased metadata that is now collected when scanning. Datastores from earlier releases of Nosey Parker cannot be used with this release; instead, the inputs will have to be rescanned with a new datastore.
-
The JSON and JSONL output formats for the
report
command have changed slightly. In particular, the.matches[].provenance
field is now an array of objects instead of a single object, making it possible to handle situations where a blob is discovered multiple ways. Theprovenenance
objects have some renamed fields, and contain significantly more metadata than before. -
Existing rules were modified to reduce both false positives and false negatives:
- Generic Password (double quoted)
- Generic Password (single quoted)
-
The default size of match snippets has been increased from 128 bytes before and after to 256. This typically gives 4-7 lines of context before and after each match.
-
When a Git repository is cloned, the default behavior is to match
git clone --bare
instead ofgit clone --mirror
. This new default behavior results in cloning potentially less content, but avoids cloning content from forks from repositories hosted on GitHub. -
The command-line help has been refined for clarity.
-
Scanning performance has been improved on particular workloads by as much as 2x by recording matches to the datastore in larger batches. This is particularly relevant to heavy multithreaded scanning workloads where the inputs have many matches.
-
Python is no longer required as a build-time dependency for
vectorscan-sys
. -
A typo was fixed in the Okta API Key rule that caused it to truncate the secret.
-
The
scan
command now correctly reports the number of newly-seen matches when reusing an existing datastore.
v0.13.0 (2023-04-24)
A prebuilt multiplatform Docker image for this release is available for x86_64 and ARM64 architectures:
docker pull ghcr.io/praetorian-inc/noseyparker:v0.13.0
-
Nosey Parker now statically links against a bundled version of Vectorscan for regular expression matching instead of Hyperscan (#5). This makes building from source simpler, particularly for ARM-based platforms. This also simplifies distribution, as a precompiled
noseyparker
binary now has no runtime library dependencies on non-default libraries. -
Several existing rules were modified to reduce false positives and false negatives:
- Generic API Key
- Telegram Bot Token
-
New rules have been added:
- Generic Username and Password (quoted)
- Generic Username and Password (unquoted)
- Generic Password (double quoted)
- Generic Password (single quoted)
- Grafana API Token
- Grafana Cloud API Token
- Grafana Service Account Token
- Postman API Key
-
References have been added for several rules:
- Twilio API Key
- Dynatrace Token
- The Docker image now has the
git
binary installed. Previously this was missing, causing thescan
command to fail when the--git-url
,--github-user
, or--github-organization
input specifiers were used (#38).
v0.12.0 (2023-03-02)
A prebuilt Docker image for this release is available for x86_64 architectures:
docker pull ghcr.io/praetorian-inc/noseyparker:v0.12.0
-
The
scan
command can now be given Git https URLs, GitHub usernames, and GitHub organization names as inputs, and will enumerate, clone, and scan as appropriate (#14). -
Nosey Parker now has rudimentary support for enumerating repositories from GitHub users and organizations (#15). The new
github repos list
command uses the GitHub REST API to enumerate repositories belonging to one or more users or organizations. An optional GitHub Personal Access Token can be provided via theNP_GITHUB_TOKEN
environment variable. -
Nosey Parker now has an optional
rule_profiling
crate feature that causes performance-related statistics to be collected and reported when scanning. This feature imposes some performance cost and is only useful to rule authors, and so is disabled by default. -
Many new rules have been added:
- Adobe OAuth Client Secret
- Age Identity (X22519 secret key)
- Age Recipient (X25519 public key)
- crates.io API Key
- DigitalOcean Application Access Token
- DigitalOcean Personal Access Token
- DigitalOcean Refresh Token
- Figma Personal Access Token
- GitLab Personal Access Token
- GitLab Pipeline Trigger Token
- GitLab Runner Registration Token
- Google OAuth Client Secret (prefixed)
- New Relic API Service Key
- New Relic Admin API Key
- New Relic Insights Insert Key
- New Relic Insights Query Key
- New Relic License Key
- New Relic License Key (non-suffixed)
- New Relic Pixie API Key
- New Relic Pixie Deploy Key
- New Relic REST API Key
- NPM Access Token (fine-grained)
- OpenAI API Key
- Segment Public API Token
- Shopify Access Token (Custom App)
- Shopify Access Token (Legacy Private App)
- Shopify Access Token (Public App)
- Shopify App Secret
- Shopify Domain
- RubyGems API Key
- Telegram Bot Token
These rules match token formats that are well-specified fixed-length strings with notable prefixes or suffixes, and so should produce very few false positives.
-
Several existing rules were modified to improve signal-to-noise:
- Azure Connection String
- Credentials in ODBC Connection String
- PyPI Upload Token
-
The
report
command now offers rudimentary SARIF support (#4). Thank you @Coruscant11!
-
Several default rules have been revised to improve performance of the matching engine and to produce fewer false positives. In particular, several rules previously had avoided using a trailing
\b
anchor after secret content which could include a literal-
character, due to a matching discrepancy between Hyperscan and Rust'sregex
library. These have been revised to use a more complicated but functional anchoring pattern. -
The
JSON Web Token (base64url-encoded)
rule has been changed to only produce a single match group instead of three. -
The
Google Client Secret
rule has been improved to detect additional occurrences and has been renamed toGoogle OAuth Client Secret
. -
Blobs are now deduplicated at enumeration time when first enumerating a Git repository, rather than only at scan time. This results in more accurate progress bars.
-
When scanning, Git repositories are now opened twice: once at input enumeration time, and once at scanning time. This drastically reduces the amount of memory required to scan a large number of Git repositories.
- When scanning, the datastore is now explicitly excluded from filesystem enumeration. This ensures that files used internally for Nosey Parker's operation are not inadvertently scanned (#32).
v0.11.0 (2022-12-30)
This is the first released version of Nosey Parker.
Its scan
, summarize
, and report
commands are functional.
It is able to scan files, directories, and the complete history of Git repositories at several hundred megabytes per second per core.
It comes with 58 rules.
A prebuilt Docker image for this release is available for x86_64 architectures:
docker pull ghcr.io/praetorian-inc/noseyparker:v0.11.0