- Introduction
- How To Use
- Data Overview
- Data
- Metadata
- Obfuscation
- License
- Directory Structure
- Benchmark
- Citation
- How to Get Involved
- How to Contact
CredData (Credential Dataset) is a set of files including credentials in open source projects. CredData includes suspicious lines with manual review results and more information such as credential types for each suspicious line.
CredData can be used to develop new tools or improve existing tools. Furthermore, using the benchmark result of the CredData, users can choose a proper tool among open source credential scanning tools according to their use case. We sincerely hope that CredData will help minimize credential leaks.
We offer the dataset for download as metadata files and script that would collect required code files based on the metadata.
To download data please use:
- Linux based OS
- Python 3.7.10 or higher version
Then run:
pip install PyYAML
python download_data.py --data_dir data
That Linux based OS is required due to NTFS filenames issue. Some files that would be downloaded and processed have names that are invalid in Windows/NTFS systems (such as
version->1.2.js
).
Using python with lower versions may result in
OSError: [Errno 40] Too many levels of symbolic links
exception.
Resulting dataset has no invalid filenames and can be used on Windows OS.
tmp
directory can be removed after dataset generated.
Dataset consists of 19,459,282 lines of code extracted from 11,408 files in 297 different repositories. Total 73,842 lines of code are labeled, of which 4,583 (6%) labeled as True. Labeled data divided into 8 major categories according to their properties.
Language | Total | Labeled | True | Language | Total | Labeled | True |
---|---|---|---|---|---|---|---|
Text | 85,144 | 10,896 | 1,626 | No Extension | 48,645 | 1,132 | 49 |
Go | 838,816 | 6,515 | 459 | Config | 7,920 | 340 | 46 |
YAML | 74,643 | 2,781 | 344 | AsciiDoc | 27,803 | 448 | 36 |
JavaScript | 742,704 | 4,130 | 340 | Shell | 42,019 | 1,340 | 31 |
Python | 351,494 | 5,643 | 260 | Haskell | 5,127 | 90 | 31 |
Markdown | 186,099 | 3,752 | 253 | Java Properties | 1,878 | 124 | 30 |
JSON | 15,036,863 | 14,430 | 238 | XML | 57,377 | 1,409 | 25 |
Java | 178,326 | 1,751 | 148 | reStructuredText | 38,267 | 531 | 21 |
Ruby | 186,196 | 4,669 | 145 | SQLPL | 16,808 | 612 | 20 |
Key | 8,803 | 364 | 116 | Objective-C | 19,840 | 211 | 14 |
TypeScript | 151,832 | 2,533 | 79 | TOML | 2,566 | 239 | 12 |
PHP | 113,865 | 1,936 | 76 | Scala | 9,564 | 163 | 12 |
Other | 1,226,683 | 7,803 | 172 |
Category | True credentials |
---|---|
Password | 1,395 |
Generic Secret | 1,056 |
Private Key | 992 |
Generic Token | 333 |
Predefined Pattern | 327 |
Authentication Key & Token | 67 |
Seed, Salt, Nonce | 39 |
Other | 374 |
In order to collect various cases in which credentials exist, we selected publicly accessible repositories on Github through the following process:
-
We wanted to collect credentials from repositories for various languages, frameworks, and topics, so we primarily collected 181 topics on Github.
In this process, to select widely known repositories for each topic, we limited repositories with more than a certain number of stars. 19,486 repositories were selected in this process.
-
We filtered repositories which have the license that can not be used for dataset according to the license information provided by Github.
In some cases, the provided license was inaccurate. So we conducted with manual review.
-
Filtering was carried out by checking whether strings related to the most common credentials such as 'password' and 'secret' among the result repositories are included and how many are included. After that, we executed several open source credential scanning tools.
-
For the results of No.3, we manually reviewed the detection results of all tools. Please check Ground Rules for Labeling Suspected Credential Information for the method used in the review.
As a result, we selected 297 repositories containing lines that we suspect had at least one credential value.
It is difficult to know whether a line included in the source code is a real credential. However, based on human cognitive abilities, we can expect the possibility that the detected result contains actual credential information. We classify the detection results to the three credential type.
- True : It looks like a real credential value.
- False : It looks like a false positive case, not the actual credential value.
- Template : It seems that it is not an actual credential, but it is a placeholder. It might be helpful in cases such as ML.
In order to compose an accurate Ground Truth set, we proceed data review based on the following 'Ground Rules':
-
All credentials in test (example) directories should be labeled as True.
-
Credentials with obvious placeholders (
password = <YOUR_PASSWORD>;
) should be labeled as Template. -
Function calls without string literals (
password=getPass();
) and environmental variable assignments (password=${pass}
) should be labeled as False. -
Base64 and other encoded data should be labeled as False. If it is a plaintext credential just encoded to Base64, that should be labeled as True.
-
Package and resource version hash is not a credential, so common hash string (
integrity sha512-W7s+uC5bikET2twEFg==
) is False. -
Be careful about filetype when checking variable assignment:
In .yaml file row (
password=my_password
) can be a credential but in .js or .py it cannot. This languages require quotations (' or ") for string declaration (password="my_password"
). -
Check if the file you are labeling is not a localization file. For example
config/locales/pt-BR.yml
is not a credentials, just a translation. So those should be labeled as False.
We could see that many credentials exist in directories/files that have the same test purpose as test/tests. In the case of these values, people often judge that they contain a real credential, but we do not know whether this value is an actual usable credential or a value used only for testing purposes. We classify those values as True in order to prevent the case of missing real usable credentials. Since it may be necessary to separate the values in the future, we have separated the files for testing and the files that are not. (Check metadata or data set)
Metadata includes Ground Truth values and additional information for credential lines detected by various tools.
Name of property | Data Type | Description |
---|---|---|
ID | Integer | Credential ID |
FileID | String | Filename hash. Used to download correct file from a external repo |
Domain | String | Domain of repository. (ex. Github) |
RepoName | String | Project name that credential was found |
FilePath | String | File path where credential information was included |
LineStart:LineEnd | Integer:Integer | Line information, it can be single(2:2) or multiple(ex. 2:4 means 2 to 4 inclusive) |
GroundTruth | String | Ground Truth of this credential. True / False or Template |
ValueStart | Integer | Index of value on the line. always nan if GroundTruth is False. |
ValueEnd | Integer | Index of character right after value ends in the line. |
InURL | Boolean | Flag to indicate if credential is a part of a URL, such as "http://user:pwd@site.com" |
CharacterSet | String | Characters used in the credential (NumberOnly, CharOnly, Any) |
CryptographyKey | String | Type of a key: Private or Public |
PredefinedPattern | String | Credential with defined regex patterns (AWS token with AKIA... pattern) |
VariableNameType | String | Categorize credentials by variable name into Secret, Key, Token, SeedSalt and Auth |
Entropy | Float | Shanon entropy of a credential |
WithWords | Boolean | Flag to indicate word(https://github.com/first20hours/google-10000-english) is included on the credential |
Length | Integer | Value length, similar to ValueEnd - ValueStart |
Base64Encode | Boolean | Is credential a base64 string? |
HexEncode | Boolean | Is credential a hex encoded string? (like \xFF or FF 02 33 ) |
URLEncode | Boolean | Is credential a url encoded string? (like one%20two ) |
Category | String | Labeled data divided into 8 major categories according to their properties. see Category. |
Labeled data divided into 8 major categories according to their properties.
Name | Description |
---|---|
Password | Short secret with entropy <3.5 or Password keyword in variable name |
Generic Secret | Secret of any length with high entropy |
Private Key | Private cryptographic key |
Predefined Pattern | Credential detected based on defined regex, such as Google API Key/JWT/AWS Client ID |
Seed, Salt, Nonce | Credential with seed, salt or nonce in variable name |
Generic Token | Credential with Token in VariableNameType and not covered by other categories |
Authentication Key & Token | Credential with Auth in VariableNameType and not covered by other categories |
Other | Any credentials that is not covered by categories above |
You can see metadata files in the meta directory. A single metadata file contains rows including line location, value index and GT(GroundTruth) information about the suspect credential information for a specific repository.
Let's look at the meta/02dfa7ec.csv. file as an example.
Id,FileID,Domain,RepoName,FilePath,LineStart:LineEnd,GroundTruth,WithWords,ValueStart,ValueEnd,...
34024,61ed9af5,GitHub,02dfa7ec,data/02dfa7ec/test/n.example,83:83,True,Secret,HighEntropy,0,31,73,...
Convert the above line with only essential columns into a table format:
... | RepoName | FilePath | LineStart:LineEnd | GroundTruth | ... | ValueStart | ValueEnd | ... |
---|---|---|---|---|---|---|---|---|
... | 02dfa7ec | data/02dfa7ec/test/n.example | 83:83 | True | ... | 31 | 73 | ... |
This line means that the credential line exists in the 83th line of the data/02dfa7ec/test/n.example
file that downloaded and obfuscated output after running the download_data.py
script.
You can find the n.example
file in the output generated by the download_data.py script execution by referring to the above path.
When you check the 83th line of the file, you can see that the following line exists.
GITHUB_ENTERPRISE_ORG_SECRET=hfbpozfhvuwgtfosmo2imqskc73w04jf3313309829
In here, you can see the credential based on the values of ValueStart and ValueEnd in the metadata.
hfbpozfhvuwgtfosmo2imqskc73w04jf3313309829
That the corresponding value is an obfuscated value, not an actual credential value. For that part, please refer to the following Obfuscation section.
If the suspicious lines are included in the dataset as it is, the credential values actually used in real may be leaked, which is not a desirable result.
To avoid such cases we proceeded:
- Credential values obfuscation in files.
- Directory & file name and directory hierarchy obfuscation.
To prevent leakage of the actual credential value in the file, we can mask the line that is supposed to be credential or change it to a random string. However, this masking and changing to a random string can make the side effects to the detection performance of several tools. We have applied other methods to substitute the actual credential values within the limits of ensuring the detectability of these various tools.
- Replacing the real value to a example value for a case where a fixed pattern is clear (ex. AWS Access Key)
- Replacing the entire file with credential information to a example file. (ex. X.509 Key)
- Random key generation using regex pattern from the character set of real string and length.
Even if the line suspected of having a credential in the file is obfuscated, you can easily check the original credential value and where it came from by the information of the repository (repo name, directory structure, file name).
To prevent this from happening, we obfuscated the directory structure and file names.
Files with lines suspected of being credential / detected by the tool.
It is saved in the form of data/RepoID/(src|test)/FileID.FILE_EXTENSION
.
RepoID and FileID are not opened separately, but original repository information can be provided only for research purposes and the actual repository owner through separate contact.
For contact information, please refer to the How to Contact section.
Each file is under the existing project's license. It may be difficult to check the license of an existing project due to obfuscation of file and directory structure, etc., but you can check the actual license through the license corresponding to the RepoID in the license directory.
📦CredData
┃
┣ 📂benchmark
┃
┣ 📂data ---- To be generated by `download_data.py`
┃ ┗ 📂A
┃ ┣ 📂src
┃ ┃ ┗ 📜a.c ---- Source File
┃ ┃
┃ ┣ 📂test
┃ ┃ ┗ 📜b.c ---- Source File but in the test/tests.
┃ ┃
┃ ┣ 📂other
┃ ┃ ┗ 📜c ---- File has no extension or Readme
┃ ┃
┃ ┗ 📜LICENSE(COPYING) ---- License File for repo A
┃
┣ 📂meta
┃ ┗ 📜A.csv
┃
┣ 📜snapshot.yaml ---- URL and commit info for used repositories
┃
┣ 📜README.md
┃
┣ 📜download_data.py
┃
┗ 📜LICENSE
- Python3.7 or higher
$ virtualenv venv
$ source venv/bin/activate
$ pip install -qr requirements.txt
usage: python -m benchmark [-h] --scanner [SCANNER]
optional arguments:
-h, --help show this help message and exit
--scanner [SCANNER] scanner name to benchmark (support: ['credsweeper', 'credential_digger',
'detect_secrets', 'gitleaks', 'shhgit', 'trufflehog3', 'wraith'])
Such as:
$ python -m benchmark --scanner credsweeper
A table of performance metrics for each tool tested based on CredData. The content will be updated in detail with the release of our tool in October. For the tools used, see the Used Tools for Benchmarking section below. (The benchmark test was conducted by combining training and test data.)
Name | TP | FP | TN | FN | FPR | FNR | Accuracy | Precision | Recall | F1 |
---|---|---|---|---|---|---|---|---|---|---|
CredSweeper | 3,701 | 337 | 19,454,362 | 882 | 0.0000173223 | 0.1924503600 | 0.9999373564 | 0.9165428430 | 0.8075496400 | 0.8586010904 |
credential-digger | 479 | 4,871 | 19,449,828 | 4,104 | 0.0002503765 | 0.8954833079 | 0.9995387805 | 0.0895327103 | 0.1045166921 | 0.0964461895 |
detect-secrets | 1,748 | 10,599 | 19,444,100 | 2,835 | 0.0005448041 | 0.6185904429 | 0.9993096354 | 0.1415728517 | 0.3814095571 | 0.2064973420 |
gitleaks | 1,120 | 1011 | 19,453,688 | 3,463 | 0.0000519669 | 0.7556185904 | 0.9997700840 | 0.5255748475 | 0.2443814096 | 0.3336312183 |
shhgit | 330 | 306 | 19,454,393 | 4,253 | 0.0000157288 | 0.9279947633 | 0.9997657159 | 0.5188679245 | 0.0720052367 | 0.1264610079 |
truffleHog | 42 | 126 | 19,454,573 | 4,541 | 0.0000064766 | 0.9908356971 | 0.9997601659 | 0.2500000000 | 0.0091643029 | 0.0176804883 |
truffleHog3 | 2,507 | 14,235 | 19,440,464 | 2,076 | 0.0007316998 | 0.4529783984 | 0.9991617882 | 0.1497431609 | 0.5470216016 | 0.2351230950 |
wraith(gitrob) | 898 | 3,099 | 19,451,600 | 3,685 | 0.0001592931 | 0.8040584770 | 0.9996513746 | 0.2246685014 | 0.1959415230 | 0.2093240093 |
- Terminology
- TP(True Positive) : Values that are Actually Positive and Predicted Positive.
- FP(False Positive): Values that are Actually Negative but Predicted Positive.
- TN(True Negative): Values that are Actually Negative and Predicted Negative.
- FN(False Negative): Values that are Actually Positive but Predicted Negative.
- FPR(False Positive Rate): False Positive / Negative, FNR and FPR should be low.
- FNR(False Negative Rate): False Negative / Positive
- Accuracy = True Positive + True Negative / Total, Accuracy, Precision and Recall should always be high.
- Precision = True Positive / Predicted Positive
- Recall = True Positive / Actually Positive
- F1 = 2 * (Precision * Recall) / (Precision + Recall)
Name | URL |
---|---|
CredSweeper | https://github.com/Samsung/CredSweeper |
credential-digger | https://github.com/SAP/credential-digger |
detect-secrets | https://github.com/Yelp/detect-secrets |
gitLeaks | https://github.com/zricethezav/gitleaks |
shhgit | https://github.com/eth0izzle/shhgit |
truffleHog | https://github.com/trufflesecurity/truffleHog |
wraith(gitrob) | https://gitlab.com/gitlab-com/gl-security/security-operations/gl-redteam/gitrob |
You can use CredData on your research.
@misc{sr-cred21,
author = {JaeKu Yun and ShinHyung Choi and YuJeong Lee and Oleksandra Sokol and WooChul Shim and Arkadiy Melkonyan and Dmytro Kuzmenko},
title = {Project CredData: A Dataset of Credentials for Research},
howpublished ={ \url{https://github.com/Samsung/CredData}},
year = {2021}
}
In addition to developing under an Open Source license, A use an Open Source Development approach, welcoming everyone to participate, contribute, and engage with each other through the project.
A recognizes the following formal roles: Contributor and Maintainer. Informally, the community may organize itself and give rights and responsibilities to the necessary people to achieve its goals.
A Contributor is anyone who wishes to contribute to the project, at any level. Contributors are granted the following rights, to:
- Can suggest
- Change in Ground Truth for currently added/ newly added codes
- New open repository to be included
- Report defects (bugs) and suggestions for enhancement
- Participate in the process of reviewing contributions by others
Contributors are required to:
- Must follow below rules when updating additional credential dataset
- Additional data must be individual from the original data, they must not effect(change/remove/conflict) with the original data
- Additional data must not include valid/real credential data to prevent further exposure of the credential, they must be transformed by the obfuscation rule guided in README.md, or changed through other process which has similar obfuscation effect.
- To contribute and reflect changes, Contributors receive the approval of the maintainer.
Contributors who show dedication and skill are rewarded with additional rights and responsibilities. Their opinions weigh more when decisions are made, in a fully meritocratic fashion.
A Maintainer is a Contributor who is also responsible for knowing, directing and anticipating the needs of a given a Module. As such, Maintainers have the right to set the overall organization of the source code in the Module, and the right to participate in the decision-making. Maintainers are required to review the contributor’s requests and decide whether to accept or not the contributed data.
Please post questions, issues, or suggestions into Issues. This is the best way to communicate with the developer.