garak LLM vulnerability scanner v0.9.0.13 ( https://github.com/leondz/garak ) at 2024-05-10T15:18:18.123569 usage: python -m garak [-h] [--verbose] [--report_prefix REPORT_PREFIX] [--narrow_output] [--parallel_requests PARALLEL_REQUESTS] [--parallel_attempts PARALLEL_ATTEMPTS] [--seed SEED] [--deprefix] [--eval_threshold EVAL_THRESHOLD] [--generations GENERATIONS] [--config CONFIG] [--model_type MODEL_TYPE] [--model_name MODEL_NAME] [--generator_option_file GENERATOR_OPTION_FILE | --generator_options GENERATOR_OPTIONS] [--probes PROBES] [--probe_option_file PROBE_OPTION_FILE | --probe_options PROBE_OPTIONS | --probe_tags PROBE_TAGS] [--detectors DETECTORS] [--extended_detectors] [--buffs BUFFS] [--taxonomy TAXONOMY] [--plugin_info PLUGIN_INFO] [--list_probes] [--list_detectors] [--list_generators] [--list_buffs] [--list_config] [--version] [--report REPORT] [--interactive] [--generate_autodan] [--interactive.py] LLM safety & security scanning tool options: -h, --help show this help message and exit --verbose, -v add one or more times to increase verbosity of output during runtime --report_prefix REPORT_PREFIX Specify an optional prefix for the report and hit logs --narrow_output give narrow CLI output --parallel_requests PARALLEL_REQUESTS How many generator requests to launch in parallel for a given prompt. Ignored for models that support multiple generations per call. --parallel_attempts PARALLEL_ATTEMPTS How many probe attempts to launch in parallel. --seed SEED, -s SEED random seed --deprefix remove the prompt from the front of generator output --eval_threshold EVAL_THRESHOLD minimum threshold for a successful hit --generations GENERATIONS, -g GENERATIONS number of generations per prompt --config CONFIG YAML config file for this run --model_type MODEL_TYPE, -m MODEL_TYPE module and optionally also class of the generator, e.g. 'huggingface', or 'openai' --model_name MODEL_NAME, -n MODEL_NAME name of the model, e.g. 'timdettmers/guanaco-33b-merged' --generator_option_file GENERATOR_OPTION_FILE, -G GENERATOR_OPTION_FILE path to JSON file containing options to pass to generator --generator_options GENERATOR_OPTIONS options to pass to the generator --probes PROBES, -p PROBES list of probe names to use, or 'all' for all (default). --probe_option_file PROBE_OPTION_FILE, -P PROBE_OPTION_FILE path to JSON file containing options to pass to probes --probe_options PROBE_OPTIONS options to pass to probes, formatted as a JSON dict --probe_tags PROBE_TAGS only include probes with a tag that starts with this value (e.g. owasp:llm01) --detectors DETECTORS, -d DETECTORS list of detectors to use, or 'all' for all. Default is to use the probe's suggestion. --extended_detectors If detectors aren't specified on the command line, should we run all detectors? (default is just the primary detector, if given, else everything) --buffs BUFFS, -b BUFFS list of buffs to use. Default is none --taxonomy TAXONOMY specify a MISP top-level taxonomy to be used for grouping probes in reporting. e.g. 'avid-effect', 'owasp' --plugin_info PLUGIN_INFO show info about one plugin; format as type.plugin.class, e.g. probes.lmrc.Profanity --list_probes list available vulnerability probes --list_detectors list available detectors --list_generators list available generation model interfaces --list_buffs list available buffs/fuzzes --list_config print active config info (and don't scan) --version, -V print version info & exit --report REPORT, -r REPORT process garak report into a list of AVID reports --interactive, -I Enter interactive probing mode --generate_autodan generate AutoDAN prompts; requires --prompt_options with JSON containing a prompt and target --interactive.py Launch garak in interactive.py mode See https://github.com/leondz/garak