Skip to content
Mapping Corelight or Zeek data to Elastic Common Schema fields
Shell
Branch: master
Clone or download
Latest commit 2570779 Nov 18, 2019
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
automatic_install _path added Nov 18, 2019
automatic_install_reduced _path added Nov 18, 2019
manual_install _path added Nov 18, 2019
manual_install_reduced _path added Nov 18, 2019
LICENSE Initial commit Oct 18, 2019
readme.md Add files via upload Oct 18, 2019

readme.md

Corelight Elastic Common Schema Mapping

Overview

The Elastic Common Schema (https://github.com/elastic/ecs) is a way to unify field names across multiple data sources in Elastic. This mapping connects either Corelight or Zeek data types to relevant Elastic Common Schema field names, using Elastic's ingest node pipelines. A few notes on how the mapping works:

  • Field names are replaced in this operation (i.e. ECS does not support having both the original field name and the ECS field name for the same record).
  • Both open source Zeek and Corelight source types are supported. If you have changed your source type you will need to edit these mapping files to account for that.
  • The mapping can be done using either an ElasticSearch ingest node or directly in Kibana, as outlined below.
  • Once done, the mapping applies to new data only and should be done using a new index. This is because (due to how Elastic works) if the ingestion is done in a mixed index it will cause problems with field name conflicts between old and new field names. If this is a particular concern, you can work around it by having two exports to Elastic - one to a new index with the direct Elastic integration (which will have the new ECS field names) and one to the old index using the JSON exporter (which will have the original Zeek field names).
  • Using the ECS mapping will increase CPU consumption within Elastic due to the real time field mapping.
  • ECS does not yet support many Corelight/Zeek fields (in particular for VLANs, some DNS field types, and MIME types). As a result, those fields will retain their original naming.

This mapping is Version 1, which is current as of ECS 1.0, Zeek 3.0, and Corelight v17. As always, we are looking forward to any suggestions and ideas for improvement you have!

License

The mapping files and automation script are open-source under a BSD license. See COPYINGfor details.

Installation

There are three main steps for a successful installation: (1) load the ECS template into Elasticsearch (with a new index), (2) load the Corelight templates into Elasticsearch and (3) configure the Corelight sensor to export to the new index.

  1. Load the ECS template into Elasticsearch. Installation instructions are available on the ECS GitHub page (https://github.com/elastic/ecs)

  2. Import index template

    2.a) Goto the Dev console in Kibana and run the command from the file "index_template_corelight". The index pattern name is "ecs-corelight*", which you can change if needed.

    2.b) Import the default pipeline with the command from the file "corelight_main_pipeline" (this essentially maps the ECS datasets (groups of field names) to the appropriate Corelight mapping file).

    2.c) Import all other pipelines from the files "corelight_conn_pipeline", "corelight_dce_rpc_pipeline", etc. You can also import all the pipelines automatically with the custom script (below).

    For automatic installation of pipelines on your Elasticsearch instance you can use the script pipelines_import.sh:

    • Use the folder automatic_install. Edit index name (or names) in file 'template_corelight' according to your environment.
    • Copy all files to a Linux host.
    • Run the script using the command # bash pipelines_import.sh
    • Using an interactive menu, install all pipelines to your environment.

    Note that the script will not work from a Mac as it relies on a direct TCP connection (please comment that line out of the script if you want to use it from a Mac). Also, if you want to do cert based authentication you need to change the curl command to best fit your environment.

    Also, for either the script or the manual installation if you are using Corelight's "reduced" logs you should use the mapping folder labeled "reduced." If you are only doing a subset of the reduced logs, replace the specific log pipeline file in the "reduced" folder with the appropriate mapping file from the original (i.e. non-reduced) pipeline folder. We will work to combine these in to a single pipeline ingest folder with nested logic for the next update of this mapping!

  3. Configure your sensor to send events to the new elasticsearch index. This is documented in the Corelight manual; for Zeek you have likely written your own export mechanism so configure that as appropriate for your environment.

You can’t perform that action at this time.