Skip to content

(WIP) Log parser and forwarder for your cloud foundry app logs

License

Notifications You must be signed in to change notification settings

cloudfoundry-rs/cf-app-log-parser

Repository files navigation

Cloudfoundry App Log Parser

Open in Gitpod

Are you in need for a simple, extendable but robust solution to parse and forward your logs from clouf foundry? This repository has you covered (probably).

The scripts in this repository provide simple, ready to use container images to parse and store your cloud foundry apps logs and very low ressource usage using vector

If you're only interested in the deployment examples, head to Deployment

Table of contents

Usage options

There are (mainly) two options of usage:

  • Use it as a middle man to parse the basic elements of the cf log
  • Parse and push the final result to your data store
  • (or you can go full DIY and just copy the parsing logic)

Parser Only

Using this services as a middle man to only parse the cf app logs and forward the parsed events to another vector service. Making sure:

  • You're using the latest parser without worrying
  • You do not f***up the whole service while implementing your own sink
  • Giving you the option to parse even more, specific to your needs

If you're into how the general app behaves have a look at the paragraph The App

To follow an example deployment have a look at examples/parser/

The parser will use the vector sink, in other words your (second) custom service will have to implement the vector source. An example for that is provided in examples/custom

If you would like to it this way but wanna make sure your data is stored somewhere for reliability reasons have a look at prebuilt images

Parse and Store

Instead of sending the output to another vector instance you can opt in to use the same service to handle the storing to other services.

  • You may use a prebuilt image, allowing you to go ahead a be ready within minutes for a prebuilt (set of) sink(s).
  • Use the generic image to customize your sink setup as you start the app
  • Or extend the basic image to add your custom logic such as entrypoint scripts or sinks.

Prebuilt images

Prebuilt images are the easiest way to get started with this log parser. Each images is setup to use a unique collection of sinks (or one sink).

While deploying you can specify the prebuilt collection via the tag <version>-<collection-name>.

i.e. 0.1.0-elasticsearch

The following collections are available:

  • elasticsearch
  • elasticsearch-s3
  • vector
  • vector-s3
  • s3

Open an issue if you're in need for another sink implementation or sink combination.

If you would like to know more about the deployment steps go ahead and have a look at the deployment paragraph or follow examples/prebuilt

Generic image

The generic image is a great way to have a custom combination of sinks which are not prebuilt atm. (or you're badass and have a central management console which manages this)

Using the generic sink implementation it will prepare for any combination of sinks to be combined at start of the image.

You may specify the environment variable INCLUDED_SINKS to your needs. i.e. vector,s3,elasticsearch. This will include the sinks for vector, s3 and elasticsearch in your vector toml file and run the setup scripts of each of them.

Be aware that this way of generating your own collection of sinks may result in some milliseconds to seconds delay at startup

If you would like to know more about the deployment steps go ahead and have a look at the deployment paragraph or follow examples/generic

Extending the base image

Be aware that a container image registry is needed i.e. gcr, ghcr, docker.io

If you would like to add a custom entrypoint script i.e. a script to get credentials or configurations from a vault service or would like to add a custom sink cause the implemented ones in this repository are not what you need. You need to extend the image.

You can choose one of the follwing image types:

  • basic, only including the source and parsing logic
  • one of the prebuilt images
  • the generic image

For the sake of simplicity the images are based on alpine. All images are built for linux/arm64 and linux/amd64

You can extend the container image by a custom Dockerfile. You can do what ever you would like. The most common options are:

  • Copy a shell file inside the /entrypoints directory to add a script which will run before vector starts
  • Copy and install a sink

With that you can create your own Dockerfile

FROM chweicki/cf-vector:0.1.0 # Uses the basic image with only the parsing

# if you need additional dependencies
RUN apk add --no-cache wget ...

# if you need to add a custom entrypoint
COPY custom-entrypoint.sh /entrypoint/0_custom.sh

# if you need to install a sink
COPY custom-sink/ /src/custom-sinks/my-sink

# run the install script of your sink
RUN sh /src/custom-sinks/my-sink/install.sh

For a full example have a look at the examples/extended example.

Deployment

The follwing deployment example will use vector forwarding the results to elasticsearch and s3. For more examples have a look at examples/

For this to work we will need to create the two services

Creating your services will highly depend on your host. it will look something like this:

cf create-service elasticsearch huge vector-elasticsearch

Here comes s3:

cf create-service s3 fast vector-s3

And lastly our login credentials:

cf create-user-provided-service vector-auth -p '{"username":"admin","password":"pa55woRD"}'

Depending on your provider, user provided service instances may not be the best way to handle credentials

Done that, you can go ahead and create your deployment manifest.

Your manifest.yml should look something like this:

applications:
    - name: vector
      memory: 64MB # depending on your setup 64MB should be enough to cover sudden spikes
      instances: 3 # to your liking and load, a autoscaler can be added
      docker:
        image: chweicki/cf-vector:0.1.0-elasticsearch-s3 # uses the prebuilt image
      env:
        INGEST_AUTH_STRATEGY: "basic" # see the paragraph "Deployment" for details
        INGEST_AUTH_SERVICE_NAME: "vector-auth" # see the paragraph "Deployment" for details
        S3_BUCKET_NAME: "logs"
        ELASTICSEARCH_INDEX: "logs-%Y-%m-%d"
      services:
        - vector-s3 # s3 service
        - vector-elasticsearch # elastcisearch service
        - vector-auth # see the paragraph "Deployment" for details
      routes:
       - route: vector.example.io # exposed at this route
      health-check-type: port # vector offers a healthcheck endpoint but not on the port the parser exposes

To deploy that, save it and run:

cf push

Once the app is deployed and ready you can go ahead and register it as a log destination in your space.

cf create-user-provided-service vector-log-drain -l https://admin:pa55woRD@vector.example.io

Now you are able to bind that service to your own app

cf bind-service my-app vector-log-drain

Congrats 🍺 you deployed an instance of this parser and your apps logs will be forwarded to it by cloudfoundry

A word about scaling: Yes it does!

Follow these steps to deploy an autoscaler

Cloud foundry supports auto scaling via the autoscaler service

You need to create a autoscaler service, depending on your provider this service may be called a bit different.

cf create-service autoscaler free vector-scaler

You can edit the following configuration to your liking:

{
  "instance_min_count": 1,
  "instance_max_count": 10,
  "scaling_rules": [
    {
      "metric_type": "memoryutil",
      "breach_duration_secs": 600,
      "threshold": 30,
      "operator": "<",
      "cool_down_secs": 300,
      "adjustment": "-1"
    },
    {
      "metric_type": "memoryutil",
      "breach_duration_secs": 600,
      "threshold": 80,
      "operator": ">=",
      "cool_down_secs": 300,
      "adjustment": "+1"
    }
  ]
}

Depending on your host you can use cpu as a metric or throughput. If you are in the need to add them you should already be an advanced user and aware of your deployment environment. Since these two parameter are not always allowed and highly depend on your infrastructure giving here a general template which works out of the box is impossible. Use your cf knowlage.

That config can be applied by binding the service:

cf bind-service vector vector-scaler -c ./vector-policy.json

Done 🎉

The App

App Start

You may store configurations or keys in other services or need to parse something to configure this image.

At the start of the container the entrypoint will execute all bash files in /entrypoints/. For example have a look at scripts/setup.sh or and other file in the scripts folder.

If you are using the generic image, that where all the sinks will be bundeled together

Ingest

Cloud foundry only allows https which is terminated by your cf ingres and forwarded as http traffic. The Parser exposes a http endpoint at 0.0.0.0:$PORT which is defined by the cf runtime.

The ingest endpoint is either protected by bearer token validation or basic authentication.

Use the following env variables for setup:

  • Use INGEST_AUTH_STRATEGY to set it up as basic or bearer

If you would like to configure your crendentials manually via env variables:

  • Use INGEST_AUTH_PASSWORD and INGEST_AUTH_USERNAME to set the password and username for basic authentication
  • Use INGEST_AUTH_TOKEN to set the token for bearer token validation with bearer authentication

Or if you would like to use a cf service (such as a credential service):

  • Use INGEST_AUTH_SERVICE_NAME to specify the name of the service

Parsing

Data received will then be parsed by the main parsing logic as described on the cloud foundry docs page for app logging

Your cf instance send the following to vector:

2020-01-13T16:12:25.86-0800 [APP/PROC/WEB/0] OUT app instance exceeded log rate limit (100 log-lines/sec) set by platform operator
2016-06-14T13:44:38.14-0700 [CELL/0]     OUT Successfully created container
2016-06-14T14:16:11.49-0700 [SSH/0]      OUT Successful remote access by 192.0.2.33:7856
2016-06-14T14:10:15.18-0700 [APP/0]      OUT Exit status 0
2016-06-14T14:10:05.36-0700 [API/0]      OUT Updated app with guid cdabc600-0b73-48e1-b7d2-26af2c63f933 ({"name"=>"spring-music", "instances"=>1, "memory"=>512, "environment_json"=>"PRIVATE DATA HIDDEN"})
2016-06-14T14:10:27.91-0700 [STG/0]      OUT Staging...

Once parsed the following will be forwarded:

// TODO export example output

Result forwarding

After processing your data will be forwarded to a sink. Configuring your sinks highly dependy on their configuration options.

To make sure you're good with your setup head to your sinks README file i.e. the s3 sink readme

Releases

No releases published

Packages

No packages published