Skip to content
Go to file

Latest commit

* Use AxiosHttpClient in ServiceFabrikClient
Fix unit test failures

* Use AxiosHttpClient in DeploymentHookClient
Fix unit test failures

Wrap axios.request promise in bluebird promise
to support .tap() in clients

* Use AxiosHttpClient in CloudControllerClient
Fix unit test failures

* Use AxiosHttpClient in ServiceBrokerClient
Fix unit test failures

* Use AxiosHttpClient in BoshDirectorClient.
Fix unit test failures.

Enhance error handling in AxiosHttpClient.
Configure rejectUnauthorized flag during http request.
Workaround for axios/axios/issues/907

* changes in quota clients and uaa client

* Fix test case #makeRequestWithConfigWithUAA of BoshDirectorClient

* Use AxiosHttpClient in MeteringClient.
Fix unit test failures.

* Use AxiosHttpClient in Agent.
Fix unit test failures.

* Rename AxiosHttpClient to HttpClient.
Remove request from package.json
Fix HttpClient unit test failures.

* Fix test case failures,
- $makeRequestWithConfigWithUAA case in BoshDirectorService
- #update cases in DirectorService

Co-authored-by: i350504 <>

Git stats


Failed to load latest commit information.
Latest commit message
Commit time

Coverage Status Travis Build Status Go Report Card codecov lerna

Service Fabrik Broker aka Interoperator

Service Fabrik Broker is an OSBAPI compliant broker which is deployed using BOSH and supports BOSH based and Docker based service deployment. The OSB broker is written in an operator pattern so that it can be used also as a broker for exposing Kubernetes native services written based on operator based pattern in an OSBAPI based way. This flavor of the broker is deployed using helm-chart on Kubernetes and is called as Interoperator. This broker was very initially inspired by the cf-containers-broker. Read the Big Picture behind Service Fabrik Broker.

In the next sections we first discuss about the Deployment layout of BOSH based flavor, Service Fabrik Broker and then the architecture of Interoperator, deployed on Kubernetes.

Service Fabrik Broker Deployment Layout

Missing Above is the component diagram of Service Fabrik. Service Fabrik has been re-designed to a model which is event driven and based on decoupled components which participate and communicate via an APIServer managed by Service Fabrik.

The concept is based on Event Sourcing where the Event Store is point of coordination for different components. This facilitates easy onboarding of external components and modules into the service Fabrik eco-system.

To facilitate plugging in external components, we intend to model service Fabrik on a resource and control loop based programming model.

This allows capabilities like provisioning and operations on provisioned instances to be built independently and plugged into the Service Fabrik APIServer based on specific requirements.

Steps to Integrate new provisioners are mentioned in here

Interoperator Architecture

Interoperator Architecture is the heart of the OSB framework of Service Fabrik. The details are discussed here.

Deploy Interoperator


Interoperator uses helm charts for deployment. Interoperator requires helm version >= 3.0.0, and is not supported by helm 2.

Deploy and run your operator

You have to deploy and run your own operator. Currently SF-Inter-Operator supports only in-cluster operators, so you have to deploy it in the same cluster where SF-Inter-Operator is deployed.

Deploy Interoperator(Fresh deploy)

Deploy using publicly published helm chart (Recommended)

To add service fabrik interoperator helm chart repo

helm repo add interoperator-charts
helm repo update

Deploy SF Interoperator using helm

helm install --set< clusterdomain > --namespace interoperator [--version <helm chart version>] interoperator interoperator-charts/interoperator

Deploy using helm chart repo

To deploy using helm chart repo

git clone
helm install --set< clusterdomain > --namespace interoperator interoperator helm-charts/interoperator

NOTE: should be within the 63 character limit.

Upgrade existing Interoperator deployment

To check how you can upgrade existing interoperator deployment, please check this document here.

Deploy SFClusters, SFServices and SFPlans and Register with Interoperator

Please create sfcluster CRs and add reference to kubeconfig. In the example below, resource_v1alpha1_sfcluster.yaml is the sfcluster CR which refers to secret kubeconfig-secret.yaml, which has the current cluster kubeconfig as the base64 encoded data.

For multi-cluster support, all corresponding sfcluster CRs need to be created and their kubeconfig needs to be supplied in the corresponding secret.

Please note that sfservice, sfplans and sfcluster (along with the secret it refers to) need to be deployed in the same namespace where SF is deployed (default is interoperator).

To understand the CRs and their structures, please check the Architecture. The different templates are described in interoperator-templates

Deploying multiple interoperator in the same cluster

Multiple instances of interoperator can be deployed on a single cluster. But each instance must be deployed in a separate namespace. Only one instance of interoperator can be deployed in one namespace. The the custom resources like sfservice, sfplans and sfcluster (along with the secret it refers to) related on deployment of interoperator must be created in the namespace where interoperator is deployed.

Deploy an instance of interoperator on a cluster using

helm install --set< clusterdomain >  --namespace < namespace > [--version <helm chart version>] interoperator interoperator-charts/interoperator

Register with the platform

You can either register with service manager which then exposes it to consuming environments or register with service catalog to expose in local kubernetes cluster.

Other Features

Checkout other features like enabling Service Instance sharing, custom service dashboard and quota management here.


Local Development Setup (Ubuntu)

Certainly when you are a broker developer, but also if you are a service developer, you may want to work locally with the broker without packaging it into a Bosh release and deploying it. As a pre-requisites, you would need to have the following installed.

Installing NVM and Running Node Processes

NVM helps you manage node versions and isolated environments.

curl -o- | bash
source ~/.nvm/
nvm install node
nvm use node
  • Clone this repo (assuming your working directory is ~/workspace)
cd ~/workspace
  • Clone and setup fork and git-secrets ( requires hub and git secrets to be installed)
    sh -c "$(curl -fsSL"
    cd service-fabrik-broker/broker
    git checkout -b my-new-feature
    # make code changes
    git push <github_username> my-new-feature
  • Install dependencies
npm install
  • Optional: To locally run all unit test To run all the unit tests:
npm run -s test

To run only unit tests for specific processes like broker, deployment_hooks

# help
npm run -s help
  • Launch the broker process APIServer is a prerequisite for the Service Fabrik. To get more details about the APIs, this can be looked into.

Assuming that apiserver is already deployed on boshlite, to start provisioning, broker process has to be started.

If apiserver is not deployed already, please follow this guide.

Useful prerequisites: When working with the broker, install curl (sudo apt-get install curl), jq, and yaml2json.

Dependencies on other processes: broker process is dependent on deployment hooks process which has to be running for broker to run any lifecycle operation if actions are present in service catalogs.

Before starting broker process SETTINGS_PATH env variable has to be set.

export SETTINGS_PATH=$(pwd)/config/settings.yml

If you need to change the settings.yml configuration you should copy the file and point the broker to your settings file via the environment variable SETTINGS_PATH.

# env vars you may like to set to different than these default values
# export NODE_ENV=development ## For bosh2.0, use the environment boshlite2, as the passwords and BOSH IP are different.
# cp $(pwd)/broker/config/settings.yml $(pwd)/broker/config/my-settings.yml
# export SETTINGS_PATH=$(pwd)/config/my-settings.yml
npm run -s start

Check endpoint with curl

curl -sk -u broker:secret -H "X-Broker-Api-Version: 2.9" | jq .

If predeployment actions are present in service catalog then deployment hook process has to be running. Before starting deployment hooks process, SETTINGS_PATH env variable has to be set.

export SETTINGS_PATH=$(pwd)/deployment_hooks/config/settings.yml

If you need to change the settings.yml configuration you should copy the file and point the deployment_hooks to your settings file via the environment variable SETTINGS_PATH.

# env vars you may like to set to different than these default values
# export NODE_ENV=development ## For bosh2.0, use the environment boshlite2, as the passwords and BOSH IP are different.
# cp $(pwd)/deployment_hooks/config/settings.yml $(pwd)/deployment_hooks/config/my-settings.yml
# export SETTINGS_PATH=$(pwd)/deployment_hooks/config/my-settings.yml
node $(pwd)/applications/deployment_hooks/src/HookServer.js
  • Launch the Operators For running lifecycle operations, corresponding operator processes have to be started. Currently Service Fabrik Broker supports Bosh Director based and Docker Based services.

Assuming that all required env variables (SETTINGS_PATH, NODE_ENV) are already set. Both bosh operator and docker operator can be launched.

node applications/operators/StartBoshOperators.js #to start bosh operator
node applications/operators/StartDockerOperators.js #to start docker operator

Installing Golang Prerequisites and Running Golang Processes

version=2.2.0 # latest stable version

# download the release
curl -L -O${version}/kubebuilder_${version}_darwin_${arch}.tar.gz

# extract the archive
tar -zxvf kubebuilder_${version}_darwin_${arch}.tar.gz
sudo mv kubebuilder_${version}_darwin_${arch} /usr/local/kubebuilder

# update your PATH to include /usr/local/kubebuilder/bin
export PATH=$PATH:/usr/local/kubebuilder/bin

cd interoperator

# install mockgen
go get
go get
go install

# update your PATH to include $GOPATH/bin
export PATH=$PATH:$GOPATH/bin
Get dependencies
go mod download
Generate the crds and mocks
make generate manifests
Generate clients for crds
go mod vendor
bash vendor/ client "osb:v1alpha1 resource:v1alpha1" --go-header-file hack/boilerplate.go.txt

A step by step series of examples that tell you how to get a development env running

Installing the CRDs

make install

Run it locally

export KUBECONFIG=<path to kubeconfig>
export CLUSTER_ID="1"
export POD_NAMESPACE="interoperator" # namespace should exist in the k8s cluster

# To run the provisioner
make run_provisioners

# To run the scheduler
make run_schedulers

# To run the multiclusterdeploy
make run_mcd

In a new terminal, create an instance of the CRD and check if controller picks it up

kubectl apply -f config/samples/interoperator_v1alpha1_serviceinstance.yaml

Installing Docker

The famous Docker will be required for a local start of the broker. You can avoid this by removing any Docker service defintion from the broker settings/configuration/catalog.

If you're using a Mac, we recommend to use Docker for MAC.

Post installation of docker make modifications to manage docker as a non-root user

Register the Broker

You have to do this only once or whenever you modify the catalog. Then of course, use update-service-broker instead of create-service-broker.

  • Registration
cf create-service-broker service-fabrik-broker broker secret # host IP reachable from within the Vagrant box
cf service-brokers # should show the above registered service broker
curl -sk -u broker:secret -H "X-Broker-Api-Version: 2.9" | jq -r ".services[].name" | xargs -L 1 -I {} cf enable-service-access {}
cf service-access # should show all services as enabled, cf marketplace should show the same

Run a Service Lifecycle

You will need a Cloud Foundry application, let's call it my-app (see below). If you have no specific one, you can use our blueprint-app.

cf create-service blueprint v1.0-container my-service
cf bind-service my-app my-service
# take a look at the generated binding with cf env my-app
cf restart my-app # do this a.) to make binding information available in environment of the app and b.) to activate the security group created with the service
# verify the application sees the service; if you have deployed the above app, run curl -skH "Accept: application/json" ""
cf unbind-service my-app my-service
cf delete-service -f my-service

How to obtain support

If you need any support, have any question or have found a bug, please report it in the GitHub bug tracking system. You can also reach us out on our Slack Channel.


This project is licensed under the Apache Software License, v. 2 except as noted otherwise in the LICENSE file.


Cloud Foundry service broker which provisions service instances as Docker containers and BOSH deployments.




No packages published
You can’t perform that action at this time.