Skip to content
Switch branches/tags
Go to file


Failed to load latest commit information.
Latest commit message
Commit time


You can find official yangcatalog website in here.

These scripts in this repository serve as a backend to add, update, remove and manage yang modules files in yangcatalog. It is composed out of:

  • scripts that run every day as a cron job,
  • API which lets users to add, remove or find modules they expect to
  • scripts that parse and populate yangcatalog database.

This repository works directly with yangModels/yang repository. That repository contains all the different modules structered in vendors (Cisco, Huawei and others modules) and SDOs (IETF, IEEE, MEF, BBF and others modules)

Parse and Populate

The most important module in this repository is called ParsedAndPopulate. This module contains parsing scripts to parse all the modules on given directory and therefore we receive all the metadata about the module according to draft-clacla-netmod-model-catalog-03. Parsed metedata are subsequently populated to confd datastore using confd REST request. All the new modules that were not yet in the datastore are sent to metadata-update script of yangcatalog using REST request which populates a MySQL database with all the new modules. This database is used for yang-search purposes of

We can parse module either with using sdo option which will go through given directory and will parse all the yang modules in there one by one, or without this option which will try to find platform-metadata.json file in the directory which contains paths to capability.xml files and it will parse all the modules according to those files with vendor metadata added.

To find all the modules missing or wrong revisions, namespaces, imports, include or modules that according to capability.xml file should be in folder but are missing, we can use runCapabilities script with __integrity__ option.


The API module runs as a UWSGI emperor vassal (using the yang-catalog.ini file) and contains several endpoints. Most of the endpoints serves to find modules in different ways. This is described deeper in API documentation. If the user is registered, she/he can add modify or delete modules based on pre-approved path. Once user has filled in registration form, one of yangcatalog admin users needs to use validate script which will walk him through the validation process to give the user specific rights so he is able to add, remove or update only certain modules.

Some of the requests may take a longer period of time to process. For this matter a sender and receiver was made. These scripts use rabbitMQ to communicate. API will use sender to send a job to the receiver. While receiver is processing this job, user will receive a job-id. User can check his job at any time if it has been completed or not. Once a receiver is done it will update a job status to either Failed of Finished successfully.

Note about rabbitMQ: on some Linux, you need to add HOSTNAME=localhost in file /etc/rabbitmq/rabbitmq-env.conf....

Yangcatalog API is also used by some automated jobs. Every time new modules are merged in yangModels/yang repository a job is triggered to populate all new modules to yangcatalog database.

The backend API also receives IETF Yang models every day and if there are any new drafts it will automatically populate yangcatalog database and update the repository with all the new IETF modules if travis job passed successfully.

Please note that UWSGI cache is used to improve the performance as compared to the ConfD request. At the load of the UWSGI, the cache is pre-populated by issueing one ConfD request per module; during this initial load time, the API will probably time-out and the NGINX server will return a 50x error.


There are several cron jobs that are running every day.

  • Statistics job under statistic module which goes through all the modules that are in yangcatalog and generates an HTML file which has information about what vendors and SDOs modules do we have and amount of modules that we have.
  • Resolve expiration job that checks all the IETF draft modules and their expiration date and update its metadata accordingly.
  • Remove unused job that removes data on the server that are not used anymore.
  • User reminder script that will be triggered twice a year to show us what users we have in our database.
  • In ietfYangDraftPull directory there are three jobs.
  1. adds new modules to yangModels/yang repository if there are any new modules.
  2. goes through all ietf drafts and rfcs and populates yangcatalog if there are any new modules.
  3. populates all the new openconfig yang modules from their own repository to the yangcatalog.
  • Recovery script which pulls all the data from confd and creates a json file which is saved in server as backup. If we loose all the data for some reason we can use this script to upload them back with no loss of data.


For every new module that has been added to yangcatalog database, yang admin users are informed about this using a Cisco Webex teams room and by email.



ConfD Premium has to be accessible

API code

Since this is just a small part of the whole functional environment you need to build a docker-compose file from deployment folder Then the catalog_backend_api:latest image can be used to run a docker container where everything will start as it is suppose to


See the file in the documentation/ directory.

Fill the ConfD database

Using the backend/recovery/ --type load /var/yang/cache/<latest>.json.

NGINX Configuration

To be localized to your configuration.

        location /doc {
            alias /home/yang/slate/build;

        location /api {
            rewrite /api(/.*)$ $1 break;
            include uwsgi_params;
            uwsgi_read_timeout 900;


All backend functions of




No releases published


No packages published