OCL client for OpenMRS is a web service that will allow OpenMRS users to manage their dictionaries in the cloud, allow re-usability of the existing concepts and allow one to create custom concepts to complement existing ones while building their dictionaries.
Runs the app in the development mode.
Open http://localhost:8080 to view it in the browser.
The page will reload if you make edits.
You will also see any lint errors in the console.
Launches the test runner in the interactive watch mode.
See the section about running tests for more information.
npm run build
Builds the app for production to the
It correctly bundles React in production mode and optimizes the build for the best performance.
The build is minified and the filenames include the hashes.
Native Environment Setup
Clone a forked version of the repository to your computer;
git clone https://github.com/yourusername/openmrs-ocl-client.git
In the root directory of the project install all the dependencies;
Running the Docker Image
Build the Docker image using the provided Docker file in the root folder.
Run the docker image using docker-compose
Navigate to http://localhost:8080/.
Choosing an API server
The API accessed by this UI falls back to the OpenMRS demo environment.
To specify your own API server create a file called
env-config.js under the public directory containing;
var OCL_API_HOST = "<your endpoint, e.g https://api.qa.openconceptlab.org>"; var TRADITIONAL_OCL_HOST = "<your endpoint, e.g https://qa.openconceptlab.org>";
These variables will automatically be injected into the app at runtime. This is the same strategy we use while deploying the application.
Contributing to this project
- Visit the project wiki for a layout of the project and how you can start contributing.
Things we've pulled our hair out for so you don't have to
- Have an env-config file before running the e2e tests
- Talk to us at OpenMRS Talk