Skip to content

A series of tools for working with PamGene PamChip kinome arrays

License

Notifications You must be signed in to change notification settings

kinome/kinome_toolbox

Repository files navigation

Welcome to the Kinome Toolbox

A platform for analysis and visualization of PamGene PamChip peptide arrays.

This toolbox has 3 distinct parts.

  1. File Parsing A small bit of node code that can be utilized to automatically parse a new file into all the different components and add them to the database.
  2. Hosting server A server component made to host your own data. This is set up to run with just about any flavor of linux, it is the exact server configuration that we utilize (CentOS 7-3 and Ubuntu 16 tested).
  3. Web Toolbox A series of client side JavaScript scripts that create the pages for the toolbox.

Getting started (client side)

There are a lot of functions and tools you can use to interact with our data. What follows is our 'hello world' example.

Start by loading in some data. The data in our pending publication can be loaded using this shortened URL.

http://bit.kinomecore.com/?p1_1.0.0

From here you can either use the developer console or create your own script, loading it in by adding the &code="<script_url>" to the resolved url, or the url above. More info...

After that here is the "hello world script" that gets all the lvl 1.0.0 data and logs one data point to the console and on a new analysis tab.

(function () {
    'use strict';

    //Get all level 1.0.0 data in here
    var dataArr = KINOME.get({level: '1.0.0'});

    //Get the data list information explicitly, 
    // these can also be listed with a single dataArr.list().
    var ids = dataArr.list('ids');
    var cycles = dataArr.list('cycles');
    var exposures = dataArr.list('exposures');
    var peptides = dataArr.list('peptides');

    //Just get the first one from all of these things
    var onePoint = dataArr.get({
        id: ids[0],
        cycle: cycles[0],
        exposure: exposures[0],
        peptide: peptides[0]
    });

    //Log it to the developers console
    console.log(onePoint);

    //create div
    var myDiv = KINOME.addAnalysis('Print One');

    //pretty print the object to the div minus functions
    myDiv.html(
        JSON.stringify(onePoint, null, 2)
            .replace(/\ /g, '&nbsp;')
            .replace(/\n/g,'<br />')
    );

}());

The KINOME based functions and many more are described below Web Toolbox.

The dataArr functions and many more are described below Kinome Data Objects.

Getting started (server side)

Skip to Web Toolbox if you are just interested in using the client side developer tools.

Download the repository with:

git clone https://github.com/kinome/kinome_toolbox.git

To get our updates you will have to issue a pull command from within the folder you downloaded the code base to periodically.

git pull

There are only three things missing from a full installation. All of these have to be installed globally with root privileges. The monogdb instance will likely require some work to get the dependencies set up correctly, this is out of the scope of this readme, but can be found on their website.

  1. Node and npm, (these come packaged together).
  2. (Only required if you are hosting your own data.) A mongodb instance. Once this is installed you will have to start it up and make sure it runs on reboot.
  3. (Only required if you are hosting your own data for external use.) Either forever or pm2 to ensure the server keeps running through restarts and non-fatal errors.

Once those dependencies are met, you need to build the npm modules. Navigate to the folder that you have downloaded with the clone then type:

npm install

This will download all the local repositories that are needed for a working server and for the server side parsing. Unfortunately these are not currently seperable.

In summary the steps are:

  1. Download NodeJS/NPM (The programming language)
  2. Download and start Mongo (The database)
  3. Download Forever/pm2 (Programs designed to keep a server running continuously)
  4. Git clone this repository (Above)
  5. Navigate to the directory you just created where server.js sits.
  6. npm install to install all the required packages
  7. forever start server.js

From there you can see the below on how to add files to your own database. The toolbox will be running at localhost:8000. You can utilize your server settings to port this to outside traffic. Of note server2.js is set up to utilize an auth component and would take significant work to impliment that is not included nor documented here.

Just be sure to occasionally do a git pull to add any updates to the toolbox, and if you make any improvements or add any features please submit them as a pull request.

Server Side File Parsing

Skip to Web Toolbox if you are just interested in using the client side developer tools.

After following the instructions in Getting Started you will be able to get this component up quickly. This is used by pointing index.js at your file of interest.

node index.js -h

Prints help options which are:

Flag Description
-b (Required.) Link to the background crosstab file as exported from bionavigator. When exporting, select to export all possible meta data on both peptides and samples.
-s (Required.) Link to the signals crosstab file as exported from bionavigator. When exporting, select to export all possible meta data on both peptides and samples.
-o (Highly recommended) Base location and file name to create all the mongodb and json files that are created by the series of command prompts. This is recommended because without it all files are created in the current working directory.
-d Database name to upsert all documents into a mongodb.

This script expects that you have exported your bionavigator crosstab files seperately as median signal and median background. Both of these files are required for parsing to take place. This toolbox does not accept median signal - background combination files.

The _id field for mongo is generated for each of these files so it is the same for 'name' as it is for each level. This prevents multiple documents being created for a given _id in each collection.

This generates .json and .mdb files. The only difference is that the .json is an array so it can be imported directly into the toolbox, and the .mdb is documents seperated by newlines so it can be directly imported into the mongodb database with the following commands:

mongoimport --db <db_name> --collection <collection_name> --file <file_for_collection_name>

With older versions of node the Promise system is not well garbage collected. This can cause memory issues. To fix this run node with more memory:

node --max-old-space-size=4096 index.js -...

Monogo with Restify

Skip to Web Toolbox if you are just interested in using the client side developer tools.

After following the instructions in Getting Started you can make sure the server will work by typing:

node server.js

If that works for you then just use your favorite forever program and type:

forever <or pm2> start server.js

And you have a database server set up.

Web Toolbox

This toolbox is avaliable at: (http://toolbox.kinomecore.com) unfortunately for now you must be inside of UAB's network in order to gain access to the data. This will be changed soon.

There are three major components to the web toolbox:

  1. Default Global Functions and Packages Functions availabe for use on every page created.
  2. URL Parameter Options The URL itself is used to load data into the pages. This data determines the default packages loaded in. A list of these defaults along with a short description of each is avaliable here. Additionally the URL can be used to load in scripts for the purpose of development.
  3. Kinome Data Objects These objects build the infrastructure of all the pages. There are several functions added to them that make working with them as simple as possible.

Default Global Functions and Packages

Every page in this enviornment will have the following standard packages loaded in:

In addition to this there are number default global functions and objects that are created.

require(request [,type [,cache]])
  • require

    require works a lot like require in NodeJS, but instead of returning an object with properties attached, it returns a JavaScript Promise. This documentation got too verbose for this file. Please see require readme for more information and examples.
save(file_content, file_name)
  • save

    save is a simple function, the first parameter is a string with the content of the file you would like to create. The second is a string with the file name you would like to create. These will be downloaded to the users prefered downloads folder.
KINOME.addAnalysis(title)
  • KINOME.addAnalysis

    This function is passed a string and adds that string as an option to the dropdown menu 'Analyses Avaliable'. It returns a jQuery object that is the container avaliable to build your specific tool in.
KINOME.list(list_string)
  • KINOME.list

    This lists the parameters avaliable for a follow up get from KINOME.get. At this time that includes:

    • names: a list of the barcode_array combinations present
    • groups: a list of the groups loaded in (0, 1, 2, ...)
    • levels: a list of the levels loaded in (name, level_1.0.0, ...)
    • ids: a list of the unique ids loaded in

    This will all be returned as an object if no parameters are added in, if a string matching one of the above is passed in it will return an array corresponding to the requested variable.

KINOME.get(get_object)
  • KINOME.get

    The easiest way to use this function is to pass it nothing. It will then return an array of all the objects currently loaded in. These objects will be enriched. (See KINOME.enrich for more). Beyond this an object may be passed in with property names corresponding to the properties from KINOME.list. Each property may be passed either an array or a string and get will return an array of values that matches what was passed in.
KINOME.error(error_obj, error_message)
  • KINOME.error

    This will console.error the error object and the error message. Additionally it creates a dismisable error message based on the error message passed in. This message may be either a string, html or jquery objects.
KINOME.warn(warn_message)
  • KINOME.warn

    This will console.warn the message. Additionally it creates a dismisable warning message based on the warning passed in. This message may be either a string, html or jquery objects.
KINOME.alert(alert_message)
  • KINOME.alert

    This will open a dismissible modal with the message you provide. If the modal is already open, it just updates the message. If it is closing as the message is passed in (or shortly after the message is passed in) it will re-open with the new message.
KINOME.formatEquation(equation_str)
  • KINOME.formatEquation

    This takes a jqMath string and converts it into the appropriate HTML for appending to the page.
KINOME.enrich(data_object)
  • KINOME.enrich

    This is passed a JSON kinome object as outlined at Swagger Hub KINOME. It returns the same object but with an additional series of functions. These functions include list, get, clone, stringify, and depending on the level: put and level_up. More information below at Kinome Data Objects.
KINOME.loadData(urls)
  • KINOME.loadData

    This is a specialized require function. It will accepts urls as an array of strings. The urls must be to arrays of kinome objects (lvl1, lvl2 or names). It has the added benifits of adding the results the the KINOME.params object so that KINOME.get and KINOME.list recognize it. Additionally it will KINOME.enrich the returned object. It returns a JavaScript Promise which will pass the resultant object as an array to the then function.
KINOME.params
  • KINOME.params

    This is the object that is parsed from the URL parameters passed in. It has a data array that has all data loaded in by group. A list of the scripts loaded in by the URL and an array of the strings loaded in my URL. For more on this see URL Parameter Options.
KINOME.page
  • KINOME.page

    jQuery div for the contents of the 'home page' (This is the page that shows all currently loaded in information as a table and builds links to other data levels.)

URL Parameter Options

We have two base URLs, one for the toolbox and one for images. They both accept query parameters. They are as follows:

http://toolbox.kinomecore.com/ ?params (click "?params" to jump to details)

http://toolbox.kinomecore.com/image ?params (click "?params" to jump to details)

Toolbox Parameters

Starting from: http://toolbox.kinomecore.com/ we can add a large number of parameters. As is is the standard, the first parameter is indicated by a '?', all subsequent ones are seperated by an '&'.

  • &data=*[ data_url1 [; data_url2 [; ... ] ] ]*

    You can add any number of &data objects. They expect a list of data urls seperated by ';'. Spaces should be ignored, but it is best to just leave them out. All of these data urls should be Kinome Data Objects.

    If multiple &data=*[...]* tags are utilized then those group numbers will be indicated on the data objects loaded in and as the group number from the KINOME.list function.

  • &data=*[ data_url1 [| data_url2 [| ... ] ] ]*

    This notation for data_urls works identically to the above expect if all URLs are from the same base then you can indicate only the end of the URL. For example: &data=*[http://example.com/direc1/?param=param1;http://example.com/direc2/?param=param2;http://example.com/direc1/?param=param3]* is identical to &data=*[http://example.com/direc1/?param=param1|/direc2/?param=param2|?param=param3]*.

  • &code="<script_url>"

    Based on loading preferences this will be the last thing that loads. So all indicated data and text will be avaliable when this loads. This will allow you to develop an add on/plug in for the toolbox. Which can then be formalized by submitting the add on as a pull request. All add ons should be submitted as part of the plugin's folder.

  • &text="<text_url>"

    Our reccomendation is to utilize the (require function)[#default-global-functions-and-packages] in your script for this purpose, however this is supported and the results will be added to the KINOME.params.strings array. This is blocking, &code will not load until this resolves.

  • &style="<css_url>"

    Our reccomendation is to utilize the (require function)[#default-global-functions-and-packages] in your script for this purpose, however this is supported and the results will loaded into the system as part of <head>. This is non-blocking as style will only change the appearance of the page not the functionality.

Image Parameters

Starting from: http://toolbox.kinomecore.com/ we can add parameters as described below.

  • ?img="<image_file_name>"

    This will load in the image file indicated and alter the image for viewing online. It also provides a link to the original image file since the web version is low quality. Example

Kinome Data Objects

There are currently three major data types: names; level 1; level 2. The actual data models are described here: Swagger Hub KINOME. However the power of this toolbox is that you do not need to know what these data objects look like. You only need the following series of functions and a general description of each data type.

Data Array

This is an enriched array containing some number of enriched KINOME objects. While this could have multiple types in theory, it has only been tested on individual types. The easiest way to get one of these is with a KINOME.get() command, which returns an enriched array. The enriched array has the following functions:

data_arr.get(get_object)
  • data_arr.get(get_object)

    This can take as entry any of the parameters that return from an argumentless data_arr.list() call. It will iteratively call get functions on all members that match the main object filters (ie name or id) and return an array with all matching data points as objects. This and all data level get functions take an object with parameters pointing to either strings or arrays of strings and return an array will all matching points. The client side getting started provides an example of this.

    Each returned object can have functions attached, these are specifically described in the 'get' function for each type below.

data_arr.list(list_string)
  • data_arr.list(list_string)
    This will create an array for the matching list string option based on all possible options for a get function. If it is not given a list string it will return an object with all possible lists. The client side getting started provides an example of this function in action.
data_arr.clone()
  • data_arr.clone()
    This iteratively calls .clone() on all member objects. This clone is special in that it preserves the enrich functions and the non-enumerable properties.
data_arr.stringify()
  • data_arr.stringify()
    This calls JSON.stringify on all member objects, but only after going through and rounding all numbers to a six digit precision. The idea being that beyond that the data is not useful and the storage cost increases greatly.

The following are all specific entries in the data_arr above. Essentially if the first object in a data_arr (data_arr[0]) is of type 'Name' then it will have the functions described below on it. Additionally it will have a properperty data_arr[0].level === 'name'.

Name

This is just the meta data and peptide list for each data object loaded in. This data exists for quick load and for the building of down the line analytic groups. It has the following functions attached:

name.get(get_object)
  • name.get(get_object)
    This can take as entry any of the parameters that return from an argumentless name.list() call. It will return an array with all matching data points as objects. This and all data level get functions take an object with parameters pointing to either strings or arrays of strings and return an array will all matching points. The client side getting started provides an example of this function in action on a data_arr. It works identically here, just with different options.
name.list(list_string)
  • name.list(list_string)
    This will create an array for the matching list string option based on all possible options for a get function. If it is not given a list string it will return an object with all possible lists. The client side getting started provides an example of this function in action on a data_arr. It works identically here, just with different options.
name.clone()
  • name.clone()
    This clone is special in that it preserves the enrich functions and the non-enumerable properties.
name.stringify()
  • name.stringify()
    This calls JSON.stringify, but only after going through and rounding all numbers to a six digit precision. The idea being that beyond that the data is not useful and the storage cost increases greatly.

Level 1

Level 1 data is essentially base level data. It has not been parameterized and the only additions or alterations are based on the level 1.X.Y indications. These levels will be described in the supplement of a pending publication and in a future youtube video. (Links coming soon)

lvl1.get(get_object)
  • lvl1.get(get_object)

    This can take as entry any of the parameters that return from an argumentless lvl1.list() call. It will return an array with all matching data points as objects. This and all data level get functions take an object with parameters pointing to either strings or arrays of strings and return an array will all matching points. The client side getting started provides an example of this function in action on a data_arr. It works identically here, just with different options.

    Each returned object has 2 functions on it: A set function that sets the value of specific parameters in the base object and a more function that gives more information about the point itself.

    • set(key, value), sets the indicated key to the indicated value. Only works for 'background', 'background_valid', 'signal' and 'signal_valid'.
    • more(), returns the image index for the point and the full meta data array for the indicate peptide.
lvl1.list(list_string)
  • lvl1.list(list_string)
    This will create an array for the matching list string option based on all possible options for a get function. If it is not given a list string it will return an object with all possible lists. The client side getting started provides an example of this function in action on a data_arr. It works identically here, just with different options.
lvl1.level_up(equation_string)
  • lvl1.level_up(equation_string)
    This creates an empty level 2 object from the level 1 object scaffolding. It does not require, but it is highly recommended to pass it the equation string used for the parameterization of kinetic curves. Example is found here, but remove the extra spaces and comments using: eq_string.replace(/\/\/[^\n]*/g, "").replace(/\s+/g, ' '). Relevent metadata and other crucial information is passed over. This resultant object can then be built up using lvl2.put({put_object}).
lvl1.put(put_object)
  • lvl1.put(put_object)
    This will add a data point to a level 1 object. It requires 6 parameters passed in as object properties. They are as follows
        {
            valid: "boolean, true/false",
            value: "value to set, a number",
            type: "either signal or background, string",
            peptide: "peptide the value is for, string",
            cycle: "cycle number for value, null === post wash, number or null",
            exposure: "camera exposure time for value, null === cycle series, number or null"
        }

As expected this will add that value/valid combination into the object. This works regardless of if the point exists already. However if it does, this will add it in using a lvl1.get call then a pnt.set call.

lvl1.clone()
  • lvl1.clone()
    This clone is special in that it preserves the enrich functions and the non-enumerable properties.
lvl1.stringify()
  • lvl1.stringify()
    This calls JSON.stringify, but only after going through and rounding all numbers to a six digit precision. The idea being that beyond that the data is not useful and the storage cost increases greatly.

Level 2

Level 2 data is parameterized level 1 data. The slight variations on it are based on the origin level 1 data. These levels will be described in a future youtube video. (Links coming soon)

lvl2.get(get_object)
  • lvl2.get(get_object)

    This can take as entry any of the parameters that return from an argumentless lvl2.list() call. It will return an array with all matching data points as objects. This and all data level get functions take an object with parameters pointing to either strings or arrays of strings and return an array will all matching points. The client side getting started provides an example of this function in action on a data_arr. It works identically here, just with different options.

    Each returned object has 2 functions on it: A set function that sets the value of specific parameters in the base object and a more function that gives more information about the point itself.

    • set(key, value), sets the indicated key to the indicated value. Only works for background and signal as these are the only point determined parameters. You must set the entire object ({R^2, WW, [params]} for kinetic or {R^2, [params]} for linear).
    • more(), returns the image index for the point and the full meta data array for the indicate peptide.
lvl2.list(list_string)
  • lvl2.list(list_string)
    This will create an array for the matching list string option based on all possible options for a get function. If it is not given a list string it will return an object with all possible lists. The client side getting started provides an example of this function in action on a data_arr. It works identically here, just with different options.
lvl2.put(put_object)
  • lvl2.put(put_object)
    This will add a data point to a level 2 object. It requires 4 parameters passed in as object properties. They are as follows
        {
            type: "required, describes type of fit, either linear or kinetic, string",
            peptide: "required, peptide the value is for, string",            
            

            signal: "optional [either this or background required], fit object",
            background: "optional [either this or signal required], fit object",

            
            exposure: "required if type === kinetic, otherwise ignored, camera exposure time for value, null === cycle series, number or null",
            cycle: "required if type === linear, otherwise ignored, cycle number for value, null === post wash, number or null"
        }

As expected this will add that background/signal object into the main object. This works regardless of if the point exists. However if it does exist, this will add it in using a lvl2.get call then a pnt.set call.

lvl2.clone()
  • lvl2.clone()
    This clone is special in that it preserves the enrich functions and the non-enumerable properties.
lvl2.stringify()
  • lvl2.stringify()
    This calls JSON.stringify, but only after going through and rounding all numbers to a six digit precision. The idea being that beyond that the data is not useful and the storage cost increases greatly.

About

A series of tools for working with PamGene PamChip kinome arrays

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages