Optimal Data Engine (ODE) Build 005.002.003
Copyright 2015 OptimalBI - Licensed under the GNU GENERAL PUBLIC LICENSE Version 3.0 (GPL-3.0)
Good architecture is modular, separating concerns across its components. This helps ensure each component does one job well, which in turn helps the components work together, delivering a valuable experience to users. Modular architecture is also easier to maintain. A useful data warehouse does not attempt to take data directly from source systems to end-users in one hit, instead it uses modular components we call data layers. While these modular components have many specific names (e.g. “staging”, “foundation”, “presentation”) what ties them together is their layered nature: each one does a single job well, before passing data on to another layer. Many of the ideological battles of the past (e.g. Inmon vs Kimball) were founded on an assumption that one methodology must rule them all. As soon as we think about a data warehouse in terms of layers, we are free to choose the optimal methodology for the job each layer is doing.
What Optimal Data Engine (ODE) does:
Using the Data Vault methodology, ODE breaks each complex part of your data processing into layers to make it easier and faster to manage and implement. It's a true agile methodology for building a Data Warehouse. ODE breaks the complexity into layers - to find out more visit http://www.ode.ninja/data-layers/
Requirements (All users):
To Install ODE, you need to have:
- Access to a running Microsoft SQL Server (for example, Express Edition)
- SQL Server Management Studio installed on a computer which you can access
- Administrative rights to be able to create databases
If you wish to develop ODE further, we recommend:
- Visual Studio 2015 Community Edition installed on a computer which you can access. This can be downloaded from https://www.visualstudio.com/en-us/mt171547.aspx
The related project "optimal-data-engine-mssql-scripts" in the OptimalBI repository contains a set of scripts, which can be downloaded and used to assist in making use of ODE. https://github.com/OptimalBI/optimal-data-engine-mssql-scripts
Currently, ODE has two Branches available:
- master and
Master contains code which is known to be running in a production environment.
Develop contains the latest, locally tested version of the codebase.
Download and build instructions:
If you wish to develop the code base further, we recommend that you use the Visual Studio solution which is provided.
If you simply wish to build an ODE instance and use it, the following instructions will direct you:
- Download a copy of the zip and extract to a temporary folder
- Open SQL Server Management studio and load ode_to_mssql_Create.sql from the extracted zip file. You will find it in the ReleaseScript folder.
- Within SQL Server Management Studio > Click Query Menu > SQLCMD Mode
- Within the script optionally change the DatabaseName and DefaultFilePrefix in the code to the preferred Configuration database name; default is ode_to_mssql. ODE_Config is recommended.
- Click Execute from the toolbar. This should run successfully with a result of 'Update complete' on the Message panel
- The Results panel of Management Studio query execution window should show 13 rows, which are the contents of the log4.JournalControl table.
- This method expects that the user has a good understanding of Visual Studio Data Tools.
- Download the Optimal Data Engine code from GitHub, into a directory, e.g. "C:\Git\OptimalDataEngine"
- Using Visual Studio, open the ode_to_mssql.sln solution.
- Publish the Database to your server - preferred name for the database is ODE_Config.
How to start using ODE for Data Vault modeling
- Check http://ode.ninja/create-a-data-vault-with-ode-v5/ for the next steps
Details of the current functionality can be found here http://www.ode.ninja/category/features/
- Untested on SQL Server editions prior to 2014. Release script is compiled for SQL Server 2016 edition.
Feedback, suggestions, bugs, contributions:
Please submit these to GitHub issue tracking or join us in developing by forking the project and then making a pull request!
Find out more:
Visit http://ode.ninja/ - this is where we keep our guides and share our knowledge. To find out more about OptimalBI and the work we do visit http://www.optimalbi.com or check out our blogs at http://optimalbi.com/blog/tag/data-vault/ for all the latest on our Data Vault journey. If you want to get in touch, you can email us at email@example.com
Build 005.002.003 on 20181123 * Fixed retired columns and enabled hyphen in tables name Build 005.002.002 on 20180302 * Bug fixes Build 005.002.001 on 20171028 * Added connectivity to the Oracle as a data source via BIML scripts Build 005.001.001 on 20170911 * Added BIML scripts to generate SSIS package that loads data from another server * Implemented CDC (change data capture) load type * Removed limitation on release script length Build 004.001.001 on 20170301 * Added satellite columns * Added link columns and reorganised the hub-link relationship to enable Same-as links * Reorganised source table functionality * Added satellite functions feature for the calculated fields * Added table reconciliation functionality (left-right object match) Build 002.002.001 on 20160819 * Added the ability to pass a parameter from the scheduler, to a Staging Procedure, telling the Procedure what type of run to perform (Full or DELTA). Build 002.001.001 on 20160805 * Added the "is_retired" switch on dv_hub, dv_link, dv_sat, dv_source_table, dv_source_system and dv_column tables. This column is currently documentary only. Build 001.001.001 on 20151020 * Automated Install Script. * Added new schema (dv_config) which holds helper scripts to populate some configuration tables. Build 001.001.001 on 20150928 * Initial Build.