-
Notifications
You must be signed in to change notification settings - Fork 0
2014 March Status Phase 1
OpenM++ phase 1 completed in March 2014 with following results (enumeration corresponds to OpenM++ design document, which also describes tasks):
- OpenM++ compiler (2.1 priority1): alpha version, working beta version with 60% functionality coverage
- types (classifications, ranges, partitions)
- parameters (exogeneous)
- agents
- variables (25% complete)
- inter-agent links
- events
- cross tabulation (except margins)
- meta-information (except labels & groups)
- OpenM++ controller for MPI cluster (2.2 priority1): beta version
- OpenM++ modelling library (3.1 priority1): beta version
- case-based and time-based models
- agent & event lifecycle
- event queue
- on-the-fly cross-tabulation updating
- Modgen-equivalent random number generators for exact output comparability
- OpenM++ model data storage design (3.2 priority1): beta version
- OpenM++ data library (3.3. priority1): beta version
- OpenM++ execute library (3.4 priority1): beta version
On top of essential phase 1 Roadmap tasks following items completed:
- compatibility layer for Modgen source model code .mpp files (2.3 priority2): alpha version
- OpenM++ output result viewers and model analysis tools, import/export into R (1.2 priority2): beta version
Deferred items mentioned in phase 1 Roadmap:
- all optional items (except two listed above) due to limited resources
- compatibility converter for Modgen parameters .dat files (2.3 priority1) postponed after extensive design discussions.
- components of OpenM++ compiler (2.1 priority1) due to limited resources
- agent collections
- parameters (endogenous)
- variables (75% remaining)
- cross-tabulation (margins)
- meta-information (labels & groups)
- derived tables
- other miscellaneous functionality
Overall results of OpenM++ phase 1 cover most of existing Modgen desktop functionality (excluding GUI).
OpenM++ foundation created as result of phase 1 project and it is opens up following four streams for subsequent development:
- model stream: ongoing work to move existing Modgen models onto OpenM++ platform
- cloud stream: build openM++ cloud PaaS and/or SaaS stack, emphasizing on scalability. Time: 11 months
- tools stream: creating openM++ desktop GUI based on Visual Studio, Eclipse or similar for model developers and users. Time: 9 months
- core stream: enhance openM++ core functionality, for example, modelling results post-processing and analysis.
Tools and cloud stream partially described in OpenM++ Design and Model Architecture documents.
Core stream task list is very flexible because it is generally include OpenM++ small core enhancements required for other development streams.
For example, as it is today, beta version of OpenM++ supports only SQLite as model database storage and cloud version of OpenM++ most likely require at least one of MySQL, PostgreSQL, Oracle, MSSQL or DB2 support. Due to flexible nature of core stream development it can be done incrementally as long as resources available, however it is very important strictly follow OpenM++ Design documents to make sure we are proceeding to the right direction and avoid common "creeping featurism" mistake.
Following tasks are required to be completed before or during OpenM++ cloud or desktop GUI development (enumeration corresponds to OpenM++ design):
- OpenM++ output converters:
- 2.5 priority 1: export into .csv for parameters and output results. Time: 10 days
- 2.5 priority 2: export into .xml for model data or user-defined subset of model data. Time: 16 days
- OpenM++ SQL loaders. Time: 4 weeks + 10 days for each db-vendor
- 2.4 priority 1: MS SQL, MySQL / MariaDB
- 2.4 priority 2: generic SQL99
- 2.4 priority 3: PostgreSQL, Oracle, DB2, Apache Derby, H2 or HSQL
- extend data library to support MySQL, PostgreSQL, Oracle, MSSQL or DB2 (3.3 priority3). Time: 3-4 weeks for each db-vendor
- completion of OpenM++ core support for i18n / L10n in runtime library. Time: 3 weeks
- Modgen .dat files compatibility converter (2.3 priority 1): required design decision. Time: from 10 days to 6 weeks.
- exploratory subsamples suite for OpenM++ models (see below). Time: between 5-9 weeks
Their is no fixed order in the list above, it can be implemented as required by other project or OpenM++ users.
This is high priority component which defined in OpenM++ design document (2.3 priority 1) as command-line utility to convert existing Modgen models data into OpenM++ format. It was originally planned for phase 1 development, but deferred due to unresolved design dependency with other parts of OpenM++, i.e. cloud model publusher or SQL loaders mentioned above.
There are two important aspects of .dat-convertor design:
- language complexity of .dat file syntax, which is in fact c++ initializers syntax with Modgen extensions
- environmental complexity of .dat-convertor use cases
Environmental complexity actually means variety of .dat-convertor use case scenarios in not yet well defined runtime environment. Please look at explanation on OpenM++ model use cases in Model Architecture document for more details.
Some examples may include:
-
developer:
- uses local Windows or Linux PC with GUI
- often recreate SQLite database and load input data hundred times to debug the model
- eventually need to pack model executable and data files and send it to researcher
-
researcher:
- HPC cluster (large or small) or local Windows, Linux workstation without GUI
- run the model thousand times loading wide variety of input data from prepared .dat files
- do not have admin privileges, especially on cluster, as result, can not install or adjust runtime environmemnt
- often need to pack model .dat files to publish it, move from local PC to HPC cluster or share with other researchers
-
institutional user:
- uses web UI to run the model in cloud, on HPC cluster or other powerful server environment
- have absolutely no access to actual server environment
- receives initial set of input .dat files from developer or researcher and want to upload it into cloud database
- cloud database most likely one of: MySQL, Oracle, MSSQL, PostgreSQL, DB2
From examples above you can see following requirements to model input data tools:
- it must be portable and can not assume specific OS or database
- user may have no access to actual model database (i.e. model developer have no access to cloud instance)
Possible solutions for .dat-files converter in context of above requirements:
- due to language complexity of .dat files it is nice to use OpenM++ compiler (omc) to parse it
- omc read .dat files and saves as:
- c++ values compiled into model executable, which in turn, saves it into target database during first run
- pro: everything in one file, ideal for consistency and transfer
- cons: model executable is huge, which increase compilation and execution time
- pro/cons: it is not possible to change model input data without re-compilation
- SQLite database
- pro: compact storage
- pro: ideal for model developer (or even researcher) as no any other steps required to run the model
- pro: there are many standard utilities to browse or edit the data
- cons: extra tool required to import from SQLite into actual database in cloud environment
- sql script files
- pro: portable and human-readable format
- pro: no any other tools required to transfer data from one environment into another
- cons: least compact storage, size of input data files are largest of all
- some other text format, i.e.: .csv or .xml files
- pro: portable and human-readable format
- cons: some custom tools required to load the data from such files into model database
- c++ values compiled into model executable, which in turn, saves it into target database during first run
We must keep in mind when choosing .dat-converter solution couple of other items from OpenM++ design document:
- OpenM++ must have SQL-loader utilities to facilitate data export into different model databases
- OpenM++ must have utilities to export input (and output) data into other formats, i.e.: text .csv and .xml files
That means we can relay on presence such OpenM++ utilities in foreseeable future.
Current OpenM++ model subsamples design is Modegn-compatible. It was done on purpose to provide Modgen model developers and users familiar concepts and even ability to reuse existing tools within OpenM++. However, there is fundamental limitation in that design, which became obvious when OpenM++ model runs in HPC cluster environment.
For example, if we have 16,000 CPUs cluster then it may be make sense to prepare 1000 different sets of input parameters, submit model job with those 1000 inputs * 16 subsamples each to use all 16,000 CPUs and analyse results to find optimal set of model parameters. It is possible to do in OpenM++ now by specifying working set of input parameters for and run the model 1000 times by submitting 1000 jobs to the cluster. However it would be nice to have such capability incorporated in OpenM++ runtime to allow simply submit single job with 1000 different sets of parameters and 16 subsamples each.
To implement such feature following changes in OpenM++ required:
- execution library: organize model MPI groups to effectively broadcast input parameters to slave modelling processes
- model database schema: allow multiple sets of input parameters for each model run (Modgen allow only single)
- model database schema: store relationship between input set and output results inside of single modelling job
- data library: redesign output results aggregation to do it over related output values (now it is done across all subsamples)
That feature should significantly increase model users productivity and allow more effective HPC cluster resource usage. It is recommended to have it for OpenM++ cloud version.
- Windows: Quick Start for Model Users
- Windows: Quick Start for Model Developers
- Linux: Quick Start for Model Users
- Linux: Quick Start for Model Developers
- MacOS: Quick Start for Model Users
- MacOS: Quick Start for Model Developers
- Model Run: How to Run the Model
- MIT License, Copyright and Contribution
- Model Code: Programming a model
- Windows: Create and Debug Models
- Linux: Create and Debug Models
- MacOS: Create and Debug Models
- MacOS: Create and Debug Models using Xcode
- Modgen: Convert case-based model to openM++
- Modgen: Convert time-based model to openM++
- Modgen: Convert Modgen models and usage of C++ in openM++ code
- Model Localization: Translation of model messages
- How To: Set Model Parameters and Get Results
- Model Run: How model finds input parameters
- Model Output Expressions
- Model Run Options and ini-file
- OpenM++ Compiler (omc) Run Options
- OpenM++ ini-file format
- UI: How to start user interface
- UI: openM++ user interface
- UI: Create new or edit scenario
- UI: Upload input scenario or parameters
- UI: Run the Model
- UI: Use ini-files or CSV parameter files
- UI: Compare model run results
- UI: Aggregate and Compare Microdata
- UI: Filter run results by value
- UI: Disk space usage and cleanup
- UI Localization: Translation of openM++
-
Highlight: hook to self-scheduling or trigger attribute
-
Highlight: The End of Start
-
Highlight: Enumeration index validity and the
index_errors
option -
Highlight: Simplified iteration of range, classification, partition
-
Highlight: Parameter, table, and attribute groups can be populated by module declarations
- Oms: openM++ web-service
- Oms: openM++ web-service API
- Oms: How to prepare model input parameters
- Oms: Cloud and model runs queue
- Use R to save output table into CSV file
- Use R to save output table into Excel
- Run model from R: simple loop in cloud
- Run RiskPaths model from R: advanced run in cloud
- Run RiskPaths model in cloud from local PC
- Run model from R and save results in CSV file
- Run model from R: simple loop over model parameter
- Run RiskPaths model from R: advanced parameters scaling
- Run model from Python: simple loop over model parameter
- Run RiskPaths model from Python: advanced parameters scaling
- Windows: Use Docker to get latest version of OpenM++
- Linux: Use Docker to get latest version of OpenM++
- RedHat 8: Use Docker to get latest version of OpenM++
- Quick Start for OpenM++ Developers
- Setup Development Environment
- 2018, June: OpenM++ HPC cluster: Test Lab
- Development Notes: Defines, UTF-8, Databases, etc.
- 2012, December: OpenM++ Design
- 2012, December: OpenM++ Model Architecture, December 2012
- 2012, December: Roadmap, Phase 1
- 2013, May: Prototype version
- 2013, September: Alpha version
- 2014, March: Project Status, Phase 1 completed
- 2016, December: Task List
- 2017, January: Design Notes. Subsample As Parameter problem. Completed
GET Model Metadata
- GET model list
- GET model list including text (description and notes)
- GET model definition metadata
- GET model metadata including text (description and notes)
- GET model metadata including text in all languages
GET Model Extras
GET Model Run results metadata
- GET list of model runs
- GET list of model runs including text (description and notes)
- GET status of model run
- GET status of model run list
- GET status of first model run
- GET status of last model run
- GET status of last completed model run
- GET model run metadata and status
- GET model run including text (description and notes)
- GET model run including text in all languages
GET Model Workset metadata: set of input parameters
- GET list of model worksets
- GET list of model worksets including text (description and notes)
- GET workset status
- GET model default workset status
- GET workset including text (description and notes)
- GET workset including text in all languages
Read Parameters, Output Tables or Microdata values
- Read parameter values from workset
- Read parameter values from workset (enum id's)
- Read parameter values from model run
- Read parameter values from model run (enum id's)
- Read output table values from model run
- Read output table values from model run (enum id's)
- Read output table calculated values from model run
- Read output table calculated values from model run (enum id's)
- Read output table values and compare model runs
- Read output table values and compare model runs (enun id's)
- Read microdata values from model run
- Read microdata values from model run (enum id's)
- Read aggregated microdata from model run
- Read aggregated microdata from model run (enum id's)
- Read microdata run comparison
- Read microdata run comparison (enum id's)
GET Parameters, Output Tables or Microdata values
- GET parameter values from workset
- GET parameter values from model run
- GET output table expression(s) from model run
- GET output table calculated expression(s) from model run
- GET output table values and compare model runs
- GET output table accumulator(s) from model run
- GET output table all accumulators from model run
- GET microdata values from model run
- GET aggregated microdata from model run
- GET microdata run comparison
GET Parameters, Output Tables or Microdata as CSV
- GET csv parameter values from workset
- GET csv parameter values from workset (enum id's)
- GET csv parameter values from model run
- GET csv parameter values from model run (enum id's)
- GET csv output table expressions from model run
- GET csv output table expressions from model run (enum id's)
- GET csv output table accumulators from model run
- GET csv output table accumulators from model run (enum id's)
- GET csv output table all accumulators from model run
- GET csv output table all accumulators from model run (enum id's)
- GET csv calculated table expressions from model run
- GET csv calculated table expressions from model run (enum id's)
- GET csv model runs comparison table expressions
- GET csv model runs comparison table expressions (enum id's)
- GET csv microdata values from model run
- GET csv microdata values from model run (enum id's)
- GET csv aggregated microdata from model run
- GET csv aggregated microdata from model run (enum id's)
- GET csv microdata run comparison
- GET csv microdata run comparison (enum id's)
GET Modeling Task metadata and task run history
- GET list of modeling tasks
- GET list of modeling tasks including text (description and notes)
- GET modeling task input worksets
- GET modeling task run history
- GET status of modeling task run
- GET status of modeling task run list
- GET status of modeling task first run
- GET status of modeling task last run
- GET status of modeling task last completed run
- GET modeling task including text (description and notes)
- GET modeling task text in all languages
Update Model Profile: set of key-value options
- PATCH create or replace profile
- DELETE profile
- POST create or replace profile option
- DELETE profile option
Update Model Workset: set of input parameters
- POST update workset read-only status
- PUT create new workset
- PUT create or replace workset
- PATCH create or merge workset
- DELETE workset
- POST delete multiple worksets
- DELETE parameter from workset
- PATCH update workset parameter values
- PATCH update workset parameter values (enum id's)
- PATCH update workset parameter(s) value notes
- PUT copy parameter from model run into workset
- PATCH merge parameter from model run into workset
- PUT copy parameter from workset to another
- PATCH merge parameter from workset to another
Update Model Runs
- PATCH update model run text (description and notes)
- DELETE model run
- POST delete model runs
- PATCH update run parameter(s) value notes
Update Modeling Tasks
Run Models: run models and monitor progress
Download model, model run results or input parameters
- GET download log file
- GET model download log files
- GET all download log files
- GET download files tree
- POST initiate entire model download
- POST initiate model run download
- POST initiate model workset download
- DELETE download files
- DELETE all download files
Upload model runs or worksets (input scenarios)
- GET upload log file
- GET all upload log files for the model
- GET all upload log files
- GET upload files tree
- POST initiate model run upload
- POST initiate workset upload
- DELETE upload files
- DELETE all upload files
Download and upload user files
- GET user files tree
- POST upload to user files
- PUT create user files folder
- DELETE file or folder from user files
- DELETE all user files
User: manage user settings
Model run jobs and service state
- GET service configuration
- GET job service state
- GET disk usage state
- POST refresh disk space usage info
- GET state of active model run job
- GET state of model run job from queue
- GET state of model run job from history
- PUT model run job into other queue position
- DELETE state of model run job from history
Administrative: manage web-service state
- POST a request to refresh models catalog
- POST a request to close models catalog
- POST a request to close model database
- POST a request to delete the model
- POST a request to open database file
- POST a request to cleanup database file
- GET the list of database cleanup log(s)
- GET database cleanup log file(s)
- POST a request to pause model run queue
- POST a request to pause all model runs queue
- PUT a request to shutdown web-service