Skip to content

CM interface and automation recipes to access, manage, prepare, run and reproduce research projects from AI, ML and Systems conferences

License

Notifications You must be signed in to change notification settings

ctuning/cm4research

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

86 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Common CM interface to make it easier to prepare, run and reproduce experiments from research projects

About

This CM repository contains a unified Collective Mind interface (MLCommons CM) to access, run and reproduce experiments from research projects and benchmarks (ACM, IEEE, NeurIPS, ICML, MLCommons, MLPerf ...) in a unified and automated way using the artifact evaluation methodology from ACM/IEEE/cTuning, MLCommons and NeurIPS.

Why?

While working with the community to reproduce or replicate 150+ research papers during artifact evaluation, we have seen that reviewers spend most of their time at the kick-the-tires phase deciphering various READMEs and scripts to figure out how to prepare and run experiments.

This experience motivated us to develop a simple, technology-agnostic and human-friendly interface and automation language (MLCommons Collective Mind) to provide the same common interface to prepare, run and visualize experiments from any paper or research project.

The goal is to make it easier for the community and evaluators to start reproducing/replicating research results and even fully automate this process in the future.

License

MIT

How to use

Install the CM automation language as described here.

Install common research automations from MLCommons via CM

cm pull repo mlcommons@cm4mlops --branch=dev

Install this repository

cm pull repo ctuning@cm4research

List available interfaces for research projects

cm find script --tags=reproduce,project

Set up virtual environment before running experiments

If you use python, we suggest you to set up Python virtual environment via CM as follows:

cm run script "install python-venv" --name=reproducibility
export CM_SCRIPT_EXTRA_CMD="--adr.python.name=reproducibility"

It will be automatically activated for all further CM commands while keeping your Python installation intact.

Run experiments

Each CM script wrapper should have a CM script with the following variations: install_deps, run, analyze, plot, validate, reproduce

It's possible to run them as follows:

cm find script --tags=reproduce,project
cm run script {name from above list} --tags=_{variation}

Check/clean CM cache

Many CM commands download or produce new artifacts that are stored in CM cache (including CM Python virtual environments) to avoid polluting user system.

You can view current CM cache as follows:

cm show cache
cm show cache --tags=python

You can clean CM cache and start from scratch at any time using the following command:

cm rm cache -f

Add CM interface for new projects and papers

You can just copy any CM script and update alias in _cm.yaml. You can then extend it based on the following examples:

After manually updating meta description of CM artifacts (_cm.yaml or _cm.json) you must reindex them as follows:

cm reindex repo

We are updating a tutorial to add CM interface to new projects and papers and will share it soon!

Contact us

Don't hesitate to open tickets here or contact the cTuning foundation and cKnowledge.org (developers of the MLCommons CM automation framework).

About

CM interface and automation recipes to access, manage, prepare, run and reproduce research projects from AI, ML and Systems conferences

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published