-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Improved industrialization support: persisting of model (configuration), reapeated model scoring, model monitoring #127
Comments
Hi @sborms, @ZlaTanskY, @nicolasmorandi and @c-morey! Before we can finish implementing this issue entirely and have a better support for industrializing Cobra models, I think we need to discuss a few things. Proposed use cases (and processes) throughout industiralization & model usage phase that we could support better in Cobra:
Integrations necessary for the above:
Additional task: documenting the thoughts above Am I missing interesting use cases or integrations above? Feel free to suggest. Also: we cannot implement everything right now, and not even in the coming years, but must pick the most interesting things at each time, just adding the use cases and integrations just on-the-go as we are industrializing Cobra for clients with different demands and infrastructure. I've also gathered the files from the Brico pull request and started structuring it a bit, so we can integrate their efforts into Cobra, see the draft pull request mentioned below on this page (only FYI). But I'd like to first discuss the above thoughts before proceeding on the gathered code, so we agree on what we want to do. |
Nicolás is interested in the investigation of MLFlow, that investigation could fit in this issue. See details on MLFlow's github, all 4 of the MLFlow components (Tracking for storing model parameters, Projects for reproducible runs, Models for easy deployment and Model registry to track model evolution throughout the model's lifecycle) are very interesting to build integrators for within Cobra. Up to you to decide @nicolasmorandi @pietrodantuono. |
Add functionality to write away model metadata
It would be nice to have a function for writing away the date and time of each new modeling attempt, which variables were selected, which preprocessing was done and which was the resulting score.
Task Description
This could comprise:
Provide the code for extracting this metadata, but allow a data scientist/engineer to write a plugin function to do the actual writing of the metadata to the database/filestore of choice.
The text was updated successfully, but these errors were encountered: