Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Capture whether a Tool implement an ensemble strategy #259

Open
tschaffter opened this issue Sep 5, 2021 · 3 comments
Open

Capture whether a Tool implement an ensemble strategy #259

tschaffter opened this issue Sep 5, 2021 · 3 comments
Labels
Enhancement New feature or request Priority: Low

Comments

@tschaffter
Copy link
Member

tschaffter commented Sep 5, 2021

This proposal comes after reading the article On the “usefulness” of the Netflix Prize. In this article, it is highlighted that complicated ensemble methods may not be suitable for production ready application. We have experienced a similar situation with the Digital Mammography DREAM Challenge where the final method was an ensemble of 11 tools/docker images. This strategy is actually adopted in most DREAM challenges, where the final model published is an ensemble of the best performing models submitted during the competitive or collaborative phase.

There are two bits of information that we may want to capture, possibly as properties of the Tool schemas:

  • Whether the "tool" - the docker image submitted - is an ensemble of the output of multiple algorithms.
    • This may also flag tools that may take a long time to run, though we will ultimately capture and report on tool runtime in the future.
  • Whether the submitted tool can be trained, re-trained and/or fine-tuned
    • Distinguish between a tool that has not yet been trained and must absolutely be trained before being used for inference, and a tool that has been previously trained and can be further re-trained or fine-tuned, for example periodically on new data.
    • Enabling submitted tools to train on private data from data sites should be possible before the end of 2021.
@tschaffter tschaffter added Enhancement New feature or request Priority: Low labels Sep 5, 2021
@tschaffter
Copy link
Member Author

tschaffter commented Sep 6, 2021

Curious to hear what are your thoughts regarding the additional tool metadata mentioned above. @cascadianblue @yy6linda @ymnliu @chepyle

@tschaffter
Copy link
Member Author

I'm wondering if there are existing schemas that describe the "trainability"/training state of a model that we could reuse for the NLP Sandbox Tool schema.

@boyleconnor
Copy link
Contributor

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Enhancement New feature or request Priority: Low
Projects
None yet
Development

No branches or pull requests

2 participants