Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Establish Concrete Metrics for Pipeline Tasks #25

Open
Micky774 opened this issue Sep 13, 2021 · 2 comments
Open

Establish Concrete Metrics for Pipeline Tasks #25

Micky774 opened this issue Sep 13, 2021 · 2 comments

Comments

@Micky774
Copy link
Collaborator

It may be worth codifying exactly what the goal of the pipeline is. Roughly speaking, it can be described as creating a potent representation for downstream tasks, however classification can be considered one manifestation of this task (e.g. a simple ML model using the representation as features). What other tasks can we use to act as a test of representation power? And what qualitative features are desirable? What baseline classification scores can be achieved with "null models" e.g. random forest or simple conv-net?

TL;DR

  • How can we directly calculate "representation power"?
  • What downstream tasks can demonstrate "representation power"?
  • What baseline performance can we generate on such downstream tasks?
@Micky774 Micky774 added this to the Create Baseline Models milestone Sep 14, 2021
@rmattson1008
Copy link
Contributor

We should check robustness to noise. I don’t think it would be too hard to compare this on “downstream” tasks like the ones in the last scipy paper.

@Micky774
Copy link
Collaborator Author

Agreed, that would be a great addition

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants