Skip to content

Conversation

@dimitris-athanasiou
Copy link
Contributor

This commit extends the set of default metrics for the
data frame analytics evaluation API to all available metrics.
The motivation is that if the user skips setting an explicit
set of metrics, they get most of the evaluation offering.

This commit extends the set of default metrics for the
data frame analytics evaluation API to all available metrics.
The motivation is that if the user skips setting an explicit
set of metrics, they get most of the evaluation offering.
@elasticmachine
Copy link
Collaborator

Pinging @elastic/ml-core (:ml)

@dimitris-athanasiou
Copy link
Contributor Author

@szabosteve Perhaps it would be worth adding to our docs an explanation about which metrics are included by default. If you agree let's discuss it offline.

@dimitris-athanasiou dimitris-athanasiou merged commit 5f4f8b2 into elastic:master Oct 21, 2020
@dimitris-athanasiou dimitris-athanasiou deleted the extend-default-metrics-to-all-available branch October 21, 2020 06:00
dimitris-athanasiou added a commit to dimitris-athanasiou/elasticsearch that referenced this pull request Oct 21, 2020
…#63939)

This commit extends the set of default metrics for the
data frame analytics evaluation API to all available metrics.
The motivation is that if the user skips setting an explicit
set of metrics, they get most of the evaluation offering.

Backport of elastic#63939
dimitris-athanasiou added a commit that referenced this pull request Oct 21, 2020
…#63965)

This commit extends the set of default metrics for the
data frame analytics evaluation API to all available metrics.
The motivation is that if the user skips setting an explicit
set of metrics, they get most of the evaluation offering.

Backport of #63939
pugnascotia pushed a commit to pugnascotia/elasticsearch that referenced this pull request Oct 21, 2020
This commit extends the set of default metrics for the
data frame analytics evaluation API to all available metrics.
The motivation is that if the user skips setting an explicit
set of metrics, they get most of the evaluation offering.

private static List<EvaluationMetric> defaultMetrics() {
return Arrays.asList(new MeanSquaredError(), new RSquared());
return Arrays.asList(new MeanSquaredError(), new RSquared(), new Huber());
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Was it on purpose not to include MSLE here? Is it because we cannot set offset automagically for the user?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It was intentional indeed. The reason is because we cannot calculate MSLE for negative values.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants