yocto-gl icon indicating copy to clipboard operation
yocto-gl copied to clipboard

[FR] Mlflow Evaluation method to provide feasibility to subset the list of evaluation metrics

Open khojarohan4 opened this issue 2 years ago • 4 comments

Willingness to contribute

Yes. I would be willing to contribute this feature with guidance from the MLflow community.

Proposal Summary

Currently, based on the type of the model be it classification or regression, with evaluators as default, Mlflow evaluate method outputs a defined list of evaluation metrics. This feature request is to check if we could achieve a kind of filtering in the evaluation metric so the user can select which of the default metrics is required. In some case, not all metrics are required and this provides the user with more flexibility.

Motivation

What is the use case for this feature?

User to select the list of evaluation metrics from the list present when default evaluator is used.

Why is this use case valuable to support for MLflow users in general?

More flexibility

Why is this use case valuable to support for your project(s) or organization?

User can only look at the list of metrics that are required

Why is it currently difficult to achieve this use case?

Details

No response

What component(s) does this bug affect?

  • [ ] area/artifacts: Artifact stores and artifact logging
  • [ ] area/build: Build and test infrastructure for MLflow
  • [ ] area/docs: MLflow documentation pages
  • [ ] area/examples: Example code
  • [ ] area/model-registry: Model Registry service, APIs, and the fluent client calls for Model Registry
  • [ ] area/models: MLmodel format, model serialization/deserialization, flavors
  • [ ] area/pipelines: Pipelines, Pipeline APIs, Pipeline configs, Pipeline Templates
  • [ ] area/projects: MLproject format, project running backends
  • [ ] area/scoring: MLflow Model server, model deployment tools, Spark UDFs
  • [ ] area/server-infra: MLflow Tracking server backend
  • [ ] area/tracking: Tracking Service, tracking client APIs, autologging

What interface(s) does this bug affect?

  • [ ] area/uiux: Front-end, user experience, plotting, JavaScript, JavaScript dev server
  • [ ] area/docker: Docker use across MLflow's components, such as MLflow Projects and MLflow Models
  • [ ] area/sqlalchemy: Use of SQLAlchemy in the Tracking Service or Model Registry
  • [ ] area/windows: Windows support

What language(s) does this bug affect?

  • [ ] language/r: R APIs and clients
  • [ ] language/java: Java APIs and clients
  • [ ] language/new: Proposals for new client languages

What integration(s) does this bug affect?

  • [ ] integrations/azure: Azure and Azure ML integrations
  • [ ] integrations/sagemaker: SageMaker integrations
  • [ ] integrations/databricks: Databricks integrations

khojarohan4 avatar Aug 02 '22 08:08 khojarohan4

Hi @khojarohan4. Thank you for your feature request; this sounds like a good idea. I think that adding a metrics configuration key to the evaluator_config dictionary for the default evaluator whose value is a list of metrics to compute seems like a good way to approach it, and we would be excited to review a pull request that implements this approach. Please let me know if you have any questions.

dbczumar avatar Aug 03 '22 23:08 dbczumar

Hi @khojarohan4 , just following up here - are you still interested in contributing this feature?

dbczumar avatar Aug 09 '22 17:08 dbczumar

Hi @dbczumar , Apologies I havent had a chance to even look at the codebase yet. I'm really willing to and just been on a tight deadline. If its urgent, I would take a step back. Also, I havent contributed anytime so if you have any doc for the process or any reference, I can atleast take a stab but this isnt likely soon. Let me know.

khojarohan4 avatar Aug 10 '22 10:08 khojarohan4

@dbczumar Please reply to comments.

mlflow-automation avatar Aug 25 '22 00:08 mlflow-automation