yocto-gl icon indicating copy to clipboard operation
yocto-gl copied to clipboard

We'd like your feedback on MLflow Experiment Tracking

Open ridhimag11 opened this issue 2 years ago • 1 comments

MLflow Experiments lets you easily track your experiments and compare runs. We're looking to further improve the tracking and comparison UX and would love to get your feedback on some prototypes we are exploring.

If you're interested in chatting with us and giving us feedback, please sign up here 🙌

ridhimag11 avatar Jul 27 '22 23:07 ridhimag11

@BenWilson2 @dbczumar @harupy @WeichenXu123 Please assign a maintainer and start triaging this issue.

mlflow-automation avatar Aug 04 '22 00:08 mlflow-automation

Hello team, nice registry for the tracking...

KeyanatGiggso avatar Apr 26 '23 10:04 KeyanatGiggso

In Table View, the metrics that are displayed are the ones from the last epoch. Would be nice to have the minimum / maximum achieved for that model instead of the last metric obtained.

dnsleu avatar May 18 '23 13:05 dnsleu

In Table View, the metrics that are displayed are the ones from the last epoch. Would be nice to have the minimum / maximum achieved for that model instead of the last metric obtained.

Or even the number from a custom step?

lester1027 avatar May 23 '23 01:05 lester1027

One difficulty I find in the UI is it only loading 100 runs at a time. The charts in "Chart View" are only initially representative of a small portion of the dataset. One has to repeatedly click "load more" to see the charts fill out. It would be nice to be able to load all trials, or at least a larger number at once.

Another improvement would be if the Charts built in "Chart View" were persistent. As well as transferable to other experiments, since often I'm always building the same charts repeatedly.

Anyway, hope these recommendations are helpful. I enjoy using MLFlow, keep up the good work.

roark avatar Jun 15 '23 02:06 roark

Bug - in "Chart View", a tooltip shows on mouseover, however clicking on the relevant model (sometimes) causes it to disappear. Therefore, the buttons on the tooltip are unreachable. image (it seems this is an intermittent bug. Tough to pin down a consistent way to reproduce but it seems to happen about 20-30% of the time)

Feedback - it would be great to be able to set a different sort metric for bar charts. "First N" seems to be the only option - would be great to get "Top N". We can also only select N from a list of [5, 10, 20, 100, 250, 500] - a "nice to have" would be the ability to set a custom value. image

jmcapra avatar Jun 21 '23 14:06 jmcapra

Feedback - add a "hide most recent run" button for the Parallel Coordinates chart - the most recent (currently training) run tends to populate this chart with null values, greatly reducing the resolution of the true parameter values.

Also, a related bug - filtering, e.g. by val_accuracy > 0.6, does not remove this run where val_accuracy is null.

jmcapra avatar Jun 21 '23 14:06 jmcapra

Another feature request - some way to show calculated summary statistics, e.g. average accuracy across all runs. E.g. a new chart type, "Statistics" or "Table", which updates based on the SQL-style filter in the search box and can show the average of any metrics.

jmcapra avatar Jun 21 '23 15:06 jmcapra

Thanks for sharing your this feedback @jmcapra. Would you be up for a brief chat about this issues and any other feedback on MLflow? For context, I'm a designer working on MLflow. You can book some time with me here.

ridhimag11 avatar Jun 21 '23 20:06 ridhimag11

Hi @ridhimag11, I'd be happy to. I've booked some time with you next week.

Another tiny bug: I have a run named "ConvNet - Augmented - FFT/Wavelet" but it is interpreted as a file path and just shows as "Wavelet" in some sections. image

Another feature request: The ability to sort the z-level of different runs in the parallel coordinates plot (and line charts). It's not clear how they are currently sorted, however I think a better default would be to sort by the performance metric, so the best performing models are on top/in front. That way, it would be much easier to judge which parameter values are performing better. Currently, the only way I've been able to bring them to the front is by selecting by mouseover/click in the final section where they move from the last parameter value to the metric (top-down). Naturally, all of the lines between the other parameter values overlap and make it difficult to analyse performance from the bottom up.

jmcapra avatar Jun 28 '23 00:06 jmcapra

how to remove the banner 'We’ve made several improvements to the new runs comparison experience.... If you prefer to use this experience, we'd love to know more.' ?

nkrot avatar Jul 25 '23 06:07 nkrot

I can see the following experiment being run on localhost:5000 but there are no files saved at the given location. Experiment ID: 1 Artifact Location: /home/ubuntu/notebooks/07-project/training_tracking_registry/mlruns/1

What could be the possible issue ? I'm using mlflow 2.3.1 on python3.10.0

khushal2911 avatar Jul 31 '23 12:07 khushal2911

It would be nice to be able to group experiments via project.

samthompsonAQ avatar Aug 02 '23 12:08 samthompsonAQ

Thanks for this issue. Is there any reason why the plots are not displayed by default like the chat_view when comparing? It is pretty hard to open and analyse each metrics and impossible to compare metrics of the same run. Am I missing something?

image

bsridatta avatar Sep 05 '23 15:09 bsridatta

Just upgraded to the version with chart view, and it's very cool and nice to be able to see comparisons side by side now :) A bit of feedback: More color diversity would be great, a lot of my charts end up looking like this by accident (two grays, two blues) image On a cold start of the chart page, a lot of charts are generated, would be convenient if there was an option to "clear the dashboard". Would be very useful if selecting an experiment from the sidebar would highlight the corresponding chart elements, seeing as occlusion sometimes prevents clicking the element directly in the chart. Another useful feature would be a paired zoom, e.g. focusing on the same step range across all charts.

Bonnevie avatar Sep 20 '23 09:09 Bonnevie

It would be nice to be able to "Show Differences Only" in the Table View among the Parameters/Metrics as it was possible in the older versions (i.e. 1.23.1) to distinguish the runs at a glance.

claudiofernandez avatar Oct 10 '23 13:10 claudiofernandez

I think it would be helpful to have a datasets section of the run comparison, and be able to see the dataset hashes in the comparison view so we can easily see what, if any, datasets have changed between runs!

aboomer07 avatar Oct 26 '23 18:10 aboomer07

+1 @claudiofernandez ; please add back "Show diffs". Very important feature in practice. @harupy

ReHoss avatar Nov 03 '23 10:11 ReHoss

I'd really appreciate dark mode.

phelps-matthew avatar Nov 13 '23 06:11 phelps-matthew

pls hide the artifactory location s3://aml....

Screenshot 2023-11-19 at 7 30 57 PM

a3sha2 avatar Nov 20 '23 00:11 a3sha2

pls hide the artifactory location s3://aml....

Screenshot 2023-11-19 at 7 30 57 PM

I personally find it very useful to see where data is stored!

ReHoss avatar Nov 21 '23 18:11 ReHoss

I find the new useless. It lacks the ability to select runs (there's only option to select 5, 10, 100 etc). Since I can't select runs to analyze it's of no use for me.

emsi avatar Nov 22 '23 17:11 emsi

I find the new useless. It lacks the ability to select runs (there's only option to select 5, 10, 100 etc). Since I can't select runs to analyze it's of no use for me.

@emsi drive-by comment: you can use "Eye" icon in order to add/remove runs from the comparison.

hubertzub-db avatar Nov 24 '23 14:11 hubertzub-db

@emsi drive-by comment: you can use "Eye" icon in order to add/remove runs from the comparison.

There are selctbox which are NOT honored when entering comparison and I end up in comparison with dozens to uncheck with eye now? So instead of checking 3 I have to un-eye 50? That's not convenient.

emsi avatar Nov 24 '23 14:11 emsi

Hi, is it possible to add token numbers to all parts of prompts? Currently we can only see the total number of tokens, but in addition to the total number it would be useful to see the tokens for each part separately. image

alena-m avatar Jan 10 '24 13:01 alena-m

Is it possible to add the ability to change the row width in the Evaluation table? Sometimes the output is too long for cells, but I still want to compare them side by side. image

alena-m avatar Jan 10 '24 13:01 alena-m

Is it possible to automatically fill in variable values from the previous run when we do "Duplicate run" for Prompt Engineering? If we have many variables it is tedious to fill them again and again. image

alena-m avatar Jan 10 '24 13:01 alena-m

I think showing the min values of train_loss and val_loss in the main table would be more helpful.

kyoungrok0517 avatar Jan 17 '24 22:01 kyoungrok0517

I miss being able to see the datasets used in a specific run. I find it quite difficult to filter a dataset in order to find which runs used it. The previous run view (with a dataset dropdown) was friendlier.

mendelson avatar Feb 01 '24 13:02 mendelson

Hi @mendelson - can you share a screenshot of what you're seeing regarding the missing datasets. AFAIK, you should still be able to see the dataset dropdown in the list view.

ridhimag11 avatar Feb 01 '24 18:02 ridhimag11