Marigold icon indicating copy to clipboard operation
Marigold copied to clipboard

[CVPR 2024 - Oral, Best Paper Award Candidate] Marigold: Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation

Results 50 Marigold issues
Sort by recently updated
recently updated
newest added

Hi, nice work! I always want to compare the predicted depth with the colorful unprojected point cloud. I compared Marigold, ZoeDepth, OmniDatav2. I tried the following image. ![example_5](https://github.com/prs-eth/Marigold/assets/50771152/e8f50bf2-70b1-4d5b-a4f1-81759c4f4e4b) Marigold: ![image](https://github.com/prs-eth/Marigold/assets/50771152/31e56261-d2e4-47bb-b274-c13e32d17f67)...

I was told that Marigold maxing out my 24gb VRAM isn't supposed to happen and was told to post my settings here. I have a bat file that I run...

Thanks for sharing your code and model. The depth visualization is really awesome, especially sharp edges. I noticed that both training datasets(Hypersim and Virtual KITTI) were synthetic datasets. Have you...

In the current implementation, we have the following problem: If the user input an image of odd shape (either not divible by 8 or aspect ratio which don't allow to...

Thanks for your great work! This effect is currently the best and most effective model I have ever seen. I am not familiar with depth estimation work, so how can...

Can you share how the depth maps were used to generate the point clouds? I checked this issue ( https://github.com/prs-eth/Marigold/issues/6 ), but it doesn't specify any method on how to...

Hi I want to know if it's possible to add vector displacement ? it's allow to get better result than depth map on 3d objects depth vs vector : ![VectorDisp02](https://github.com/prs-eth/Marigold/assets/157552424/e26820cf-4afc-455c-8ba2-c72e0855a727)

I was requesting the model in another format because I cannot convert it without the proper model configuration file (I've tried) Need them in onnx or .pt format specifically for...

During fine-tuning for depth estimateaion conditioned on input image, how to deal with the text prompt required in the original pre_trained text-to-image Stable Diffusion model?

As the title said, why the paper was named Marigold? I'm curious about this.