TextZoom
TextZoom copied to clipboard
Production of data sets自制数据集
wenjia,你好,我想参照你的方法自己制作数据集,用长焦距拍摄图像做HR,短焦距做LR。请问你是如何预处理图像的,具体是如何crop小焦距和大焦距图像的同一个区域且大小一样来做成图像对的?非常期待您的回复,谢谢
你好yuan-anan,建议你看一下SRRAW和realSR的github,他们有根据传统算法进行角点匹配进行HR-LR两张图匹配抠图的代码,但是不算非常准确,会有十几到几十个像素的偏移。
Hi Jason, I am trying to use TPL for documents super-resolution. But I am not able to get most of the things from this repo, can you help me out to include TPL Loss in my model...
Thanks,
Hi Jason, following YUAN-ANAN question. Maybe you can add to the repo the code you used to create the dataset from the original images ? (detect text, crop and detect same text and crop in second image). Thanks
Hi Jason, I am trying to use TPL for documents super-resolution. But I am not able to get most of the things from this repo, can you help me out to include TPL Loss in my model...
Thanks,
TPL, I think it is not so useful. I quit OCR now.
Hi Jason, following YUAN-ANAN question. Maybe you can add to the repo the code you used to create the dataset from the original images ? (detect text, crop and detect same text and crop in second image). Thanks
It is very easy. I did not detect the text bboxes by detection networks. I crop the text from manual annotation.
Hi Jason, I am trying to use TPL for documents super-resolution. But I am not able to get most of the things from this repo, can you help me out to include TPL Loss in my model... Thanks,
TPL, I think it is not so useful. I quit OCR now.
Thanks for reply.
But I have tried all approaches (SRGAN,ESRGAN,UNET models) ,and also different kinds of losses.
But , model is able to reconstruct blurry images, but when it comes to come up with some letters when the source is almost distorted completely, it cannot,(AS MODEL IS NOT AWARE OF ENGLISH LETTERS)
Like, this is source image...

And this is model output:
But in this image, you can see for blurry, it can do something:


So, I think if model has any sense of English letters, it can come up with something... Thanks, for any help....