Shishira R Maiya
Shishira R Maiya
Hi, Do you know how to recover the graph/table structure from the model predictions ?
@akshowhini @rmporsch Had a doubt regarding the annotations. x1,x2,y1,y2 with respect to what ? Original PDF image ? or the cropped image ? Both do not seem to match. I...
@rmporsch I saw that. So basically there are chunks in the chunk folder and once loaded they are transformed. But I visualized the both the table and pdf (as image)...
@rmporsch Is this a preprocessing function you apply in utils.py ? And I assume they refer to xmin,xmax,ymin,ymax of the table image (Not the pdf image).
@rmporsch Okay. Her is my code for relation predictions. `for d in dataset: outputs = model(d.nodes, d.edges, d.adj, d.incidence); output_rel = outputs.max(dim=1)[1]; ` Now, your post processing needs to be...
@rmporsch @akshowhini Sorry for the silly questions. Also what exactly do the coordinates in .chunk files represent ? I tried visualizing them for both pdf and table image. The coordinates...
I think you should follow the error metric provided in compute_ap_range. As for the second problem, I suggest keep it at 1024*1024.
The objective is to track the same car across different camera angles. Hence, we would like our model to learn features which are invariant to different views.