ddf-pipeline icon indicating copy to clipboard operation
ddf-pipeline copied to clipboard

Slow running of P070+29

Open vmahat opened this issue 1 year ago • 12 comments

Hello,

I've been running the ddf-pipeline with standard LoTSS settings on P070+29.

It seems to hang, however, in one of the DI steps during deconvolution (image_ampphase1_di). The longest I've left this running is about 4 days during which nothing is written on disk -- I could wait a bit longer but this is probably an issue.

I've been running the pipeline on the Herts cluster on one of the large nodes.

I've attached the latest log file and my config file for reference -- I noticed that the nearby P067+29 has been processed before. The only difference between the cfg file used for that pointing and P070+29 is the addition of NVSS for flux bootstrapping. There was a warning of diverging flux at the end of some of the major cycles, so I may re-run with NVSS information included as for P067+29. Note both of these fields contain the extremely bright 3C 123. Update on this will be posted, but any tips on the deconvolution hanging without progressing would be appreciated.

image_ampphase1_di.log tier1-jul2018_vijay.log

vmahat avatar Jan 19 '24 15:01 vmahat

Here are the corresponding bootstrap images with my P070+29 run without NVSS (left) and that from the P067+29 with NVSS in /beegfs/car/mjh/P067+29/ (right), centred on 3C 123. image

vmahat avatar Jan 19 '24 15:01 vmahat

Hi Vijay, well I think it's clear why the deconvolution doesn't play nicely... 3C123's calibration is sufficiently screwed up that the mask is most likely filled with garbage.

The question is whether there's anything you can do to rescue this. Does the initial image (image_dirin_SSD_m.app.restored.fits) look comparable in quality to the one from the adjacent field?

mhardcastle avatar Jan 19 '24 16:01 mhardcastle

It looks like this (my one on the left), so I guess the error doesn't originate from the flux corrections image

vmahat avatar Jan 19 '24 17:01 vmahat

Yes, looks like the error is there in initial calibration. Could be that the sky model used for 3C123 in init cal is just not good enough...

mhardcastle avatar Jan 19 '24 17:01 mhardcastle

Is this an external model or that from image_dirin_SSD.app.model.fits ? The latter looks sensible to me

vmahat avatar Jan 19 '24 18:01 vmahat

I meant the sky model used in initial calibration with LINC (TGSS model probably). We redo the DI calibration but there's only so much you can fix...

mhardcastle avatar Jan 22 '24 10:01 mhardcastle

There are a few things I want to try so will keep this open until they're resolved

vmahat avatar Jan 22 '24 16:01 vmahat

I think this LINC change that is coming soon (https://git.astron.nl/RD/LINC/-/merge_requests/175) might be pretty good for this field. Alternatively supplying your own model to LINC.

twshimwell avatar Jan 22 '24 17:01 twshimwell

Ah yes, this came about due to some improvements in the long baseline imaging with that fix. I'll have a go with that when it's merged

vmahat avatar Jan 23 '24 18:01 vmahat

Small update on this: hearing some rumours on the problems of demixing TauA, I processed the data through LINC without demixing TauA, and the resulting image_dirin_SSD_m.app.restored.fits looks appropriate (left) vs with demixing (right). image There are still artefacts/ripples across the field, that are probably due to the presence of TauA so this isn't the optimal solution -- given I'm interested in the long baselines I think this is okay for now.

vmahat avatar Feb 05 '24 16:02 vmahat

@vmahat There a simple option now in kMS to subtract all ATeam(+Sun) sources in one go, would you be willing to give it a try and compare?

cyriltasse avatar Feb 05 '24 18:02 cyriltasse

Sure, how does one specify this in the config?

vmahat avatar Feb 06 '24 10:02 vmahat