software-submission
software-submission copied to clipboard
ncompare
Submitting Author: Daniel Kaufman (@danielfromearth)
All current maintainers: (@danielfromearth)
Package Name: ncompare
One-Line Description of Package: ncompare
compares two netCDF files at the command line, by generating a report of the matching and non-matching groups, variables, and attributes.
Repository Link: https://github.com/nasa/ncompare
Version submitted: 1.4.0
Editor: @tomalrussell
Reviewer 1: @cmarmo
Reviewer 2: @cmtso
Archive: 10.5281/zenodo.10625407
JOSS DOI: TBD
Version accepted: 1.7.2
Date accepted (month/day/year): 02/06/2024
Code of Conduct & Commitment to Maintain Package
- [x] I agree to abide by pyOpenSci's Code of Conduct during the review process and in maintaining my package after should it be accepted.
- [x] I have read and will commit to package maintenance after the review as per the pyOpenSci Policies Guidelines.
Description
This tool ("ncompare") compares the structure of two Network Common Data Form (NetCDF) files at the command line. It facilitates rapid, human-readable, multi-file evaluation by generating a formatted display of the matching and non-matching groups, variables, and associated metadata between two NetCDF datasets. The user has the option to colorize the terminal output for ease of viewing. As an option, ncompare
can save comparison reports in text and/or comma-separated value (CSV) formats.
Scope
-
Please indicate which category or categories. Check out our package scope page to learn more about our scope. (If you are unsure of which category you fit, we suggest you make a pre-submission inquiry):
- [ ] Data retrieval
- [ ] Data extraction
- [ ] Data processing/munging
- [ ] Data deposition
- [x] Data validation and testing
- [ ] Data visualization[^1]
- [ ] Workflow automation
- [ ] Citation management and bibliometrics
- [ ] Scientific software wrappers
- [ ] Database interoperability
Domain Specific & Community Partnerships
- [ ] Geospatial
- [ ] Education
- [ ] Pangeo
Community Partnerships
If your package is associated with an existing community please check below:
- [ ] Pangeo
- [ ] My package adheres to the Pangeo standards listed in the pyOpenSci peer review guidebook
[^1]: Please fill out a pre-submission inquiry before submitting a data visualization package.
-
For all submissions, explain how the and why the package falls under the categories you indicated above. In your explanation, please address the following points (briefly, 1-2 sentences for each):
- Who is the target audience and what are scientific applications of this package?
The target audience is anyone who manages the generation, manipulation, or validation of netCDF files. This package can be applied to to these netCDF file tasks in any scientific discipline; although it would be most relevant to applications with large multidimensional datasets, e.g., for comparing climate models, for Earth science data reanalyses, and for remote sensing data.
- Are there other Python packages that accomplish the same thing? If so, how does yours differ?
The ncdiff
function in the nco
(netCDF Operators) library, as well as ncmpidiff
and nccmp
, compute value
differences, but --- as far as we are aware --- do not have a dedicated function to show structural differences between netCDF4 datasets. Our package, ncompare
provides a light-weight Python-based tool for rapid visual comparisons of group & variable structures, attributes, and chunking.
- If you made a pre-submission enquiry, please paste the link to the corresponding issue, forum post, or other discussion, or
@tag
the editor you contacted:
Pre-submission inquiry #142
Technical checks
For details about the pyOpenSci packaging requirements, see our packaging guide. Confirm each of the following by checking the box. This package:
- [x] does not violate the Terms of Service of any service it interacts with.
- [x] uses an OSI approved license.
- [x] contains a README with instructions for installing the development version.
- [x] includes documentation with examples for all functions.
- [x] contains a tutorial with examples of its essential functions and uses.
- [x] has a test suite.
- [x] has continuous integration setup, such as GitHub Actions CircleCI, and/or others.
Publication Options
- [x] Do you wish to automatically submit to the Journal of Open Source Software? If so:
JOSS Checks
- [x] The package has an obvious research application according to JOSS's definition in their submission requirements. Be aware that completing the pyOpenSci review process does not guarantee acceptance to JOSS. Be sure to read their submission requirements (linked above) if you are interested in submitting to JOSS.
- [x] The package is not a "minor utility" as defined by JOSS's submission requirements: "Minor ‘utility’ packages, including ‘thin’ API clients, are not acceptable." pyOpenSci welcomes these packages under "Data Retrieval", but JOSS has slightly different criteria.
- [(NOT YET)] The package contains a
paper.md
matching JOSS's requirements with a high-level description in the package root or ininst/
. - [(NOT YET)] The package is deposited in a long-term repository with the DOI:
Note: JOSS accepts our review as theirs. You will NOT need to go through another full review. JOSS will only review your paper.md file. Be sure to link to this pyOpenSci issue when a JOSS issue is opened for your package. Also be sure to tell the JOSS editor that this is a pyOpenSci reviewed package once you reach this step.
Are you OK with Reviewers Submitting Issues and/or pull requests to your Repo Directly?
This option will allow reviewers to open smaller issues that can then be linked to PR's rather than submitting a more dense text based review. It will also allow you to demonstrate addressing the issue via PR links.
- [x] Yes I am OK with reviewers submitting requested changes as issues to my repo. Reviewers will then link to the issues in their submitted review.
Confirm each of the following by checking the box.
- [x] I have read the author guide.
- [x] I expect to maintain this package for at least 2 years and can help find a replacement for the maintainer (team) if needed.
Please fill out our survey
- [x] Last but not least please fill out our pre-review survey. This helps us track submission and improve our peer review process. We will also ask our reviewers and editors to fill this out.
P.S. Have feedback/comments about our review process? Leave a comment here
Editor and Review Templates
Thanks @danielfromearth for making the full submission!
Here are the initial editor-in-chief checks.
Almost everything we need to start the review is there, but there's one I don't see.
- [ ] User-facing documentation that overviews how to install and start using the package.
I think you can add this at the top of the single page docs you have now? Basically repeating what you have in the README.
Please let me know if you can add that.
We're ready to find an editor and reviewers once you do.
Editor in Chief checks
- [x] Installation The package can be installed from a community repository such as PyPI (preferred), and/or a community channel on conda (e.g. conda-forge, bioconda).
- [x] The package imports properly into a standard Python environment
import package
.
- [x] The package imports properly into a standard Python environment
- [x] Fit The package meets criteria for fit and overlap.
- [ ] Documentation The package has sufficient online documentation to allow us to evaluate package function and scope without installing the package. This includes:
- [ ] User-facing documentation that overviews how to install and start using the package.
- [x] Short tutorials that help a user understand how to use the package and what it can do for them.
- [x] API documentation (documentation for your code's functions, classes, methods and attributes): this includes clearly written docstrings with variables defined using a standard docstring format.
- [x] Core GitHub repository Files
- [x] README The package has a
README.md
file with clear explanation of what the package does, instructions on how to install it, and a link to development instructions. - [x] Contributing File The package has a
CONTRIBUTING.md
file that details how to install and contribute to the package. - [x] Code of Conduct The package has a
Code_of_Conduct.md
file. - [x] License The package has an OSI approved license. NOTE: We prefer that you have development instructions in your documentation too.
- [x] README The package has a
- [x] Issue Submission Documentation All of the information is filled out in the
YAML
header of the issue (located at the top of the issue template). - [x] Automated tests Package has a testing suite and is tested via a Continuous Integration service.
- [x] Repository The repository link resolves correctly.
- [x] Package overlap The package doesn't entirely overlap with the functionality of other packages that have already been submitted to pyOpenSci.
- [ ] Archive (JOSS only, may be post-review): The repository DOI resolves correctly.
- [ ] Version (JOSS only, may be post-review): Does the release version given match the GitHub release (v1.0.0)?
- [x] Initial onboarding survey was filled out We appreciate each maintainer of the package filling out this survey individually. :raised_hands: Thank you authors in advance for setting aside five to ten minutes to do this. It truly helps our organization. :raised_hands:
Editor comments
- The package imports but there is nothing in the
ncompare
namespace because nothing is imported in ncompare/init.py. I understand the package to be designed principally for use through the command-line, so this doesn't affect use
👋 Hi @NickleDave, thanks for the info!
I've now updated the GitHub Pages to include the installation and basic usage section from the README.md
.
Great, thank you @danielfromearth -- sorry I didn't reply sooner.
Just letting you know that @tomalrussell has graciously volunteered to be the editor for this review. We are now seeking reviewers and @tomalrussell will reply here with next steps once we have found them.
Hi @danielfromearth, thanks for the package submission - we've now identified reviewers and can move on to the next steps..
👋🏻 welcome, @cmarmo and @cmtso! Thank you for volunteering to review for pyOpenSci ☺️ The review process is outlined in the guides below, and I'll aim to check in occasionally over the next weeks as we go.
Please fill out our pre-review survey
Before beginning your review, please fill out our pre-review survey. This helps us improve all aspects of our review and better understand our community. No personal data will be shared from this survey - it will only be used in an aggregated format by our Executive Director to improve our processes and programs.
- [x] reviewer 1 survey completed.
- [x] reviewer 2 survey completed.
The following resources will help you complete your review:
- Here is the reviewers guide. This guide contains all of the steps and information needed to complete your review.
- Here is the review template that you will need to fill out and submit here as a comment, once your review is complete.
Please get in touch with any questions or concerns! Your review is due in ~3 weeks: 13 December 2023
Reviewers: @cmarmo @cmtso Due date: 13 December 2023
Hi @cmarmo , @cmtso just checking in - thanks for filling out the pre-review surveys! Are you both on track to post reviews next week? Do raise any questions here or drop me a line.
@tomalrussell , @danielfromearth , sorry for the silence! I'm glad to help with this review, and yes I'm going to take care of the review during the week! I had some commitments to honor with closer deadlines, but I'm available now.
Hi @tomalrussell, it's been a busy two weeks. But I should be able to make the deadline next week.
Package Review
Please check off boxes as applicable, and elaborate in comments below. Your review is not limited to these topics, as described in the reviewer guide
- [x] As the reviewer I confirm that there are no conflicts of interest for me to review this work (If you are unsure whether you are in conflict, please speak to your editor before starting your review).
Documentation
The package includes all the following forms of documentation:
- [ ] (partial) A statement of need clearly stating problems the software is designed to solve and its target audience in README.
- [x] Installation instructions: for the development version of the package and any non-standard dependencies in README.
- [x] Vignette(s) demonstrating major functionality that runs successfully locally.
- [x] Function Documentation: for all user-facing functions.
- [x] Examples for all user-facing functions.
- [ ] Community guidelines including contribution guidelines in the README or CONTRIBUTING.
- [ ] Metadata including author(s), author e-mail(s), a url, and any other relevant metadata e.g., in a
pyproject.toml
file or elsewhere.
Readme file requirements The package meets the readme requirements below:
- [x] Package has a README.md file in the root directory.
The README should include, from top to bottom:
- [x] The package name
- [x] Badges for:
- [ ] Continuous integration and test coverage,
- [ ] Docs building (if you have a documentation website),
- [ ] A repostatus.org badge,
- [ ] Python versions supported,
- [x] Current package version (on PyPI / Conda).
NOTE: If the README has many more badges, you might want to consider using a table for badges: see this example. Such a table should be more wide than high. (Note that the a badge for pyOpenSci peer-review will be provided upon acceptance.)
- [x] Short description of package goals.
- [x] Package installation instructions
- [x] Any additional setup required to use the package (authentication tokens, etc.)
- [x] Descriptive links to all vignettes. If the package is small, there may only be a need for one vignette which could be placed in the README.md file.
- [x] Brief demonstration of package usage (as it makes sense - links to vignettes could also suffice here if package description is clear)
- [x] Link to your documentation website.
- [x] If applicable, how the package compares to other similar packages and/or how it relates to other packages in the scientific ecosystem.
- [ ] Citation information
Usability
Reviewers are encouraged to submit suggestions (or pull requests) that will improve the usability of the package as a whole. Package structure should follow general community best-practices. In general please consider whether:
- [x] Package documentation is clear and easy to find and use.
- [x] The need for the package is clear
- [x] All functions have documentation and associated examples for use
- [x] The package is easy to install
Functionality
- [x] Installation: Installation succeeds as documented.
- [x] Functionality: Any functional claims of the software been confirmed.
- [x] Performance: Any performance claims of the software been confirmed.
- [ ] Automated tests:
- [ ] All tests pass on the reviewer's local machine for the package version submitted by the author. Ideally this should be a tagged version making it easy for reviewers to install.
- [ ] Tests cover essential functions of the package and a reasonable range of inputs and conditions.
- [ ] Continuous Integration: Has continuous integration setup (We suggest using Github actions but any CI platform is acceptable for review)
- [x] Packaging guidelines: The package conforms to the pyOpenSci packaging guidelines.
A few notable highlights to look at:
- [x] Package supports modern versions of Python and not End of life versions.
- [ ] Code format is standard throughout package and follows PEP 8 guidelines (CI tests for linting pass)
For packages also submitting to JOSS
- [ ] The package has an obvious research application according to JOSS's definition in their submission requirements.
Note: Be sure to check this carefully, as JOSS's submission requirements and scope differ from pyOpenSci's in terms of what types of packages are accepted.
The package contains a paper.md
matching JOSS's requirements with:
- [ ] A short summary describing the high-level functionality of the software
- [ ] Authors: A list of authors with their affiliations
- [ ] A statement of need clearly stating problems the software is designed to solve and its target audience.
- [ ] References: With DOIs for all those that have one (e.g. papers, datasets, software).
Final approval (post-review)
- [ ] The author has responded to my review and made changes to my satisfaction. I recommend approving this package.
Estimated hours spent reviewing: 3
Review Comments
The package looks very good: package structure, documentation, and tests looks very good. It is a simple-to-use command-line tool based on python to compare the structure of two netCDF file. Note this tool only examines the structure of the files--so if the two files have different values but have the same fields/structure, this tool will not tell them apart.
I have tested the package with all the example tests provided, as well as some of my own data. However, there is no CI in place at the moment and may be something to work on in the future.
Hello everybody!
Thanks for giving me the opportunity to review ncompare
!
Please find my comments following.
Package Review
Please check off boxes as applicable, and elaborate in comments below. Your review is not limited to these topics, as described in the reviewer guide
- [x] As the reviewer I confirm that there are no conflicts of interest for me to review this work (If you are unsure whether you are in conflict, please speak to your editor before starting your review).
Documentation
The package includes all the following forms of documentation:
- [x] A statement of need clearly stating problems the software is designed to solve and its target audience in README.
- [x] Installation instructions: for the development version of the package and any non-standard dependencies in README.
- [x] Vignette(s) demonstrating major functionality that runs successfully locally: note that only
ncompare
is exemplified, as the user is not meant to use low-level functions. This comment is applicable to any other documentation reference. - [x] Function Documentation: for all user-facing functions.
- [x] Examples for all user-facing functions.
- [x] Community guidelines including contribution guidelines in the README or CONTRIBUTING.
- [x] Metadata including author(s), author e-mail(s), a url, and any other relevant metadata e.g., in a
pyproject.toml
file or elsewhere.
Readme file requirements The package meets the readme requirements below:
- [x] Package has a README.md file in the root directory.
The README should include, from top to bottom:
- [x] The package name
- [ ] Badges for:
- [ ] Continuous integration and test coverage,
- [ ] Docs building (if you have a documentation website),
- [ ] A repostatus.org badge,
- [ ] Python versions supported,
- [x] Current package version (on PyPI / Conda).
NOTE: If the README has many more badges, you might want to consider using a table for badges: see this example. Such a table should be more wide than high. (Note that the a badge for pyOpenSci peer-review will be provided upon acceptance.)
- [x] Short description of package goals.
- [x] Package installation instructions
- [x] Any additional setup required to use the package (authentication tokens, etc.)
- [x] Descriptive links to all vignettes. If the package is small, there may only be a need for one vignette which could be placed in the README.md file.
- [x] Brief demonstration of package usage (as it makes sense - links to vignettes could also suffice here if package description is clear)
- [ ] Link to your documentation website.
- [x] If applicable, how the package compares to other similar packages and/or how it relates to other packages in the scientific ecosystem.
- [ ] Citation information
Usability
Reviewers are encouraged to submit suggestions (or pull requests) that will improve the usability of the package as a whole. Package structure should follow general community best-practices. In general please consider whether:
- [x] Package documentation is clear and easy to find and use.
- [x] The need for the package is clear
- [x] All functions have documentation and associated examples for use
- [x] The package is easy to install
Functionality
- [x] Installation: Installation succeeds as documented.
- [x] Functionality: Any functional claims of the software been confirmed.
- [ ] Performance: Any performance claims of the software been confirmed (Does not apply).
- [ ] Automated tests:
- [x] All tests pass on the reviewer's local machine for the package version submitted by the author. Ideally this should be a tagged version making it easy for reviewers to install.
- [ ] Tests cover essential functions of the package and a reasonable range of inputs and conditions:
ncompare/core.py
andncompare/printing.py
may benefit of an improved test coverage, both files have a coverage less than 75% and they are the core of the application.
- [x] Continuous Integration: Has continuous integration setup (We suggest using Github actions but any CI platform is acceptable for review): there is no integrated build for the documentation though.
- [x] Packaging guidelines: The package conforms to the pyOpenSci packaging guidelines.
A few notable highlights to look at:
- [x] Package supports modern versions of Python and not End of life versions: this is claimed in
pyproject.toml
, but the CI is run against 3.9 only. - [x] Code format is standard throughout package and follows PEP 8 guidelines (CI tests for linting pass)
- [x] Package supports modern versions of Python and not End of life versions: this is claimed in
For packages also submitting to JOSS
- [ ] The package has an obvious research application according to JOSS's definition in their submission requirements.
Note: Be sure to check this carefully, as JOSS's submission requirements and scope differ from pyOpenSci's in terms of what types of packages are accepted.
The package contains a paper.md
matching JOSS's requirements with:
- [ ] A short summary describing the high-level functionality of the software
- [ ] Authors: A list of authors with their affiliations
- [ ] A statement of need clearly stating problems the software is designed to solve and its target audience.
- [ ] References: With DOIs for all those that have one (e.g. papers, datasets, software).
Final approval (post-review)
- [ ] The author has responded to my review and made changes to my satisfaction. I recommend approving this package.
Estimated hours spent reviewing: 2h
Review Comments
Below a summary of my comments.
- As
ncompare
is a shell command line it would be informative to have a--version
option printing the current version. - I am not used to poetry, the tests can indeed run with
poetry run pytest tests
butpytest tests
from a pythonvenv
works too, it could be nice to have this clarified in theREADME
as both installations are previously described. - As python 3.12 is out it could be interesting to have CI for the last python version.
- The license is partially available in the README and as pdf in the
license
directory: as this is a specific NASA license and probably less known by the users, I suggest to have its name specified in the README perhaps with its link at the OSI website, for better discoverability. - I was able to run
ncompare
on the test data shipped with the package and other files, but the notebook inncompare/example
(which is also used in the documentation) uses two files which I was unable to download (permission denied): I strongly recommend to build example with public data, so they can be reproducible. - I cannot find a documentation build in the CI, am I wrong? This will make easier having consistent notebooks and example for the users.
- The documentation only documents
ncompare
, which is fine because this is the main utility: the CONTRIBUTING file suggests "functions should contain a docstring, though short or trivial function may contain a 1-line docstring", it is unclear to me (as a possible future contributor) how to make the difference between trivial and non-trivial functions, perhaps private and public can be a more consistent definition?
Hi folks, a quick update: I'm wrapping up some final tweaks to respond to the review comments, and will also hopefully be able to add a JOSS paper to the repo next week.
My responses to the comments and blank check-boxes from @cmtso's review can be found next to the green checkmarks (✅) below:
Documentation
The package includes all the following forms of documentation:
- [ ] (partial) A statement of need clearly stating problems the software is designed to solve and its target audience in README.
- [ ] Community guidelines including contribution guidelines in the README or CONTRIBUTING.
- [ ] Metadata including author(s), author e-mail(s), a url, and any other relevant metadata e.g., in a
pyproject.toml
file or elsewhere.
- ✅ A statement of need can now be found as the "Why ncompare?" section in the
README.md
- ✅ Community guidelines are now present in the
CONTRIBUTING.md
file. - ✅ pyproject.toml now includes author, author email, and "repository" url.
The README should include, from top to bottom:
[x] Badges for:
- [ ] Continuous integration and test coverage,
- [ ] Docs building (if you have a documentation website),
- [ ] A repostatus.org badge,
- [ ] Python versions supported,
- ✅ Badges for CI and test coverage, docs, repostatus, and python versions have been added to the
README.md
- [ ] Citation information
- ✅ This has been added here:
CITATION.cff
Functionality
[ ] Automated tests:
- [ ] All tests pass on the reviewer's local machine for the package version submitted by the author. Ideally this should be a tagged version making it easy for reviewers to install.
- [ ] Tests cover essential functions of the package and a reasonable range of inputs and conditions.
[ ] Continuous Integration: Has continuous integration setup (We suggest using Github actions but any CI platform is acceptable for review)
- ✅ Automated tests are part of the main Continuous Integration workflow defined here, and these tests are run as part of GitHub Actions shown here. The tests cover essential functions, and test coverage can now be seen via Codecov.
[x] Packaging guidelines: The package conforms to the pyOpenSci packaging guidelines. A few notable highlights to look at:
- [x] Package supports modern versions of Python and not End of life versions.
- [ ] Code format is standard throughout package and follows PEP 8 guidelines (CI tests for linting pass)
- ✅ The code format is further standardized and uses
ruff
for linting, as defined here in the CI workflow.
For packages also submitting to JOSS
- [ ] The package has an obvious research application according to JOSS's definition in their submission requirements.
Note: Be sure to check this carefully, as JOSS's submission requirements and scope differ from pyOpenSci's in terms of what types of packages are accepted.
The package contains a
paper.md
matching JOSS's requirements with:
- [ ] A short summary describing the high-level functionality of the software
- [ ] Authors: A list of authors with their affiliations
- [ ] A statement of need clearly stating problems the software is designed to solve and its target audience.
- [ ] References: With DOIs for all those that have one (e.g. papers, datasets, software).
- ✅ The JOSS manuscript with those sections has been added to the
joss-paper
branch in thispaper.md
file. A pdf of this initial manuscript upload can be found as this build artifact.
Estimated hours spent reviewing:
3
Review Comments
The package looks very good: package structure, documentation, and tests looks very good. It is a simple-to-use command-line tool based on python to compare the structure of two netCDF file. Note this tool only examines the structure of the files--so if the two files have different values but have the same fields/structure, this tool will not tell them apart.
I have tested the package with all the example tests provided, as well as some of my own data. However, there is no CI in place at the moment and may be something to work on in the future.
- ✅ Thank you very much @cmtso for your comments and the time and attention you devoted to this review!
My responses to the comments and blank check-boxes from @cmarmo's review can be found next to the green checkmarks (✅) below:
The README should include, from top to bottom:
[ ] Badges for:
- [ ] Continuous integration and test coverage,
- [ ] Docs building (if you have a documentation website),
- [ ] A repostatus.org badge,
- [ ] Python versions supported,
- [x] Current package version (on PyPI / Conda).
- ✅ Badges for CI and test coverage, docs, repostatus, and python versions have been added to the
README.md
- [ ] Link to your documentation website.
- ✅ A link is now in the
docs: passing
badge.
- [ ] Citation information
- ✅ This has been added here:
CITATION.cff
Functionality
- [ ] Automated tests:
- [ ] Tests cover essential functions of the package and a reasonable range of inputs and conditions:
ncompare/core.py
andncompare/printing.py
may benefit of an improved test coverage, both files have a coverage less than 75% and they are the core of the application.
- ✅ Automated tests are part of the main Continuous Integration workflow defined here, and these tests are run as part of GitHub Actions shown here. The tests cover essential functions, and the improved test coverage (
core.py
is now at 90% andprinting.py
at 87%) can now be seen via Codecov.
For packages also submitting to JOSS
- [ ] The package has an obvious research application according to JOSS's definition in their submission requirements.
Note: Be sure to check this carefully, as JOSS's submission requirements and scope differ from pyOpenSci's in terms of what types of packages are accepted.
The package contains a
paper.md
matching JOSS's requirements with:
- [ ] A short summary describing the high-level functionality of the software
- [ ] Authors: A list of authors with their affiliations
- [ ] A statement of need clearly stating problems the software is designed to solve and its target audience.
- [ ] References: With DOIs for all those that have one (e.g. papers, datasets, software).
- ✅ The JOSS manuscript with those sections has been added to the
joss-paper
branch in thispaper.md
file. A pdf of this initial manuscript upload can be found as this build artifact.
Estimated hours spent reviewing:
2h
Review Comments
Below a summary of my comments.
- As
ncompare
is a shell command line it would be informative to have a--version
option printing the current version.
- ✅ This option has now been added to the command line interface.
- I am not used to poetry, the tests can indeed run with
poetry run pytest tests
butpytest tests
from a pythonvenv
works too, it could be nice to have this clarified in theREADME
as both installations are previously described.
- ✅ This is a good suggestion, and it is now part of the README.
- As python 3.12 is out it could be interesting to have CI for the last python version.
- ✅ Done.
- The license is partially available in the README and as pdf in the
license
directory: as this is a specific NASA license and probably less known by the users, I suggest to have its name specified in the README perhaps with its link at the OSI website, for better discoverability.
- ✅ The suggested information has been added to the "Contributing" section of the README.md
- I was able to run
ncompare
on the test data shipped with the package and other files, but the notebook inncompare/example
(which is also used in the documentation) uses two files which I was unable to download (permission denied): I strongly recommend to build example with public data, so they can be reproducible.
- ✅ The notebook has been updated to use publicly accessible data files.
- I cannot find a documentation build in the CI, am I wrong? This will make easier having consistent notebooks and example for the users.
- ✅ Documentation has been changed from GitHub pages to use ReadTheDocs, and this is now part of the CI build.
- The documentation only documents
ncompare
, which is fine because this is the main utility: the CONTRIBUTING file suggests "functions should contain a docstring, though short or trivial function may contain a 1-line docstring", it is unclear to me (as a possible future contributor) how to make the difference between trivial and non-trivial functions, perhaps private and public can be a more consistent definition?
-
✅ I understand how this may lead to some uncertainty, although I don't think complicated private functions should have 1-line docstrings either. There will unavoidably be some subjectivity. Therefore, this description has been modified to "...private functions (that are exceptionally short)...".
-
✅ Thank you very much @cmarmo for these additional comments and the time and attention you devoted to reviewing
ncompare
!
Hi @tomalrussell, thanks for shepherding and providing editorial review of this submission thus far.
We have made changes to the repository to address reviewer recommendations, and responded directly to the reviewers' comments above. In addition, please notice that we have now included a JOSS manuscript in the repository (in the joss-paper
branch). Cheers.
Thanks for the super clear responses @danielfromearth !
@cmarmo and @cmtso - over to you to respond to the changes. Discussion is encouraged, we're in the "review follow-up" phase.
Could you pay particular attention to the For packages also submitting to JOSS
section, now that the manuscript is up on the repository (paper.md
and paper.bib
)? It may be helpful to refer to the JOSS guidelines for more detail and background on the expectations here.
As ever, do ask or drop me a line if you need anything ☺️
Responses due in ~2 weeks: 31 January.
Hello @danielfromearth , thank you for all your work and your answers to my comments! I have checked all your improvements and I have still some comments about the documentation:
- in the
README.md
the link to the license inside the package is not rendered (I cannot understand why at first sight... ) - In the
README.md
and the documentation the link to the example jupyter notebook gives a 404 error - The file https://gpm1.gesdisc.eosdis.nasa.gov/data/GPM_L3/GPM_3IMERGDL.06/2024/01/3B-DAY-L.MS.MRG.3IMERG.20240110-S000000-E235959.V06.nc4 used in the notebook still asks me for user and password.... am I missing something?
I'm going to check the paper for JOSS in the next week.... sorry for the two steps answer....
Hello @danielfromearth, I realized that you are merging your modifications in a develop
branch rather than in main
.
I see that the link to the notebook and to the example files are fixed: thank you!
It looks like the link to the local version of the license in the readme still contains spaces and it is still not rendered.
@tomalrussell after this nit fixing I consider the author has responded to my review and made changes to my satisfaction. I recommend approving this package.
I have also checked the paper for JOSS: I'm copying here the checklist related to JOSS.
For packages also submitting to JOSS
- [x] The package has an obvious research application according to JOSS's definition in their submission requirements: I believe
ncompare
can be considered as a tool to "extract knowledge from large data sets". However, note that JOSS states that "single-function packages are not acceptable." It is not clear to me how to interpret "single-function" and ifncompare
can be considered as one.
Note: Be sure to check this carefully, as JOSS's submission requirements and scope differ from pyOpenSci's in terms of what types of packages are accepted.
The package contains a paper.md
matching JOSS's requirements with:
- [x] A short summary describing the high-level functionality of the software
- [x] Authors: A list of authors with their affiliations: the affiliation of the author is missing
- [x] A statement of need clearly stating problems the software is designed to solve and its target audience.
- [x] References: With DOIs for all those that have one (e.g. papers, datasets, software).
For packages also submitting to JOSS
- [ ] The package has an obvious research application according to JOSS's definition in their submission requirements: I think this is a hugely useful tool to the research community. At the moment, it may be regarded as a minor 'utility' package. Perhaps a clearer roadmap of future feature extensions to cover a slightly larger (and more specific) scope and providing research use case in the paper will help with the JOSS submission.
Note: Be sure to check this carefully, as JOSS's submission requirements and scope differ from pyOpenSci's in terms of what types of packages are accepted.
The package contains a paper.md
matching JOSS's requirements with:
- [ *] A short summary describing the high-level functionality of the software
- [ *] Authors: A list of authors with their affiliations
- [ *] A statement of need clearly stating problems the software is designed to solve and its target audience.
- [ *] References: With DOIs for all those that have one (e.g. papers, datasets, software).
Final approval (post-review)
- [ *] The author has responded to my review and made changes to my satisfaction. I recommend approving this package.
With regard to the ncompare
repository:
@cmarmo — Thanks for taking another look through this software, and I see what you're talking about! For some reason the link to the license appears incorrect in the README's preview on the develop
branch homepage, but navigating to the README.md file itself displays it correctly (see here). I suspect that once it's merged into main
, then the README preview will display correctly — and I will make note of that here once it's merged.
@cmtso — Thank you for your additional effort looking back through this software!
With regard to the JOSS scope:
@cmarmo:
- [x] ...I believe ncompare can be considered as a tool to "extract knowledge from large data sets".
✅ Like @cmarmo, we also consider ncompare
as a tool to "extract knowledge from large data sets", but even more so as a tool that "supports the functioning of research instruments or the execution of research experiments" (from JOSS's what-we-mean-by-research-software). We believe this could be made clearer in the manuscript; therefore, we are planning revisions that will include more description of how ncompare
has been utilized for the preparations and evaluation of satellite mission data to research Earth's atmosphere.
@cmarmo and @cmtso, respectively:
...However, note that JOSS states that "single-function packages are not acceptable." It is not clear to me how to interpret "single-function" and if ncompare can be considered as one.
...At the moment, it may be regarded as a minor 'utility' package. Perhaps a clearer roadmap of future feature extensions to cover a slightly larger (and more specific) scope and providing research use case in the paper will help with the JOSS submission.
✅ Although we also do not know JOSS's criteria for "single-function", we do not consider ncompare
as a single-function package, because its command line interface includes several options for customizing the comparison and the generation of reports. Some of the feature roadmap under consideration is visible in the GitHub issues.
We note that ncompare
has been intentionally designed not to be a "toolbox" of netCDF functions and it is not in our future roadmap, as toolbox libraries already exists. Instead, ncompare
is designed to fill an open source software gap, with a focus on comparison, visualizing comparisons, and generating the resulting reports. Furthermore, we consider ncompare
to not be a utility package or thin API, but rather a fully fledged tool that is enabling research now, and its functions are extendable so as to be built further upon.
We also appreciate the suggestion that "providing research use case in the paper will help with the JOSS submission", and as mentioned above, we now plan to revise the manuscript to include descriptions of how ncompare
has been used in preparations and evaluations for an atmospheric science-focused satellite instrument.
@cmtso:
...I think this is a hugely useful tool to the research community.
✅ Thanks! We think so too :)
Note that because of the timeline of one of the satellite missions I refer to above, we may not be able to submit the revised manuscript to JOSS until April, when the mission data become public.
Hi @danielfromearth, thanks for the thorough response.
On the question of scope for JOSS - in the end this is up to the JOSS editorial team, but do refer to this pyOpenSci review (as for example in this recent JOSS submission openjournals/joss-reviews#5332). I think you have a strong case for ncompare
as enabling tool that extracts useful data from large datasets. Revising the manuscript as you describe sounds worthwhile. I'm happy to recommend it for submission when you're ready.
On the last few nits - they all look to be resolved on develop
, and merging and creating a release should be part of the wrap-up, so with that -
🎉 ncompare
has been approved by pyOpenSci! Thank you @danielfromearth for submitting ncompare
and many thanks again to @cmarmo and @cmtso for reviewing this package! 😸
Author Wrap Up Tasks
There are a few things left to do to wrap up this submission:
- [x] Activate Zenodo watching the repo if you haven't already done so.
- [x] Tag and create a release to create a Zenodo version and DOI.
- [x] Add the badge for pyOpenSci peer-review to the README.md of
ncompare
. The badge should be[](https://github.com/pyOpenSci/software-review/issues/146)
. - [x] Please fill out the post-review survey. All maintainers and reviewers should fill this out.
JOSS submission
Here are the next steps:
- [x] <Editor Task> Once the JOSS issue is opened for the package, we strongly suggest that you subscribe to issue updates. This will allow you to continue to update the issue labels on this review as it goes through the JOSS process.
- [x] Login to the JOSS website and fill out the JOSS submission form using your Zenodo DOI. When you fill out the form, be sure to mention and link to the approved pyOpenSci review. JOSS will tag your package for expedited review if it is already pyOpenSci approved.
- [x] Wait for a JOSS editor to approve the presubmission (which includes a scope check).
- [x] Once the package is approved by JOSS, you will be given instructions by JOSS about updating the citation information in your README file.
- [ ] When the JOSS review is complete, add a comment to your review in the pyOpenSci software-review repo here that it has been approved by JOSS. An editor will then add the JOSS-approved label to this issue.
Editor Final Checks
Please complete the final steps to wrap up this review. @tomalrussell reminder to do the following:
- [x] Make sure that the maintainers filled out the post-review survey
- [x] Invite the maintainers to submit a blog post highlighting their package. Feel free to use / adapt language found in this comment to help guide the author.
- [x] Change the status tag of the issue to
6/pyOS-approved6 🚀🚀🚀
. - [x] Invite the package maintainer(s) and both reviewers to slack if they wish to join.
- [ ] If the author submits to JOSS, please continue to update the labels for JOSS on this issue until the author is accepted (do not remove the
6/pyOS-approved
label). Once accepted add the label9/joss-approved
to the issue. Skip this check if the package is not submitted to JOSS. - [ ] If the package is JOSS-accepted please add the JOSS doi to the YAML at the top of the issue.
If you have any feedback for us about the review process please feel free to share it here. We are always looking to improve our process and documentation in the peer-review-guide.
Author Wrap Up Tasks
There are a few things left to do to wrap up this submission:
- [x] Activate Zenodo watching the repo if you haven't already done so.
- [x] Tag and create a release to create a Zenodo version and DOI.
- [x] Add the badge for pyOpenSci peer-review to the README.md of
ncompare
. The badge should be[](https://github.com/pyOpenSci/software-review/issues/146)
.- [x] Please fill out the post-review survey. All maintainers and reviewers should fill this out.
JOSS submission
Here are the next steps:
- [ ] Login to the JOSS website and fill out the JOSS submission form using your Zenodo DOI. When you fill out the form, be sure to mention and link to the approved pyOpenSci review. JOSS will tag your package for expedited review if it is already pyOpenSci approved.
- [ ] Once the JOSS issue is opened for the package, we strongly suggest that you subscribe to issue updates. This will allow you to continue to update the issue labels on this review as it goes through the JOSS process.
- [ ] Wait for a JOSS editor to approve the presubmission (which includes a scope check).
- [ ] Once the package is approved by JOSS, you will be given instructions by JOSS about updating the citation information in your README file.
- [ ] When the JOSS review is complete, add a comment to your review in the pyOpenSci software-review repo here that it has been approved by JOSS. An editor will then add the JOSS-approved label to this issue.
✅ ncompare
has been archived on Zenodo, with this DOI: 10.5281/zenodo.10625407
Ace! I'll keep an eye on this issue for any updates on your JOSS submission whenever you can make the next steps.
You are also invited to write a short blog post about ncompare
for pyOpenSci! If this sounds interesting, have a look at a couple of examples from movingpandas
and pandera
. There's a markdown example you could use to help draft a post. This is completely optional, but if you have time, we'd love to help promote your work.
hey team. i'm checking in on this issue. did this issue / package get fastraacked through joss? if it did we need to do a few followup steps!
- cross link this issue with the joss review issue. i see the pre-review but not the final review
- add the JOSS doi to the header
- update the label to say joss accepted (in addition to pyos accepted!)
many thanks y'all!!
Hey @lwasser thanks for checking in!
The package v1.9.0
has been accepted into JOSS, with review issue https://github.com/openjournals/joss-reviews/issues/6490
DOI updated in the issue header: https://doi.org/10.21105/joss.06490
Fantastic! Thank you @tomalrussell !! We can close it then I think as complete! Thank you all for supporting our open review process!!