-
-
Notifications
You must be signed in to change notification settings - Fork 38
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[REVIEW]: RSMTool: a Python Package for facilitating research on automated scoring models #33
Comments
/ cc @openjournals/joss-reviewers - would anyone be willing to review this submission? If you would like to review this submission then please comment on this thread so that others know you're doing a review (so as not to duplicate effort). Something as simple as Reviewer instructions
Any questions, please ask for help by commenting on this issue! 🚀 |
maybe a tentative ✋ I am reviewing this? If nobody else speaks up this week, I will take it on to do a review next week (July 5-8). Full disclosure: Nitin (the submitting author) and I are friendly (and see each other, mostly at conferences, about every 18 months). I don't know anything about this specific work of his. |
Thanks, Jeremy!
|
OK that sounds great @jkahn 👍 |
Discovered missing when comparing to output of the `lice` tool. part of JOSS review: openjournals/joss-reviews#33
General checks and referencesAll the outer formatting looks good. Might be nice to find a DOI for Loukina 2015 but I can't figure out how to get DOIs from ACLweb. |
Statement of needI'm not 100% clear what the statement of actual need is here. I get that this is a useful tool within the ETS, but I'm not certain about what the contribution of this package is, and who the researchers (outside of the ETS itself) would be that would find this to be a useful tool in hand. I'm not saying that those researchers don't exist -- I might even be one of them -- but I don't think the statement of need clearly reflects what a imaginary researcher Professor X would use these tools for. Perhaps a tutorial walking through Professor X's thought process would help clarify what problem this solves -- and if the tutorial was part of the documentation, so much the better. I suspect is meeting >1 need and some researchers may have only a subset of those needs, and that is okay, but those are not clear. (See upcoming comment about multiply-situated work.) |
Undocumented entry points confuse new usersThere are at least three different command-line tools (endpoints, in the Python jargon) that all seem to take the same argument structure (a config file) but presumably have different formats expected in the config files. There's exactly one example use, scraped from a Kaggle competition, but it only uses one of the CLI endpoints ( This doesn't help your statement of need much, either. |
Please see the "available documentation" section in the main README. All There's only example, that's true. That's because we expect the "rsmtool"
|
Separating the
|
Okay, some more digging around in Is this a tool for new feature development? For comparing human raters? for comparing human raters to existing features? for comparing existing features to each other? For designing new notebooks that have lots of the existing work already done? All of the above and more? I think this is a configuration-based approach to desktop evaluation of how different schemes for combining numeric features improve (or hurt) the correlation with human scorers, but as such it's practically an IDE, which is why I am suggesting a clearer breakdown of the sub-responsibilities. |
MiscellaneousThe authors make no particular performance claims, and the software runs in reasonable time (a few seconds) on the sample data and produces plausible-looking HTML documentation of correlations among users and features. I'm happy to check off the corresponding boxes there. |
Recommendation: accept - Minor revisions.
Separately, I have a few further quibbles that should not block publication but should probably be
|
Thank you for such a detailed review, Jeremy! We'll go over your suggestions with Nitin. |
Indeed. Thanks for the careful review, Jeremy!
|
💯 - yes thanks for this excellent review @jkahn. @aloukina & @desilinguist - please let me know when you've had a chance to update your submission based upon @jkahn's feedback. |
Thanks again for the very useful review! It helped us come up with something a lot better, we think. We have just released v5.1.0 of RSMTool that addresses the suggestions made in the review. Specifically:
|
Well, I am delighted to see this. @desilinguist and @aloukina, the new documentation is actually enjoyable to read, with well thought out hyperlinking and walkthroughs describing real user scenarios. It's much less of a stretch for me to imagine a non-ETS researcher using this tool now, which was the unarticulated heart of my documentation objections before. I'm glad to hear you're exploring pip/wheel installations and I hope you'll publish wheels or sdist tarballs on PyPi periodically as part of your release cycle. I give this a 👍 and defer to the editors as to when/if/how you should mint a new DOI. |
Excellent, thanks @jkahn. @desilinguist - is there an associated DOI for the |
Hi @arfon, here is the DOI for the Please let me know if that's okay or if I need to do something else. |
Perfect. Thanks! @desilinguist - your paper is now accepted into JOSS and you DOI is http://dx.doi.org/10.21105/joss.00033 🎉 🚀 💥 Thanks for the great review @jkahn |
Submitting author: @desilinguist (Nitin Madnani)
Repository: https://github.com/EducationalTestingService/rsmtool
Version: v5.1.0
Editor: @arfon
Reviewer: @jkahn
Archive: 10.5281/zenodo.58851
Status
Status badge code:
Reviewer questions
Conflict of interest
General checks
Functionality
Documentation
Software paper
Paper PDF: 10.21105.joss.00033.pdf
paper.md
file include a list of authors with their affiliations?The text was updated successfully, but these errors were encountered: