- I wonder if the "move" action is difficult for the model to learn to use well. The model sees token location as positional encodings in the embedding, not sparse character offsets. Would be interesting to see something more like "jump to next/previous [token or set of tokens]". Or maybe a find/replace like most coding harness edit tools use?
- I'd move the exact training data generation details to an appendix. Could be summarized to improve the flow of the paper.
My goal is to submit the paper to a conference such as ACL, EMNLP, ICML, or a similar venue, and I would really appreciate technical feedback on things like:
- Boldness/strength of the claims - Weaknesses - Quality of the results, or if I should include other results
Paper: https://github.com/Sean-Diab/Reviser/blob/main/main.pdf
I would really value any feedback on what I should improve before submitting.
I am also looking for an arXiv endorsement for cs.CL. If anyone here is eligible and feels comfortable helping, my endorsement link is: https://arxiv.org/auth/endorse?x=ISRSI8
Thank you very much.