Hi.
From your preprint, Nexus achieved 100% accuracy on non-self-verifying VerilogEval V1 Human (≒ VerilogEval V2).
I'd like to reproduce the result, but the only script I found is eda.py, which I believe was not used for the evaluation described in the paper.
Could you share the full evaluation setup for VerilogEval?