Adding thinking effort to the deduplication logic#276
Conversation
|
Codex usage limits have been reached for code reviews. Please check with the admins of this repo to increase the limits by adding credits. |
|
|
There was a problem hiding this comment.
Code Review
This pull request introduces a thinking-aware deduplication hash to ensure that changes in thinking configuration (effort, timeout, and metric) correctly partition the server-side fit cache. The fit method was refactored to resolve these configurations earlier and apply the new hashing logic. Feedback suggests improving the stability of the hash by casting the timeout to a float and using consistent key names for the wire format. Additionally, a redundant check in the configuration resolution logic was identified.
|
Thanks, @eliott-kalfon ! Using the following reproducer on your branch: I'm getting Did you see the same during your debugging? |
|
Thanks again for the great catch! It turns out this issue is better solved in the server-side change detection logic, which I've just shipped out. So closing the client PR. Happy if you can retest though if you have your scripts handy :D |
Hi @ggprior, tested with my scripts again, works for me! Thank you |
Change Description
Including the thinking effort in the cache deduplication logic. So that users get different results after generating predictions for medium effort, and then trying high effort.
If you used new dependencies: Did you add them to
requirements.txt? No dependency addedBreaking changes
No breaking changes.
If you made any breaking changes, please update the version number.
Breaking changes are totally fine, we just need to make sure to keep the users informed and the server in sync.
Does this PR break the API? If so, what is the corresponding server commit?
Does this PR break the user interface? If so, why?
Please do not mark comments/conversations as resolved unless you are the assigned reviewer. This helps maintain clarity during the review process.