Replies: 3 comments 6 replies
-
The reasoning tokens are discarded from context for the end user to prevent poisoning the context but the tokens that make the chain are certainly saved and used to train. |
Beta Was this translation helpful? Give feedback.
-
We seem to have counter claims of multi turn or single shot. |
Beta Was this translation helpful? Give feedback.
-
It doesn´t seem possible to me for it to be a single shot, how would they limit the amount of compute if so? Regarding this post, the image looks weird because it seems unlikely to me that the model would be able to catch itself in it´s errors if it only saw the ouptut and not it´s reasoning |
Beta Was this translation helpful? Give feedback.
-
https://x.com/IntuitMachine/status/1835256547179413672
Beta Was this translation helpful? Give feedback.
All reactions