zlacker

[return to "We gave 5 LLMs $100K to trade stocks for 8 months"]
1. sethop+W[view] [source] 2025-12-04 23:13:11
>>cheese+(OP)
> Testing GPT-5, Claude, Gemini, Grok, and DeepSeek with $100K each over 8 months of backtested trading

So the results are meaningless - these LLMs have the advantage of foresight over historical data.

◧◩
2. itake+g1[view] [source] 2025-12-04 23:14:59
>>sethop+W
> We time segmented the APIs to make sure that the simulation isn’t leaking the future into the model’s context.

I wish they could explain what this actually means.

◧◩◪
3. nullbo+p3[view] [source] 2025-12-04 23:25:57
>>itake+g1
Overall, it does sound weird. On the one hand, assuming I properly I understand what they are saying is that they removed model's ability to cheat based on their specific training. And I do get that nuance ablation is a thing, but this is not what they are discussing there. They are only removing one avenue of the model to 'cheat'. For all we know, some that data may have been part of its training set already...
[go to top]