Have you ever attempted to fill up one of those monster context windows up with useful context and then let the model try to do some useful task with all the information in it?
I have. Sometimes it works, but often it’s not pretty. Context window size is the new MHz, in terms of misleading performance measurements.
I find there comes a point where, even with a lot of context, the AI just hasn’t been trained to solve the problem. At that point it will cycle you round and round the same few wrong answers until you give up and work it out yourself.
Have you ever attempted to fill up one of those monster context windows up with useful context and then let the model try to do some useful task with all the information in it?
I have. Sometimes it works, but often it’s not pretty. Context window size is the new MHz, in terms of misleading performance measurements.
I find there comes a point where, even with a lot of context, the AI just hasn’t been trained to solve the problem. At that point it will cycle you round and round the same few wrong answers until you give up and work it out yourself.