Discussion about this post

User's avatar
koiredoun's avatar

It’s great that someone finally studied this. The findings don’t surprise me, though. If the base model was trained purely to predict the next token given the previous ones, without any deeper scientific grounding, it makes sense that the last part of the input, the one immediately before the model starts generating, would have a disproportionately strong influence on the response. This effect seems inherent to next-token prediction models that use a decoder-only architecture. I suspect that diffusion-based language models could mitigate this issue to some extent, though they likely won’t eliminate it entirely.

Expand full comment
AI Must Die's avatar

this result is obvious to anyone who knows anything about language models. studies like these are conducted in bad faith to further legitimize the technology for these use cases under the (wrong) assumption that fundamental limitations can be satisfactorily corrected with further investment

Expand full comment

No posts