this paper deserves a deep breath and a slow exhaling “what the fuck”
who even talks about compression in OCR models?
who tries to spin an OCR model as a SOTA LLM? isn’t OCR a solved problem? what?
but oddly, i feel like they got something here, idk
they’re talking about unlimited context..
this paper deserves a deep breath and a slow exhaling “what the fuck”
View original threadi’ve always thought forgetting was a crucial missing piece, and they’ve presented a solution
idk, this paper doesn’t stand on its own
taken alone, there’s nothing here, for sure. but if you take it to where they’re hinting at, they might have just taken the lead
idk, this paper doesn’t stand on its own
taken alone, there’s nothing here, for sure. but if you take it to where they’re hinting at, they might have just taken the lead
11
oh. yeah. that.
13
i forget how much of, e.g. chemistry or even finance, is NOT text
15