Mastodon Feed: Post

Mastodon Feed

glyph ("Glyph") wrote:

I need a graph of context window growth vs. resources like inference compute / RAM over time. Has anyone already made this?

And does the model's size itself put limits on context windows, or is that purely a feature of the hardware you're doing inference on?