Towards Infinite LLM Context Windows

It all started with GPT having an input context window of 512 tokens. After only 5 years the newest LLMs are capable of handling 1M+ context inputs. Where’s the limit?

Krzysztof K. Zdeb
Towards Data Science
9 min readApr 28, 2024

--

I like to think of the LLMs (specifically, of the models’ parameters, i.e., their weights of the neural network layers and…

--

--

Chief Delivery Officer @ BitPeak, Data Science and Data Engineering geek 🤓