This article looks like it might end up being relatively important. They seem to have found a way to both avoid the need for tokenization in LLMs and also avoid the small context windows that result in LLMs short memories. Andrej Karpathy has a post going into more detail
Favorites May 14th 2023
Favorites May 14th 2023
Favorites May 14th 2023
This article looks like it might end up being relatively important. They seem to have found a way to both avoid the need for tokenization in LLMs and also avoid the small context windows that result in LLMs short memories. Andrej Karpathy has a post going into more detail