@peterkaminski "LCMs employ a hierarchical structure, mirroring human reasoning processes"
Not a good sign if they think our reasoning is hierarchical.
"reduces sequence length compared to token-level processing, addressing the quadratic complexity of standard Transformers and enabling more efficient handling of long contexts"
Now seeing "concepts" as a kind of compression (of strings of tokens), which I've seen articulated before as a way of understanding much of what's happening with LLMs.