@freemo I'm not even saying that. Something simpler.
LLMs typically work in syllables, not whole words. That's good for LLMs, because then it learns patterns of conjugation and pluralization naturally from seeing those patterns used in its training.
But what if you ran it on words, not syllables, and you fed it a well-structured initial dataset (i.e. a dictionary) to seed its initial tokens. It would understand that "humans" and "human" are related words and how.
1/2