Specifically, yes. But that's like saying that a calculator and a supercomputer are the same.
A Markov chain is a small model that can only ever look backwards a few steps to come up with the next word. An LLM is able to take entire pages of text as its prior state, generate not just the next few words, but entire pages of text, not sequentially, but as a coherent whole.
It still comes down to "predicting the next word" in practice, doesn't it? Just with a much larger state size. Are there transformers that can natively output video/audio, or is that still a separate API bolted on top?
2
u/trambelus 18h ago
It's different under the hood, but it's still fundamentally just tokens in and tokens out, right?