zlacker

[parent] [thread] 1 comments
1. xpe+(OP)[view] [source] 2023-09-02 03:25:00
ChatGPT is a hybrid system; it isn't "just" an LLM any longer. What people associate with "LLM" is fluid. It changes over time.

So it is essential to clarify architecture when making claims about capabilities.

I'll start simple: Plain sequence to sequence feed-forward NN models are not Turing complete. Therefore they cannot do full reasoning, because that requires arbitrary chaining.

replies(1): >>famous+ike
2. famous+ike[view] [source] 2023-09-07 00:51:33
>>xpe+(OP)
cGPT is exactly "just" an LLM though. a sparse MoE architecture is not an ensemble of experts.
[go to top]