Discussion about this post

User's avatar
Alex's avatar

Regarding Scenario 1 in the internal/external distinction, wouldn't an external approach ultimately be able to extract internal features (unless we don't allow it to see anything that reflects the starting conditions - without which we could not arrive at the analytical solution using the 'internal' approach either)?

Given identical input - like the first few frames of a video of the coin being flipped, I think a sequence model could definitely achieve more than 50% accuracy/the causal factors are able to be reconstructed from the provided data. Maybe I misunderstood the terminology here?

In that sense, I'm not sure I understand why fundamentally this would be a barrier to recreating something like human thought (from text for example). Although there is definitely added 'complexity'/training needed for the model to reconstruct those internal factors.

Also I'm curious as to your thoughts on computational complexity regarding the storage size of such a model. I have no experience with how this is usually done for truly large models, but from my initial understanding, memory speed can be a large contributing factor. At even 1 byte per synapse that's 600 TB of data. This seems non-trivial to get running even using H100's, but I would imagine even if you do the effective FLOPS values might be different.

Interesting read! Honestly would be really cool to even see a simulated insect brain - I think we could get a lot of interesting insights on the scaling/viability of such an approach.

Expand full comment
3 more comments...

No posts