Put another way: Alex is basically telling Claude 3 ("Opus") that he's running a test on it, & is excited when Claude (a system for analyzing & producing human-plausible representations of similar text) "recognizes" a needle-testing prompt and produces text that's plausibly consistent with needle-testing.
What one SHOULD then do is remake one's tests (or better-sandbox the model). Instead, Alex leaps to concluding the model is self-aware.
https://twitter.com/alexalbert__/status/1764722513014329620