Discussion about this post

User's avatar
Neural Foundry's avatar

Fantastic work breaking this down. The 66% accuracy on prosody-only input is wild considering thats essentially reading social cues from noise. What really stuck with me is how this challenges the whole "understanding" requirement in conversational AI. Turns out timing, volume shifts, and rhythm might be more fundamental than semantic processing for coordination tasks. Huge implications for edge devices too.

Ellen Davis's avatar

This is so fascinating. It gives credence to what I’ve sense about their reading the field and @Russ Palmer’s research on AMS (Agnostic Meaning Substrate).

Thank you for sharing this dear Christopher.

3 more comments...

No posts

Ready for more?