I repurposed an old Google #AIY Voice Kit to use @openai Realtime API with #WebRTC. Vibecoding failed, so I did some old school research+trial&error.
My main learnings:
· Pass all the same WSS messages via Data Channel
· Ephemeral tokens not always needed
· API is a lot like WHIP (no SDK needed)
· Best practices for managing gUM and RTCPeerConnection still apply (not in docs)
· Use speech events for voice functions
Full code walkthrough on webrtcHacks here: https://webrtchacks.com/the-unofficial-guide-to-openai-realtime-webrtc-api