I've been researching this one for quite a bit before but haven't found any solid leads.
Previously, I had a Wowza/Flash app with video chatroom functionality (let's say 4 or 8 video/audio streams) and would like to combine the streams server-side into one video/audio stream in order to be sent to a live Youtube channel or say a CDN like AWS CloudFront. In addition, I'm wondering how this would be implemented in WebRTC with a central server.
I've found a couple projects such as jMixer and some helpful keywords such as "vision mixer" to help with my search but looking for any previous experience or new ideas. Now including WebRTC I'm seeing "multipoint control unit" to centralize the streams, and "mashup".
The other option is building something like it myself with a commercial video decoding/encoding library to raw frames, stitching the frames together, then encoding it. I was originally going down this route but put project on hold.
What are some ideas, keywords, existing software (open source preferred), or even articles to take those live streams and combine them into one in real-time? Or is coding it myself like I was the required route?
Not looking for a coded solution but more of a generally efficient (in time and resources, not computation) solution applicable to all users.
Thanks!