Advanced Digital Video (Part 4)

Q-SYS Quantum Level 1 Training (Online) : Advanced Digital Video

10 ) Control Troubleshooting

9m 52s

Video Transcript

Advanced Digital Video (Part 4) 7m 4s
00:07
Welcome back! Let’s wrap up with video codecs.
00:11
We found that there are a lot of codecs out there that can deliver nice looking video
00:16
but in doing so they're using up loads of unneeded bandwidth.
00:20
That’s why we created our own codec.
00:22
Q-SYS Shift is true to its name, in that it changes based off the content that’s being transported.
00:28
For example, content with relatively little motion or changes occurring can be transmitted with very little bandwidth.
00:35
However, when full motion video begins the shift codec can adjust and increase the bandwidth as needed
00:42
to send full motion video without any compression artifacts.
00:46
We started with a DCT (or Discrete Cosine Transform) as our base compression algorithm,
00:52
which is also used by the most popular codecs out there today,
00:56
along with both temporal and spacial compression to give you a codec that can adjust to any situation.
01:03
Q-SYS Shift also has a few other features like RFI and RTSP. Let's take a look at all of these in detail.
01:10
DCT is less intense with computations
01:14
and widely accepted when compared to DWT-based codecs which use discrete wavelet transport.
01:21
Spatial compression reduces the amount of data within a frame.
01:25
Think of it like a JPEG. You can take a RAW image from a camera
01:30
and the algorithm will reduce it without any negligible compression artifacts.
01:34
We do this with every frame in video.
01:38
We know that video is really just a sequence of still frames played at a rate
01:42
so our eyes cannot detect the change in frames.
01:45
The change between frames is where the temporal compression is used.
01:50
For example, the image below shows a few frames of a boomerang flying through the air.
01:56
Using temporal compression, we only send the data that is changing between frames.
02:01
Note that the background does not change so that data does not need to be resent.
02:06
Only the data about the boomerang.
02:08
What we get as a result of these two compression techniques is a native Q-SYS device
02:14
well suited for the conference room environment,
02:16
although the NV series is capable of being in any environment.
02:21
For example, a user has a slide deck or a spreadsheet which doesn’t change very often,
02:27
you can see significantly less bandwidth usage in that kind of content.
02:31
Compared to a JPEG 2000 codec, Q-SYS Shift has far less network bandwidth usage with static content.
02:38
However, Q-SYS Shift can easily be deployed where full motion video is required,
02:43
the codec will just shift or adjust based on the incoming content
02:47
and use the right amount of bandwidth that is required for full motion video.
02:51
RFI stands for Reference Frame Interval.
02:54
RFI can be a useful setting within Q-SYS that can be used to help with network recovery
03:00
with increased bandwidth or reduce bandwidth with a reduced network recovery.
03:05
In general the codec uses I-Frames, B-frames and P-Frames.
03:10
The Shift codec doesn’t exactly use these but the idea is the same.
03:15
The I-frame contains the full frame containing information for the full image.
03:19
The P-frame only encodes the changes between the current frame and the previous frame.
03:25
Like the P-Frame, the B-frame encodes the changes between the current frame and the previous frame,
03:30
but also encodes the changes between the current frame and the future frame.
03:35
So, how does this apply to RFI?
03:38
The RFI decides more or less how frequently we send a full frame of data.
03:43
If there are network errors, an I-frame will clean those up.
03:47
However, I-frames require more bandwidth than a P-frame or a B-frame.
03:53
So, how frequently should an I-frame be inserted? The lower the RFI, the more I-frames get inserted,
04:01
which means you get better error recovery at the expense of increased bandwidth.
04:06
The higher the RFI, the worse error recovery but and a decrease in bandwidth usage.
04:13
What does that look like? Well, here's a real-world test using a computer desktop, a static image, at 4K.
04:20
With an RFI of 200 we see a bandwidth usage of around 7 megabits per second.
04:26
That same content with an RFI of 30 will yield a bandwidth of approximately 40 Mbps.
04:33
If you open the NV Series component in Q-SYS Designer Software,
04:37
you’ll will find the RFI setting, with a range between 5 and 255 and a default setting of 30.
04:43
You will also note that you can set the bitrate here.
04:46
This is the maximum allowable bitrate that the algorithm will try and use.
04:51
This is equally applied across all HDMI inputs.
04:54
The NV Series is capable of encoding either a single 4K source or 3 simultaneous 1080p sources.
05:03
If you set the encoder to a bitrate of 600 Mbps and you set it to 4K60 max mode
05:10
then you can transmit a single 4K stream with a maximum bitrate of 600 Mbps.
05:16
Now, if the encoder is set to 1080p60 max mode
05:21
then you are able to share that 600 Mbps across all three inputs.
05:27
300 Mbps for 2 streams, 200 Mbsp for 3 streams. Math!
05:32
One last thing. If an encoder in 4K60 mode attempts to encode more than one 4K60 stream
05:40
then a hot plug event will be triggered and this will change all input EDIDs to 1080p,
05:46
thus allowing multiple streams at a resolution of 1080p or less.
05:51
So, if a 4K max mode is required
05:54
it is highly recommended that the encoder only wire a single AV stream to any decoder,
06:01
try and not wire multiple 4K sources on a single encoder.
06:05
If an encoder is in the 1080p60 max mode then there are no worries for a hot plug event,
06:11
you can stream all 3 inputs simultaneously without worry.
06:14
Real time streaming protocol or (RTSP) is another useful tool in establishing
06:20
and controlling the media during a streaming session.
06:23
The decoder requests the media from the source device and the source will send data as unicast or multicast.
06:31
Also known as RTSP pull streaming.
06:34
This means that video data is not broadcast over the network for no reason.
06:39
Encoders do not send video into the network until the decoder requests the video stream.
06:45
If all decoders are set to a graphic image then no video data is sent onto the network.
06:52
And that’s it for Digital Video. Phew, you made it! Congratulations!
06:57
Thanks for watching and we’ll see you next time.

Downloads and Links

Advanced Digital Video (Part 4) 7m 4s