NSC3™-Platform Protocols

NSC3™-Platform - When a second is a lifetime

NSC3-Core™ is a proprietary data management engine inside NSC3™ Platform built on C, C++ and JavaScript. It utilizes the most optimum codecs, standards and TCP protocol to deliver Ultra-Realtime communication between the node applications. It is designed to control the video and data traffic from end to end. It specially optimizes first but also last and second hop at the networks. Users connecting via Applications built within NSC3™-SDK,Chrome, Firefox, MsEDGE, Safari and any browser that is built on Chromium open source project can communicate directly through their browsers —enabling sub-150 millisecond latency for up –and downlink simultaneously.

Audio Codecs: OpusData Codecs:H.264, WebP, Codec-agnostic Protocols supported at the edge applications:

  • RTMPS (Real-Time Messaging Protocol with (encrypted over SSL)
  • RTSP (Real-Time Streaming Protocol)
  • WebRTC (Web Real-Time Communications)

Playback Compatibility:

  • NSC3™SDK embedded applications in iOS and Android
  • Chrome, Chromium, EDGE and Firefox support without any plugin, Safari from latest release 14.0

Benefits:

  • Full end to end control, not only the first or the last mile
  • Network agnostic, system adjusts automatically bandwidth changes
  • Ultra-fast, application and browser based UX
  • Super scalable, from small teams to large organizations
  • Ultra-secured Data, on the go through unique AES256 Ciphering with TLS1.2 channel

Drawbacks:

  • Designed for mission critical data, video and audio communication not for conferencing and streaming media services. 

Latency:

  • Sub-150-millisecond delivery and return in 4G and faster networks
  • Sub-500-millisecond delivery and return in 3Gnetworks
  • Sub-2000-millisecond delivery and return in 2G networks

What Is a Protocol?

A protocol is a set of rules governing how data travels from one communicating system to another. These are layered on top of one another to form a protocol stack. That way, protocols at each layer can focus on a specific function and cooperate with each other. The lowest layer acts as a foundation, and each additional layer adds complexity. You’ve likely heard of an IP address, which stands for Internet Protocol. This protocol structures how devices using the internet communicate. The Internet Protocol sits at the network layer and is typically overlaid by the Transmission Control Protocol (TCP) at the transport layer, as well as the Hypertext Transfer Protocol (HTTP) at the application layer.

The seven layers — which include physical, data link, network, transport, session, presentation, and application— were defined by the International Organization for Standardization’s (IS0’s) Open Systems Interconnection model, as depicted above.

What Is a Streaming Protocol?

Each time you watch a live stream or video on demand, streaming protocols are used to deliver data over the internet. These can sit in the application, presentation, and session layers. Online video delivery uses both streaming protocols and HTTP-based protocols. Streaming protocols like Real Time Messaging Protocol (RTMP) enable speedy video delivery using dedicated streaming servers, whereas HTTP-based protocols rely on regular web servers to optimize the viewing experience and quickly scale. Finally, a handful of emerging HTTP-based technologies like the Common Media Application Format(CMAF) and Apple’s Low-Latency HLS seek to deliver the best of both options to support low-latency streaming at scale.

UDP vs. TCP: A Quick Background

User Datagram Protocol (UDP) and Transmission Control Protocol (TCP) are both core components of theinternet protocol suite, residing in the transport layer. The protocols used for streaming sit on top of these. UDP and TCP differ in terms of quality and speed, so it’s worth taking a closer look.

The primary difference between UDP and TCP hinges on the fact that TCP requires a three-way handshake when transporting data. The initiator (client) asks the accepter (server) to start a connection, the accepter responds, and the initiator acknowledges the response and maintains a session between either end. For this reason, TCP is quite reliable and can solve for packet loss and ordering. UDP, on the other hand, starts without requiring any handshake. It transports data regardless of any bandwidth constrains, making it speedier and riskier. Because UDP doesn’t support retransmissions, packet ordering, or error-checking, there’s potential for a network glitch to corrupt the data en route.

Protocols like Secure Reliable Transport (SRT) often use UDP, whereas HTTP-based protocols use TCP.

Considerations When Choosing a Streaming Protocol

Selecting the right protocol starts with defining what you’re trying to achieve. Latency, playback compatibility, and viewing experience can all be impacted. What’s more, content distributors don’t always stick with the same protocol from capture to playback. Many broadcasters use RTMP to get from the encoder to server and then transcode the stream into an adaptive HTTP-based format.

What Are the Most Common Protocols Used for Streaming?

  • Traditional Streaming Protocols RTMP (Real-Time Messaging Protocol)
  • RTSP (Real-Time Streaming Protocol)/RTP (Real-Time Transport Protocol)

HTTP-Based Adaptive Protocols

  • Apple HLS (HTTP Live Streaming)
  • Low-Latency HLS
  • MPEG-DASH (Moving Picture Expert Group Dynamic Adaptive Streaming over HTTP)
  • Low-Latency CMAF for DASH (Common Media Application Format for DASH)
  • Microsoft Smooth Streaming
  • Adobe HDS (HTTP Dynamic Streaming)

Traditional Streaming Protocols

Traditional streaming protocols, such as RTSP and RTMP, support low-latency streaming. But they aren’t supported on all endpoints (e.g., iOS devices). These work best for streaming to a small audience from a dedicated media server.

As shown above, RTMP delivers video at roughly the same pace as a cable broadcast — in just over five seconds. RTSP/RTP is even quicker at around two seconds. These protocols achieve this by transmitting the data using a firehose approach rather than requiring local download or caching. But because very few players support these protocols, they aren’t optimized for great viewing experiences at scale. Many broadcasters choose to transport live streams to the media server using a stateful protocol like RTMP and then transcode it into anHTTP-based technology for multi-device delivery.

Adobe RTMP

Adobe designed the RTMP specification for the transmission of audio and video data between technologies like a dedicated streaming server and the Adobe Flash Player. Reliable and low-latency, this worked great for live streaming. But open standards and adaptive bitrate streaming eventually edged RTMP out. The writing on the wall came when Adobe announced the death of Flash — slated for 2020.

While Flash’s end-of-life date is overdue, the same cannot be said for using RTMP for video contribution. RTMP encoders are still a go-to for many content producers even though the proprietary protocol has fallen out of favor for last-mile delivery.

  • Audio Codecs: AAC, AAC-LC, HE-AAC+ v1 & v2, MP3, Speex, Opus, Vorbis
  • Video Codecs: H.264, VP8, VP6, Sorenson Spark®, Screen Video v1 & v2
  • Playback Compatibility: Not widely supported (Flash Player, Adobe AIR, RTMP-compatible players)
  • Benefits: Low-latency and requires no buffering
  • Drawbacks: Not optimized for quality of experience or scalability
  • Latency: 5 seconds
  • Variant Formats: RTMPT (tunneled through HTTP), RTMPE (encrypted), RTMPTE (tunneled and encrypted), RTMPS (encrypted over SSL), RTMFP (travels over UDP instead of TCP)

RTSP/RTP

Like RTMP, RTSP/RTP describes a stateful protocol used for video contribution as opposed to multi-device delivery. While RTMP is a presentation-layer protocol that lets end users command media servers via pause and play capabilities, RTP is a transport protocol used to move said data. It’s supported by UDP at this same layer. Android and iOS devices don’t have RTSP compatible players out of the box, making this another protocol that’s rarely used for playback.

  • Audio Codecs: AAC, AAC-LC, HE-AAC+ v1 & v2, MP3, Speex, Opus, Vorbis
  • Video Codecs: H.265 (preview), H.264, VP9, VP8
  • Playback Compatibility: Not widely supported (Quicktime Player and other RTSP/RTP-compliant players,VideoLAN VLC media player, 3Gpp-compatible mobile devices)
  • Benefits: Low-latency and requires no buffering
  • Drawbacks: Not optimized for quality of experience and scalability
  • Latency: 2 seconds
  • Variant Formats: The entire stack of RTP, RTCP (Real-Time Control Protocol), and RTSP is often referred to asRTSP

Adaptive HTTP-Based Streaming Protocols

Streams deployed over HTTP are not technically “streams.” Rather, they’re progressive downloads sent via regular web servers. Using adaptive bitrate streaming, HTTP-based protocols deliver the best video quality and viewer experience possible — no matter the connection, software, or device. Some of the most common HTTP-based protocols include MPEG-DASH and Apple’s HLS.

Apple HLS

Since Apple’s a major player in the world of internet-connected devices, it follows that Apple’s HLS protocol rules the digital video landscape. For one, the protocol supports adaptive bitrate streaming, which is key to viewer experience. More importantly, a stream delivered via HLS will play back on the majority of devices — thereby expanding your audience.While HLS support was initially limited to iOS devices such as iPhones and iPads, native support has since been added to a wide range of platforms. All Google Chrome browsers, as well as Android,Linux, Microsoft, and MacOS devices can play streams delivered using HLS.

  • Audio Codecs: AAC-LC, HE-AAC+ v1 & v2, MP3
  • Video Codecs: H.265, H.264
  • Playback Compatibility: Great (All Google Chrome browsers; Android, Linux, Microsoft, and MacOS devices; several set-top boxes, smart TVs, and other players)
  • Benefits: Adaptive bitrate and widely supported
  • Drawbacks: Quality of experience is prioritized over low latency
  • Latency: 6-30 seconds (lower latency only possible when tuned)
  • Variant Formats: Low-Latency HLS (see below), PHLS (Protected HTTP Live Streaming)

Low-Latency HLS

Mid 2019, Apple announced an extension to their HLS protocol designed to drive latency down at scale. The protocol achieves this using HTTP/2 PUSH delivery combined with shorter media chunks.Unlike standard HLS, Apple Low-Latency HLS doesn’t yet support adaptive bitrate streaming — but it is on the roadmap.

  • Playback Compatibility: Any players that aren’t optimized for Low-Latency HLS can fall back to standard (higher-latency) HLS behavior
  • Benefits: Low latency meets HTTP-based streaming
  • Drawbacks: As an emerging spec, vendors are still implementing support
  • Latency: 3 seconds or less

MPEG-DASH

When it comes to MPEG-DASH, the acronym spells out the story. The Moving Pictures Expert Group(MPEG), an international authority on digital audio and video standards, developed Dynamic AdaptiveStreaming over HTTP (DASH) as an industry-standard alternative to HLS. Basically, with DASH you get an open-source option. But because Apple tends to prioritize its proprietary software, support forDASH plays second fiddle.

  • Audio Codecs: Codec-agnostic
  • Video Codecs: Codec-agnostic
  • Playback Compatibility: Good (All Android devices; most post-2012 Samsung, Philips, Panasonic, and Sony TVs; Chrome, Safari, and Firefox browsers)
  • Benefits: Vendor independent, international standard for adaptive bitrate
  • Drawbacks: Not supported by iOS or Apple TV
  • Latency: 6-30 seconds (lower latency only possible when tuned)
  • Variant Formats: MPEG-DASH CENC (Common Encryption)

Low-Latency CMAF for DASH

The Common Media Application Format, or CMAF, is in itself is a media format. But when paired with chunked encoding and chunked transfer encoding for delivery over DASH, it should support sub-three-second delivery. While its transfer setup differs from that of Low-Latency HLS, the use of shorter data segments is quite similar.

  • Playback Compatibility: Any players that aren’t optimized for low-latency CMAF for DASH can fall back to standard (higher-latency) DASH behavior
  • Benefits: Low latency meets HTTP-based streaming
  • Drawbacks: As an emerging spec, vendors are still implementing support
  • Latency: 3 seconds or less

Microsoft Smooth Streaming

Microsoft developed Microsoft Smooth Streaming for use with Silverlight player applications. It enables adaptive delivery to all Microsoft devices.

  • Audio Codecs: AAC, MP3, WMA
  • Video Codecs: H.264, VC-1
  • Playback Compatibility: Good (Microsoft and iOS devices, Xbox, many smart TVs)
  • Benefits: Adaptive bitrate and supported by iOS
  • Drawbacks: Proprietary technology
  • Latency: 6-30 seconds (lower latency only possible when tuned)

Adobe HDS

HDS was developed for use with Flash Player applications as the first adaptive bitrate protocol. Because the end of-life date for Flash is looming, it’s fallen out of favor.

  • Audio Codecs: AAC, MP3
  • Video Codecs: H.264, VP6
  • Playback Compatibility: Not widely supported (Flash Player, Adobe AIR)
  • Benefits: Adaptive bitrate technology for Flash
  • Drawbacks: Proprietary technology with lacking support
  • Latency: 6-30 seconds (lower latency only possible when tuned)

Other New Technologies

  • SRT (Secure Reliable Transport)
  • WebRTC (Web Real-Time Communications)

Emerging Technologies

Other New technologies like WOWZ, WebRTC, and SRT were designed with latency in mind. Similar to low latency CMAF for DASH and Apple Low-Latency HLS, these protocols continue to evolve.

SRT

This open-source protocol is recognized as a proven alternative to proprietary transport technologies — helping to deliver reliable streams, regardless of network quality. From recovering lost packets to preserving timing behavior, SRT was designed to solve the challenges of video contribution and distribution across the public internet. Although it’s quickly taking the industry by storm, SRT is most often used for first-mile contribution rather than last-mile delivery.

  • Audio Codecs: Codec-agnostic
  • Video Codecs: Codec-agnostic
  • Playback Compatibility: Limited (currently used for contribution)
  • Benefits: High-quality, low-latency video over suboptimal networks
  • Drawbacks: Playback support is still in the works
  • Latency: 3 seconds or less

WebRTC

WebRTC is a combination of standards, protocols, and JavaScript APIs that enables real-time communications (RTC, hence its name). Users connecting via Chrome, Firefox, or Safari can communicate directly through their browsers — enabling sub-500 millisecond latency. According to Google, “with Chrome, Edge, Firefox, and Safari supporting WebRTC, more than 85% of all installed browsers globally have become a client for real-time communications on the internet.”

  • Audio Codecs: Opus, iSAC, iLBC
  • Video Codecs: H.264, VP8, VP9
  • Playback Compatibility: Chrome, Firefox, and Safari support WebRTC without any plugin
  • Benefits: Super fast and browser-based
  • Drawbacks: Designed for video conferencing and not scale
  • Latency: Sub-500-millisecond delivery

What Protocols Are Not: Codecs

Conclusion: How to Choose the Right Protocol

Protocols differ in the following areas:

  • Scalability
  • Latency
  • Quality of experience (adaptive bitrate enabled, etc.)
  • Use (first-mile contribution vs. last-mile delivery)• Playback support
  • Proprietary vs. open source
  • Codec requirements

By prioritizing the above considerations, it’s easy to narrow down what’s best for you.RTMP and SRT are great bets for first-mile contribution, while both DASH and HLS lead the way when it comes to playback. That’s why we’re especially excited to see low-latency CMAF for DASH and Low-Latency HLS takeoff. But you may be looking to deploy a one-to-few conference, in which case WebRTC would be better suited.