Welcome to the NAB’s 2020 Broadcast Engineering and Information Technology (BEIT) Conference Proceedings. The papers offered here have been presented at the annual BEIT Conference at NAB Show, the world’s largest trade show for the media content creation and distribution industry.
The BEIT Conference program is established each year by the NAB Broadcast Engineering and Information Technology Conference Committee, a rotating group of senior technologists from NAB member organizations, along with representatives from the Society of Broadcast Engineers (SBE). The 2020 BEIT Conference Committee roster is available here.
The content available in the BEIT Conference Proceedings is covered under copyright provisions listed here.
2020 Proceedings Topics
- Advanced Advertising Technologies
- Advanced Emergency Alerting
- Artificial Intelligence Applications for Media Production
- Broadcast Facility Design
- Broadcast Workflows
- Converting to Ultra HD Television Broadcasting
- Cybersecurity for Broadcast
- Designing Cloud-based Facilities
- Emerging Radio Technologies — On-air and Online
- Improving OTT Video Quality
- IP Conversion: Broadcasters’ Research & Recommendations
- Managing HDR and SDR Content
- Media over IP: Security and Timing
- New Directions in IP-based Media Systems
- New Spectrum Issues
- New Technologies for Sports Coverage
- Next Gen Academy I: The Broadcast/Broadband Revolution
- Next Gen Academy II: Transport Matters
- Next Gen Academy III: Next Steps
- Next Gen Academy IV: Planning for SFNs
- Next Gen Academy V: Implementing SFNs
- Next Gen Academy VI: PHY Layer Issues
- Next Gen TV Audio
- Optimizing the OTT User Experience
- Refining Radio Delivery
- TV Repack and Next Gen Transition Preparation
- Using 5G Broadcasting for Content Delivery
- Using 5G Technologies for Media Production
- Using Artificial Intelligence for Closed Captioning
- Using the Cloud for Live Production
Other Proceedings
Artificial Intelligence: Transforming the Live Sports Landscape - $15
Date: April 26, 2020Topics: 2020 BEITC Proceedings, New Technologies for Sports CoverageThe proposed paper focuses on a media recognition AI engine that is trained to catalog an exhaustive range of elements for live sports in near real-time. It helps sports producers cut down the time and effort involved in highlight creation by 60% and enables OTT platforms to deliver high levels of interactivity and personalization with a create-your-own-highlights experience and a powerful live search option. The engine identifies storytelling graphics and events in the footage along with game content to stitch together an end-to-end story in highlights, and also provides finesse with smooth visual transitions and automatic audio leveling – just like a professional editor. The said engine helps sports content creators enhance viewer engagement, increase monetization and achieve greater scale and speed in live sports.
Adrish Bera | Prime Focus Technologies | Burbank, CA, United States
Amer Saleem
Audio Streaming: A New Approach - $15
Date: April 26, 2020Topics: 2020 BEITC Proceedings, Refining Radio DeliveryMost of your streaming audience wants something that just works reliably; they don’t care about the tech behind it. Many audio streams currently in service fall short. However, it is up to the content providers to make it work. If you stream audio, you need to know about this. This information is targeted to streaming professionals.
This paper/presentation describes a standards-based method of quality live streaming audio with dramatically decreased operating costs, increased reliability, professional features, and a better user experience. Streaming audio has now evolved past legacy protocols allowing reliable content delivery to mobile and connected car dashboards, where all the audience growth continues.
Greg Ogonowski | StreamS HiFi Radio/Modulation Index, LLC | Diamond Bar, CA USA
Nathan Niyomtham | StreamS HiFi Radio/Modulation Index, LLC | Diamond Bar, CA USA
Audio-Specific Metadata To Enhance the Quality of Audio Streams and Podcasts - $15
Date: April 26, 2020Topics: 2020 BEITC Proceedings, Emerging Radio Technologies -- On-air and OnlineAudio-specific metadata was envisioned several years ago in the MPEG-D standard for Dynamic Range Control. The application of this metadata to online content awaited a newer audio codec and the current generation of mobile operating systems. Now that both are becoming widely available, this paper explains how audio content providers can offer new consumer benefits as well as a more compelling listening experience.
This paper and presentation will explain the types of audio-specific metadata that monitor key characteristics of audio content, such as loudness, dynamic range and signal peaks. From simple to large-scale producers, the metadata sets are added to the content during encoding for real-time distribution?as in streams, or for file storage?as with podcasts.
In the playback device or system, this metadata is decoded along with the audio data frames. Through diagrams the decoder operations are described, providing benefits such as loudness matching across different audio content?ending annoying blasting from some audio and reaching for the volume. It will be shown that audio dynamic range can be controlled according to the noise environment around the listener?quiet parts of a performance can be raised to audibility, but only for those who need it.
An audio demonstration is planned to allow the audience to hear the same encoded program over a range of playout conditions with the same device, from riding public transit to full dynamic range for listeners who want highest fidelity. The workflows in production and distribution to add audio-specific metadata are explained, showing how content producers need to make only one target level for all listeners, rather than one for smart speakers, another for fidelity-conscious listeners, etc.
John Kean | Cavell Mertz & Associates Inc. | Manassas, Virginia, USA
Alex Kosiorek | Central Sound at Arizona PBS | Phoenix, Arizona, USA
Automated Brand Color Accuracy for Real-Time Video - $15
Date: April 26, 2020Topics: 2020 BEITC Proceedings, New Technologies for Sports CoverageSports fans know their team?s colors and recognize inconsistencies in brand color display when their beloved team?s colors show up incorrectly. Repetition and consistency build brand recognition and that equity is valuable and worth protecting. Accurate display of specified colors on screen is impacted by many factors including the capture source, the display screen technology, and the ambient light which can range broadly throughout the event due to changes in time and weather (for outdoor fields) or mixed-artificial lighting (for indoor facilities). Changes to any of these factors demand adjustments of the output in order to maintain visual consistency. According to the industry standard, color management is handled by a technician who manually corrects footage from up to two dozen camera feeds in real-time to adjust for consistency across camera feeds. In contrast, the AI-powered ColorNet system ingests live video, adjusts each video frame with a trained machine learning model, and outputs a color corrected feed in real-time. This system is demonstrated using Clemson University?s orange specification, Pantone 165.? The machine learning model was trained using a dataset of raw and color-corrected videos of Clemson football games. The footage was manually color corrected using Adobe Premiere Pro. This trains the model to target specific color values and adjust only the targeted brand colors pixel-by-pixel without shifting any other surrounding colors in the frame, generating localized corrections while adjusting automatically to changes in lighting and weather. The ColorNet model successfully reproduces the manually-created corrections masks while being highly computationally efficient both in terms of prediction fidelity and speed. This approach has the ability to circumvent human error when color correcting while constantly adjusting for negative impacts caused by lighting or weather changes on the display of brand colors. Current progress is being made to beta test this model in live-time broadcast streaming during a live sporting event with large-format screens in partnership with Clemson University Athletics.
Emma Mayes | Clemson University | Clemson, South Carolina, United States
John Paul Lineberger | Clemson University | Clemson, South Carolina, United States
Michelle Mayer | Clemson University | Clemson, South Carolina, United States
Andrew Sanborn | Clemson University | Clemson, South Carolina, United States
Hudson Smith | Clemson University | Clemson, South Carolina, United States
Erica Walker | Clemson University | Clemson, South Carolina, United States
Bit-Rate Evaluation of Compressed HDR Using SL-HDR1 - $15
Date: April 26, 2020Topics: 2020 BEITC Proceedings, Managing HDR and SDR ContentFrom a signal standpoint, what differentiates High Dynamic Range (HDR) content from Standard Dynamic Range (SDR) content is the mapping of the pixel samples to actual colors and light intensity. Video compression encoders and decoders (of any type) are agnostic to that ? the encoder will take a signal, compress it, and at the other side, the decoder will re-create something that is ?about the same? as the signal fed to the encoder. Existing encoders may have been optimized to provide good results with SDR, but HDR will still flow through, albeit possibly requiring higher bit rates for good reproduction.
Some of the HDR standards include dynamic metadata to help a display device render the content based on its capabilities. Some standards transmit a native HDR signal with metadata that allows the creation of anything from the original HDR all the way down to SDR. SL-HDR1, on the other hand, does the opposite: An SDR signal is transmitted, with metadata to inverse tone map to HDR. This metadata adds overhead to the video elementary stream.
This paper focuses on the required bit rates to produce a final HDR signal over a compressed link. We compare encoding SMPTE-2084 PQ HDR signals directly versus using SL-HDR1 to generate an SDR signal plus dynamic metadata. The comparison is done objectively by comparing the PSNR of the decoded signal. The SMPTE-2084 HDR signal is used as a reference at a fixed bit rate, and the bit rate of the SL-HDR1 encoded signal is varied until it matches the PSNR, over a range of source material. The evaluation is done for both AVC (H.264) and HEVC (H.265). This is similar to the work by Touze and Kerkhof published in 2017, but using commercial equipment.
Ciro Noronha | Cobalt Digital Inc. | Champaign, Illinois, USA
Kyle Wilken | Cobalt Digital Inc. | Champaign, Illinois, USA
Ryan Wallenberg | Cobalt Digital Inc. | Champaign, Illinois, USA