Topics
- 2025 BEITC Proceedings
- Enhancing Video Streaming Quality and Efficiency
- 5G in Broadcast Spectrum and Video Quality Metrics
- Getting the Most out of ATSC 3.0
- AI Applications: Captions, Content Detection and Advertising Management
- Immersive Audio, Satellite and OTT Delivery
- Innovations in Live Production and Broadcast Workflows
- IP Networks and the Broadcast Chain: Fast Friends
- AI Applications: Sports, Newsrooms and Archives
- Making ATSC 3.0 Better than Ever
- AM Radio: Measurements and Modeling
- Making Radio Better Than Ever
- Brigital: Integrating Broadcast and Digital
- Production Advancements: Avatars and Immersive Content
- Broadcast Positioning System (BPS): Resilience and Precision
- Resilience, Safety and Protection for Broadcast Service
- Cybersecurity for Broadcasters
- Streaming Improvements: Low Latency and Multiview
- Embracing the Cloud: Transforming Broadcast Operations with ATSC 3.0 and Broadband Technologies
- 2024 BEITC Proceedings
- 2023 BEITC Proceedings
- 2022 BEITC Proceedings
- 2021 BEITC Proceedings
- 2020 BEITC Proceedings
2025 BEITC Proceedings
Transferring traceable time to BPS-enabled ATSC 3.0 station - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, Broadcast Positioning System (BPS): Resilience and PrecisionBroadcast Positioning System (BPS) necessitates precise time synchronization across TV stations, utilizing a shared time reference and their fixed locations for triangulation and reliable positioning, with time over fiber solutions connected to trusted timing sources serving as a viable method for maintaining accurate synchronization. In this paper, we present a concept to link BPS leader stations to timescales maintained by National Metrology Institutes (NMIs) using High Accuracy White Rabbit synchronization through optical networks.
Francisco Girela Lopez, Ramki Ramakrishnan | Safran Electronics and Defense | Rochester, N.Y., United States
Transforming Team Collaboration with Avatars: Enhancing Productivity in Content Creation in Broadcasting - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, Production Advancements: Avatars and Immersive ContentIn the fast-paced world of broadcast and media, teams must collaborate efficiently and create content quickly to remain competitive. Avatar technology offers an innovative solution, enabling team members to stay connected even when they are unable to be physically present. By stepping in to summarize past meetings and capture key insights, avatars bridge communication gaps and ensure continuity, allowing teams to resume their work without the need for lengthy catchups. This approach enhances decision-making and streamlines workflows, keeping teams aligned and productive. In addition to avatars, the integration of intuitive tools for video editing, content creation, and subtitling empowers teams to produce high-quality content without specialized skills. These tools support multiple languages, enabling broadcasters to reach a diverse, global audience. With simplified video clipping, merging, and subtitling processes, teams can adapt and release content more quickly, making their workflows more efficient and responsive to industry demands. This combination of avatar-driven collaboration and user-friendly content tools optimizes production time, enhances engagement, and ensures a smooth workflow, even when team schedules do not align. It allows broadcasters to work smarter, integrating technology into the creative process. This approach empowers teams to create exceptional content, collaborate seamlessly, and meet the growing demands of today’s dynamic media landscape.
Ira Sharma, Kaushal Dadi, Anesh Madapoosi, Shyam Kapur | TipTop Technologies | Sunnyvale, Calif., United States
TSENet Video Super Resolution for Broadcast Television - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, Production Advancements: Avatars and Immersive ContentThe emergence of Video Super Resolution (VSR) has advanced the media industry. It holds great potential for enhancing broadcast and streaming, improving visual quality, and boosting user engagement. While there are benefits, applying VSR technology to broadcast presents challenges, as it requires a high-quality solution that maintains consistent quality across diverse broadcast standards, handles live content with minimal latency, and ensures compatibility with existing broadcast infrastructure. This demands sophisticated AI models, often needing high-performance GPUs, which can increase costs. In this paper, we introduce a method to upscale low-resolution, compressed video frames based on an enhanced version of Equivalent Transformation and Dual Stream Network (ETDS)[1], called a Temporally Stabilized ETDS Network (TSENet) that performs equivalent or better compared to more compute extensive solutions. This paper examines various VSR models, emphasizing the key performance indicators identified by the AIM2024 Challenge for Efficient VSR[2]. The focus is on deploying these models in workflows, considering visual quality through both subjective and objective analyses, compression artifact reduction, and real-time processing capabilities.
Surbhi Madan | Intel Corp | San Diego, Calif., United States
Onur Barut | Intel Corp | Berlin, Mass., United States
Anand V Bodas | Intel Corp | Bangalore, India
Christopher A. Bird | Intel Corp | Chandler, Az., United States
Jerry Dong, Lin Xie | Intel Corp | Beijing, China
Two-Pass Encoding for Live Video Streaming - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, Enhancing Video Streaming Quality and Efficiency2025 NAB BEIT Conference Proceedings Best Student Paper Award Winner
Live streaming has become increasingly important in our daily lives due to the growing demand for real-time content consumption. Traditional live video streaming typically relies on single-pass encoding due to its low latency. However, it lacks video content analysis, often resulting in inefficient compression and quality fluctuations during playback. Constant Rate Factor (CRF) encoding, a type of single-pass method, offers more consistent quality but suffers from unpredictable output bitrate, complicating bandwidth management. In contrast, multi-pass encoding improves compression efficiency through multiple passes. However, its added latency makes it unsuitable for live streaming. In this paper, we propose OTPS, an online two-pass encoding scheme that overcomes these limitations by employing fast feature extraction on a downscaled video representation and a gradient-boosting regression model to predict the optimal CRF for encoding. This approach provides consistent quality and efficient encoding while avoiding the latency introduced by traditional multi-pass techniques. Experimental results show that OTPS offers 3.7% higher compression efficiency than single-pass encoding and achieves up to 28.1% faster encoding than multi-pass modes. Compared to single-pass encoding, encoded videos using OTPS exhibit 5% less deviation from the target bitrate while delivering notably more consistent quality.
Mohammad Ghasempour, Hadi Amirpour, Christian Timmerer | Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität | Klagenfurt, Austria
Verifying Video Signals Using Computer Vision and Machine Learning Techniques - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, Enhancing Video Streaming Quality and EfficiencyHigh-quality video streams pass through various processes inside media workflows, including pre-processing, transcoding, editing, post-processing, etc. The media transformation in these processes can alter the properties of the intermediate video, potentially disrupting subsequent processes or degrading the final output video quality experienced by viewers. Reliable content delivery necessitates verification of video data to ensure an acceptable quality of experience (QoE). The verification of video data is not just limited to the measurement of degradation in perceived quality on the viewing screen but also considers validation of video parameters affecting the proper functioning of media devices. For example, the Y, Cb, and Cr values are altered after lossy compression that can result in out-of-gamut RGB data and hence will lead to erroneous functioning of display devices. Similarly, parameters like light levels, black bar widths, color bar types, telecine patterns, photosensitive epilepsy (PSE) levels and patterns, field order, scan types (interlaced or progressive), etc. also need validation before distribution. This paper explores critical video properties and demonstrates how computer vision, image processing, and machine learning techniques can measure and validate these properties and detect defects. Experiments utilizing machine learning (ML) techniques to quantify the quality degradation, due to lossy compression of video data, will also be discussed. A discussion of challenges and future directions to enhance the accuracy of measurement and detection is also included here.
Shekhar Madnani, Raman Kumar Gupta, Siddharth Gupta, Saurabh Jain | Interra Systems, Inc. | Cupertino, Calif., United States
Vision and Language Models for Enhanced Archive Video Management - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, AI Applications: Sports, Newsrooms and ArchivesArchival video collections contain a wealth of historical and cultural information. Managing and analyzing this data can be challenging due to the lack of metadata and inconsistent formatting across different sources. In particular, identifying and separating individual stories within a single archived tape is critical for efficient indexing, analysis and retrieval. However, manual segmentation is time-consuming and prone to human error. To address this challenge, we propose a novel approach that combines vision and language models to automatically detect transition frames and segment archive videos into distinct stories. A vision model is used to cluster frames of the video. Using recent robust automatic speech recognition and large language models, a transcript, a summary and a title for the story are generated. By leveraging computed features from the previous transition frames detection, we also propose a fine-grained chaptering of the segmented stories. We conducted experiments on a dataset consisting of 50 hours of archival video footage. The results demonstrated a high level of accuracy in detecting and segmenting videos into distinct stories. Specifically, we achieved a precision of 93% for an Intersection over Union threshold set at 90%. Furthermore, our approach has shown to have significant sustainability benefits as it is able to filter and remove approximately 20% of the content from the 50 hours of videos tested. This reduction in the amount of data that needs to be managed, analyzed and stored can lead to substantial cost savings and environmental benefits by reducing energy consumption and carbon emissions associated with data processing and storage.
Khalil Guetari, Yannis Tevissen, Frederic Petitpont | Moments Lab Research | Boulogne-Billancourt, France
VVC Broadcast Deployment Update - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, Making ATSC 3.0 Better than EverThe VVC video coding standard was finalized in July 2020, and is thus almost five years old. In light of this approaching anniversary, VVC’s deployment status is reviewed. Major milestones in VVC deployment are detailed, and a comparison with the preceding HEVC codec are provided at an equivalent point in time.
Justin Ridge | Nokia | Dallas, Tx., United States
Lukasz Litwic | Ericsson | Warsaw, Poland
Why do you need proactive video delivery network observability? - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, Immersive Audio, Satellite and OTT DeliveryThe rapid adaptation of IP workflows in production, contribution, cloud, and distribution is always challenging. Organizations spend millions of dollars training the workforce, deploying monitoring solutions and troubleshooting “network issues,” which takes an enormous portion of their tasks. Time and effort are wasted on old techniques that were suitable for standard file delivery versus today’s high throughput and low latency demands of live production.
In this paper we would like to draw your attention to a different way of thinking, observing and gaining in-depth understanding of the lifeline of the broadcast industry: the network. The traditional reactive way of monitoring events and alarms must evolve to use proactive observability. Unlike standard observability that just collects data from elements and passive probes, proactive observability involves injecting small amounts of traffic into the network to discover the network routes and to support the continuous measurement of the performance of the traffic to gain valuable insights into the experience of the traffic.
Adi Rozenberg | AlvaLinks. Ltd | Neve Yamin, Israel