Topics
- 2025 BEITC Proceedings
- Enhancing Video Streaming Quality and Efficiency
- 5G in Broadcast Spectrum and Video Quality Metrics
- Getting the Most out of ATSC 3.0
- AI Applications: Captions, Content Detection and Advertising Management
- Immersive Audio, Satellite and OTT Delivery
- Innovations in Live Production and Broadcast Workflows
- IP Networks and the Broadcast Chain: Fast Friends
- AI Applications: Sports, Newsrooms and Archives
- Making ATSC 3.0 Better than Ever
- AM Radio: Measurements and Modeling
- Making Radio Better Than Ever
- Brigital: Integrating Broadcast and Digital
- Production Advancements: Avatars and Immersive Content
- Broadcast Positioning System (BPS): Resilience and Precision
- Resilience, Safety and Protection for Broadcast Service
- Cybersecurity for Broadcasters
- Streaming Improvements: Low Latency and Multiview
- Embracing the Cloud: Transforming Broadcast Operations with ATSC 3.0 and Broadband Technologies
- 2024 BEITC Proceedings
- 2023 BEITC Proceedings
- 2022 BEITC Proceedings
- 2021 BEITC Proceedings
- 2020 BEITC Proceedings
AI Applications: Captions, Content Detection and Advertising Management
Enhancing Instream Shoppable Brand and Product Detection in Broadcast, OTT, and VOD Content through Multi-Model Object Detection and Real-Time SCTE-35/SEI/VMAP Integration - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, AI Applications: Captions, Content Detection and Advertising ManagementThe increasing adoption of shoppable video content in Broadcast, OTT, and VOD has transformed how consumers engage with digital media. Traditional advertising models struggle to create seamless, interactive, and real-time brand engagement, leading to a need for AI-driven solutions. Emerging technologies now enable real-time product detection, metadata embedding, and interactive instream commerce integration, making video content viewable and instantly shoppable. The paper introduces an AI-based platform that instantly detects brands and products in video streams before implementing metadata structures for interactive shopping platforms. By leveraging deep learning-based object detection models and metadata signaling protocols, the framework ensures synchronized, real-time engagement between consumers and brands, unlocking new monetization opportunities for advertisers.
Multiple AI detection engines under this framework based on YOLO, Mask R-CNN, ResNet, and MobileNet SSD precisely identify products and brand placements inside real-time and pre-recorded content. The detections from the system link up with SCTE-35 (live streams), SEI (compressed video streams), and VMAP (VOD-based ad scheduling), which enable precise frame-level interactivity for product engagement. Based on AI technology, the system allows moment-by-moment brand interactions and stream-based e-commerce capabilities, and advertisers can use it for content monetization. The framework enhances content creator and advertiser success while delivering improved consumer engagement since it automates product detection combined with advertisement synchronization features. This framework redefines media commerce by combining AI-based detection, metadata injection, and HTML-based interactivity, allowing for scalable, real-time, and seamless brand engagement in video content.
Chaitanya Mahanthi | Google, YouTube | New York, N.Y., United States
Media campaign workflow using AI agents - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, AI Applications: Captions, Content Detection and Advertising ManagementThe increasing adoption of shoppable video content in Broadcast, OTT, and VOD has transformed how consumers engage with digital media. Traditional advertising models struggle to create seamless, interactive, and real-time brand engagement, leading to a need for AI-driven solutions. Emerging technologies now enable real-time product detection, metadata embedding, and interactive instream commerce integration, making video content viewable and instantly shoppable. The paper introduces an AI-based platform that instantly detects brands and products in video streams before implementing metadata structures for interactive shopping platforms. By leveraging deep learning-based object detection models and metadata signaling protocols, the framework ensures synchronized, real-time engagement between consumers and brands, unlocking new monetization opportunities for advertisers.
Multiple AI detection engines under this framework based on YOLO, Mask R-CNN, ResNet, and MobileNet SSD precisely identify products and brand placements inside real-time and pre-recorded content. The detections from the system link up with SCTE-35 (live streams), SEI (compressed video streams), and VMAP (VOD-based ad scheduling), which enable precise frame-level interactivity for product engagement. Based on AI technology, the system allows moment-by-moment brand interactions and stream-based e-commerce capabilities, and advertisers can use it for content monetization. The framework enhances content creator and advertiser success while delivering improved consumer engagement since it automates product detection combined with advertisement synchronization features. This framework redefines media commerce by combining AI-based detection, metadata injection, and HTML-based interactivity, allowing for scalable, real-time, and seamless brand engagement in video content.
Matt Ferris | Decentrix Inc. | Denver, Colo., United States
Role of AI in Quality Checking of Captions - $15
Date: March 21, 2025Topics: 2025 BEITC Proceedings, AI Applications: Captions, Content Detection and Advertising ManagementIn today’s ultra-competitive media and entertainment industry, captioning demands exceptional precision. However, when quality control (QC) is conducted manually, it is very labor-intensive and prone to errors, which can lead to compromises in quality. For example, the QC process of ensuring adherence to Federal Communications Commission (FCC) guidelines — specifically, those relating to sync, accuracy, and completeness — requires multiple reviews. Furthermore, it involves verification of segmentation, reading speed, display duration, and layout metrics like row and column count. Caption placement also needs careful adjustment to avoid obstructing important visual elements, while global deliveries necessitate multilingual quality checks to meet diverse audience standards. Additionally, profanity censoring is critical. With all these requirements, performing general checks — such as ensuring captions aren’t delayed during critical or suspenseful moments in a scene — can often be overlooked. This paper will explore how Artificial Intelligence (AI) is streamlining these complex QC tasks and freeing up human resources, enabling media companies to focus on the more creative aspects of their workflows.
Manik Gupta, Sana Afsar, Jeff Ross | Interra Systems | Cupertino, Calif., United States