Topics
- 2024 BEITC Proceedings
- Application of 5G in Broadcasting
- Application of Large Language Model (LLM) in Media
- Applications of ATSC 3.0 Technology
- BPS as the Complementary PNT Solution
- Broadcast Facility Design
- Content Creation and Delivery Technology
- Cybersecurity for Broadcasters
- Data Delivery
- Digital Online Operations
- Emerging Technologies in Media Delivery
- Generative AI for Media
- Generative AI Uses and Video Transcoding
- Quantifying Quality in Video Technology
- Radio Topics
- Society of Cable Telecommunications Engineers
- Striving for Efficiency in Video Technology
- The NMCS Concept
- Timing Solutions for Broadcasters
- Video Encoding and Codecs
- Video Technology - Miscellaneous Topics
- 2023 BEITC Proceedings
- 2022 BEITC Proceedings
- 2021 BEITC Proceedings
- 2020 BEITC Proceedings
- Uncategorized
2024 BEITC Proceedings
Making Your Assets Mean Something: Evolving Asset Management Systems Using Semantic Technologies - $15
Date: April 3, 2024Topics: 2024 BEITC Proceedings, Application of Large Language Model (LLM) in MediaProduce more content! Faster! Cheaper! Make it available in different places! And never throw anything away! It is seemingly impossible for media enterprises to keep up with these escalating and often conflicting business pressures. How can media asset management systems evolve to provide the right content based on the intent of the user? Can a creator produce content faster using knowledge hidden within the content? How can assets be easily accessed from distributed locations to enrich the experience? And does content growth exist without exploding costs? The answer to all these questions is yes. It is derived by leveraging tools that provide contextual meaning to the content, so that more effective utilization of the content can be performed. It goes beyond providing rich searching mechanisms to now providing the ability to get insights into the content. It relies on tapping into richer information sources that go beyond logged metadata for a piece of content, to now harvest knowledge from scripts, stories, and dialogue. The answer lies in creating a connected environment between data silos where a central brain has knowledge not just where content is located, but of what is in the content. The answer is in the ability for a Creator to use their native language and express in natural conversational form what they want to achieve when searching for, analyzing, and manipulating the content. The answer lies in tools understanding a user’s creative intent to perform appropriate actions, versus the creative having to learn and manipulate complex user interfaces to achieve the same intent. This paper discusses how certain artificial intelligence (AI) technologies such as Semantic Embeddings, Knowledge Graphs, and Multimodal Large Language Models are coming together under the umbrella of Knowledge Management to solve this problem of meaningfully extracting and using information from content that takes us beyond what we can do today with Asset Management systems.
Shailendra Mathur | Avid Technology | Montreal, Quebec, Canada
Rob Gonsalves | Avid Technology | Burlington, Mass., United States
Roger Sacilotto | Avid Technology | Burlington, Mass., United States
MC-IF VVC technical guidelines - $15
Date: April 3, 2024Topics: 2024 BEITC Proceedings, Video Encoding and CodecsVersatile Video Coding (VVC or H.266), standardized by ISO/IEC MPEG and ITU-T VCEG in 2020, offers best-in-class compression performance and has been selected (or is currently being considered) for use in next-generation broadcast and streaming standards around the world. These standards typically define VVC-based profiles and corresponding receiver capabilities. How the service is realized and the impact of the codec’s operational parameters on delivered compression performance is not in scope. The Media Coding Industry Forum (MC-IF) has developed – with input from broadcasters, encoder vendors and others in the community – technical guidelines that serve as a reference for VVC configuration choices to address operational, interoperability, and regulatory needs while achieving optimal compression performance. A community review phase concluded in late 2023, and the guidelines were made publicly and freely available at the start of 2024. This paper provides an overview of MC-IF VVC technical guidelines version 1.0. It gives an overview of the guidelines development process, scope, and contents. The paper provides a summary of VVC’s support in media transport and systems standards and follows with VVC adoption status in broadcast and streaming application specifications. A review of VVC-based profiles and receiver capabilities is presented. The paper concludes with an analysis on operational bitrates which can be expected with VVC for selected service deployment scenarios.
Lukasz Litwic | Ericsson | Gdańsk, Poland
Justin Ridge | Nokia Technologies | Dallas, Texas, United States
Alan Stein | InterDigital Communications, Inc. | Princeton, N.J., United States
MV-HEVC: How to optimize compression of immersive 3D content - $15
Date: April 3, 2024Topics: 2024 BEITC Proceedings, Video Encoding and CodecsMultiview High Efficiency Video Coding (MV-HEVC) is an HEVC extension focused on efficiently coding spatially related images, such as left eye and right eye views of 3D stereoscopic content. MV-HEVC was released in the second version of HEVC back in October 2014 [1]. MV-HEVC was created alongside another extension named SHVC, which is already employed in ATSC 3.0 use cases since 2018 [2]. The two extensions share many fundamentals, codified in an annex for common specifications for multi-layer extensions. Their base principle is scalability: A traditional HEVC decoder can rely on the base layer only, hence providing backward compatibility, but a more elaborate decoder can use the second layer as well to improve the rendition.
The MV-HEVC format gained widespread attention with Apple announcing support for 3D movies on the Apple Vision Pro at WWDC23 [3] [4]. Apple provided public guidelines on stereo video encoding [5], packaging [6], and streaming [7] as well as on providing parallax information to improve caption placement [8].
This presentation will describe the MV-HEVC standard and how it works in theory. An implementation of MV-HEVC [9] will also be studied to assess performance in practice in industrial settings. Deployment aspects such as packaging and captioning will be discussed. Finally, 3D-specific optimization strategies will be analyzed, including software architecture and rate control.
Thomas Guionnet| ATEME | Rennes, France
Khaled Jerbi | ATEME | Rennes, France
Thomas Burnichon| ATEME | Rennes, France
Mickaël Raulet | ATEME | Rennes, France
New Data Management Techniques and Technologies to Accelerate Live Broadcast Events - $15
Date: April 3, 2024Topics: 2024 BEITC Proceedings, Striving for Efficiency in Video TechnologyToday’s live event broadcast workflows present unique challenges for storage solutions and data management. These multi-camera environments need high performance storage to keep up with the high bandwidth of streaming data, which can total anywhere from 30TB to hundreds of TBs per event. Not to mention that data then needs to be captured in real time, stored, and moved to different locations, throughout the post-production and delivery process. Factor in ever-increasing video capture resolutions and frame rates, as well as infrastructure and bandwidth variations from location to location—it’s no surprise many professionals in the industry are finding data management an increasingly difficult task. Especially as this type of production environment has rigid and often short timelines that the crew has to operate within. In short, a new kind of data storage solution is long overdue.
To address this need, broadcast teams are turning to Data-Transfer-As-A-Services (DTaaS) to support these live environment productions.
Competitive DTaaS services with enterprise removable storage options enable customers to pay for high capacity, 100 terabyte storage arrays with high performance that meet data challenges head on, enabling users to store data anywhere – while only paying for the hardware each specific project needs – and physically transport that data securely to their post-production landing destination of choice, allowing users to choose secure physical transport or rapid ingest to any S3 destination. Not only does an on-demand, consumption-as-a-service model simplify users’ device management, preventing unnecessary on-set IT costs, it also gives video production teams the flexibility they need to alter the numbers of devices they deploy to complete certain projects, as storage needs change. By circumnavigating the headaches that come with owning data storage infrastructure outright, like maintenance fees and technology upgrades, DTaaS-based strategies give users the freedom to expand their production, accelerate their timeline, and even reallocate their budget without having to worry about your data.
This technical paper will explore how DTaaS accelerates live capture workflows resulting in cost savings and streamlined content collaboration.
Preliminary research and application results include:
DAS performance can enable in-field, direct editing and transcoding when logistics prevent the standard workflow
Cloud import services quickly and securely move event media from camera to the cloud of choice for collaborative post-production workflows
Secure, encrypted, rugged solution, ensure customers won’t lose or leak content during transport for greater peace of mind
Transfer content in days, not weeks: high-performance storage arrays on-set can save up to 12-15 hours per week of overtime costs across multiple departmentsIn an industry as fast paced as broadcast streaming, data storage capacity, security, and mobility should be the least of a production team’s concerns. With DTaaS, IT professionals can trust their enterprise-level video media will get where they need to go, right on time.
Jonathan Bauder | Seagate Technology | Fremont, Calif., United States
Next-Generation Token to Fight Piracy and CDN Leeching - $15
Date: April 3, 2024Topics: 2024 BEITC Proceedings, Cybersecurity for BroadcastersContent Delivery Networks (CDNs) are widely used to stream video efficiently across the Internet, but they face persistent challenges from unauthorized access attempts and content piracy, a phenomenon known as CDN leeching. Traditional token-based authentication mechanisms, often relying on long-lived tokens with fixed expiration times, have proven inadequate in addressing these security threats. In response, we propose a novel randomized token rejection mechanism to bolster the security of CDN architectures against unauthorized access attempts. By dynamically adjusting rejection probabilities based on token properties and operational policies, our mechanism enhances CDN resilience while minimizing operational overhead. Furthermore, we present refinements to our approach, including geographic probability of rejection, content-sensitivity, and frequency analysis, to tailor security policies to specific content access patterns. Evaluation results demonstrate the efficacy of our randomized token rejection mechanism in mitigating piracy risks while maintaining operational efficiency. While further research is needed to implement fully transparent token renewal schemes, the current technologies, typically based on the Common Access Token (CAT), already enable the implementation of the randomized token rejection mechanism.
Gwendal Simon | Synamedia | Rennes, France
Gwenaël Doërr | Synamedia | Rennes, France
Objective Evaluation System for 2K Video Quality Transcoded from 4K Source using Dynamic Thresholds and Statistical Methods - $15
Date: April 3, 2024Topics: 2024 BEITC Proceedings, Quantifying Quality in Video TechnologyObjective evaluation methodologies can be applied only to comparing images of the same resolutions, leaving several problems in assessing video quality across different resolutions. Our proposed method employs a spatial two-dimensional filter to enable pixel-by-pixel comparisons for automating the evaluation of 2K video quality transcoded from 4K. Detection thresholds are adapted for spatially and temporally complex videos, minimizing false alarms. For noise detection in specific regions, a small-area PSNR metric is applied with a statistical approach for enhanced accuracy. This automated approach matches or exceeds human review accuracy, offering cost reduction and improved quality assurance in 4K/2K productions.
Nariaki Takahashi | Japan Broadcasting Corporation | Shibuya, Tokyo, Japan
Keita Kataoka | Japan Broadcasting Corporation | Shibuya, Tokyo, Japan
Satellite/Internet Hybrid Content Delivery (Ku-Band Without Rain Fade!) - $15
Date: April 3, 2024Topics: 2024 BEITC Proceedings, Content Creation and Delivery TechnologySatellite distribution is the ideal way to send the same content to many locations that are geographically distributed. However, the traditional C-band distribution is now impacted by the new 5G cellular network, which operates mostly in the 3.3 to 3.6 GHz band (which partially overlaps with the Space-to-Earth spectrum). Ku-band is not impacted by 5G, but unfortunately it is subject to rain fade. One possible approach to solving this content delivery problem is to augment satellite delivery using the Internet. The basic idea is to use the satellite for the “heavy lifting” (transmitting as much data as viable), with the Internet to “fill in the gaps”. In other words, any data that is corrupted or lost in transit is retransmitted over the Internet, and only to the locations that need it. This way, if a region is experiencing any sort of fade or interference, only the receivers in that region need to use the Internet. Any solution to this problem needs to be designed in such a way that it can co-exist with current receivers. This means that the signal transmitted to the satellite either cannot change, or any changes to it must be backward-compatible with existing receivers. Broadcasters utilizing this solution can then gradually deploy the solution as needed, with high-priority sites being upgraded first. The Reliable Internet Stream Transport (RIST) Activity Group in the Video Services Forum (VSF) has been working on this problem, to create a common industry Specification for addressing it. Having such a Specification will allow interoperability between solutions, and not lock broadcasters into a single vendor. This Specification is expected to be published as VSF TR-06-4 Part 6, sometime in 2024. This paper is a technical overview and discussion of the techniques and algorithms which are being considered for inclusion in the VSF Satellite/Hybrid Specification.
Ciro Noronha | Cobalt Digital Inc. | Champaign, Illinois, United States
Solving Chaos in the Newsroom - $15
Date: April 3, 2024Topics: 2024 BEITC Proceedings, Content Creation and Delivery TechnologyThis presentation explores how implementing cloud-based news storytelling can streamline news operations to manage chaos and optimize ROI. It will cover trends in technology and newsroom working practices.
Stephane Guez | Dalet | New York, N.Y., United States
Solving Localization & Compliance Pain Points with AI - $15
Date: April 3, 2024Topics: 2024 BEITC Proceedings, Generative AI for MediaThis paper investigates the challenges of meeting localization and compliance needs in the global video streaming industry. It highlights the complexities of managing multiple languages, regulations, and cultural nuances. It introduces AI as a game-changer, highlighting its potential to automate processes, enhance quality, and assist with accessibility. The paper presents a comprehensive overview of the benefits of AI, including improved translation accuracy, faster content delivery, synthesized and cloned voice technologies, cost savings, and enhanced viewer engagement. It also discusses the challenges and considerations associated with implementing AI solutions, such as data privacy, regulatory compliance, and the need for human oversight. Overall, the paper provides valuable insights for content owners and video streaming providers seeking to optimize their localization and compliance strategies in a rapidly evolving global market.
Bill Admans | Ateliere Creative Technologies | Century City, Calif., United States
Dan Goman | Ateliere Creative Technologies | Century City, Calif., United States
Speech Intelligibility and Audio Monitoring in OTT - $15
Date: April 3, 2024Topics: 2024 BEITC Proceedings, Generative AI Uses and Video TranscodingThe paper details ways of controlling speech intelligibility in OTT and broadcast production, primarily based on a new metric, Loudness to Dialog Ratio (LDR), and on calibrated sound monitoring at a moderate listening level. Principles of Loudness normalization are described; as is a study to define and quantify LDR for use in broadcast and OTT. Findings may be applied during mixing or dubbing, used for default settings in NGA delivery, and in automated QC as a key machine learning parameter. The article is facts-based and free of commercial bias.
Thomas Lund | Genelec Inc | Natick, Mass., United States