Showing posts with label avc. Show all posts
Showing posts with label avc. Show all posts

Monday, June 4, 2018

ACM MMSys 2018: Multi-Codec DASH Dataset

Multi-Codec DASH Dataset

Anatoliy Zabrovskiy (Petrozavodsk State University & Alpen-Adria-Universität Klagenfurt), Christian Feldmann (Bitmovin Inc.), and Christian Timmerer (Alpen-Adria-Universität Klagenfurt & Bitmovin Inc.)

Abstract: The number of bandwidth-hungry applications and services is constantly growing. HTTP adaptive streaming of audio-visual content accounts for the majority of today's internet traffic. Although the internet bandwidth increases also constantly, audio-visual compression technology is inevitable and we are currently facing the challenge to be confronted with multiple video codecs.

This paper proposes a multi-codec DASH dataset comprising AVC, HEVC, VP9, and AV1 in order to enable interoperability testing and streaming experiments for the efficient usage of these codecs under various conditions. We adopt state of the art encoding and packaging options and also provide basic quality metrics along with the DASH segments. Additionally, we briefly introduce a multi-codec DASH scheme and possible usage scenarios. Finally, we provide a preliminary evaluation of the encoding efficiency in the context of HTTP adaptive streaming services and applications.

Tuesday, July 28, 2015

MPEG news: a report from the 112th meeting, Warsaw, Poland



This blog post is also available at at bitmovin tech blog and SIGMM records.

The 112th MPEG meeting in Warsaw, Poland was a special meeting for me. It was my 50th MPEG meeting which roughly accumulates to one year of MPEG meetings (i.e., one year of my life I've spend in MPEG meetings incl. traveling - scary, isn't it? ... more on this in another blog post). But what happened at this 112th MPEG meeting (my 50th meeting)...

  • Requirements: CDVA, Future of Video Coding Standardization (no acronym yet), Genome compression
  • Systems: M2TS (ISO/IEC 13818-1:2015), DASH 3rd edition, Media Orchestration (no acronym yet), TRUFFLE
  • Video/JCT-VC/JCT-3D: MPEG-4 AVC, Future Video Coding, HDR, SCC
  • Audio: 3D audio
  • 3DG: PCC, MIoT, Wearable
MPEG Friday Plenary. Photo (c) Christian Timmerer.
As usual, the official press release and other publicly available documents can be found here. Let's dig into the different subgroups:
Requirements

In requirements experts were working on the Call for Proposals (CfP) for Compact Descriptors for Video Analysis (CDVA) including an evaluation framework. The evaluation framework includes 800-1000 objects (large objects like building facades, landmarks, etc.; small(er) objects like paintings, books, statues, etc.; scenes like interior scenes, natural scenes, multi-camera shots) and the evaluation of the responses should be conducted for the 114th meeting in San Diego.

The future of video coding standardization is currently happening in MPEG and shaping the way for the successor of of the HEVC standard. The current goal is providing (native) support for scalability (more than two spatial resolutions) and 30% compression gain for some applications (requiring a limited increase in decoder complexity) but actually preferred is 50% compression gain (at a significant increase of the encoder complexity). MPEG will hold a workshop at the next meeting in Geneva discussing specific compression techniques, objective (HDR) video quality metrics, and compression technologies for specific applications (e.g., multiple-stream representations, energy-saving encoders/decoders, games, drones). The current goal is having the International Standard for this new video coding standard around 2020.

MPEG has recently started a new project referred to as Genome Compression which is about of course about the compression of genome information. A big dataset has been collected and experts working on the Call for Evidence (CfE). The plan is holding a workshop at the next MPEG meeting in Geneva regarding prospect of Genome Compression and Storage Standardization targeting users, manufactures, service providers, technologists, etc.

Systems


Summer in Warsaw. Photo (c) Christian Timmerer.
The 5th edition of the MPEG-2 Systems standard has been published as ISO/IEC 13818-1:2015 on the 1st of July 2015 and is a consolidation of the 4th edition + Amendments 1-5.

In terms of MPEG-DASH, the draft text of ISO/IEC 23009-1 3rd edition comprising 2nd edition + COR 1 + AMD 1 + AMD 2 + AMD 3 + COR 2 is available for committee internal review. The expected publication date is scheduled for, most likely, 2016. Currently, MPEG-DASH includes a lot of activity in the following areas: spatial relationship description, generalized URL parameters, authentication, access control, multiple MPDs, full duplex protocols (aka HTTP/2 etc.), advanced and generalized HTTP feedback information, and various core experiments:
  • SAND (Sever and Network Assisted DASH)
  • FDH (Full Duplex DASH)
  • SAP-Independent Segment Signaling (SISSI)
  • URI Signing for DASH
  • Content Aggregation and Playback COntrol (CAPCO)
In particular, the core experiment process is very open as most work is conducted during the Ad hoc Group (AhG) period which is discussed on the publicly available MPEG-DASH reflector.

MPEG systems recently started an activity that is related to media orchestration which applies to capture as well as consumption and concerns scenarios with multiple sensors as well as multiple rendering devices, including one-to-many and many-to-one scenarios resulting in a worthwhile, customized experience.

Finally, the systems subgroup started an exploration activity regarding real-time streaming of file (a.k.a TRUFFLE) which should perform an gap analysis leading to extensions of the MPEG Media Transport (MMT) standard. However, some experts within MPEG concluded that most/all use cases identified within this activity could be actually solved with existing technology such as DASH. Thus, this activity may still need some discussions...

Video/JCT-VC/JCT-3D

The MPEG video subgroup is working towards a new amendment for the MPEG-4 AVC standard covering resolutions up to 8K and higher frame rates for lower resolution. Interestingly, although MPEG most of the time is ahead of industry, 8K and high frame rate is already supported in browser environments (e.g., using bitdash 8K, HFR) and modern encoding platforms like bitcodin. However, it's good that we finally have means for an interoperable signaling of this profile.

In terms of future video coding standardization, the video subgroup released a call for test material. Two sets of test sequences are already available and will be investigated regarding compression until next meeting.

After a successful call for evidence for High Dynamic Range (HDR), the technical work starts in the video subgroup with the goal to develop an architecture ("H2M") as well as three core experiments (optimization without HEVC specification change, alternative reconstruction approaches, objective metrics).

The main topic of the JCT-VC was screen content coding (SCC) which came up with new coding tools that are better compressing content that is (fully or partially) computer generated leading to a significant improvement of compression, approx. or larger than 50% rate reduction for specific screen content.

Audio

The audio subgroup is mainly concentrating on 3D audio where they identified the need for intermediate bitrates between 3D audio phase 1 and 2. Currently, phase 1 identified 256, 512, 1200 kb/s whereas phase 2 focuses on 128, 96, 64, 48 kb/s. The broadcasting industry needs intermediate bitrates and, thus, phase 2 is extended to bitrates between 128 and 256 kb/s.

3DG

MPEG 3DG is working on point cloud compression (PCC) for which open source software has been identified. Additionally, there're new activity in the area of Media Internet of Things (MIoT) and wearable computing (like glasses and watches) that could lead to new standards developed within MPEG. Therefore, stay tuned on these topics as they may shape your future.

The week after the MPEG meeting I met the MPEG convenor and the JPEG convenor again during ICME2015 in Torino but that's another story...
L. Chiariglione, H. Hellwagner, T. Ebrahimi, C. Timmerer (from left to right) during ICME2015. Photo (c) T. Ebrahimi.



Tuesday, May 25, 2010

New release of H.264/AVC reference software JM 17.1

-- this blog post is based on the email sent by Karsten Sühring to the JVT reflector

A new release of the JM reference software is available at http://iphome.hhi.de/suehring/tml/

This is a quite large maintenance release: It contains a number of bug fixes including a decoder issue that was introduced in JM 17.0 and caused a mismatch for MBAff and constrained intra prediction.

The documentation has been updated with a description of the new MVC coding mode. Many people experienced crashes when enabling MVC with the default config file. This was caused by an incompatible setting which is now prohibited for the muti-view encoding mode. We have also added a Stereo High profile example config (encoder_stereo.cfg) and a modified default file for faster encoding (encoder_fast.cfg).

The VS 2008 and XCode workspaces have been updated for better x64 architecture support. OPENMP (for parallel deblocking) is now enabled by default (also in the Makefiles).

-------------------------------------------------------------------------
Changes in Version JM 17.1 (since JM 17.0)
-------------------------------------------------------------------------
- decoder: fix delta_pic_order_cnt[0] in IDR frames (KS, bug #57)
- decoder: Direct mode fixes (YHE)
- encoder/decoder: various cleanups and bug fixes (AT/YHE)
- encoder/decoder: rewrite of deblocking functions (AT)
- decoder : rewrite of intra prediction functions (AT)
- encoder/decoder: MVC bug fixes (PeP/AL)
- encoder/decoder: Frame Packing SEI Message (AL)
- decoder: Minor CABAC updates (AT)
- encoder/decoder: Update work spaces for VS 2008 and XCode with x64 and OPENMP support (KS)
- encoder/decoder: Update Makefiles for OPENMP support (KS)
- encoder: avoid crashes by prohibiting incompatible tool combinations (KS/YHE)

Monday, July 27, 2009

MPEG Global Conference points the way to Ultra HD online services

London meeting sees significant improvement in compression for High Performance Video Coding

London, United Kingdom – The 89th MPEG meeting was held in London, United Kingdom from the 29th of June to the 3rd of July 2009.

Highlights of the 89th Meeting

Responses for Evidence Evaluated for HVC
During its 89th meeting, MPEG evaluated responses that were received on the Call for Evidence on High-Performance Video Coding (HVC), issued to obtain evidence of video coding technology providing compression capability clearly higher than that provided by the existing AVC standard (ITU-T H.264 | ISO/IEC 14496-10). Significant gains in compression were found when an assessment was made based on information brought by the contributors. A subjective comparison was performed in a blind test with a set of video test sequences encoded by the AVC High Profile at matching rate points. Gains were demonstrated for test cases ranging from resolutions as low as 412x240 pixels (Wide QVGA) up to resolutions for ultra-high definition. MPEG has therefore concluded that the development of the next generation of video compression technology is to be started with the issuing of a formal Call for Proposals by the next meeting.

AVC Extended with New Profiles for Baseline and MVC Technologies

At the 89th meeting, the AVC standard (ITU-T H.264 | ISO/IEC 14496-10) was further extended with the issuing of a Final Draft Amendment (FDAM) ballot containing the specification of two new profiles and new supplemental enhancement information. The first of the new profiles is the Constrained Baseline Profile, which forms the maximally-interoperable set of coding tools from the most widely deployed of existing profiles (the Baseline and High Profiles). The second new profile is a special case of multivew video coding (MVC) called the Stereo High Profile. The Stereo High profile enables all of the coding tools of the High Profile along with inter-view prediction capability for two-view (stereo) video applications such as 3D entertainment video.

Additionally, a new supplemental enhancement information (SEI) message has been defined for AVC. This new message – called the frame packing arrangement SEI message – enables the encoder to indicate to the decoder how to extract two distinct views of a video scene from a single decoded frame. The message also serves as a way to support stereo-view video in applications that require full compatibility with prior decoder designs that are not capable of supporting the new Stereo High Profile.

MPEG Promotes Technologies to link Real and Virtual Worlds

At its 88th meeting, MPEG had published a new call for proposals (N10526) with updated requirements (N10235) for an extension of the Media Context and Control project.

The technical contributions related to haptic and tactile devices, emotions, and virtual goods received at its 89th meeting have enabled MPEG to build a complete framework for defining haptic properties on top of virtual objects and to control haptic devices. This is now part of ISO/IEC 23005 or MPEG-V, a standard (formerly called Information Exchange with Virtual Worlds) providing a global framework and associated data representations to enable the interoperability between different virtual worlds (e.g. a digital content provider of a virtual world, a game with the exchange of real currency, or a simulator) and between virtual worlds and the real world (sensors, actuators, robotics, travel, real estate, or other physical systems).

MPEG Progresses Media Context and Control Project

MPEG has also advanced to the Committee Draft stage four parts of MPEG-V. The first part describes the architecture of the standard. The second part, “Control Information”, provides metadata representation of device capabilities and user preferences to be used for the information exchange between a controlling device and the real actuator or sensors. The third part, “Sensory Information”, provides metadata to represent sensory effects such as temperature, wind, vibration, fog, and more. The fourth part, “Avatar Characteristics”, provides metadata to commonly represent information about Avatars for the exchange of virtual characters between virtual worlds.

MPEG Hosts MXM Developer’s Day

The first MXM Developer’s Day workshop has been hosted by MPEG during its 89th meeting. The workshop featured demonstrations by companies and organisations that are developing MXM standards and applications. MXM, currently at its Final Committee Draft stage, provides specifications of APIs and an open source implementation (released under the BSD licence) to access various MPEG standards for easy deployment of applications. In this workshop detailed information about the APIs currently under standardization has been provided and several interesting demonstrations with the potential to create new business opportunities have also been presented. More information about this workshop can be found at http://mxm.wg11.sc29.org.

Rich Media User Interface Moves toward Completion

At its 89th meeting, MPEG has also advanced MPEG Rich Media UI (ISO/IEC 23007 or MPEG-U), to the Committee Draft stage. MPEG-U standardizes widget packaging, delivery, representation and communication formats. In its current draft, MPEG-U adopts and extends the W3C widget representation to provide a complete framework that can be used also in a non-Web based environment without a browser. Additionally, this standard enables communication among widgets on the same device or different devices, and other applications to better support connected environments.

Visual Signatures Enable New Applications

MPEG’s Visual Signatures define the world’s first standardized tools for content-based identification of any visual content even in very large databases, e.g. on the web. These tools enable a range of new applications including semantic linking, library management, metadata association (e.g. title, photographer, director, etc.) and content usage tracking. In the same way that a fingerprint or signature identifies a person, a Visual Signature is a compact descriptor uniquely representing either an image or video. The descriptor is derived directly from analysis of the visual content and is robust to heavy compression and editing.

The Image Signature and Video Signature are two separate amendments to MPEG-7. Collectively the two amendments are referred to as the MPEG-7 Visual Signatures. At the London meeting, the Video Signature advanced to the Proposed Draft Amendment (PDAM) stage with a target completion date of July 2010. The Image Signature was published as an ISO/IEC standard in April 2009.

Mobile Services to Be Enhanced by New BIFS Profile
At this meeting, MPEG advanced the new BInary Format for Scenes (BIFS) profile to the Committee Draft stage by incorporating additional nodes and technologies submitted as responses to the Call for Proposals for new BIFS technologies. The requirements for this profile (provided in N10567) originated from organizations of various industries and SDOs for digital radio and mobile television broadcasting. This profile will enable the development of more efficient and enhanced interactive services for mobile broadcasting services including digital radio or mobile television on small handheld devices. Moreover, it is backward compatible with Core2D@Level1 which is widely adopted by the industry.

Contact MPEG

Digging Deeper Once Again
Communicating the large and sometimes complex array of technology that the MPEG Committee has developed is not a simple task. The experts past and present have contributed a series of white-papers that explain each of these standards individually. The repository is growing with each meeting, so if something you are interested is not there yet, it may appear there shortly – but you should also not hesitate to request it. You can start your MPEG adventure at: http://www.chiariglione.org/mpeg/mpeg-tech.htm

Ends

Further Information
Future MPEG meetings are planned as follows:
No. 90, Xian, CN, 26-30 October, 2009
No. 91, Kyoto, JP, 18-22 January, 2010
For further information about MPEG, please contact:
Dr. Leonardo Chiariglione (Convener of MPEG, Italy)
Via Borgionera, 103
10040 Villar Dora (TO), Italy
Tel: +39 011 935 04 61
Email: mailto:leonardo@chiariglione.org
or
Dr. Arianne T. Hinds
Ricoh | IBM InfoPrint Solutions Company
6300 Diagonal Highway, MS 04N
Boulder, CO 80301, USA
Tel +1 720 663 3565
Email: arianne.hinds@infoprint.com

This press release and other MPEG-related information can be found on the MPEG homepage:
http://www.chiariglione.org/mpeg
The text and details related to the Calls mentioned above (together with other current Calls) are in the Hot News section, http://www.chiariglione.org/mpeg/hot_news.htm. These documents include information on how to respond to the Calls.
The MPEG homepage also has links to other MPEG pages which are maintained by the MPEG subgroups. It also contains links to public documents that are freely available for download by those who are not MPEG members. Journalists that wish to receive MPEG Press Releases by email should contact Dr. Arianne T. Hinds using the contact information provided above.

Friday, October 17, 2008

MPEG got an Emmy!

At this week's MPEG meeting, MPEG celebrated the JVT receipt of the 2008 Primetime Emmy Engineering Award. They've received it for its work on the Advanced Video Coding (AVC) High Profile. The actual awards - three of them - are now hosted at the three main standardization bodies who worked on AVC, namely IEC, ISO, and ITU. However, somehow Gary Sullivan managed to get an additional Emmy statue for MPEG and presented it to Leonardo Chiariglione (MPEG convenor). The MPEG Emmy statue will travel from national body to national body based on MPEG's meeting schedule. That is, the next chance to see the Emmy statue live will be in Lausanne, CH at the 87th MPEG meeting.

During the Friday plenary I also got the chance to take a picture of/with the Emmy statue ;-) Another picture can be found here.