I made this tutorial at Web3D 2012 conference. It provides MPEG position to AR, technologies currently used, as well as explanations on how to set up AR applications.
DevEX - reference for building teams, processes, and platforms
MPEG Augmented Reality Tutorial
1. MPEG Augmented Reality Tutorial
Web3D Conference, August 4-5, Los Angeles, CA
Marius Preda, MPEG 3DG Chair
Institut Mines TELECOM
http://www.slideshare.net/MariusPreda/mpeg-augmented-reality-tutorial
2. MPEG Augmented Reality Tutorial
Topics of the day
What is MPEG?
MPEG offer in the Augmented Reality field
MPEG-A Part 14 Augmented Reality Reference Model
MPEG-A Part 13 Augmented Reality Application Format
3. MPEG Augmented Reality Tutorial
Topics of the day
What is MPEG?
MPEG offer in the Augmented Reality field
MPEG-A Part 14 Augmented Reality Reference Model
MPEG-A Part 13 Augmented Reality Application Format
4. What is MPEG?
A suite of ~130 ISO/IEC standards
Coding/compression of elementary media:
– Audio (MPEG-1, 2 and 4)
– Video (MPEG-1, 2 and 4)
– 2D/3D graphics (MPEG-4)
Storage and Transport
– MPEG-2 Transport
– File Format (MPEG-4)
– Dynamic Adaptive Streaming over HTTP (DASH)
Hybrid (natural & synthetic) scene description, user interaction (MPEG-4)
Metadata (MPEG-7)
Media management and protection (MPEG-21)
Sensors and actuators, virtual worlds (MPEG-V)
Advanced User interaction (MPEG-U)
Media-oriented middleware (MPEG-M)
More ISO/IEC standards under development for
– 3D Video, 3D Audio
– Coding and Delivery in Heterogeneous Environments
– …
5. What is MPEG?
Involvement, approach, deployment
A standardization activity continuing for 24 years
– Supported by several hundreds companies/organisations from ~25 countries
– ~500 experts participating in quarterly meetings
– More than 2300 active contributors
– Many thousands experts working in companies
A proven manner to organize the work to deliver useful and used standards
– Developing standards by integrating individual technologies
– Well defined procedures
– Subgroups with clear objectives
– Ad hoc groups continuing coordinated work between meetings
MPEG standards are widely referenced by industry
– 3GPP, ARIB, ATSC, DVB, DVD-Forum, BDA, EITSI, SCTE, TIA, DLNA, DECE, OIPF…
Billions of software and hardware devices built on MPEG technologies
– MP3 players, cameras, mobile handsets, PCs, DVD/Blue-Ray players, STBs, TVs,
…
6. MPEG Augmented Reality Tutorial
Topics of the day
What is MPEG?
MPEG offer in the Augmented Reality field
MPEG-A Part 14 Augmented Reality Reference Model
MPEG-A Part 13 Augmented Reality Application Format
7. MPEG technologies related to AR
1992/4 1997 1998 1999
MPEG-1/2
(AV content) VRML
MPEG-4 v.1
• Part 11 - BIFS:
-Binarisation of VRML
-Extensions for streaming
-Extensions for server command
-Extensions for 2D graphics
- Real time augmentation with
audio & video
• Part 2 - Visual:
- 3D Mesh compression MPEG-4 v.2
- Face animation • Part 2 – Visual
- Body animation
First form of broadcast signal augmentation
8. MPEG technologies related to AR
2003 2005 2007 2011
MPEG-4
• Part 16 - AFX:
- A rich set of 3D
MPEG-4
graphics tools
•AFX 2nd Edition:
- Compression of
- Animation by
geometry,
morphing
appearance,
- Multi-texturing
animation MPEG-4
• AFX 3rd Edition
- WSS for terrain
and cities
A rich set of 3D Graphics - Frame based MPEG-4
representation and animation • AFX 4th Edition
compression tools - Scalable complexity
mesh coding
9. MPEG technologies related to AR
2003 2004 2005 2007 2009 2011
MPEG-4
• Part 16 - AFX:
- A rich set of 3D
MPEG-4
graphics tools
•AFX 2nd Edition:
- Compression of
- Animation by
geometry,
morphing
appearance,
- Multi-texturing
animation MPEG-4
• AFX 3rd Edition
- WSS for terrain
and cities
MPEG-4 MPEG-4
- Frame based MPEG-4
• Part 16 • Part 25
animation • AFX 4th Edition
- X3D - Compression of
- Scalable complexity
Interactive third-party XML
mesh coding
Profile (X3D, COLLADA)
10. MPEG technologies related to AR
2011 2012 201x 201x 201x
MPEG-V - Media
Context and Control MPEG-U –
• 1st Edition Advanced
- Sensors and User Interface MPEG-V
actuators • 2nd Edition:
- Interoperability - GPS
between Virtual - Biosensors
Worlds - 3D Camera
MPEG-H
- 3D Video
• Compression
A rich set of sensors and CDVS
of video +
actuators • Feature-point based
depth
- 3D Audio descriptors for image
recognition
11. Main features of MPEG AR technologies
All AR-related data is available from MPEG standards
Real time composition of synthetic and natural objects
Access to
– Remotely/locally stored BIFS/compressed 2D/3D mesh objects
– Streamed real-time BIFS/compressed 2D/3D mesh objects
Inherent object scalability (e.g. for streaming)
User interaction & server generated scene changes
Physical context
– Captured by a broad range of standard sensors
– Affected by a broad range of standard actuators
12. MPEG vision on AR, the MPEG AR Browser
Point to a URL – no need to download new applications for each context.
The browser
– Retrieves scenario from the internet
– Starts video acquisition
– Tracks objects
– Recognizes objects from visual signatures
– Recovers camera pose
– Gets streamed 3D graphics
– Composes new scenes
– Gets inputs from various sensors
– Offers optimal AR experience by constantly adapting interaction possibilities
and objects from a remote server.
Industry
– Maximize number of customers through MPEG-compliant authoring tools and
browsers
– No need to develop a new application for each use case and device platform
13. MPEG vision on AR
Produce
Download
Compression
Authoring Tool
MPEG-4/MPEG-7/MPEG-21/
MPEG-U/MPEG-V
MPEG Player
14. Architecture
Remote Remote Local Local
Real World Sensors & Sensors & Real World
Environment Actuators Actuators Environment
AR Player User
AR file or
stream
Media Service
Servers Servers
15. MPEG ongoing work on AR
ISO/IEC 23000-14 Augmented Reality Reference Model
– WD stage, collaborating with SC24/WG9, ARStandards, OGC, Khronos,
Web3D
ISO/IEC 23000-13 Augmented Reality Application Format
– CD stage, based on MPEG standards
16. MPEG Augmented Reality Tutorial
Topics of the day
What is MPEG?
MPEG offer in the Augmented Reality field
MPEG-A Part 14 Augmented Reality Reference Model
MPEG-A Part 13 Augmented Reality Application Format
17. Augmented Reality Reference Model
WD2.0 content
Viewpoints Glossary
Community Objectives
Enterprise
Viewpoint
Abstract/Design
Information Computational
Viewpoint Viewpoint
Implementation/Development
Use cases
Engineering - Guide
Viewpoint
- Create
Technology
- Play
Viewpoint
18. Augmented Reality Reference Model
Enterprise viewpoint: global architecture and actors
Local / Remote Context
MCP
ARTC
TO Telecommunication Operator (TO)
End-User (EU)
DM MCP
AREC
AR EU
TO AR Player User
Document
Device Manufacturer (DM)
AC TO TO Middleware/Component Provider (MCP)
AR Tools Creator (ARTC)
AR Experience Creator (AREC) Online Middleware/Component
Assets Creator (AC) Media Provider (OMCP)
Service
AR Service Provider (ARSP)
Servers Servers
Assets Aggregator (AA)
AC AA OMCP ARSP
19. Augmented Reality Reference Model
Information viewpoint
Local/Remote Context Scene/Real World
Device Context • Raw image
• Device capabilities • Sensed data
Location of Device • Virtual Camera view
• Location • Detected features
• Orientation • Area of Interest/Anchors
Presentation AR Player Spatial Models
• • Coordinate Ref. Sys.
Augmentation
• (Geol)ocation
User
• Registration
AR • Styling/complexity • Projections
• • Coordinate conversion
Document Spatial Filtering, e.g. User Input
range
Tracking objects • Query
• Markers • Manipulation of
• Marker-less Presentation
• Topics of interest
• Preferences
Digital Assets
• Presentation data Media Service
• Trigger/Event rules Servers Servers
• Accuracy based
20. Augmented Reality Reference Model
Computational viewpoint
Local / Remote Context
2
1 5
AR
AR Player User
Document
3 4
Media Service
Servers Servers
21. Augmented Reality Reference Model
Computational viewpoint
Local / Remote Context
2
1 3
AR
AR Player User
Document
4 5
Media Service
Servers Servers
22. Augmented Reality Reference Model
Engineering viewpoint
Local/Remote Context
Accelero
Camera Mic Compass GPS …
meter
AR Player
User
AR Rendering Display Application
Document …
Engine (A/V/H) Engine
Media Service
Servers Servers
26. MPEG Augmented Reality Tutorial
Topics of the day
What is MPEG?
MPEG offer in the Augmented Reality field
MPEG-A Part 14 Augmented Reality Reference Model
MPEG-A Part 13 Augmented Reality Application Format
27. MPEG-A Part 13 ARAF
3 components: scene, sensors/actuators, medias
A set of scene graph nodes/protos as defined in MPEG-4 Part 11
– Existing nodes
– Audio, image, video, graphics, programming, communication, user
interactivity, animation
– New standard PROTOs
– Map, MapMarker, Overlay, ReferenceSignal,
ReferenceSignalLocation, CameraCalibration, AugmentedRegion
Connection to sensors as defined in MPEG-V
– Orientation, Position, Angular Velocity, Acceleration, GPS, Geomagnetic,
Altitude
– Local camera sensor
Compressed media
28. MPEG-A Part 13 ARAF
Scene: 63 XML Elements
Node, Protos / Elements Node, Protos / Elements
Category Sub-category Category Sub-category
name in MPEG-4 BIFS / XMT name in MPEG-4 BIFS / XMT
AudioSource AugmentationRegion
Audio Sound Background
Sound2D Background2D
Image and ImageTexture CameraCalibration
video MovieTexture Group
Textual FontStyle Inline
information Text Layer2D
Appearance Scene related Layer3D
Color information Layout
LineProperties (spatial and NavigationInfo
LinearGradient temporal OrderedGroup
Material relationships) ReferenceSignal
Elementary media Material2D ReferenceSignalLocation
Rectangle Switch
Shape Transform
SBVCAnimationV2 Transform2D
Graphics
SBBone Viewpoint
SBSegment Viewport
SBSkinnedModel Form
MorphShape OrientationInterpolator
Coordinate ScalarInterpolator
TextureCoordinate Dynamic and CoordinateInterpolator
Normal animated scene ColorInterpolator
IndexedFaceSet PositionInterpolator
IndexedLineSet Valuator
Programming Script BitWrapper
InputSensor MediaControl
Communication
SphereSensor Map
and compression
TimeSensor Maps MapOverlay
User interactivity
TouchSensor MapMarker
MediaSensor Terminal TermCap
PlaneSensor
29. MPEG-A Part 13 ARAF
Scene: the distance between ARAF and X3D is 32 (XML Elements)
Node, Protos / Elements Node, Protos / Elements
Category Sub-category Category Sub-category
name in MPEG-4 BIFS / XMT name in MPEG-4 BIFS / XMT
AudioSource AugmentationRegion
Audio Sound Background
Sound2D Background2D
Image and ImageTexture CameraCalibration
video MovieTexture Group
Textual FontStyle Inline
information Text Layer2D
Appearance Scene related Layer3D
Color information Layout
LineProperties (spatial and NavigationInfo
LinearGradient temporal OrderedGroup
Material relationships) ReferenceSignal
Elementary media Material2D ReferenceSignalLocation
Rectangle Switch
Shape Transform
SBVCAnimationV2 Transform2D
Graphics
SBBone Viewpoint
SBSegment Viewport
SBSkinnedModel Form
MorphShape OrientationInterpolator
Coordinate ScalarInterpolator
TextureCoordinate Dynamic and CoordinateInterpolator
Normal animated scene ColorInterpolator
IndexedFaceSet PositionInterpolator
IndexedLineSet Valuator
Programming Script BitWrapper
InputSensor MediaControl
Communication
SphereSensor Map
and compression
TimeSensor Maps MapOverlay
User interactivity
TouchSensor MapMarker
MediaSensor Terminal TermCap
PlaneSensor
30. ark
MPEG-ANPartoung We: P13 ARAF
am Y
-
hu Signal F
Scene:: Reference it ion
C :
s
Po m : FC
ea nal:
T e
Ars
age
I m
ce
r en
M arker Tracking
Name: Park
Chu- Young
fe
Position: FW
Team:
Re
Arsenal: FC
M arker
Ref erence Image
3D graphic Synchronized wit h
movement of marker image
38. MPEG-A Part 13 ARAF
Sensors/Actuators
MPEG-4 Player Scene
mapping of
MPEG-4 Scene captured data
MPEG-V
Sensor 1
InputSensor 1 Acceleration Sensor
MPEG-V Orientation Sensor
Compositor
Sensor 2 InputSensor 2 Screen Angular Velocity
Global Position Sensor
MPEG-V InputSensor 3 Altitude Sensor
Sensor 3
MPEG-4 Player Compositor
hw://camera/back
Camera Input RAW Decoder Compositor mapping of
Camera Stream Screen captured data
Camera Sensor
39. MPEG-A Part 13 ARAF
Sensors/Actuators:: MPEG-V
Virtual World
Sensed VW Object Sensory
Information Characteristics Effects
(5) (4) (3)
Engine
R→V Adaptation: converts Sensed V→R Adaptation: converts
Info from RW to VW Object Sensory Effects from VW into
Char/Sensed Info applied to VW Device Cmds applied to RW
Sensor Sensor Sensory Sensory
Sensed Device
Device Adaptation Effects Device
Information Commands
Capability Preferences Preferences Capability
(5) (5)
(2) (2) (2) (2)
Real World User Real World
(Sensor Device) (Sensory Device)
41. MPEG-A Part 13 ARAF
Sensors/Actuators:: MPEG-V types
Sensors Global position Actuators
Light Altitude
Bend Light
Ambient noise
Temperature Gas Flash
Dust Heating
Humidity
Body height
Distance
Body weight
Cooling
Atmospheric pressure Wind
Body temperature
Position
Body fat Vibration
Velocity
Blood type Sprayer
Acceleration
Blood pressure
Orientation Scent
Blood sugar
Angular velocity Fog
Blood oxygen
Angular acceleration
Heart rate Color correction
Force
Torque Electrograph Initialize color correction parameter
Pressure EEG , ECG, EMG, EOG , GSR Rigid body motion
Weather
Motion Tactile
Intelligent camera type Facial expression
Facial morphology Kinesthetic
Multi Interaction point
Gaze tracking Facial expression characteristics Global position command
Wind Geomagnetic
42. MPEG-A Part 13 ARAF
Compression
Media Compression tool name Reference standard
Image JPEG ISO/IEC 10918
JPEG2000 ISO/IEC 15444
Video Visual ISO/IEC 14496-2
Advanced Video Coding ISO/IEC 14496-10
Audio MP3 ISO/IEC-11172-3
Advanced Audio Coding ISO/IEC 14496-3
3D Graphics Scalable Complexity Mesh Coding ISO/IEC 14496-16
Bone-based Animation ISO/IEC 14496-16
Scenes BIFS ISO/IEC 14496-11
44. MPEG-A Part 13 ARAF
Exercises
AR Quiz Augmented Book
http://youtu.be/la-Oez0aaHE http://youtu.be/LXZUbAFPP-Y
45. MPEG-A Part 13 ARAF
AR Quiz setting, preparing the medias
images, videos, audios, 2D/3D assets
GPS location
46. MPEG-A Part 13 ARAF
AR Quiz XML inspection
http://tiny.cc/MPEGARQuiz
47. MPEG-A Part 13 ARAF
AR Quiz Authoring Tool
www.MyMultimediaWorld.com go to Create / Augmented Reality
48. MPEG-A Part 13 ARAF
Augmented Book setting
images, audios
49. MPEG-A Part 13 ARAF
Augmented Book XML inspection
http://tiny.cc/MPEGAugBook
50. MPEG-A Part 13 ARAF
Augmented Book Authoring Tool
www.MyMultimediaWorld.com go to Create / Augmented Books
51. MPEG-A Part 13 ARAF
Next Steps
Support for metadata at scene and object level
Support for usage rights at scene and object level
Collisions between real and virtual objects, partial rendering
52. ARAF distance to X3D
On Scene Graph
– 32 elements
– including 2D graphics, humanoid animation, generic
input, media control, and pure AR protos
On Sensors/Actuators
– 6 elements
On Compression
– MPEG-4 Part 25 already compresses X3D
53. Conclusions
• Joint development of AR Reference Model
– The community at large is invited to react/contribute
such as the model became a reference
– http://wg11.sc29.org/trac/augmentedreality
• MPEG promoted a first version of an integrated and
consistent solution for representing content in AR
applications and services
– Continue synchronized/harmonized development of
technical specifications with X3D, COLLADA, OGC
content models