Working
This commit is contained in:
@@ -0,0 +1,48 @@
|
||||
= 3D bookmarks and navigation aids
|
||||
|
||||
One of the uses for 3D streaming is to allow users interacting with the content while it is being downloaded.
|
||||
However, devising an ergonomic technique for browsing 3D environments through a 2D interface is difficult.
|
||||
Controlling the viewpoint in 3D (6 DOFs) with 2D devices is not only inherently challenging but also strongly task-dependent. In their review, #cite("interaction-3d-environment") distinguish between several types of camera movements: general movements for exploration (e.g., navigation with no explicit target), targeted movements (e.g., searching and/or examining a model in detail), specified trajectory (e.g., a cinematographic camera path).
|
||||
For each type of movement, specialized 3D interaction techniques can be designed.
|
||||
In most cases, rotating, panning, and zooming movements are required, and users are consequently forced to switch back and forth among several navigation modes, leading to interactions that are too complicated overall for a layperson.
|
||||
Navigation aids and smart widgets are required and subject to research efforts both in 3D companies (see #link("https://sketchfab.com")[sketchfab.com], #link("https://cl3ver.com")[cl3ver.com] among others) and in academia, as reported below.
|
||||
|
||||
Translating and rotating the camera can be simply specified by a _lookat_ point.
|
||||
This is often known as point-of-interest (POI) movement (or _go-to_, _fly-to_ interactions) #cite("controlled-movement-virtual-3d").
|
||||
Given such a point, the camera automatically moves from its current position to a new position that looks at the POI.
|
||||
One key issue of these techniques is to correctly orient the camera at destination.
|
||||
In Unicam #cite("two-pointer-input"), the so-called click-to-focus strategy automatically chooses the destination viewpoint depending on 3D orientations around the contact point.
|
||||
The more recent Drag'n Go interaction #cite("drag-n-go") also hits a destination point while offering control on speed and position along the camera path.
|
||||
This 3D interaction is designed in the screen space (it is typically a mouse-based camera control), where cursor's movements are mapped to camera movements following the same direction as the on-screen optical-flow.
|
||||
|
||||
#figure(
|
||||
image("../assets/related-work/3d-interaction/dragngo.png", width: 70%),
|
||||
caption: [Screenshot of the drag'n go interface #cite("drag-n-go") (the percentage widget is for illustration)]
|
||||
)
|
||||
|
||||
Some 3D browsers provide a viewpoint menu offering a choice of viewpoints #cite("visual-perception-3d", "showmotion").
|
||||
Authors of 3D scenes can place several viewpoints (typically for each POI) in order to allow easy navigation for users, who can then easily navigate from viewpoint to viewpoint just by selecting a menu item.
|
||||
Such viewpoints can be either static, or dynamically adapted: #cite("dual-mode-ui") report that users clearly prefer navigating in 3D using a menu with animated viewpoints than with static ones.
|
||||
|
||||
#figure(
|
||||
image("../assets/related-work/3d-interaction/burtnyk.png", width: 70%),
|
||||
caption: [Screenshot of an interface with menu for navigation #cite("showmotion")]
|
||||
)
|
||||
|
||||
Early 3D VRML environments #cite("browsing-3d-bookmarks") offer 3D bookmarks with animated transitions between bookmarked views.
|
||||
These transitions prevent disorientation since users see how they got there.
|
||||
Hyperlinks can also ease rapid movements between distant viewpoints and naturally support non-linear and non-continuous access to 3D content.
|
||||
Navigating with 3D hyperlinks is faster due to the instant motion, but can cause disorientation, as shown by the work of #cite("ve-hyperlinks").
|
||||
#cite("linking-behavior-ve") examine explicit landmark links as well as implicit avatar-chosen links in Second Life.
|
||||
These authors point out that linking is appreciated by users and that easing linking would likely result in a richer user experience.
|
||||
#cite("dual-mode-ui") developed the Dual-Mode User Interface (DMUI) that coordinates and links hypertext to 3D graphics in order to access information in a 3D space.
|
||||
|
||||
#figure(
|
||||
image("../assets/related-work/3d-interaction/dmui.png", width: 100%),
|
||||
caption: [The two modes of DMUI #cite("dual-mode-ui")]
|
||||
)
|
||||
|
||||
The use of in-scene 3D navigation widgets can also facilitate 3D navigation tasks.
|
||||
#cite("navigation-aid-multi-floor") propose and evaluate 2D and 3D maps as navigation aids for complex virtual buildings and find that the 2D navigation aid outperforms the 3D one for searching tasks.
|
||||
The ViewCube widget #cite("viewcube") serves as a proxy for the 3D scene and offers viewpoint switching between 26 views while clearly indicating associated 3D orientations.
|
||||
Interactive 3D arrows that point to objects of interest have also been proposed as navigation aids in #cite("location-pointing-navigation-aid", "location-pointing-effect"): when clicked, the arrows transfer the viewpoint to the destination through a simulated walk or a faster flight.
|
||||
@@ -0,0 +1,219 @@
|
||||
= 3D streaming
|
||||
|
||||
In this thesis, we focus on the objective of delivering large, massive 3D scenes over the network.
|
||||
While 3D streaming is not the most popular research field, there has been a special attention around 3D content compression, in particular progressive compression which can be considered a premise for 3D streaming.
|
||||
In the next sections, we review the 3D streaming related work, from 3D compression and structuring to 3D interaction.
|
||||
|
||||
== Compression and structuring
|
||||
|
||||
According to #cite("maglo20153d"), mesh compression can be divided into four categories:
|
||||
- single-rate mesh compression, seeking to reduce the size of a mesh;
|
||||
- progressive mesh compression, encoding meshes in many levels of resolution that can be downloaded and rendered one after the other;
|
||||
- random accessible mesh compression, where different parts of the models can be decoded in an arbitrary order;
|
||||
- mesh sequence compression, compressing mesh animations.
|
||||
|
||||
Since our objective is to stream 3D static scenes, single-rate mesh and mesh sequence compressions are less interesting for us.
|
||||
This section thus focuses on progressive meshes and random accessible mesh compression.
|
||||
|
||||
Progressive meshes were introduced in #cite("progressive-meshes") and allow a progressive transmission of a mesh by sending a low resolution mesh first, called _base mesh_, and then transmitting detail information that a client can use to increase the resolution.
|
||||
To do so, an algorithm, called _decimation algorithm_, starts from the original full resolution mesh and iteratively removes vertices and faces by merging vertices through the so-called _edge collapse_ operation (Figure X).
|
||||
|
||||
// \begin{figure}[ht]
|
||||
// \centering
|
||||
// \begin{tikzpicture}[scale=2]
|
||||
// \node (Top1) at (0.5, 1) {};
|
||||
// \node (A) at (0, 0.8) {};
|
||||
// \node (B) at (1, 0.9) {};
|
||||
// \node (C) at (1.2, 0) {};
|
||||
// \node (D) at (0.9, -0.8) {};
|
||||
// \node (E) at (0.2, -0.9) {};
|
||||
// \node (F) at (-0.2, 0) {};
|
||||
// \node (G) at (0.5, 0.5) {};
|
||||
// \node (H) at (0.6, -0.5) {};
|
||||
// \node (Bottom1) at (0.5, -1) {};
|
||||
//
|
||||
// \node (Top2) at (3.5, 1) {};
|
||||
// \node (A2) at (3, 0.8) {};
|
||||
// \node (B2) at (4, 0.9) {};
|
||||
// \node (C2) at (4.2, 0) {};
|
||||
// \node (D2) at (3.9, -0.8) {};
|
||||
// \node (E2) at (3.2, -0.9) {};
|
||||
// \node (F2) at (2.8, 0) {};
|
||||
// \node (G2) at (3.55, 0) {};
|
||||
// \node (Bottom2) at (3.5, -1) {};
|
||||
//
|
||||
// \draw (A.center) -- (B.center) -- (C.center) -- (D.center) -- (E.center) -- (F.center) -- (A.center);
|
||||
// \draw (A.center) -- (G.center);
|
||||
// \draw (B.center) -- (G.center);
|
||||
// \draw (C.center) -- (G.center);
|
||||
// \draw (F.center) -- (G.center);
|
||||
// \draw (C.center) -- (H.center);
|
||||
// \draw (F.center) -- (H.center);
|
||||
// \draw (E.center) -- (H.center);
|
||||
// \draw (D.center) -- (H.center);
|
||||
// \draw[color=red, line width=1mm] (G.center) -- (H.center);
|
||||
//
|
||||
// \draw (A2.center) -- (B2.center) -- (C2.center) -- (D2.center) -- (E2.center) -- (F2.center) -- (A2.center);
|
||||
// \draw (A2.center) -- (G2.center);
|
||||
// \draw (B2.center) -- (G2.center);
|
||||
// \draw (C2.center) -- (G2.center);
|
||||
// \draw (F2.center) -- (G2.center);
|
||||
// \draw (E2.center) -- (G2.center);
|
||||
// \draw (D2.center) -- (G2.center);
|
||||
// \node at (G2) [circle,fill=red,inner sep=2pt]{};
|
||||
//
|
||||
// \draw[-{Latex[length=3mm]}] (Top1) to [out=30, in=150] (Top2);
|
||||
// \draw[-{Latex[length=3mm]}] (Bottom2) to [out=-150, in=-30] (Bottom1);
|
||||
//
|
||||
// \node at (2, 1.75) {Edge collapse};
|
||||
// \node at (2, -1.75) {Vertex split};
|
||||
//
|
||||
//
|
||||
// \end{tikzpicture}
|
||||
// \caption{Vertex split and edge collapse\label{sote:progressive-scheme}}
|
||||
// \end{figure}
|
||||
|
||||
Every time two vertices are merged, a vertex and two faces are removed from the original mesh, decreasing the model resolution.
|
||||
At the end of this content preparation phase, the mesh has been reorganized into a base mesh and a sequence of partially ordered edge split operations.
|
||||
Thus, a client can start by downloading the base mesh, display it to the user, and keep downloading refinement operations (vertex splits) and display details as time goes by.
|
||||
This process reduces the time a user has to wait before seeing a downloaded 3D object, thus increases the quality of experience.
|
||||
|
||||
#figure(
|
||||
image("../assets/related-work/3d-streaming/progressivemesh.png", width: 100%),
|
||||
caption: [Four levels of resolution of a mesh]
|
||||
)
|
||||
|
||||
%These methods have been vastly researched #cite("bayazit20093", "mamou2010shape"), but very few of these methods can handle meshes with attributes, such as texture coordinates.
|
||||
|
||||
#cite("streaming-compressed-webgl") develop a dedicated progressive compression algorithm based on iterative decimation, for efficient decoding, in order to be usable on web clients.
|
||||
With the same objective, #cite("pop-buffer") proposes pop buffer, a progressive compression method based on quantization that allows efficient decoding.
|
||||
|
||||
Following these, many approaches use multi triangulation, which creates mesh fragments at different levels of resolution and encodes the dependencies between fragments in a directed acyclic graph.
|
||||
In #cite("batched-multi-triangulation"), the authors propose Nexus: a GPU optimized version of multi triangulation that pushes its performances to make real time rendering possible.
|
||||
It is notably used in 3DHOP (3D Heritage Online Presenter, #cite("3dhop")), a framework to easily build web interfaces to present 3D objects to users in the context of cultural heritage.
|
||||
|
||||
Each of these approaches define its own compression and coding for a single mesh.
|
||||
However, users are often interested in scenes that contain multiple meshes, and the need to structure content emerged.
|
||||
|
||||
To answer those issues, the Khronos group proposed a generic format called glTF (GL Transmission Format, #cite("gltf")) to handle all types of 3D content representations: point clouds, meshes, animated models, etc.
|
||||
glTF is based on a JSON file, which encodes the structure of a scene of 3D objects.
|
||||
It contains a scene graph with cameras, meshes, buffers, materials, textures and animations.
|
||||
Although relevant for compression, transmission and in particular streaming, this standard does not yet consider view-dependent streaming, which is required for large scene remote visualization and which we address in our work.
|
||||
|
||||
== Viewpoint dependency
|
||||
|
||||
3D streaming means that content is downloaded while the user is interacting with the 3D object.
|
||||
In terms of quality of experience, it is desirable that the downloaded content falls into the user's field of view.
|
||||
This means that the progressive compression must encode a spatial information in order to allow the decoder to determine content adapted to its viewpoint.
|
||||
This is typically called _random accessible mesh compression_.
|
||||
#cite("maglo2013pomar") is such an example of random accessible progressive mesh compression.
|
||||
#cite("cheng2008receiver") proposes a receiver driven way of achieving viewpoint dependency with progressive mesh: the client starts by downloading the base mesh, and from then is able to estimate the importance of the different vertex splits, in order to choose which ones to download.
|
||||
Doing so drastically reduces the server computational load, since it only has to send data, and improves the scalability of this framework.
|
||||
|
||||
In the case of streaming a large 3D scene, view-dependent streaming is fundamental: a user will only be seeing one small portion of the scene at each time, and a system that does not adapt its streaming to the user's point of view is bound to induce a low quality of experience.
|
||||
|
||||
A simple way to implement viewpoint dependency is to request the content that is spatially close to the user's camera.
|
||||
This approach, implemented in Second Life and several other NVEs (e.g., #cite("peer-texture-streaming")), only depends on the location of the avatar, not on its viewing direction.
|
||||
It exploits spatial coherence and works well for any continuous movement of the user, including turning.
|
||||
Once the set of objects that are likely to be accessed by the user is determined, the next question is in what order should these objects be retrieved.
|
||||
A simple approach is to retrieve the objects based on distance: the spatial distance from the user's virtual location and rotational distance from the user's view.
|
||||
|
||||
More recently, Google integrated Google Earth 3D module into Google Maps (@rw:google-maps).
|
||||
Users are now able to go to Google Maps, and click the 3D button which shifts the camera from the aerial view.
|
||||
Even though there are no associated publications to support this assertion, it seems clear that the streaming is view-dependent: low resolution from the center of the point of view gets downloaded first, and higher resolution data gets downloaded for closer objects than for distant ones.
|
||||
|
||||
#figure(
|
||||
image("../assets/related-work/3d-streaming/googlemaps.png", width: 80%),
|
||||
caption: [Screeshot of the 3D interface of Google Maps]
|
||||
)<rw:google-maps>
|
||||
|
||||
Other approaches use level of details.
|
||||
Level of details have been initially used for efficient 3D rendering #cite("lod").
|
||||
When the change from one level of detail to another is direct, it can create visual discomfort to the user.
|
||||
This is called the _popping effect_ and level of details have the advantage of enabling techniques, such as geomorhping #cite("hoppe-lod"), to transition smoothly from one level of detail to another.
|
||||
Level of details have then been used for 3D streaming.
|
||||
For example, #cite("streaming-hlod") propose an out-of-core viewer for remote model visualization based by adapting hierarchical level of details #cite("hlod") to the context of 3D streaming.
|
||||
Level of details can also be used to perform viewpoint dependant streaming, such as #cite("view-dependent-lod").
|
||||
|
||||
== Texture streaming
|
||||
|
||||
In order to increase the texture rendering speed, a common technique is the _mipmapping_ technique.
|
||||
It consists in generating progressively lower resolutions of an initial texture.
|
||||
Lower resolutions of the textures are used for polygons which are far away from the camera, and higher resolutions for polygons closer to the camera.
|
||||
Not only this reduces the time needed to render the polygons, but it can also reduce the aliasing effect.
|
||||
Using these lower resolutions can be especially interesting for streaming.
|
||||
#cite("mipmap-streaming") proposes the PTM format which encode the mipmap levels of a texture that can be downloaded progressively, so that a lower resolution can be shown to the user while the higher resolutions are being downloaded.
|
||||
|
||||
Since 3D data can contain many textures, #cite("simon2019streaming") propose a way to stream a set of textures by encoding them into a video.
|
||||
Each texture is segmented into tiles of a fixed size.
|
||||
Those tiles are then ordered to minimize dissimilarities between consecutive tiles, and encoded as a video.
|
||||
By benefiting from the video compression techniques, the authors are able to reach a better rate-distortion ratio than webp, which is the new standard for texture transmission, and jpeg.
|
||||
|
||||
== Geometry and textures
|
||||
|
||||
As discussed in Chapter~\ref{f:3d}, most 3D scenes consist in two main types of data: geometry and textures.
|
||||
When addressing 3D streaming, one must handle the concurrency between geometry and textures, and the system needs to address this compromise.
|
||||
|
||||
Balancing between streaming of geometry and texture data is addressed by #cite("batex3"), #cite("visual-quality-assessment"), and #cite("mesh-texture-multiplexing").
|
||||
Their approaches combine the distortion caused by having lower resolution meshes and textures into a single view independent metric.
|
||||
#cite("progressive-compression-textured-meshes") also deals with the geometry / texture compromise.
|
||||
This work designs a cost driven framework for 3D data compression, both in terms of geometry and textures.
|
||||
The authors generate an atlas for textures that enables efficient compression and multi-resolution scheme.
|
||||
All four works considered a single mesh, and have constraints on the types of meshes that they are able to compress.
|
||||
Since the 3D scenes we are interested in in our work consist in soups of textured polygons, those constraints are not satisfied and we cannot use those techniques.
|
||||
% All four works considered a single, manifold textured mesh model with progressive meshes, and are not applicable in our work since we deal with large and potentially non-manifold scenes.
|
||||
|
||||
|
||||
== Streaming in game engines
|
||||
|
||||
In traditional video games, including online games, there is no requirement for 3D data streaming.
|
||||
Video games either come with a physical support (CD, DVD, Blu-Ray) or they require the downloading of the game itself, which includes the 3D data, before letting the user play.
|
||||
However, transferring data from the disk to the memory is already a form of streaming.
|
||||
This is why optimized engines for video games use techniques that are reused for streaming such as level of details, to reduce the details of objects far away for the point of view and save the resources to enhance the level of detail of closer objects.
|
||||
|
||||
Some other online games, such as #link("https://secondlife.com")[Second Life], rely on user generated data, and thus are forced to send data from users to others.
|
||||
In such scenarios, 3D streaming is appropriate and this is why the idea of streaming 3D content for video games has been investigated.
|
||||
For example, #cite("game-on-demand") proposes an online game engine based on geometry streaming, that addresses the challenge of streaming 3D content at the same time as synchronization of the different players.
|
||||
|
||||
== NVE streaming frameworks
|
||||
|
||||
An example of NVE streaming framework is 3D Tiles #cite("3d-tiles"), which is a specification for visualizing massive 3D geospatial data developed by Cesium and built on top of glTF.
|
||||
Their main goal is to display 3D objects on top of regular maps, and their visualization consists in a top-down view, whereas we seek to let users freely navigate in our scenes, whether it be flying over the scene or moving along the roads.
|
||||
|
||||
#figure(
|
||||
image("../assets/related-work/3d-streaming/3dtiles.png", width: 80%),
|
||||
caption: [Screenshot of 3D Tiles interface]
|
||||
)
|
||||
|
||||
3D Tiles, as its name suggests, is based on a spacial partitionning of the scene.
|
||||
It started with a regular octree, but has then been improved to a $k$-d tree (see Figure~\ref{sote:3d-tiles-partition}).
|
||||
|
||||
#grid(
|
||||
columns:(1fr, 0.1fr, 1fr),
|
||||
|
||||
figure(
|
||||
image("../assets/related-work/3d-streaming/3d-tiles-octree.png", width: 100%),
|
||||
caption: [With regular octree (depth 4)]
|
||||
),
|
||||
|
||||
[],
|
||||
|
||||
figure(
|
||||
image("../assets/related-work/3d-streaming/3d-tiles-kd-tree.png", width: 100%),
|
||||
caption: [With $k$-d tree (depth 4)]
|
||||
)
|
||||
)
|
||||
|
||||
In~\citeyear{3d-tiles-10x}, 3D Tiles streaming system was improved by preloading the data at the camera's next position when known in advance (with ideas that are similar to those we discuss and implement in Chapter~\ref{bi}, published in~\citeyear{bookmarks-impact}) and by ordering tile requests depending on the user's position (with ideas that are similar to those we discuss and implement in Chapter~\ref{d3}, published in~\citeyear{dash-3d}).
|
||||
|
||||
#cite("zampoglou") is another example of a streaming framework: it is the first paper that proposes to use DASH to stream 3D content.
|
||||
In their work, the authors describe a system that allows users to access 3D content at multiple resolutions.
|
||||
They organize the content, following DASH terminology, into periods, adaptation sets, representations and segments.
|
||||
Their first adaptation set codes the tree structure of the scene graph.
|
||||
Each further adaptation set contains both geometry and texture information and is available at different resolutions defined in a corresponding representation.
|
||||
To avoid requests that would take too long and thus introduce latency, the representations are split into segments.
|
||||
The authors discuss the optimal number of polygons that should be stored in a single segment.
|
||||
On the one hand, using segments containing very few faces will induce many HTTP requests from the client, and will lead to poor streaming efficiency.
|
||||
On the other hand, if segments contain too many faces, the time to load the segment is long and the system loses adaptability.
|
||||
Their approach works well for several objects, but does not handle view-dependent streaming, which is desirable in the use case of large NVEs.
|
||||
|
||||
@@ -0,0 +1,12 @@
|
||||
#import "../chapter.typ"
|
||||
|
||||
#chapter.chapter[Related work]
|
||||
|
||||
In this chapter, we review the part of the state of the art on multimedia streaming and interaction that is relevant for this thesis.
|
||||
As discussed in the previous chapter, video and 3D share many similarities and since there is already a very important body of work on video streaming, we start this chapter with a review of this domain with a particular focus on the DASH standard.
|
||||
Then, we proceed with presenting topics related to 3D streaming, including compression and streaming, geometry and texture compromise, and viewpoint dependent streaming.
|
||||
Finally, we end this chapter by reviewing the related work regarding 3D navigation and interfaces.
|
||||
|
||||
#include "video.typ"
|
||||
#include "3d-streaming.typ"
|
||||
#include "3d-interaction.typ"
|
||||
@@ -0,0 +1,95 @@
|
||||
= Video
|
||||
|
||||
Accessing a remote video through the web has been a widely studied problem since the 1990s.
|
||||
The Real-time Transport Protocol (RTP, #cite("rtp-std")) has been an early attempt to formalize audio and video streaming.
|
||||
The protocol allowed data to be transferred unilaterally from a server to a client, and required the server to handle a separate session for each client.
|
||||
|
||||
In the following years, HTTP servers have become ubiquitous, and many industrial actors (Apple, Microsoft, Adobe, etc.) developed HTTP streaming systems to deliver multimedia content over the network.
|
||||
In an effort to bring interoperability between all different actors, the MPEG group launched an initiative, which eventually became a standard known as DASH, Dynamic Adaptive Streaming over HTTP.
|
||||
Using HTTP for multimedia streaming has many advantages over RTP.
|
||||
While RTP is stateful (that is to say, it requires keeping track of every user along the streaming session), HTTP is stateless, and thus more efficient.
|
||||
Furthermore, an HTTP server can easily be replicated at different geographical locations, allowing users to fetch data from the closest server.
|
||||
This type of network architecture is called CDN (Content Delivery Network) and increases the speed of HTTP requests, making HTTP based multimedia streaming more efficient.
|
||||
|
||||
== DASH: the standard for video streaming
|
||||
|
||||
Dynamic Adaptive Streaming over HTTP (DASH), or MPEG-DASH #cite("dash-std", "dash-std-2") is now a widely deployed
|
||||
standard for adaptively streaming video on the web #cite("dash-std-full"), made to be simple, scalable and inter-operable.
|
||||
DASH describes guidelines to prepare and structure video content, in order to allow a great adaptability of the streaming without requiring any server side computation. The client should be able to make good decisions on what part of the content to download, only based on an estimation of the network constraints and on the information provided in a descriptive file: the MPD.
|
||||
|
||||
#heading(level: 3, numbering: none)[DASH structure]
|
||||
|
||||
All the content structure is described in a Media Presentation Description (MPD) file, written in the XML format.
|
||||
This file has 4 layers: the periods, the adaptation sets, the representations, and the segments.
|
||||
An MPD has a hierarchical structure, meaning it has multiple periods, and each period can have multiple adaptation sets, each adaptation set can have multiple representation, and each representation can have multiple segments.
|
||||
|
||||
#heading(level: 4, numbering: none)[Periods]
|
||||
Periods are used to delimit content depending on time.
|
||||
It can be used to delimit chapters, or to add advertisements that occur at the beginning, during or at the end of a video.
|
||||
|
||||
#heading(level: 4, numbering: none)[Adaptation sets]
|
||||
Adaptation sets are used to delimit content according to the format.
|
||||
Each adaptation set has a mime-type, and all the representations and segments that it contains share this mime-type.
|
||||
In videos, most of the time, each period has at least one adaptation set containing the images, and one adaptation set containing the sound.
|
||||
It may also have an adaptation set for subtitles.
|
||||
|
||||
#heading(level: 4, numbering: none)[Representations]
|
||||
The representation level is the level DASH uses to offer the same content at different levels of quality.
|
||||
For example, an adaptation set containing images has a representation for each available quality (it might be 480p, 720p, 1080p, etc.).
|
||||
This allows a user to choose its representation and change it during the video, but most importantly, since the software is able to estimate its downloading speed based on the time it took to download data in the past, it is able to find the optimal representation, being the highest quality that the client can request without stalling.
|
||||
|
||||
#heading(level: 4, numbering: none)[Segments]
|
||||
Until this level in the MPD, content has been divided but it is still far from being sufficiently divided to be streamed efficiently.
|
||||
A representation of the images of a chapter of a movie is still a long video, and keeping such a big file is not possible since heavy files prevent streaming adaptability: if the user requests to change the quality of a video, the system would either have to wait until the file is totally downloaded, or cancel the request, making all the progress done unusable.
|
||||
|
||||
Segments are used to prevent this issue.
|
||||
They typically encode files that contain two to ten seconds of video, and give the software a greater ability to dynamically adapt to the system.
|
||||
If a user wants to seek somewhere else in the video, only one segment of data is potentially lost, and only one segment
|
||||
of data needs to be downloaded for the playback to resume. The impact of the segment duration has been investigated in many work, including #cite("sideris2015mpeg", "stohr2017sweet").
|
||||
For example, #cite("stohr2017sweet") discuss how the segment duration affects the streaming: short segments lower the initial delay and provide the best stalling quality of experience, but make the total downloading time of the video longer because of overhead.
|
||||
|
||||
#heading(level: 3, numbering: none)[Content preparation and server]
|
||||
|
||||
Encoding a video in DASH format consists in partitioning the content into periods, adaptation sets, representations and segments as explained above, and generating a Media Presentation Description file (MPD) which describes this organization.
|
||||
Once the data are prepared, they can simply be hosted on a static HTTP server which does no computation other than serving files when it receives requests.
|
||||
All the intelligence and the decision making is moved to the client side.
|
||||
This is one of the DASH strengths: no powerful server is required, and since static HTTP server are mature and efficient, all DASH clients can benefit from it.
|
||||
|
||||
|
||||
#heading(level: 3, numbering: none)[Client side adaptation]
|
||||
|
||||
A client typically starts by downloading the MPD file, and then proceeds on downloading segments from the different adaptation sets. While the standard describes well how to structure content on the server side, the client may be freely implemented to take into account the specificities of a given application.
|
||||
The most important part of any implementation of a DASH client is called the adaptation logic. This component takes into account a set of parameters, such as network conditions (bandwidth, throughput, for example), buffer states or segments size to derive a decision on which segments should be downloaded next. Most of the industrial actors have their own
|
||||
adaptation logic, and many more have been proposed in the literature.
|
||||
A thorough review is beyond the scope of this state-of-the-art, but examples include #cite("chiariotti2016online") who formulate the problem in a reinforcement learning framework, #cite("yadav2017quetra") who formulate the problem using queuing theory, or #cite("huang2019hindsight") who use a formulation derived from the knapsack problem.
|
||||
|
||||
== DASH-SRD
|
||||
Being now widely adopted in the context of video streaming, DASH has been adapted to various other contexts.
|
||||
DASH-SRD (Spatial Relationship Description, #cite("dash-srd")) is a feature that extends the DASH standard to allow streaming only a spatial subpart of a video to a device.
|
||||
It works by encoding a video at multiple resolutions, and tiling the highest resolutions as shown in Figure \ref{sota:srd-png}.
|
||||
That way, a client can choose to download either the low resolution of the whole video or higher resolutions of a subpart of the video.
|
||||
|
||||
#figure(
|
||||
image("../assets/related-work/video/srd.png", width: 60%),
|
||||
caption: [DASH-SRD #cite("dash-srd")],
|
||||
)
|
||||
|
||||
For each tile of the video, an adaptation set is declared in the MPD, and a supplemental property is defined in order to give the client information about the tile.
|
||||
This supplemental property contains many elements, but the most important ones are the position ($x$ and $y$) and the size (width and height) describing the position of the tile in relation to the full video.
|
||||
An example of such a property is given in @rw:srd-xml.
|
||||
|
||||
#figure(
|
||||
align(left,
|
||||
raw(
|
||||
read("../assets/related-work/video/srd.xml"),
|
||||
block: true,
|
||||
lang: "xml",
|
||||
),
|
||||
),
|
||||
caption: [MPD of a video encoded using DASH-SRD]
|
||||
)<rw:srd-xml>
|
||||
|
||||
Essentially, this feature is a way of achieving view-dependent streaming, since the client only displays a part of the video and can avoid downloading content that will not be displayed.
|
||||
While Figure \ref{sota:srd-png} illustrates how DASH-SRD can be used in the context of zoomable video streaming, the ideas developed in DASH-SRD have proven to be particularly useful in the context of 360 video streaming (see for example #cite("ozcinar2017viewport")).
|
||||
This is especially interesting in the context of 3D streaming since we have this same pattern of a user viewing only a part of a content.
|
||||
|
||||
Reference in New Issue
Block a user