Exploiting Network Awareness to Enhance DASH Over Wireless
The introduction of Dynamic Adaptive Streaming over HTTP (DASH) helped reduce the consumption of resource in video delivery, but its client-based rate adaptation is unable to optimally use the available end-to-end network bandwidth. We consider the problem of optimizing the delivery of video content to mobile clients while meeting the constraints imposed by the available network resources. Observing the bandwidth available in the network’s two main components, core network, transferring the video from the servers to edge nodes close to the client, and the edge network, which is in charge of transferring the content to the user, via wireless links, we aim to find an optimal solution by exploiting the predictability of future user requests of sequential video segments, as well as the knowledge of available infrastructural resources at the core and edge wireless networks in a given future time window. Instead of regarding the bottleneck of the end-to-end connection as our throughput, we distribute the traffic load over time and use intermediate nodes between the server and the client for buffering video content to achieve higher throughput, and ultimately significantly improve the Quality of Experience for the end user in comparison with current solutions.
Video consumption over the internet has experienced a continuous growth in the last few years and now accounts for about two thirds of the total amount of global Internet traffic. Its share is expected to increase to up to 79% by 2018 . While video streaming services continue to raise their popularity thanks to large availability of content and reduced costs, Internet Service Providers are struggling to provide high quality services to their costumers due to their inability to allocate enough capacity to meet such demand, especially at peak hours.
To adjust the demand to the network conditions, Dynamic Adaptive Streaming over HTTP (DASH) has been introduced. DASH let the client adjust the rate of the video stream by monitoring the network conditions perceived by this stream. DASH is a client-based rate adaptation, as the server is stateless and the network is considered as a black box. The client measures the end-to-end throughput between itself and the server and modifies the rate accordingly.
While this approach has been incredibly successful and DASH-like rate adaptation now accounts for most of the video traffic (say, Netflix or YouTube), end-to-end rate adaptation is suboptimal. Indeed, end-to-end monitoring measures the lowest throughput of all the links in between the client and the server, while the available bandwidth on these links varies a lot.
As a simple illustrative example, consider the following scenario of Figure 1: the client is connected to the server by two links (say, a wireless link for the edge network, and a wired link to the server). Both of the links’ bandwidth will oscillate and in modern networks, both could be congested111It is well known for instance that Netflix bandwidth is throttled inside the network by operators with whom Netflix does not have a peering agreement.. Therefore an end-to-end mechanism will yield the minimal available bandwidth of each link. If for one unit of time, the capacity of the wireless is 1 unit of transfer, while the capacity of the wired link is 2, and for the next unit of time the capacities are reversed, then the end-to-end mechanism can only achieve a rate of 1.
However, inserting an intermediary node with storage capacity in between the two links could increase the throughput significantly. In the illustrative example, during the first time slot, the full capacity of the wired link could be used to delivery 1 unit to the client and 1 unit to the intermediate storage; in the second time slot, this extra stored unit can be delivered over the air interface, for a total delivery of 2 units. On average, the delivery is of 1.5 for each time slot, versus 1 in the end-to-end mechanism. While this example is simplistic, and does not take into account rate adaptation, it shows the benefit of a network-assisted content deliver mechanism. Here we present an intermediated rate adaptation mechanism which leverages the different time-varying capacities of the multiple links in the network.
The goal of our work is to propose a novel infrastructural approach to control the load on the network for clients with quickly changing available capacity while maintaining high quality experience for the end users. The experience is defined as a set of quality metrics, such as average bitrate, temporal variability and amount of rebuffering time. Our solution is based on two key ideas: (1) to use local caches strategically deployed into the access networks to decouple the transmission speed of the Internet long distance network from the local access network and (2)to exploit the predictability of future available infrastructural resources (at the core and edge wireless network) and the predictability of video requests to distribute the load on the network of the video streams and control the Quality of Experience for the end users.
The rest of paper is structured as follow: in Section II we will discuss our contributions in comparison to other related works; in Section III we will present the network model that we use to then design our solution; Section IV will provide a detailed experimental analysis of our design through extensive simulations and finally Section V will conclude the paper.
Ii Related work
Bianchi  has considered the role of intermediate storage in the content delivery, however in the context of ATM and without rate adaptation. They did observe a similar benefit in their application scenario. Since this work, Dynamic Adaptive Streaming over HTTP (DASH)  has been introduced as a standard for video streaming. It has been widely deployed by many major video streaming services such as Netflix, Hulu, and Youtube.
Letting the network assist in the delivery of the content is related to the ideas of the Information-Centric Network (ICN) architectures that have been proposed recently to allow the network to be aware of content semantics. Some have proposed to modify DASH over ICN, for instance in the IRTF .  examines the interaction of DASH with ICN, notes the potential issues, and identifies the synergies.  targets HTTP adaptive streaming (HAS) in Content-Centric Networking (CCN)  for Scalable Video Coding.
 has looked at how to predict the request from the users by looking a social interactions. Here, we use the natural predictability that video streaming offers .  describes a web prefetching module running on the CDN main node (controller) which downloads web contents that will be requested in the future to the LAN CDN surrogates. Our work is different as the time domain is much smaller and the prefetching utilizes the bandwidth more dynamically. In addition,  does not specifically consider video contents.  tries to optimize the utilization of the wireless channel to delivery buffered video (i.e. streaming with availability of buffer at the receiving side) conjunct with the minimization of the rebuffering time using a bandwidth prediction model, but does not consider network caching or adaptive video streaming.
 analyzes the potential benefits of CDN augmentation strategies can offer for Internet video workloads using a dataset of 30 million VOD and live sessions. It has also been observed in  that fractions of viewers typically watch only the first 10 minutes of video, around 4.5% of users are serial early quitters, and 16.6% of users consistently watch video to completion. This suggests that a user based prefetching policy should be a natural extension for our work.
 proposes a Network-Friendly DASH (NF-DASH) architecture for coordinating peer-assisted CDNs operated by ISPs and the traditional CDNs.  analyzes the waiting time and network utilization with service prioritization considering both on-demand fetching/caching and prefetching in a P2P-assisted video streaming system.  formulates the CDN assignment as an optimization problem targeting minimizing the cost based on the QoE constraints to CDN servers at different locations.  shows by measurement the shortcomings of today’s video delivery infrastructure and proposes a control plane in the network for video distribution for global resource optimization.
Hybrid P2P-CDN video streaming enhancements have also been considered. That is, serving content from dedicated CDN servers using P2P technology. and telco-CDN (CDNs operated by telecommunication companies, enabling users to reach CDN caches that are closer) federation are two emerging strategies. Telco-CDN federation can reduce the provisioning cost by 95%. Using P2P can lead up to 87.5% bandwidth savings for the CDN during peak access hours.
Proxy-assisted caching and prefetching has been widely studied in the literature. Some approaches consider the quality of the connections  and the usage of the client buffers . Approaches for transcoding proxy caching are also presented in  in which the proxy caches different versions of the content to handle the heterogeneous user requirements. Prefix caching  caches only the frames at the beginning to minimize the average initial delay. Exponential segmentation  divides the video object such that the beginning of a video is cached as smaller segment. Lazy segmentation approach  determines the segment length according to the user access record at the late time.
Proxy technologies have also been used to enhance QoE.  uses information about wireless channel quality at the base station to provide QoE and fairness among clients.  proposes WiDASH, a proxy for adaptive http streaming over wireless networks that implements a quadratic linear optimization problem to decide on the rate to use for each user/segment while giving higher priority to lower video rates.
 discusses a segment-based proxy pre-fetching for streaming delivery.  evaluates the 1-ahead, n-ahead, and priority-based segment prefetching. The results show that if the bottleneck link is between client and proxy, all prefetching schemes achieve high cache hit rate after 2-3 client requesting a video. On the other hand, if the bottleneck link is between proxy and server, no prefetching helps. Our approach considers the link between the cache and the server and makes a pre-fetching decision accordingly.
Iii Network Model and Architecture
We build our system around the network model shown in Figure 2 which follows the model presented in : in this classical scenario, a clients connects to the Internet through a local access network; the goal of the client is to retrieve video contents from a remote server that can be reached by traversing a long distance network. We focus on a scenario where mobile devices are connected to the access network through a wireless interface (or more than one); our system is designed to be technology agnostic so both wifi or cellular connectivity can be considered.
We base our video delivery model around the recently standardized Dynamic Adaptive Streaming over HTTP protocol. In this protocol each video file is divided into segments of equal duration encoded at different bitrates and all stored in one or more webservers. The choice of using DASH as the underlying technology for our system naturally derives from its popularity that has been constantly out-growing other protocols over the last few years thanks to its easiness of deployment as it relies on the already available HTTP infrastructure of webservers, proxies and caches. In a typical DASH system, a client interested in retrieving the video first has to retrieve from the server a Media Presentation Description file that contains information on the structure of the video, the available different bitrates for the segments and at which location they are stored. Once this step is completed, the client proceeds to sequentially download consecutive video segments. When a new segment has to be retrieved, the clients select an appropriate bitrate where the decision is usually based on different factors such as recently experience bandwidth  or current buffer size .
It easy to notice that this model heavily relies on the ability of the client to estimate the available bandwidth and general network resources, a task arguably very difficult under normal network conditions , and even more complex under wireless and mobile environments due to the high dynamicity caused by time-varying fading, shadowing interference and hand-off delays [35, 36]. To combat these effects we propose to move the adaptation logic where this information is more easily accessible: into the network. In our system, a caching system is made available strategically positioned inside at the edge of the architecture. An orchestration system, that from now on we will call the network controller and that could be centralized or distributed implements the adaptation logic by mean of periodically scheduling video segment requests to temporarily store the corresponding video segments into the cache; to do so, we assume that the controller, at the beginning of a periodic time window of given size, schedules the download of video segments to the available caches from where the clients will the retrieve them from. The scheduling process is based on two fundamental pieces of information to which the controller has access:
A general view of the network infrastructure resources availability both in the core network and at the edge network: as different studies have proved, it is indeed possible to predict to a certain extent the variability of users connectivity in wireless networks by either exploiting movement predictability  or by recognizing performance patterns , which is even easier if done from a convenient location inside the access network.
Information on the client playback and buffer status: assuming that the normal DASH client behavior is expected (i.e. sequential downloading of consecutive video segments), the controller maintains a detailed view of the current status of active video streams by tracking the progression of the requested segments; this includes the size of the video buffers available at each client and the playback time.
While this description provides a general view of our proposed architecture, we now analyze the details of each protocol and algorithm implemented in our system; we start from describing in Section III-A how the basic DASH model would be modified in order to exploit the new functionalities; we then provide in Section III-B details on the optimization functions involved in the scheduling process and finally, Section III-C introduces the scheduling algorithm performed by the controller.
Iii-a System Protocols Progression
With the presented infrastructure we propose a solution that integrates all the given components with the goal of maximizing the Quality of Experience for the end users while only exploiting the available resources made to the system and without overrunning them. As previously introduced we use the concept of a centralized controller that orchestrates the necessary operations; while we present it as a centralized solution, we claim that the same results could be obtained with a distributed solution. Moreover, while we refer to this orchestration system as the controller, we do not limit our solution to Software Defined Networks: any technology able to track HTTP requests can be applied; possible solutions include the use of an HTTP proxy that manipulates the traffic or obviously the use of a centralized SDN controller. Our system uniquely relies on tracking and exploiting available in-network resources (i.e. capacity and caches) and tracking and eventually modifying HTTP requests from the clients. We then abstract the required actions into three main steps.
1) Stream initialization. In order to initialize the streaming process, each client has to request the DASH MPD file from the server. The controller captures this requests and retrieve the same information either by deep packet inspection of the returned content or by requesting the same file from the webserver. With this information, the controller obtains a complete view of the video that can potentially be retrieved. To simplify the description and our notation, we assume that each video is composed by segments , each of duration seconds and available at different bitrates.
2) Bitrate Adaptation. Once the MPD file is retrieved, the client can proceed to sequentially request the video segments; while in a normal DASH setup, the client would be in charge of selecting one of the available bitrates for each requested segment, we leave this duty to the controller; this design choice is justifiable by considering the fact that the controller has the best possible view of the available resources, by accessing infrastructure components and tracking the client process through its issued requests; while it would be possible to argue about the introduction of additional complexity into the network, we argue that the complexity is sustainable if the performance gains can justify it. The controller selects among the different bitrates based on two main factors: status of the video reproduction at the client (i.e. buffer size and previous quality selection) and future availability of infrastructure resources; in particular we assume that the controller has access to information regarding residual capacity available in the network and bandwidth for any specific client for a time window of size seconds.
3) Streaming process. Following the normal DASH model, clients sequentially retrieve video segments by mean of issuing HTTP requests. We differentiate from the original model by having the clients to retrieve the video from the caching system without the need of specifying any particular version of the segment. This could be done using commonly exploited techniques as for example: having the network controller to modify the MPD originally retrieved replacing content URLs with locally resolvable ones or by transparently capture the HTTP requests (e.g. if the controller is implemented as an HTTP proxy) or 3) In order to meet the proposed goals the controller divides the delivery path into two components: from the server to the local caches available at the edge network and from the caches to the client over the wireless link. The controller uses the available information in the time slot of size seconds to provide in time delivery to the caches.
Additional details regarding the involved scheduling algorithm will be provide in section III-C, but in order to do so, we first need to introduce the Quality of Experience model employed.
Iii-B Optimization Function
Iii-B1 QoE Model
We include in our model the following typical key factors that play an active role in defining the QoE for video streaming:
Average quality: multiple works have shown how the average video quality can not be used as the sole metric in determining the quality experienced by a user. Nevertheless, we still need to have a way of factoring the proportional quality between different available bitrates; for this reason we refer to previous work  where has been widely proven that there is a direct relationship between bitrate selection and quality governed by logarithmic laws.
Temporal quality variance: different studies have proven how representation switching can factor negatively against the quality of experience; in particular,  found that only up to 0.5 quality switches per minute are considerable tolerable by users, causing exponential increase on rate of abandonments if these criteria are not met. Moreover, as suggested in the ITU standard , human memory effects can distort quality ratings if noticeable impairments occur in approximately the last 10-15s of the sequence, exponentially decaying afterward, causing past factors to relatively influence the current visualized video.
Buffering time and ratio: it has been widely demonstrated that frequency and length of video rebuffering highly affects the perceived quality of a video stream where each event increase the rate of abandonment and reduce the probability of client return.
While the third factor negatively impacts the abandonment rates from users, we will work with the initial assumption that these events are occurring very infrequently; namely we assume the available bandwidth always provide at least enough resources to deliver the lowest quality video; a rebuffering cost could then be introduced into our optimization function to factor it into the scheduling algorithm.
We initially focus our attention on the first two points in defining a model of the quality of experience perceived by a client over a video session. First we we express the quality of a video segment following the logarithmic law:
where is the minimum quality available for the segment, is the quality of the considered segment and and are specific factors that vary depending on the type of the displayed content.
Taking into consideration the second factor, we define a temporal quality variance penalty for two following video segments selected at qualities and as:
where and positive factors that determines how much changes impact the overall experience when a transition to a higher or to a lower quality representation happens.
We finally formulate our quality model as the objective function capturing the defined values, where for a sequence of N segments selected at qualities :
Iii-B2 Cost Factors Considerations
We can include other metrics in our model to account for the network incentives; for instance, a network operator may have a peering agreement such that it incurs a cost linear with the amount of bandwidth. Therefore we can introduce a cost with some price per unit of data transfer; also, there may be a cost associated with the storage of data at the edge cache. Namely, if is the amount of data in transit from the servers to the clients which is stored on the edge cache at time , we can introduce a storage cost .
We can then subtract these costs from the utility to get a joint objective which takes into account the QoE of the end-user (or the benefit to the end-user as a potential revenue to the operator) minus the costs associated with the system.
Iii-C Rate and Streaming Algorithms
We model our system, introduced in Section III, as a composition of four main blocs as shown in Figure 2: a web server, that has complete availability of all the segments composing the video; a long distance network, an access network that uses a caching system in order to support the streaming process; finally, a mobile client attached to the network through a wireless interface. While more than one access interface could be considered, we first focus our analysis on a single access interface. Resources available on the two hops between the server and the client are regulated by the available capacity in the core network for the path connecting the server to the access network where the caches are located, which we will name the core network represented by a function , and by the throughput available to the client for the wireless link giving the client access to the network, which we will name the edge network represented by a function .
Let us first introduce a supporting example to build intuition behind the presented model. In contrast to current solutions where the logic employed to perform the bitrate adaptation process is based on estimation of available resources perceived in the recent past, we consider a partially anticipative case in which a finite window of future edge and core network capacity variations are known beforehand. This information is used by the network controller to schedule which video segments to download in the upcoming future, by either transferring directly the content from the server to the client, or by using the available caches as support.
Assume that at time , we know the evolution of the available capacities and , where is the future knowledge window size, and we do not know the capacity beyond . Consider the simple scenario represented in Figure 2(a), where and is constant at 300kbps and varies from a first 2 seconds period at 100kbps to a second period at 500kbps. In this context a client wants to retrieve a video divided in segments of size 2 seconds and available at three different representations: 100kbps, 300kbps and 600kbps. Initially the video buffer is empty, so let us assume that the first segment of video will be downloaded at minimum quality to reduce the startup wait time for the client. This corresponds to downloading 200kbits of data which in our example, given the initial bottleneck of the wireless link will take 2 seconds to complete. Once this happens, the video client has 2 seconds of video available for display and it should try to download the next segment by this amount of time otherwise a rebuffering event would occur (i.e. the video client remains stuck waiting for more video data to be available). We consider the amount of time until the moment the next downloaded segment will be displayed its download deadline . in our example will then occur at the time of 4 seconds. In current scenarios, where downloads are controlled by the client, only two representations would meet the deadline: 100kbps (which would allow for more segments to be downloaded) and 300kbps. In our anticipative case, we know that even though the first download will be bandwidth limited by the current bottleneck, we are actually under-utilizing the network resources as 200kbps of unused capacity have not been exploited during the first segment download. As we are scheduling downloads in advance, when evaluating the amount of time needed to download next segment we consider the unused capacity as data that might actually have been downloaded to the local caches. The actual amount of data that could have been downloaded to a location in the network (i.e. the caches) is easily calculated from considering the cumulative versions of and , that we will call and , as:
In our particular case, is highlighted in black in Figure 2(b) and corresponds to 400kbits. Now we can assume that this amount of data could have already been transferred close to the edge network preventing the core network from being the bottleneck in the second period. This difference can again be noticed in Figure 2(b): given that the amount of downloaded data always corresponds to the minimum of the two functions and (as a client can never download more than what is allowed by the bottleneck in the network), at the time of 2 seconds this would correspond to 200kbits. Assuming no preemptive caching is applied, the amount of data that could be downloaded in the second period is again delimited by the minimum of the two lines and in this case it would correspond to the plain line representing . If we assume that the core network has moved to the caches an amount of data corresponding to the unused capacity, the minimum has now to be taken between and the dotted version of .
Our algorithm applies these concepts of deadlines and evaluations of unused capacity to explore the state space of valid combinations of segment bitrate downloads to select the one in the time window that produces the highest QoE utility value described in Section III-B without ever occurring in rebuffering events (hence always meeting the set deadlines).
Bitrate Selection Algorithm We now formalize the provided concepts; we define the streaming process as a combination of two tasks for each window of future knowledge: first the controller schedules segments downloads given the available predicted resources and the client reproduction status; second the delivery of the segments to the the mobile hosts with use of the edge caches. The core of the scheduling algorithm is based on a recursive function performed at the beginning of each time window as described in Algorithm 1; this function searches for an optimal path (sequence of segment bitrates) among all possible combinations. Starting from time and given a potential bitrate for the next required segment with index determines the effect of such representation on the download process and, once consumed the entire future knowledge window, determines the QoE of the selected path. The function calculates the download time for the given segment using the available throughput and residual capacity from previous steps. Each time the function is called from the base algorithm, the starting residual capacity is assumed to be zero. In recursive calls, the residual capacity in the core network, if available according to the known cumulative throughput functions and , may be adjusted. As long as the knowledge window limit is not reached, the recursion follows. Once reached the end of the window (or potentially the end of the video), the utility function value of the current path is calculated and compared with the best path previously found and only the better of the two is kept. In case a rebuffering time event is detected, the path is declared invalid and the function returns. Our algorithm can be summarized in four main steps:
Avoid overrunning the client buffer by eventually waiting until some space is created (lines 9 to 17).
Calculate the download deadline for the considered segment () and evaluate given the previously accumulated unused capacity if the deadline can be met by calculating the amount of time necessary to transfer the required data (); if not return the previously found best path (lines 18 to 24).
If the future window has been consumed or the end of the file is reached return the between the current path and the previously best path based on their utility value (lines 25 to 27 and lines from 35 to 37).
Otherwise recursively evaluate the same function for the next segment (index ) over its possible representations. (lines 28 to 34).
After completing the recursive process, the sequence of segments with the highest QoE value is returned and the controller can use it to instruct the other components on how to proceed (i.e. instruct the clients and the caches on how to proceed for downloading the selected segments).
Interleaving windows. The main characteristic of our algorithm is that it greedily tries to use as many resources as are available in the time window without much consideration for the following time slots. Since the QoE cost of switching to a higher bitrate is lower than the cost of switching to lower bitrates, the consequence of such behavior is the tendency to select higher bitrates toward the end of the window to consume the remaining capacity available. This is not always the optimal path in the longer run, since it might be necessary to choose a lower bitrate at the beginning of the next time window, and thus suffer the QoE drop due to switching to a lower bitrate immediately afterwards. To avoid this problem, we consider an alternative solution: while we still apply the same algorithm for the complete window of size to select the best path, we only apply the obtained optimal path until an earlier moment , where is smaller than . This way we prevent a higher bitrate from being selected in the last of the time window , avoiding possible quality drops at the beginning of the next window. After this is done, the new considered window will start from time .
Cost analysis. The cost of the algorithm can be exponential if the bandwidth considered is always bigger than video bitrate (i.e. all possible representations could be downloaded). While this obviously is an obstacle towards the deployment of this algorithm, a set of simple measures and considerations can be taken into account to have an effectively lesser actual cost in real deployments: (1) as quality transitions (in particular negative ones) negatively affect the final utility value, we can set a limit on the number of these events; (2) if the bottleneck bandwidth is always higher than a certain bitrate during the duration of a time window, all video representations with lower bitrate can be left out of consideration (3) in most cases the total number of downloaded segments will be limited as the window size is limited.
In this section, we evaluate the gains achieved by our system and joint rate adaptation through a set of MATLAB based simulations. In order to understand its potential, we implemented the core logic of our system and compared it to the behavior of common DASH implementations. While different proprietary algorithms are used in some of the available commercial solutions (e.g. Apple HLS, Adobe HDS, etc.), we implement our baseline following the behavior in the logic implemented by Netflix-like video services, which can be summarized in two main characteristics:
the DASH client downloads and keeps only video segments for the following seconds of playback at any given time (i.e. the buffer size is limited by time, not data space);
the DASH client logic adapts video quality by a moving average of the data rate estimates experienced on the previous segments delivered (we set to 5 for our experiments).
While it is easy to identify a wide selection of factors that might affect the final results of our simulations, we try to fairly compare the baseline with the two variants of our algorithm by applying for each run the same conditions (i.e. evolution of network infrastructure resources during the experiment time). In the next two subsections we will describe the model used to characterize these resources and the video data set employed in our tests.
Video Dataset and QoE Model. For this set of simulations, we used a video content of 5 minutes of length. The video is divided into 2 second long segments, with each segment available in three different bitrate representations (1 Mbps, 400 kbps and 100 kbps for the first case, 2 Mbps, 1.2 Mbps and 300 kbps for the second). While our system supports Variable Bitrate for the the video segments, we use only constant bitrates for these simulations (i.e. all segments at the same quality level have the same size). The Quality of Experience is calculated following the description provided in Section III-B, using parameters: , , = .
System Resources and Network Model. In our simulations we do not take into consideration availability of video segments at different servers; we use a single server that has the desired video content available at all times. Moreover, we do not consider any limit in the cache size of the intermediate nodes. We provide results using two different network models:
first, we model the available network bandwidth as two finite-state, discrete-time Markov chains, where transitions occur at constant times, every 2 seconds, and transitions occur only between the two nearest states; this is done to try to capture slow variations attributable to client mobility (for the wireless links) and evolution in congestion for the core network. These Markov chains are:
where represents the transition matrix for the wireless access network where for each state the corresponding bandwidth is shown in (expressed in kbps); the same values are shown for the core network in and .
second, we use a real cellular trace, collected and presented in . This trace consists of multiple days worth of data collected in a real metropolitan environment in a European city, using a HSDPA modem while commuting with public transportation. The core network is modeled similarly to the previous case, using the same transition matrix but with the following values for the available throughput: .
Results. We evaluate our system under two varying factors: video buffer size available to the DASH client and window of future knowledge of available bandwidth at the two analyzed network components. Figure 4 collects all the results obtained. For each of the data points represented, we repeated the experiment 5 times and collected the average result. This does not apply for the baseline in Figures 3(g) and 3(h) as the future window size can vary only for our algorithms; for these cases we used a single data point using client buffer size of 20 seconds. In general, the results confirm the overall benefit of our system with gains of at least 15% points in the utility value for both our algorithm versions. This not only corresponds to a more stable experience (as variations, in particular decreases in quality, strongly affect the final QoE value), but also in a higher average bitrate quality for all the experiments analyzed.
As expected, the buffer size available for the clients is not a major influencing factor for the analyzed use case, as neither algorithm uses the buffer size information to modify its adaptation logic; we expect this parameter to gain more importance for longer videos, where a low bandwidth period might be better compensated by accumulated buffer.
More interestingly, we can notice that increasing future knowledge window size also increases the computational cost of schedule delivery time deadlines, since there are now longer periods of time that are considered in the calculation. The achieved gains in this case might not justify the additional processing time costs.
Figure 5 shows the employed memory usage at the network caches. In particular, it shows the cache size for the single client experiment of the first analyzed case. This value enables us to understand the resource cost of running our system, and is of particular interest to service providers that would implement such a caching scheme as scalability is directly affected by the consumed resources. The amount of video data buffered at the cache nodes using our algorithm ranged from up to , meaning that it would scale well for larger videos and higher number of clients and not impose a high resource cost for the service provider.
In order to optimize delivering video content to clients, we have incorporated the knowledge of network throughput in a short future time window. Our model is based on DASH, that is now used in most major video streaming services, and in comparison with DASH it improves the usage of the available end-to-end throughput between the server and the client. Instead of having the entire network throughput equal to the bottleneck of the used network path, we optimize the video delivery method, utilizing the intermediate nodes between the server that has the whole video content available, and the client, as an intermediate video content cache. Using this cache enables us to increase the effective throughput in certain parts of the network path. In times when the core network has higher throughput than the edge network, we increase the cache, and then use the cached content in periods when the edge network is faster.
We evaluated the gains achieved by using our method with a set of MATLAB simulations. We implemented our algorithm and compared it with the behavior of DASH implementations against synthetic and real throughput traces. Our evaluations demonstrate that our method of video delivery improves the Quality of Experience for the end user by keeping the bitrate more stable and achieving significantly higher average bitrates than the benchmarks. While it does introduce some additional costs for the content provider, we claim that the resource costs are minimal, and the achieved gain outweighs the computational costs.
-  Cisco Visual Networking Index: Forecast and Methodology, 2013â2018. http://www.cisco.com/c/en/us/solutions/collateral/service-provider/ip-ngn-ip-next-generation-network/white_paper_c11-481360.html/.
-  Giuseppe Bianchi and Riccardo Melen. The role of local storage in supporting video retrieval services on atm networks. IEEE/ACM ToN, 5(6), 1997.
-  Iraj Sodagar. The mpeg-dash standard for multimedia streaming over the internet. MultiMedia, IEEE, 18(4):62–67, 2011.
-  Van Jacobson, Marc Mosko, D Smetters, and JJ Garcia-Luna-Aceves. Content-centric networking. Whitepaper, Palo Alto Research Center, pages 2–4, 2007.
-  Christian Dannewitz. Netinf: An information-centric design for the future internet. In Proc. 3rd GI/ITG KuVS Workshop on The Future Internet, 2009.
-  Teemu Koponen, Mohit Chawla, Byung-Gon Chun, Andrey Ermolinskiy, Kye Hyun Kim, Scott Shenker, and Ion Stoica. A data-oriented (and beyond) network architecture. In ACM SIGCOMM Computer Communication Review, volume 37, pages 181–192. ACM, 2007.
-  C Westphal (ed). Adaptive streaming over icn, irtf icnrg wg, draft-irtf-icnrg-videostreaming-02.
-  Reinhard Grandl, Kai Su, and Cedric Westphal. On the interaction of adaptive video streaming with content-centric networking. arXiv preprint arXiv:1307.0794, 2013.
-  Junghwan Lee, Jaehyun Hwang, Nakjung Choi, and Chuck Yoo. Svc-based adaptive video streaming over content-centric networking. KSII Transactions on Internet and Information Systems (TIIS), 7(10):2430–2447, 2013.
-  Francesco Malandrino, Maciej Kurant, Athina Markopoulou, Cedric Westphal, and Ulas C Kozat. Proactive seeding for information cascades in cellular networks. In IEEE INFOCOM’12, pages 1719–1727. IEEE, 2012.
-  LR Ariyasinghe, C Wickramasinghe, PMAB Samarakoon, UBP Perera, RA Prabhath Buddhika, and MN Wijesundara. Distributed local area content delivery approach with heuristic based web prefetching. In IEEE ICCSE’13, pages 377–382, 2013.
-  Zheng Lu and Gustavo de Veciana. Optimizing stored video delivery for mobile networks: the value of knowing the future. In INFOCOM, 2013 Proceedings IEEE, pages 2706–2714. IEEE, 2013.
-  Athula Balachandran, Vyas Sekar, Aditya Akella, and Srinivasan Seshan. Analyzing the potential benefits of CDN augmentation strategies for internet video workloads. In ACM IMC’13, pages 43–56, 2013.
-  Louis Plissonneau and Ernst Biersack. A longitudinal view of http video streaming performance. In ACM MSC’13, pages 203–214. ACM, 2012.
-  Alessandro Finamore, Marco Mellia, Maurizio M Munafò, Ruben Torres, and Sanjay G Rao. Youtube everywhere: Impact of device and infrastructure synergies on user experience. In ACM IMC’11, pages 345–360. ACM, 2011.
-  Zhe Li, Mohamed Karim Sbai, Yassine Hadjadj-Aoul, Annie Gravey, Damien Alliez, Jeremie Garnier, Gerard Madec, Gwendal Simon, and Kamal Singh. Network friendly video distribution. In International Conference on the Network of the Future, volume 1, 2012.
-  Fangming Liu, Bo Li, and H Jin. Peer-assisted on-demand streaming: characterizing demands and optimizing supplies. IEEE Trans on Computers, 2013.
-  Hongqiang Harry Liu, Ye Wang, Yang Richard Yang, Hao Wang, and Chen Tian. Optimizing cost and performance for content multihoming. ACM SIGCOMM CCR, 42(4):371–382, 2012.
-  Xi Liu, Florin Dobrian, Henry Milner, Junchen Jiang, Vyas Sekar, Ion Stoica, and Hui Zhang. A case for a coordinated internet video control plane. In ACM SIGCOMM’12, pages 359–370, 2012.
-  Cheng Huang, Angela Wang, Jin Li, and Keith W Ross. Understanding hybrid cdn-p2p: why limelight needs its own red swoosh. In ACM NOSSDAV’08, pages 75–80. ACM, 2008.
-  Cheng Huang, Jin Li, and Keith W Ross. Can internet video-on-demand be profitable? In ACM SIGCOMM CCR, volume 37, pages 133–144. ACM, 2007.
-  Shudong Jin, Azer Bestavros, and Arun Iyengar. Network-aware partial caching for internet streaming media. Multimedia systems, 9(4):386–396, 2003.
-  Zhourong Miao and Antonio Ortega. Proxy caching for efficient video services over the internet. In Packet Video Workshop (PVW’99), 1999.
-  Keqiu Li, Keishi Tajima, and Hong Shen. Cache replacement for transcoding proxy caching. In IEEE/WIC/ACM Web Intelligence’05, pages 500–507, 2005.
-  Wenyu Qu, Keqiu Li, Hong Shen, Yingwei Jin, and Takashi Nanya. The cache replacement problem for multimedia object caching. In Semantics, Knowledge and Grid, 2005. SKG’05. First International Conference on, pages 26–26. IEEE, 2005.
-  Subhabrata Sen, Jennifer Rexford, and Don Towsley. Proxy prefix caching for multimedia streams. In IEEE INFOCOM’99, volume 3, pages 1310–1319, 1999.
-  Seong-Ho Park, Eun-Ji Lim, and Ki-Dong Chung. Popularity-based partial caching for vod systems using a proxy server. In Proceedings of the 15th International Parallel & Distributed Processing Symposium, page 115. IEEE Computer Society, 2001.
-  Songqing Chen, Haining Wang, Xiaodong Zhang, Bo Shen, and Susie Wee. Segment-based proxy caching for internet streaming media delivery. MultiMedia, IEEE, 12(3):59–67, 2005.
-  Zhisheng Yan, Jingteng Xue, and Chang Wen Chen. QoE continuum driven HTTP adaptive streaming over multi-client wireless networks. In Multimedia and Expo (ICME), 2014 IEEE International Conference on.
-  Wei Pu, Zixuan Zou, and Chang Wen Chen. Video adaptation proxy for wireless dynamic adaptive streaming over http. In Packet Video Workshop (PV), 2012 19th International, pages 65–70. IEEE, 2012.
-  Vengatanathan Krishnamoorthi, Niklas Carlsson, Derek Eager, Anirban Mahanti, and Nahid Shahmehri. Helping hand or hidden hurdle: Proxy-assisted http-based adaptive streaming performance. In Proc. IEEE MASCOTS, 2013.
-  Thomas Stockhammer. Dynamic adaptive streaming over http–: standards and design principles. In Proceedings of the second annual ACM conference on Multimedia systems, pages 133–144. ACM, 2011.
-  Te-Yuan Huang, Ramesh Johari, Nick McKeown, Matthew Trunnell, and Mark Watson. A buffer-based approach to rate adaptation: Evidence from a large video streaming service. In Proc. ACM SIGCOMM, 2014.
-  Te-Yuan Huang, Nikhil Handigol, Brandon Heller, Nick McKeown, and Ramesh Johari. Confused, timid, and unstable: picking a video streaming rate is hard. In Proceedings of the 2012 ACM conference on Internet measurement conference, pages 225–238. ACM, 2012.
-  Christopher Müller, Stefan Lederer, and Christian Timmerer. An evaluation of dynamic adaptive streaming over http in vehicular environments. In Proceedings of the 4th Workshop on Mobile Video, pages 37–42. ACM, 2012.
-  Haakon Riiser, Håkon S Bergsaker, Paul Vigmostad, Pål Halvorsen, and Carsten Griwodz. A comparison of quality scheduling in commercial adaptive http streaming solutions on a 3g network. In Proceedings of the 4th Workshop on Mobile Video, pages 25–30. ACM, 2012.
-  Anthony J Nicholson and Brian D Noble. Breadcrumbs: forecasting mobile connectivity. In Proceedings of the 14th ACM international conference on Mobile computing and networking, pages 46–57. ACM, 2008.
-  Qiang Xu, Sanjeev Mehrotra, Zhuoqing Mao, and Jin Li. Proteus: network performance forecast for real-time, interactive mobile applications. In Proceeding of the 11th annual international conference on Mobile systems, applications, and services, pages 347–360. ACM, 2013.
-  Athula Balachandran, Vyas Sekar, Aditya Akella, Srinivasan Seshan, Ion Stoica, and Hui Zhang. A quest for an internet video quality-of-experience metric. In Proceedings of the 11th ACM Workshop on Hot Topics in Networks, pages 97–102. ACM, 2012.
-  Methodology for the subjective assessment of the quality of television pictures. Recommendation ITU-R BT.500-13, 2012.
-  Haakon Riiser, Paul Vigmostad, Carsten Griwodz, and Pål Halvorsen. Commute path bandwidth traces from 3g networks: analysis and applications. In Proceedings of the 4th ACM Multimedia Systems Conference, pages 114–118. ACM, 2013.