Next Article in Journal
An End-Process Blockchain-Based Secure Aggregation Mechanism Using Federated Machine Learning
Previous Article in Journal
The Development of a Compact Pulsed Power Supply with Semiconductor Series Connection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Scalable Video Streaming (ASViS): An Advanced ABR Transmission Protocol for Optimal Video Quality

by
Eliecer Peña-Ancavil
*,†,
Claudio Estevez
,
Andrés Sanhueza
and
Marcos Orchard
Department of Electrical Engineering, Faculty of Physical and Mathematical Sciences, Universidad de Chile, Beauchef 850, Santiago 8370456, Chile
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2023, 12(21), 4542; https://doi.org/10.3390/electronics12214542
Submission received: 11 October 2023 / Revised: 24 October 2023 / Accepted: 27 October 2023 / Published: 4 November 2023
(This article belongs to the Section Computer Science & Engineering)

Abstract

:
Multimedia video streaming, identified as the dominant internet data consumption service, brings forth challenges in consistently delivering optimal video quality. Dynamic Adaptive Streaming over HTTP (DASH), while prevalent, often encounters buffering problems, causing video pauses due to empty video buffers. This study introduces the Adaptive Scalable Video Streaming (ASViS) protocol as a solution. ASViS incorporates scalable video coding, a flow-controlled User Datagram Protocol (UDP), and deadline-based criteria. A model is developed to predict the behavior of ASViS across varying network conditions. Additionally, the effects of diverse parameters on ASViS performance are evaluated. ASViS adjusts data flow similarly to the Transmission Control Protocol (TCP), based on bandwidth availability. Data are designed to be discarded by ASViS according to video frame deadlines, preventing outdated information transmission. Compliance with RFC 8085 ensures the internet is not overwhelmed. With its scalability feature, ASViS achieves the highest possible image quality per frame, aligning with Scalable Video Coding (SVC) and the available data layers. The introduction of ASViS offers a promising approach to address the challenges faced by DASH, potentially providing more consistent and higher-quality video streaming.

1. Introduction

Over recent years, a marked transformation in the realm of digital communication has been observed, with multimedia video streaming emerging as the predominant internet data service. Some had predicted that by 2020, video traffic would account for 82% of global online traffic [1]. Contrary to these predictions, it is reported in [2] that in the first half of 2022, video streaming constituted 65.93% of total internet traffic, with Netflix and YouTube contributing 13.74% and 10.51%, respectively. Furthermore, a 23% increase in total traffic volume was observed in 2022 compared to 2021, attributed to the significant growth of various streaming services. Current data regarding video streaming consumption on mobile networks indicate that 71% of the traffic was video-related [3]. Projections suggest that this figure is expected to reach 80% by 2028. Such statistics underscore the undeniable growth trend. This evolution has been attributed not only to technological innovations but also to the integration of intelligent devices. Platforms, such as Zoom, Microsoft Teams, Netflix, YouTube, and Amazon Prime, have been recognized as significant players in this revolution. Moreover, unexpected external events, notably the Coronavirus disease pandemic, have acted as catalysts, propelling an unprecedented demand for online communication and entertainment platforms. This surge, documented as a 20–40% increase in video streaming [4,5], occasionally strained available bandwidth capacities, necessitating temporary transitions from high definition (HD) to standard definition (SD) [6] by some service providers.
Dynamic Adaptive Streaming over HTTP (DASH) is an open standard for transmitting video that adapts to both the network and device. In this evolving landscape, the DASH protocol, operating over the Transmission Control Protocol (TCP), has been identified as the primary delivery model for video streaming services [7,8,9]. This open-source standard, celebrated for its flexibility and codec-agnostic features, has established itself as an indispensable tool for most streaming entities [10,11]. Nevertheless, challenges associated with DASH, including buffering interruptions, have signaled an imperative need for refined protocols that can mitigate such disruptions [8].
In response to such challenges, focus within the research community has shifted towards optimizing Adaptive Bitrate (ABR) algorithms. One strategy involves having the client store chunks sent from the server on the local disk [7,8,9]. This stored content starts to play before the entire content is downloaded. It also helps tolerate network outages without interrupting playback by using the stored content on the local disk [12]. For clarity, the video content stored before playback on the client side is termed a Video Buffer (VB). This VB definition distinguishes it from the buffer term used in the context of the User Datagram Protocol (UDP) and Transmission Control Protocol (TCP) transmissions. The overarching objective remains clear: to develop a protocol that addresses DASH’s inherent limitations and sets the groundwork for future streaming innovations. At the forefront of these explorative endeavors is Scalable Video Coding (SVC), a technique that, although previously underutilized, has been underscored for its potential to reshape the video streaming paradigm [13].
This paper introduces the Adaptive Scalable Video Streaming (ASViS) protocol, envisioned as a groundbreaking ABR transmission approach. At its core, ASViS leverages Scalable Video Coding (SVC) to partition video data into base and enhancement layers, with selective transmission based on layer-discarding policies. This differentiation is orchestrated by the application of a deadline-sensitive criterion. Specifically, each layer possesses a distinct “curfew”, determining its transmission or omission based on the nearing of its playback deadline. For instance, as deadlines approach, higher-layer data might be excluded, ensuring the arrival of the base layers, which are more crucial for video coherence, are transmitted. ASViS, with its innovative feature set, promises not only to streamline video data transmission but also to elevate the end-user experience. The efficacy of ASViS is substantiated through comprehensive performance assessments, juxtaposed against the established Model Predictive Control (MPC) algorithm. Evaluation metrics, such as the Peak-Signal-to-Noise Ratio (PSNR) and the novel Video Multi-method Assessment Fusion (VMAF) technique by Netflix, are employed to gauge its performance.
For clarity and coherence, this paper is meticulously structured. Following this introduction, Section 2 delves into the intricacies of video coding and key quality metrics. Section 3 details the related work with the mechanics of streaming and ABR strategies. Section 4 elucidates the attributes and operational dynamics of ASViS, emphasizing its interplay with SVC. In Section 5, the simulation parameters adopted for ASViS evaluations are detailed, and the primary findings are presented. The paper concludes by drawing insights from the results, underscoring their significance in the evolving landscape of multimedia video streaming.

2. Background

2.1. Video Coding

The H.264 standard [14], also known as Advanced Video Coding (AVC) Part 10, is a predominant video compression standard in the industry [15,16]. This standard facilitates lossy compression by capitalizing on the similarity between adjacent pixels and the movement of objects across consecutive frames. Within AVC, a dependent frame structure termed a Group Of Pictures (GOP) is utilized [17]. Three types of frames are encompassed in AVC: Intra (I) frames, Predicted (P) frames, and Bidirectional (B) frames [8,18,19].
  • I frames:
    Nature: these frames encapsulate information solely from a single picture.
    Compression Method: spatial redundancy within the picture is utilized for compression.
    Size and Independence: typically larger in size, they can be decoded without relying on other frames.
    Significance: aerve as reference points, with subsequent frame types drawing on information from I frames during the decoding process.
  • P frames:
    - Nature: information from previous frames is used to predict and represent the current frame.
    - Compression Method: both spatial redundancy and temporal prediction (based on previously decoded frames) are employed.
    - Efficiency: encoding of P frames is more bit-efficient than I frames.
  • B frames:
    - Nature: these frames interpolate between preceding and subsequent frames.
    - Compression Method: they make use of both forward (from subsequent frames) and backward (from prior frames) temporal predictions.
    - Efficiency: B frames typically represent the most bit-efficient frame type in the bitstream.
Each GOP commences with an I frame and is followed by a sequence of P and B frames in a hierarchical manner. The GOP length signifies the interval between two I frames, inclusive of the initial I frame. Notably, the Transmission Order (TO) of video frames differs from the Display Order (DO), the latter of which pertains to the sequence during video playback. Dependencies are determined by hierarchy, with lower hierarchical frames depending on information from higher hierarchical frames for reconstruction. Figure 1 illustrates two types of hierarchical dependencies for a GOP of eight frames.
Progress in video coding techniques has introduced the concept of scalability. Within the AVC framework, this concept has been expanded with the introduction of SVC [16,20,21]. In SVC, frames consist of a Base Layer (BL) and one or more Enhancement Layers (EL). Each higher layer is contingent upon all the preceding layers for accurate display. If any layer is absent, subsequent layers are disregarded. SVC can be applied in terms of temporal resolution, spatial resolution, or quality. While SVC maintains the same GOP hierarchy as AVC, it introduces a parallel structure based on the number of ELs [22,23].
The term “scalability” in video streaming refers to a segmented layered coding system. This system allows for the removal of higher layers while still retaining the ability to decode the video [16,20]. Such a design minimizes information loss during packet drops, ensuring that videos maintain their integrity even with increased frame sizes.
SVC coding, in terms of the GOP hierarchy, mirrors the structure found in AVC. However, a distinguishing feature of SVC is its introduction of a parallel structure. This structure emerges from the number of ELs, all of which are contingent upon the BL [21,22,23]. A visualization of the SVC hierarchy for I–B with two layers and a GOP size of eight frames is provided in Figure 2, illustrating an IBBBBBBB pattern.
To determine the total number of layers, one must factor in the variety of frame types (I, B) and the number of SVC layers (such as BL, EL1, EL2, etc.). In the context of this research, four primary layers are discerned:
  • Base I frame (BI)
  • Enhancement I frame (EI)
  • Base B frame (BB)
  • Enhancement B frame (EB)
The base video data are housed within the BI and BB layers, while the enhancement video data reside in the EI and EB layers. For illustrative and experimental intents, this study adopts a GOP size of eight, following an IBBBBBBB (IB7) pattern, as depicted in Figure 1b. Two quality layers, BL and EL, are considered, with P frames being excluded from the analysis.
Within the framework of SVC, the quality of a video frame is intrinsically linked to the layers employed during its reconstruction [20,23]. This relationship is exemplified in Figure 3, which provides a visual exploration of the impact on spatial Y-PSNR when specific frame types are omitted.
Three distinct scenarios are considered in this test:
  • No Loss: This scenario serves as a reference, wherein Figure 3a displays the spatial Y-PSNR solely resulting from frame compression and subsequent restoration.
  • Discarding All EB: In this scenario, evident in Figure 3b, the frame exhibits regions with diminished Y-PSNR. This decrease is predominantly observed in areas undergoing pronounced color transitions, typically indicative of movement.
  • Retaining Only BI: As portrayed in Figure 3c, this scenario experiences a significant information deficit. When both the BL and EL of a frame are absent, the resultant reconstruction contains voids. While it is feasible to populate these gaps with data from preceding frames, the primary aim of this experiment is to assess the impact without invoking supplementary measures.

2.2. Video Quality Measure

Determining video quality has long been debated due to its reliance on numerous subjective factors. The challenge lies in transforming these subjective perceptions into objective metrics [24,25]. Taking into account the human eye process of translating visible light into visual data, the prominence of photoreceptors becomes apparent [26].
Photoreceptors, crucial components of the human retina, facilitate the transformation of visible light into biological signals, allowing visual representation of the external environment [27]. The majority of these photoreceptors are rod cells, making up approximately 95% and predominantly recognizing luminance. On the other hand, cone cells account for approximately 5% and are chiefly responsible for color recognition [26,27]. Given this distribution, the human retina heightened sensitivity to brightness over color has led to video quality measures often concentrating solely on the luma channel.

2.2.1. Peak-Signal-to-Noise Ratio

The PSNR, a standard measure of video quality, characterizes the relationship between a signal maximal energy and its associated noise, presented in decibels (dB) [25,28]. Although both luma and chrominance components of a frame can be evaluated by PSNR, the emphasis is predominantly placed on the luma channel. The initial step in determining PSNR involves calculating the Mean Squared Error ( M S E ), expressed in Equation (1) [29]:
M S E = 1 m n i = 0 m 1 j = 0 n 1 I ( i , j ) H ( i , j ) 2
In this equation, pixel values of the original and reconstructed frames are represented by I ( i , j ) and H ( i , j ) , respectively, with m and n indicating the frame horizontal and vertical dimensions. PSNR can then be calculated using Equation (2), with M A X I representing the frame maximum pixel value:
P S N R = 10 l o g 10 M A X I 2 M S E
While PSNR has been a foundational metric in video quality assessment, its capability to fully represent the subjective perception of end users has been a topic of scrutiny. This is due to its focus on mathematical differences at the pixel level, which does not always align with human perception. Especially in adaptive streaming systems, there is a pressing need for metrics that can more accurately reflect human experiences.

2.2.2. VMAF

In response to these limitations, Netflix introduced VMAF, a video quality metric crafted to mirror human perception more accurately [25,30]. By amalgamating various video quality evaluation methods using a Support Vector Machine (SVM), VMAF produces a perceptual quality score [12,24,31]. This integrated approach ensures the holistic representation of individual measures, bolstering its correlation with subjective assessments. Validations across a multitude of video resolutions and datasets indicate VMAF scores, ranging between 0 and 100, align closely with subjective evaluations [31,32].
To compute VMAF, spatial and temporal video quality measures are initially extracted as feature maps [12,31]. Spatial metrics are derived from the Detail Loss Metric (DLM) and Visual Information Fidelity (VIF), while the temporal metric employs Temporal Information (TI). Detail losses, representing the absence of crucial visual data impacting content clarity, are evaluated across four scales using DLM. In contrast, VIF discerns these losses and visual information fidelity across four distinct scales. TI gauges the luminance difference between frame pairs to identify temporal variations resulting from movement, encapsulated by six unique features [31]. Following this, the average value of each feature is ascertained and channeled into a pre-trained SVM, which then predicts the per-frame quality score, with 0 denoting low quality and 100 signifying the highest quality [32]. The VMAF underlying model was developed using SVM regression on the VMAF+ video database. VMAF+ is a comprehensive Video Quality Assessment (VQA) dataset, comprising 522 videos subjected to diverse scaling and compression levels [33]. During the supervised learning phase, the VMAF model ascertains optimal weights for rudimentary metrics, learning from established ground truth scores [32]. These steps are portrayed in Figure 4. Consistent with other advanced video quality metrics, VMAF predominantly assesses the luma components of the video channel [33].
In adaptive streaming systems such as DASH, accurate video quality metrics become pivotal. Such systems, in real time, determine the optimal video quality to transmit, considering network conditions and device capabilities. A metric misaligned with user perception might lead to suboptimal streaming decisions, either compromising user experience or wasting bandwidth. Thus, the incorporation of user-centric metrics is imperative for effective adaptive streaming decisions.

3. Related Work

3.1. Dash

DASH is known as an open-source standard, offering a modular approach for the delivery of video content between the client and the server. By using Hypertext Transfer Protocol (HTTP) for data transmission, firewall-associated challenges are mitigated, data delivery is assured, and compatibility with a plethora of network protocols is fostered [34,35].
On the server end, videos are typically segmented into K chunks with intervals usually ranging between 1 and 5 s [36,37]. Each chunk can be represented by multiple GOPs using conventional I–P–B or I–B frame hierarchies. Each chunk is encoded based on various quality parameters such as quantization, frame rate, and resolution. It is noted that encoding at heightened resolutions or quantizations requires increased bitrates.
From the client perspective, the ABR algorithm plays a crucial role, dynamically determining the bitrates for each chunk during playback. Once a chunk is initiated for download, transitions between chunks are not possible until the current chunk finishes. The transport layer, susceptible to variations due to fluctuating network conditions, ensures the receive buffer is populated at the rate of available throughput [38]. Achieving balance between the receive and extract rates is vital for a continuous video experience. Nonetheless, network-induced variations can lead to:
  • A lag in enhancing video quality following improved network conditions, thereby underutilizing the available throughput.
  • Potential interruptions in playout if a chunk of higher bitrate is downloaded amidst unfavorable network conditions.
A pivotal consideration in DASH remains the playout frame deadline, hereafter referred to as the deadline. This deadline pinpoints the time by which video frames should be decoded, ensuring a seamless viewing experience. Within a DASH framework, the transmission interval between the server and the client is integral.
Central to DASH, the ABR algorithm aspires to optimize the resolution per frame, facilitate smooth video playout, and ensure timely frame delivery [39]. However, achieving these objectives concurrently can be intricate given the constraints of bandwidth and ever-changing network conditions.

ABR Algorithm Approaches

Various classifications exist for ABR algorithms, including throughput-based, buffer-based, and a synthesis of the two, commonly termed the hybrid approach [40]. It is observed that a majority of ABR algorithms derive their decision-making from a balance between network conditions and video coding nuances. Throughput-focused strategies often use the bandwidth data of the application layer to navigate the choice of bitrate chunks. Techniques such as last-packet download and sliding window moving average are often employed to estimate available bandwidth [41]. However, these approaches may grapple with issues of transient fluctuations in bandwidth.
On the other hand, buffer-centric strategies adopt thresholds within the VB occupancy to make decisions [41]. These techniques underscore the significance of absorbing throughput variations and retaining sufficient content in the VB to avert interruptions [42]. Hybrid strategies typically select bitrate chunks centered around a Quality of Experience (QoE) prediction metric [12,40]. Such metrics might encompass parameters such as throughput, VB status, and even details related to employed technology.

3.2. Modern Video Streaming Solutions

Contemporary research has proposed a plethora of techniques to elevate video streaming performance [43,44]. Solutions, such as the one in [43], implement a finite state machine at the application layer. This strategy deploys bandwidth estimation, VB size on the client side, and individual SVC layer encoding rates to discern the layered blocks for transmission.
Contrastingly, techniques detailed in [44] harness multiple paths to transmit distinct blocks and layers, enhancing the probability of base layer reception. [45] introduces a decision support system, emphasizing the minimization of quality fluctuations between adjacent chunks, potentially elevating user QoE.
Other notable solutions, such as [13], capitalize on software-defined networks to enhance DASH adaptability. While a myriad of physical-application cross-layer solutions such as [46] exist, transport-application cross-layer solutions remain relatively scarce.
In [47], a collaboration between physical, Media Access Control (MAC), and application layers is observed. This alliance aligns the performance of the lower layer data rate with the application layer bitrate via machine learning. Moreover, [48] provides a discourse on the fairness challenge in TCP video transmissions.

Model Predictive Control Algorithm in Video Streaming

MPC, introduced in [49], emerges as a state-of-the-art technique for optimizing QoE in video streaming. MPC presents a hybrid approach, formulating the video ABR algorithm as a stochastic optimal control problem. By leveraging both throughput and VB strategies, it seeks an optimal balance [50]. This methodology employs a sliding look-ahead window to forecast a series of pivotal parameters. The derived predictions are subsequently employed to tackle the optimization problem, aiming to enhance QoE [50].
Several key parameters, pivotal for QoE, have been identified. These parameters encompass the size of chunk k encoded at bitrate level R, denoted as d k ( R k ) , perceived video quality (q), throughput (C), and the VB occupancy at the client end (B) [51].
QoE is articulated using a spectrum of metrics. Among these, the Average Video Quality ( A V Q ) stands as a prime metric, reflecting the overall perceived quality over all chunks:
A V Q = 1 K k = 1 K q ( R k )
The subsequent metric, average video quality variation ( A V Q V ), quantifies the shifts in video quality between sequential chunks:
A V Q V = 1 K 1 k = 1 K 1 q ( R k + 1 ) q ( R k )
Rebuffering ( R e ) stands as a crucial metric, expressing the interplay between the download time of a chunk and the occupancy of the client VB:
R e = k = 1 K ( d k ( R k ) C k B k ) +
It should be noted that rebuffering occurs when the download time of a chunk exceeds the VB occupancy level. The startup delay, represented by T s , signifies the initial waiting period before playback—a phase often utilized to preload the VB and assess the connection quality. The holistic QoE, spanning chunk 1 through K, is synthesized as:
Q o E 1 k = A V Q λ · A V Q V ( K 1 ) β · R e β s · T s
Here, weight parameters λ , β , and β s can be tailored based on the relevance of each QoE component. The notation ( x ) + = m a x x , 0 ensures that the rebuffer parameter in Equation (5) remains non-negative.

4. Proposed Cross-Layer Solution: ASViS

4.1. Overview of ASViS

ASViS, or adaptive scalable video streaming, is a cross-layer solution that bridges the gap between the transport and application layers. Within the application layer, an algorithm is implemented that manages the dispatched information, relying on a layer-discarding policy and deadline-sensitive criteria. While packet loss assists in determining the flow rate, reflecting the behavior of TCP, retransmissions are not essential. Based on the layer-discarding policy, certain packets might be omitted. Utilizing UDP, the system detects missing frame fragments at the application layer. Since there is no requirement to track sequence numbers or acknowledge every packet, UDP becomes a preferred choice. The flow rate is managed by adjusting the number of packets dispatched over a time span. This approach, similar to the congestion window of TCP, is termed here as the flow window or fwnd. ASViS supports Selective Acknowledgment (SACK); however, it is dispatched only when deemed essential. SACK packets are vital for measuring round-trip time ( R T T ), which is required for estimating available throughput. Lost packets might be retransmitted if the time and priority permit. The flow window is principally employed to fulfill with RFC 8085 requirements, which emphasize preventing network saturation [52].
Scalability is an important factor in the growth of any technology. When referring to ASViS, scalability is characterized by several factors: 1. Since ASViS is based on SVC, which itself is part of H.264 (and newer), it can easily evolve to higher resolutions (8 K, 16 K, etc.). 2. SVC does not have a fixed amount of layers; to prove our concept, this work uses four SVC layers, but more can be used. As technology progresses and resolutions increase, it is beneficial to also increase the amount of SVC layers used, since it would create more granularity (i.e., more quality options to choose from). There are other aspects not covered in this work, in which ASViS can scale (such as GOP size, block compression size, etc.) that are inherent to the compression standards. All these degrees of scalability allow ASViS to maintain itself relevant in the future.
Compared to traditional video streaming methods based on TCP, which often rely heavily on retransmissions and sequence acknowledgments, ASViS, with its refined layer-discarding policy and strategic use of UDP, offers a more efficient and adaptable approach. This methodology not only reduces overhead but also aligns closely with the RFC 8085 guidelines. These guidelines are a series of recommendations on UDP usage and best practices designed to prevent network congestion. By complying with these guidelines, stability and efficiency in network operations are ensured. The adherence of ASViS to this standard is underscored by its significance.
A deadline-sensitive approach is employed, which operates on a time-based metric. This metric is focused on the residual time to meet the playout frame deadline. In SVC codification, by default, all layers of a frame share the same deadline. This approach introduces differentiation, suggesting the omission of certain layers based on pre-set criteria using the deadline as a reference. A significant challenge in video streaming emerges when throughput is below the video bitrate. Stream prioritization becomes essential, with low-priority data being discarded as deemed necessary. The inherent structure of SVC layers, where higher layers rely on foundational ones, designates a higher priority to base layers. As the deadline becomes imminent, high-layer data are omitted, and if the situation demands, additional layers are also removed.
The proposed layer-discarding policy suggests having differentiated deadlines for distinct layers. To distinguish it from the frame deadline, the layer deadline is referred to as curfew. Higher layers (lower priority) have earlier curfews, and lower layers (higher priority) have more lenient curfews. The curfew spacing is an important design parameter. The algorithm may accept a low-priority layer at the cost of running out of time to process a subsequent higher-priority layer if a late curfew (short present-to-deadline period) is used. However, if the curfew is too early, the algorithm response may be premature, and higher-layer (enhancement) packets can be unnecessarily discarded. The decision-making process is a critical part of the proposed solution. It differentiates discarding criteria to specific layers for an SVC video with two layers (BI, BB, EI, and EB), as shown in  Figure 5 where all layers of a frame have a defined threshold.
  • Data received after the deadline are discarded, regardless of their priority.
  • Only BI layers are considered for data that arrive between the deadline and the first curfew.
  • At the second threshold, both BI and BB layers are accepted.
  • When the third threshold is reached, BI, BB, and EI layers are permitted.
  • Beyond the third threshold, all layers, inclusive of EB, are accepted.
Figure 5. Thresholds for an SCV video with 2 layers.
Figure 5. Thresholds for an SCV video with 2 layers.
Electronics 12 04542 g005

Deadline-Sensitive Criteria Algorithm (DSCA)

The decision-making process, shown in Algorithm 1, consists of reading a packet and extracting the frame (I or B) and layer (base or enhancement) information. Based on this, it computes the curfew, which considers the current time, frame deadline, travel time, and priority. The curfew is computed only on packets that are inside the current fwnd, as later packets depend on current decisions. Higher-priority packets have curfews closer to the deadline, while lower-priority packets are discarded earlier to increase the chances of a higher-priority packet arriving on time. If a packet inside the current fwnd is discarded, a later packet is pushed to an earlier time, increasing its probability of successfully arriving before the deadline. This is the fundamental principle, increasing the chances of successfully arriving before the deadline.
Algorithm 1 Packet Decision Process
1:
while process not ended do
2:
    ObtainNextPacketInfo( frame , layer )
3:
    if  ArrivesBeforeDeadline ( frame , layer )  then
4:
        if  IsLastPacket ( )  then
5:
           EndProcess()
6:
        else
7:
           SendPacket( frame , layer )
8:
        end if
9:
    else
10:
        DiscardPacket( frame , layer )
11:
    end if
12:
end while

4.2. Modeling ASViS

The quality of received video is contingent on available bandwidth. ASViS is designed to manage packet discards when bandwidth is limited. At the application layer, flow control is implemented, similar to TCP behavior. However, unlike TCP, ASViS allows the sender to decide on packet discards, which is why UDP is utilized. The receiver responds with SACK, enabling the sender to adjust the flow upon detecting packet losses.
The initial step involves mapping video data to packets. The simulation assumes that each video frame is transmitted in an individually assigned packet, or multiple packets if the frame size exceeds the Maximum Segment Size ( M S S ). However, multiple video frames are not encapsulated in a single packet to avoid losing multiple frames with a single packet loss. To maintain clarity, we use the term “frame” strictly in the context of a video image, distinguishing it from a data-link layer unit.
The video bitstream is denoted as b S = d T t T , where d T represents the total data to be transmitted, including all layers, and t T is the total encoding time. Total data can be expressed as:
d T = f = 1 F T l = 1 L d f , l .
F T is the total amount of frames and L is the total amount of layers per frame. d f , l is the amount of data needed to transmit the l t h layer of the f t h frame. The bitstream is varied according to the available bandwidth, considering discarded packets, so the effective bitstream is:
b F = d F t T and d F = f = 1 F T l = 1 l F d f , l .
l F is the number of layers transmitted for frame f, where l F could have a value of 0 if the whole frame is discarded. Since the transport layer is flow-controlled, it is necessary to determine the number of packets required. The flow can be controlled on a packet or byte basis. In most cases, protocols operate on a packet basis. The amount of packets necessary to transmit all the frames that are not discarded (arrive successfully) is:
S f , l = d f , l M S S .
To simplify the frame and layer indexing to a single indexing variable, the following conversion can be made: q = L ( f 1 ) + l , where L is the total amount of layers, l is the layer index per frame spanning from 1 to L, and f is the frame index varying from 1 to F T . Therefore, q spans from 1 to L F T , and (3) becomes S q = d q / M S S . The fwnd varies with each R T T cycle depending on whether packets arrive successfully at the client side. The fwnd is represented by a vector where each entry corresponds to an R T T cycle w, where W is the total number of cycles, i.e., f w n d = { f w n d 1 , . . . , f w n d w , . . . , f w n d W } ; w N .
Given the packet q, the function C q outputs the R T T cycle w where the last packet of layer q is found. For better accuracy, a packet-acceptance pattern should be included, which is a set of ones and zeros representing true and false statements, respectfully, responding to whether the packet successfully arrived at the destination and before the deadline. This is represented by the vector X { 0 , 1 } . There is a correlation between the number of ones in X and the overall quality of the video.
To model the behavior of the packet-acceptance pattern, it is necessary to perform an iterative computation to determine if that layer is to be discarded or retained.
The first step involves determining the deadline and curfew of the layer. To determine this, the R T T cycle is needed, and it is given by C q :
C q = arg min w n = 1 w f w n d n q ^ = 1 q S q ^ X q ^ > 0
S q is the number of packets required to send layer q. X q is the q t h value of the packet-acceptance pattern vector, but for this stage, we assume X q = 1 . All values earlier than X q 1 (inclusive) are known.
Second step. With the R T T cycle C q information, the estimated arrival time of the packet q (or t A q ) is computed as:
t A q = R T T · ( C q | X q = 1 1 / 2 )
Third step. The deadline of frame f (or D f ) is a straight-forward computation obtained using the buffer time t B and the frame rate r, hence D f = f r + t B . It is beneficial to obtain a deadline expression in terms of the packet index q (or D q ). Since all layers of a frame have the same deadline, the relation f = q / L is sufficient to compute D q . Substituting gives:
D q = q L 1 r + t B
Fourth step. To retrieve the priority of packet q (or p q ) for P levels of priority, it is necessary to know the GOP pattern. As mentioned earlier and shown in  Figure 1, the GOP pattern is IBBBBBBB. To obtain the priority level of this pattern as a function of layer q, the following expression is used:
p q = 2 ( q 1 ) if q L 2 Π ( L ) if q > L & Π ( L G ) < L 2 Π ( L ) + 1 if q > L & Π ( L G ) L ,
where Π ( λ ) = m o d ( q L 1 , λ ) and G is the GOP size, in this case, G = 8 . p q spans from 0 to P 1 , where the lower values have higher priorities. Note: q is in the coding order (see  Figure 1).
Fifth step. Once the priority is found, the curfew Γ q is determined on Equation (14), where the number of priority levels is obtained by P = L T where T is the number of frame types. Here, we work with I and B frames, so T = 2 .
Γ q = m o d ( p q , P ) · τ
τ is a positive value in units of R T T that determines the time spacing between priority levels, and it is a design parameter. The greater the value, the more conservative the algorithm, and the lower is riskier. Increasing τ increases the chances of having a smooth video stream, lowering τ attempts to transmit a higher quality (more layers) stream but with a higher probability of loss due to limited bandwidth.
Final iteration step. The Packet-Acceptance Pattern (PAP) set can be obtained as the transmission conditions are met using:
X q = 1 if D q t A q Γ q 0 0 if o t h e r w i s e .
The value of X q , which at the beginning of the iteration was assumed to be 1, is now determined. This value is used for the subsequent iteration.

Optimization Method for τ Configuration

In the study presented in [53], a proof-of-concept is demonstrated in which video quality is maintained without re-buffering. However, an in-depth formal analysis for optimizing the parameter τ , representing curfew gaps, is not provided. An exploration into varying τ values within a reasonable range reveals that the maximum Y-PSNR is indicative of optimal video quality.
For optimization scenarios aiming at a single objective such as a root, maximum, or minimum, the bisection method is commonly employed [54,55]. However, its limitation in potentially identifying local maxima instead of the global maximum is acknowledged. Therefore, a multi-section variant of the bisection method is utilized in this study.
A Multi-dimensional Multi-section (MM) method is introduced, extending the bisection method for multi-dimensional optimization while minimizing the likelihood of identifying local maxima. Unlike traditional bisection methods, which divide the search space into two [54,55], a more expansive search technique is adopted here to mitigate the risk of mistaking local maxima for global ones.
Various applications show a monotonic increase up to a maximum, followed by a monotonic decrease, making traditional methods effective. However, for systems exhibiting oscillatory magnitudes, a more cautious approach is warranted. In this study, multiple points per dimension are probed in each iteration, with the step size and search region being adjusted based on the identified maximum. Achieving satisfactory result precision within a reasonable computational time is a non-trivial task, particularly due to the discrete nature of the data. The algorithmic details of the multi-probe bisection method are presented in Algorithm 2.
Algorithm 2 Multi-dimensional Multi-section Method (3D example)
1:
[ m 1 , m 2 , m 3 ] [ 0 , 0 , 0 ]
2:
[ M 1 , M 2 , M 3 ] upper limit of τ 1 , τ 2 , τ 3
3:
r a n g e x ( M x + m x ) / 2 for   x = 1 , 2 , 3
4:
N number   of   steps
5:
s t e p x r a n g e x / ( N 1 ) for   x = 1 , 2 , 3
6:
Q empty   matrix   ( 1 × 0 )
7:
l i s t τ 1 , τ 2 , τ 3 empty   matrix   ( 3 × 0 )
8:
while error tolerance ( init i 0 before loop )
9:
for   τ 1   from   m 1   to   M 1   step   s t e p 1
10:
for τ 2   from   m 2   to   M 2   step s t e p 2
11:
for   τ 3   from   m 3   to   M 3   step s t e p 3
12:
if   [ τ 1 , τ 2 , τ 3 ]   is   found   in   l i s t τ 1 , τ 2 , τ 3   then   skip
13:
else   ( increase i i + 1 )
14:
add   [ τ 1 , τ 2 , τ 3 ]   to   l i s t τ 1 , τ 2 , τ 3   in   row   i
15:
run   test of video quality
16:
add quality result to   Q   in   row i
17:
end   if
18:
end   for
19:
end   for
20:
end   for
21:
r a n g e x r a n g e x / 2 for   x = 1 , 2 , 3
22:
s t e p x s t e p x / 2 for   x = 1 , 2 , 3
23:
Q M A X max ( Q )
24:
i M A X indices of Q where Q   =   Q M A X
25:
l i s t τ 1 , τ 2 , τ 3 M A X   extract   rows   i M A X   from   l i s t τ 1 , τ 2 , τ 3
26:
m x min ( τ x ) r a n g e x for   x = 1 , 2 , 3
27:
where τ x   belongs   to   l i s t τ 1 , τ 2 , τ 3 M A X
28:
M x max ( τ x ) + r a n g e x for   x = 1 , 2 , 3
29:
where   τ x   belongs   to   l i s t τ 1 , τ 2 , τ 3 M A X
30:
end   while

5. Experimental Setups and Results

Experiments were conducted in a structured manner to evaluate the performance of the ASViS algorithm in comparison to conventional methods. Each investigation was targeted towards elucidating specific attributes and behaviors. The performance of ASViS is influenced by several specific parameters, which will be elaborated upon in the subsequent sections. Notably, network conditions, such as R T T , packet loss, and buffer conditions, have been identified. Additionally, variations in τ parameters across different layers and alterations in the size of each layer have been observed to impact performance. A thorough examination of these factors and their implications on ASViS functionality is included in the following discussions. The first experiment juxtaposed a theoretical model of ASViS, particularly in assessing the anticipated arrival time of frames at the client side, against an empirical counterpart. The purpose was to determine the accuracy of the theoretical model in simulating network dynamics. The second experiment delved into the perceived video quality by modulating the parameter τ . By setting a continuum of equidistant benchmarks, the Y-PSNR was gauged, revealing a complex interrelation between the τ values and the Y-PSNR attained. In the third analysis, the objective was to discern the optimal curfew gaps, represented as the set of taus, suitable for static initial network scenarios. Here, the MM method paved the way to unearthing superior video quality. Unlike assuming uniform curfew gaps, this study hypothesized better outcomes from distinct-sized intervals, with the Y-PSNR being the primary metric of evaluation. The fourth exploration sought to validate the theory that layer-based deadline gaps echo the proportions of the layer sizes. The culminating experiment instituted a face-off between two ABR protocols, ASViS and MPC. Using a spectrum of metrics, from VB behavior and throughput to video quality, this comparative study decoded the operational nuances of ASViS juxtaposed with the MPC algorithm.
The video utilized for simulations comes from the Xiph video repository [56], specifically the bus 352 × 288 .yuv file, encoded at 30 fps. The Joint Scalable Video Model (JSVM) facilitated video coding and decoding on SVC. Video details and parameters are provided in Table 1. Insights into parameters of coded video files are provided by the JSVM trace file, which facilitates video reconstruction using specific layers or packets.The ASViS algorithm was implemented using MATLAB and JSVM. Different network conditions were simulated based on the layer information extracted from the JSVM trace file, determining whether packets were transmitted or discarded. Furthermore, MATLAB was employed in conjunction with JSVM for tasks associated with video encoding/decoding. Additionally, the results were organized and processed using MATLAB. JSVM works using the operating system command prompt. It is easy to link JSVM to MATLAB using the matlab system command.
The video quality of different ABR algorithms was gauged using PSNR, and for VMAF testing, version 0.6.1 was employed from github (https://github.com/Netflix/vmaf), incorporating default pre-trained machine learning coefficients.

5.1. Results

The objective of the first experiment is to compare the network behavior outcomes of a theoretical ASViS model with those of an experimental test. Both scenarios employ UDP transmission rules equipped with flow control to prevent congestion. In the theoretical model, the determination of which packets are discarded is based on the Estimated Packet Size (EPS) and network conditions as utilized by the ASViS algorithm. The EPS values are informed by the mean layer size information derived from a trace file, which is created for JSVM. The size in M S S for each layer is presented in Table 2. For the experimental model, the actual size of each layer, influenced by its content, whether it is BL or EL, and its classification as I or B, is utilized. This results in a variety of sizes for each layer. The size in [kB] for each layer is presented in Table 2. Both models operate with the same network transmission parameters: R T T 0.05 s, packet loss rate 0.01, and buffer 3 s. Results are calculated based on 100 pseudo-random seeds.
Within the ASViS model, the behavior can shift from conservative to risky by adjusting the τ values. For a video comprising four layers, it might be presumed that there would be four distinct τ values representing BI, BB, EI, and EB. However, in this experiment, the τ value for the B I layer is set at 0 due to its importance. This ensures its transmission until the deadline is reached. Seven configurations, each with an equidistant and incremental τ gap ( τ G) based on layer priorities, are outlined in Table 3.
The results, both theoretical and experimental, appear in Figure 6. This representation showcases the estimated arrival time of each frame at the client end. The gradient of the slope conveys the transmission rate; a gentler slope indicates a higher rate, whereas a steeper one suggests a slower rate. A significant gap between theoretical and experimental results is observed for τ G 7 . This discrepancy can be attributed to various factors, including differences in average vs. actual packet size and video complexity variations.
To determine the accuracy of the correlation between models, the Mean Absolute Percentage Error (MAPE) is employed. Table 4 provides a detailed breakdown, pointing out a maximum prediction error of approximately 14.2% for τ G 6 . Given the inherent variability of the practical experiments due to packet losses and the intricate characteristics of each frame, along with the nuances of the various layers and complexities associated with network conditions in empirical analyses, the accuracy of the theoretical model is underscored. It is noteworthy that high fidelity is achieved by the theoretical model, aligning well with the experimental results. This contributes to a reduction in the need for exhaustive and computationally demanding empirical tests, thereby facilitating more streamlined behavioral predictions based on the theoretical model. It is important to stress that this does not replace empirical benchmarks but can significantly reduce the amount of energy, resources, and time devoted to more practical tests.
The aim of the second experiment is to analyze how the video quality performance of ASViS is influenced by varying τ values, placing emphasis on the experimental outcomes. The network transmission parameters for this investigation are as follows: R T T 0.05 s, packet loss rate 0.01, and buffer 1 s. Results are calculated based on 100 pseudo-random seeds for each τ G . The configuration of τ G adheres to the format provided in Table 3. This study also assesses the behavior of SVC in the absence of ASViS, termed the “no protocol” condition.
Figure 7 delineates the arrival time at the client side against each frame for every specified τ G . The scenario devoid of a protocol displays results closely aligned with the deadline. Nevertheless, the Y-PSNR performance for each τ G is illustrated in Figure 8. This representation indicates superior performance in the “no protocol” scenario. It is also observed that there exists a significant correlation between increased τ G and decreased Y-PSNR.
At an initial assessment, it may be surmised that SVC without ASViS offers superior results. However, upon closer analysis, this assumption is challenged. One primary objective of ABR algorithms is the prevention of rebuffering. Data presented in Figure 9, which portray the fluctuation in VB size relative to video playout, reveal stalls occurring after the 50% playout in scenarios lacking a protocol. In situations with ASViS, there are no observable stalls for any τ G , and a rise in the VB size is witnessed as the τ G value increases.
In the third experiment, the objective is to pinpoint optimal τ values to maximize specific criteria. For this experiment, the optimization objective is the performance of ASViS, based on the average amount of sent layer outcomes, using the MM method. It is crucial to note that these optimal values might not always be integers, suggesting a broad range of potential τ values. The foundational setup for this experiment involved fixed parameters such as an R T T of 0.05 s, a packet loss rate of 1/200, and a buffer of 1 s. For BI, a fixed value of τ 1 at 0 s (representing the deadline) was applied. In contrast, for the layers BB, EI, and EB, the values τ 2 , τ 3 , and τ 4 were adjusted, respectively. Performance was evaluated based on the maximum number of layers transmitted, with a combined limit of 300 layers: 19 for BI, 19 for EI, 141 for BB, and 141 for EB. This evaluation determined the local maxima within each quadrant. For each point, ten pseudo-random seeds were probed, and the average of the maximums was obtained. A consistent search range, spanning from 0 to 8 s, was employed for each τ .
The visualization of these results is provided in a 3D plot, as seen in Figure 10. The axes, labeled as G 2 , G 3 , and G 4 , represent the Gaps (G) or intervals between consecutive τ values, excluding τ 1 . These represent the intervals from the deadline to τ 2 , from τ 2 to τ 3 , and finally from τ 3 to τ 4 . It is crucial to recognize that these intervals are sequential and do not intersect. The overall time span between the deadline and I G 4 is the collective sum of τ 2 , τ 3 , and τ 4 . Based on the fixed conditions of the R T T and packet loss, optimal performance metrics were identified as τ 1 0 R T T , τ 2 0 R T T , τ 3 1 R T T , and τ 4 2 R T T for the layers BI, BB, EI, and EB, respectively.
For experiment 4, a hypothesis was formulated stating that the gaps in layered-based deadlines (represented by τ values) are dependent on and proportional to the layer sizes. The video detailed in Table 1 features relatively small base layers and large enhancement layers due to quantization in each layer. To test this hypothesis, the configuration was reversed (i.e., small enhancement layers and large base layers) for testing purposes, and the network conditions from experiment 3 were used. The same three-dimensional maximum value search using the MM method was then performed. The details of each Layer Size Configuration (LSC) can be found in Table 5. This table represents the average accumulated size per layer in a GOP structure of IB × 7. L S C 1 corresponds to experiment 3, while L S C 2 is the same video from Table 1 but with a quantization of 20 for both BL and EL.
Performance, based on the percentage of sent layer outcomes and derived using the MM method for L S C 2 , is presented in Figure 11. A preliminary comparison of the two 3D coordinates results, Figure 10 and Figure 11, reveals significant differences in the optimal τ values between the two LSCs. Detailed coordinates, referencing τ 2 , τ 3 , and τ 4 , are provided in Table 6.
A thorough examination of the results between L S C 1 and L S C 2 , and the corresponding layer sizes, has revealed a distinct association between the τ values and the sizes of layers. As depicted in Figure 12, the normalized layer sizes and the τ values for scenarios L S C 1 and L S C 2 are juxtaposed. It has been discerned from the blue markers that layer 2 holds a predominant role in both configurations. Conversely, the red markers have indicated that elevated τ values are associated with more diminutive layers. A noticeable inverse correlation between the layer sizes and τ values is evident: as the magnitude of a layer amplifies, the optimal τ for video quality is observed to wane. The pivotal influence of layer sizes on video quality, in the context of τ , is underscored by this relationship.
In the fifth experiment, a comparison was made between the performance of the ABR algorithms, MPC and ASViS, using various parameters. Traditional TCP behavior, which is typical for ABR algorithms, was exhibited under the network conditions for MPC. In contrast, User Datagram Protocol (UDP) with flow control was employed by ASViS, consistent with previous experiments. Network transmission parameters from experiment 3 were replicated, and the τ G values for the ASViS protocol were derived from the same experiment. A total of 100 simulations were executed for both ABR algorithms, and the results presented were the mean values obtained from all simulations.
A chunk size of 4 GOP was designated for MPC, equivalent to 32 frames or roughly 1 s of playout, given a frame rate of 30 fps. Four bitrate levels, labeled as R, were used, each integrating various layers. Comprehensive details of each R can be accessed in Table 7. An average across all chunks served as a reference for AVQ of Y-PSNR in this MPC implementation, accompanied by a chunk size ( d ( R ) ) specified in bits. The initial T s spanned from 0.1 s to 5 s, incremented by 0.2 s. The estimated throughput (C) was calculated as the harmonic mean throughput of the previous five chunks, with weight assignments of λ = 1 , β = 3000 , and β s = 3000 , aligning with the recommendations in [49].
As illustrated in Figure 13, the relationship between the VB size variation and video playout was mapped. Results indicated that a mean VB size of 1.5 s with a standard deviation of 0.5 s was achieved by MPC, whereas ASViS achieved a mean of 0.45 s with a deviation of 0.08 s. Notably, greater stability after 20% of video playout was observed with ASViS, despite its reduced mean VB size.
Figure 14 portrays the anticipated bitrate distribution for both MPC and ASViS. This visualization aids in comprehending the adaptability of each algorithm to network conditions. An in-depth analysis confirmed that the bitrate estimated distribution for ASViS predominantly focuses on values higher than those for MPC, a pattern that contrasts with its mean and standard deviation provided in Table 8.
Subsequent evaluations explored video quality throughout the entire video playback. The Y-PSNR findings are presented in Figure 15, whereas VMAF results can be found in Figure 16. Both algorithms exhibited comparable video quality from frames 40 to 130. However, at the commencement and conclusion of the video playback, the quality of MPC was observed to be inferior to that of ASViS. This discrepancy is attributed to the conservative approach of MPC and its slower adaptation to network conditions compared to ASViS.
Furthermore, a noteworthy distinction between the results of frames 74 and 84 was observed in Figure 16, owing to the capability of VMAF to discern details overlooked by Y-PSNR. Detailed analysis revealed that frames 74 and 84, corresponding to a B4 frame as cross-referenced with Figure 1b, were dropped by ASViS.
The video quality metrics, presented in Figure 15 and Figure 16, were analyzed using a distinctive approach involving a boxplot and its corresponding histogram. The boxplot delineates the median, the 25th percentile at the bottom, and the 75th percentile at the top. Whiskers extend to cover values not deemed outliers. Notably, since no outliers were observed, there are no ‘+’ markers. The results for Y-PSNR and VMAF are displayed in Figure 17 and Figure 18, respectively. Although Y-PSNR and VMAF are founded on different philosophies, both metrics consistently show the superiority of ASViS over MPC. For a comprehensive understanding of the mean and deviation values of Y-PSNR, VMAF, and bitrate, one can refer to Table 8. It becomes evident that, in terms of VMAF, Y-PSNR, and bitrate, the ASViS algorithm holds an edge over MPC.

5.2. Discussion

The initial experiment demonstrates a close alignment between theoretical and observed behaviors of ASViS, with mean absolute percentage errors ranging between 6.5% and 14.2%. These findings substantiate the capability of the model to predict the experimental behavior of ASViS, thus eliminating the need for time-consuming emulations or tests.
Outcomes of the second experiment indicate superior video quality in a no-protocol scenario (default SVC) compared to any τ gap configuration scenario. However, this scenario exhibits video stalls, with rebuffering incidents accounting for over half the video playback duration. In contrast, all τ G x scenarios display variations in video buffering without complete depletion. Video quality can vary based on configuration, offering insights into how ASViS modulates video quality to ensure uninterrupted playback.
The third experiment leverages the MM method to discern specific τ coordinates that yield improved video quality under given network conditions. It reveals that optimal video quality can span a broad range of τ values and might be envisioned as a multi-dimensional solution. In the fourth experiment, a pronounced inverse correlation is found between the τ values and layer sizes. This discovery is pivotal, suggesting that service providers can sidestep the task of pinpointing optimal values for individual videos. They can instead modify layer deadlines in proportion to layer sizes, ensuring nearly optimal outcomes.
Findings of the fifth experiment spotlight the superiority of ASViS over MPC in multiple aspects. ASViS records a 5.8% and 4.6% higher Y-PSNR and VMAF, respectively, than MPC. Moreover, ASViS displays a consistently compact video buffer size compared to MPC. It also boasts more efficient network utilization, with a bitrate surpassing that of MPC, and steadier performance, reflected in a reduced standard deviation. In conclusion, these promising results position ASViS as a skillful ABR algorithm.
While the potential of ASViS has been demonstrated in specific contexts through our experiments, its performance in broader or more challenging scenarios remains an intriguing area of speculation. In 5G mobile networks, known for high download speeds and reduced latency, it is hypothesized that ASViS could deliver ultra-high-definition video streaming with minimal interruptions. Conversely, in regions with underdeveloped internet infrastructure or in satellite networks with inherent high latency, challenges may arise regarding how ASViS adapts video flow and manages fluctuations. It might be theorized that smooth but lower-quality video playback could prevail. Further research is warranted in these areas. Future directions are aimed at exploring and validating these scenarios to ensure the resilience of ASViS beyond ideal conditions. For future directions, it is proposed that more exhaustive tests be conducted, involving videos with varying resolutions and different frame rates. Additionally, it has been identified that simulations need to be carried out under realistic network conditions, such as those of mobile networks.

6. Conclusions

In the current landscape, robust ABR algorithms are vital to meet the vast demands of video streaming on global networks. These algorithms must ensure superior video quality, prevent stalls, and adapt swiftly to network fluctuations. The introduction of the novel concept of ASViS has been demonstrated through experiments to efficiently control video stream flow in bandwidth-limited scenarios. When SVC integrates with ASViS, it not only enhances the user QoE but also reduces network congestion. Such improvements arise from aligning the encoding rate with available throughput and discarding less critical layers.
Based on the tests conducted and the information gathered, it is suggested that the integration of DASH with ASViS enables the server to better adapt to real-time conditions, transmitting only the layers that the client can handle according to their bandwidth capacity. The ASViS proposal is found to be dynamically adaptable to network conditions and device capabilities, allowing for a more efficient use of bandwidth and an enhanced user experience, reducing start-up times and rebuffering. Furthermore, in terms of storage, it is necessary for the server to store only a single version of the scalable video. DASH is agnostic in terms of encoding, so the integration of SVC does not necessitate significant changes to the existing DASH infrastructure. Likewise, the new generation of video codec solutions, such as H.265, Video Predictor 9, and AOMedia Video 1, inherently possess the scalable encoding model.
In the conclusions drawn, ASViS is portrayed as a transformative solution to ABR algorithms, with inherent limitations of DASH being addressed. Through the experiments conducted, the proficiency of ASViS in managing video stream flow, enhancing user QoE, and mitigating network congestion was validated. A noteworthy alignment was observed between theoretical and experimental behaviors, as evidenced by the mean absolute percentage errors, further emphasizing the reliability of ASViS. With its dynamic adaptability, bandwidth utilization was optimized, video quality was elevated, and interruptions were minimized, setting it apart from algorithms such as MPC. In the tested scenarios, the efficacy of ASViS was demonstrated, marking a significant advancement in video streaming.

Author Contributions

E.P.-A. contributed on the design, implementation, data collection, and analysis of the proposed work. C.E. contributed on the design, implementation, and analysis of the proposed work. A.S. contributed on the design of the proposed work. M.O. contributed on the analysis of the proposed work. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially funded by the Comision Nacional de Investigacion Cientifica y Tecnologica (CONICYT) Doctorate Scholarship 2018-21181143, Fondo Nacional de Desarrollo Cientifico y Tecnologico (FONDECYT) Chile Grant 1210031, and the Advanced Center for Electrical and Electronic Engineering (AC3E), Agencia Nacional de Investigacion y Desarrollo (ANID) Basal Project FB0008.

Data Availability Statement

Videos are found at [56], this is a public site not managed by the authors. Please contact the correspondence author for any other data of interest.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. El Meligy, A.O.; Hassan, M.S.; Landolsi, T. A Buffer-Based Rate Adaptation Approach for Video Streaming over HTTP. In Proceedings of the 2020 Wireless Telecommunications Symposium (WTS), Virtual Event, 22–24 April 2020; pp. 1–5. [Google Scholar] [CrossRef]
  2. Sandvine. Sandvine’s 2023 Global Internet Phenomena Report. 2023. Available online: https://www.sandvine.com/hubfs/Sandvine_Redesign_2019/Downloads/2023/reports/Sandvine%20GIPR%202023.pdf (accessed on 26 October 2023).
  3. Ericsson. Ericsson Mobility Report. 2023. Available online: https://www.ericsson.com/49dd9d/assets/local/reports-papers/mobility-report/documents/2023/ericsson-mobility-report-june-2023.pdf (accessed on 26 October 2023).
  4. Comcast. COVID-19 Network Update; Comcast Corp: Philadelphia, PA, USA, 2020; Available online: https://corporate.comcast.com/covid-19/network/may-20-2020 (accessed on 26 October 2023).
  5. Mercat, A.; Mäkinen, A.; Sainio, J.; Lemmetti, A.; Viitanen, M.; Vanne, J. Comparative Rate-Distortion-Complexity Analysis of VVC and HEVC Video Codecs. IEEE Access 2021, 9, 67813–67828. [Google Scholar] [CrossRef]
  6. Pearce, J. Netflix and YouTube to Reduce Stream Quality in Europe Due to Coronavirus. IBC, 2020. Available online: https://www.ibc.org/publish/netflix-and-youtube-to-reduce-stream-quality-in-europe-due-to-coronavirus/5615.article (accessed on 26 October 2023).
  7. Zhao, J.; Liu, J.; Zhang, C.; Cui, Y.; Jiang, Y.; Gong, W. MPTCP+: Enhancing Adaptive HTTP Video Streaming over Multipath. In Proceedings of the 2020 IEEE/ACM 28th International Symposium on Quality of Service (IWQoS), Hangzhou, China, 15–17 June 2020; pp. 1–6. [Google Scholar] [CrossRef]
  8. Duan, J.; Zhang, M.; Wang, J.; Han, S.; Chen, X.; Yang, X. VCC-DASH: A Video Content Complexity-Aware DASH Bitrate Adaptation Strategy. Electronics 2020, 9, 230. [Google Scholar] [CrossRef]
  9. Bukhari, S.M.A.H.; Afandi, W.; Khan, M.U.S.; Maqsood, T.; Qureshi, M.B.; Fayyaz, M.A.B.; Nawaz, R. E-Ensemble: A Novel Ensemble Classifier for Encrypted Video Identification. Electronics 2022, 11, 4076. [Google Scholar] [CrossRef]
  10. Matsumoto, T.; Yamamoto, M. Fairness improvement by combination of ABR and TCP algorithms in ABR video streaming. IEICE Commun. Express 2021, 10, 225–230. [Google Scholar] [CrossRef]
  11. Hodroj, A.; Ibrahim, M.; Hadjadj-Aoul, Y. A Survey on Video Streaming in Multipath and Multihomed Overlay Networks. IEEE Access 2021, 9, 66816–66828. [Google Scholar] [CrossRef]
  12. Martinez-Caro, J.M.; Cano, M.D. On the Identification and Prediction of Stalling Events to Improve QoE in Video Streaming. Electronics 2021, 10, 753. [Google Scholar] [CrossRef]
  13. Zhong, H.; Wu, F.; Xu, Y.; Cui, J. QoS-Aware Multicast for Scalable Video Streaming in Software-Defined Networks. IEEE Trans. Multimed. 2021, 23, 982–994. [Google Scholar] [CrossRef]
  14. ITU. H.264: Advanced Video Coding for Generic Audiovisual Services, August, 2021. Available online: https://www.itu.int/rec/T-REC-H.264-202108-I (accessed on 26 October 2023).
  15. Xue, Y.; Zhou, J.; Zeng, H.; Zhong, P.; Wen, J. An adaptive steganographic scheme for H.264/AVC video with distortion optimization. Signal Process. Image Commun. 2019, 76, 22–30. [Google Scholar] [CrossRef]
  16. Liu, Y.; Jiang, B.; Guo, T.; Sitaraman, R.K.; Towsley, D.; Wang, X. Grad: Learning for Overhead-Aware Adaptive Video Streaming with Scalable Video Coding. In Proceedings of the 28th ACM International Conference on Multimedia, New York, NY, USA, 12–16 October 2020; pp. 349–357. [Google Scholar] [CrossRef]
  17. Skupin, R.; Bartnik, C.; Wieckowski, A.; Sanchez, Y.; Bross, B.; Hellge, C.; Schierl, T. Open GOP Resolution Switching in HTTP Adaptive Streaming with VVC. In Proceedings of the 2021 Picture Coding Symposium (PCS), Bristol, UK, 29 June–2 July 2021; pp. 1–5. [Google Scholar] [CrossRef]
  18. Pourreza, R.; Cohen, T. Extending Neural P-frame Codecs for B-frame Coding. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Los Alamitos, CA, USA, 10–17 October 2021; pp. 6660–6669. [Google Scholar] [CrossRef]
  19. Wu, Y.; Bai, X.; Hu, Y.; Chen, M. A Novel Video Transmission Latency Measurement Method for Intelligent Cloud Computing. Appl. Sci. 2022, 12, 12884. [Google Scholar] [CrossRef]
  20. Xu, C.; Ren, W.; Yu, L.; Zhu, T.; Choo, K.K.R. A Hierarchical Encryption and Key Management Scheme for Layered Access Control on H.264/SVC Bitstream in the Internet of Things. IEEE Internet Things J. 2020, 7, 8932–8942. [Google Scholar] [CrossRef]
  21. Omiyi, P.E.; Nasralla, M.M.; Rehman, I.U.; Khan, N.; Martini, M.G. An Intelligent Fuzzy Logic-Based Content and Channel Aware Downlink Scheduler for Scalable Video over OFDMA Wireless Systems. Electronics 2020, 9, 1071. [Google Scholar] [CrossRef]
  22. Zhang, X.; Lin, H.; Chen, M.; Kang, B.; Wang, L. MEC-Enabled Video Streaming in Device-to-Device Networks. IET Commun. 2020, 14, 2453–2461. [Google Scholar] [CrossRef]
  23. Zhang, Y.; Zhang, J.; Huo, Y.; Xu, C.; El-Hajjar, M.; Hanzo, L. Scalable Panoramic Wireless Video Streaming Relying on Optimal-Rate FEC-Coded Adaptive QAM. IEEE Trans. Veh. Technol. 2020, 69, 11206–11219. [Google Scholar] [CrossRef]
  24. Topiwala, P.; Dai, W.; Pian, J.; Biondi, K.; Krovvidi, A. Video Quality Analysis: Steps towards Unifying Full and No Reference Cases. Standards 2022, 2, 402–416. [Google Scholar] [CrossRef]
  25. Antsiferova, A.; Yakovenko, A.; Safonov, N.; Kulikov, D.; Gushchin, A.; Vatolin, D. Applying Objective Quality Metrics to Video-Codec Comparisons: Choosing the Best Metric for Subjective Quality Estimation. In Proceedings of the 31st International Conference on Computer Graphics and Vision, Nizhny Novgorod, Russia, 27–30 September 2021; pp. 199–210. [Google Scholar] [CrossRef]
  26. Lee, E.; Hsu, L.; Chen, E.; Lee, C. Cross-Resolution Flow Propagation for Foveated Video Super-Resolution. In Proceedings of the 2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Los Alamitos, CA, USA, 2–7 January 2023; pp. 1766–1775. [Google Scholar] [CrossRef]
  27. Wang, X.; Chen, C.; Zhu, L.; Shi, K.; Peng, B.; Zhu, Y.; Mao, H.; Long, H.; Ke, S.; Fu, C.; et al. Vertically integrated spiking cone photoreceptor arrays for color perception. Nat. Commun. 2023, 14, 3444. [Google Scholar] [CrossRef]
  28. Dziembowski, A.; Mieloch, D.; Stankowski, J.; Grzelka, A. IV-PSNR—The Objective Quality Metric for Immersive Video Applications. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 7575–7591. [Google Scholar] [CrossRef]
  29. Keleş, O.; Yılmaz, M.A.; Tekalp, A.M.; Korkmaz, C.; Dogan, Z. On the Computation of PSNR for a Set of Images or Video. arXiv 2021, arXiv:2104.14868. [Google Scholar] [CrossRef]
  30. Saha, A.; Pentapati, S.K.; Shang, Z.; Pahwa, R.; Chen, B.; Gedik, H.E.; Mishra, S.; Bovik, A.C. Perceptual video quality assessment: The journey continues! Front. Signal Process. 2023, 3, 1193523. [Google Scholar] [CrossRef]
  31. Bampis, C.G.; Li, Z.; Bovik, A.C. Spatiotemporal Feature Integration and Model Fusion for Full Reference Video Quality Assessment. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 2256–2270. [Google Scholar] [CrossRef]
  32. Luo, Z.; Zhu, C.; Huang, Y.; Xie, R.; Song, L.; Kuo, C.C.J. VMAF Oriented Perceptual Coding Based on Piecewise Metric Coupling. IEEE Trans. Image Process. 2021, 30, 5109–5121. [Google Scholar] [CrossRef]
  33. Chen, L.H.; Bampis, C.G.; Li, Z.; Sole, J.; Bovik, A.C. Perceptual Video Quality Prediction Emphasizing Chroma Distortions. IEEE Trans. Image Process. 2021, 30, 1408–1422. [Google Scholar] [CrossRef] [PubMed]
  34. Yousef, H.; Feuvre, J.L.; Storelli, A. ABR prediction using supervised learning algorithms. In Proceedings of the 2020 IEEE 22nd International Workshop on Multimedia Signal Processing (MMSP), Tampere, Finland, 21–24 September 2020; pp. 1–6. [Google Scholar] [CrossRef]
  35. de Fez, I.; Belda, R.; Guerri, J.C. New objective QoE models for evaluating ABR algorithms in DASH. Comput. Commun. 2020, 158, 126–140. [Google Scholar] [CrossRef]
  36. Cong Thang, T.; Ho, Q.D.; Kang, J.; Pham, A. Adaptive streaming of audiovisual content using MPEG DASH. IEEE Trans. Consum. Electron. 2012, 58, 78–85. [Google Scholar] [CrossRef]
  37. Hodroj, A.; Ibrahim, M.; Hadjadj-Aoul, Y. Parallel Streaming for a Multi-homed Dash client. In Proceedings of the 2020 International Wireless Communications and Mobile Computing (IWCMC), Limassol, Cyprus, 15–19 June 2020; pp. 63–68. [Google Scholar] [CrossRef]
  38. Chen, J.; Wu, J. Dynamic Adaptive Streaming based on Deep Reinforcement Learning. J. Phys. Conf. Ser. 2019, 1237, 022124. [Google Scholar] [CrossRef]
  39. Qin, Y.; Jin, R.; Hao, S.; Pattipati, K.R.; Qian, F.; Sen, S.; Yue, C.; Wang, B. A Control Theoretic Approach to ABR Video Streaming: A Fresh Look at PID-Based Rate Adaptation. IEEE Trans. Mob. Comput. 2020, 19, 2505–2519. [Google Scholar] [CrossRef]
  40. Wang, B.; Xu, M.; Ren, F.; Wu, J. Improving Robustness of DASH Against Unpredictable Network Variations. IEEE Trans. Multimed. 2021, 24, 323–337. [Google Scholar] [CrossRef]
  41. Khan, M.J.; Harous, S.; Bentaleb, A. Client-driven Adaptive Bitrate Techniques for Media Streaming over HTTP: Initial Findings. In Proceedings of the 2020 IEEE International Conference on Electro Information Technology (EIT), Tampere, Finland, 21–24 September 2020; pp. 53–59. [Google Scholar] [CrossRef]
  42. Wang, B.; Ren, F. Improving Robustness of DASH Against Network Uncertainty. In Proceedings of the 2019 IEEE International Conference on Multimedia and Expo (ICME), Shanghai, China, 8–12 July 2019; pp. 748–753. [Google Scholar] [CrossRef]
  43. Kumar, S.; Devaraj, R.; Sarkar, A.; Sur, A. Client-side QoE management for SVC video streaming: An FSM supported design approach. IEEE Trans. Netw. Serv. Manag. 2019, 16, 1113–1126. [Google Scholar] [CrossRef]
  44. Elgabli, A.; Liu, K.; Aggarwal, V. Optimized preference-aware multi-path video streaming with scalable video coding. IEEE Trans. Mob. Comput. 2018, 19, 159–172. [Google Scholar] [CrossRef]
  45. Elgabli, A.; Aggarwal, V.; Hao, S.; Qian, F.; Sen, S. LBP: Robust rate adaptation algorithm for SVC video streaming. IEEE/ACM Trans. Netw. 2018, 26, 1633–1645. [Google Scholar] [CrossRef]
  46. Lu, H.; Zhang, M.; Gui, Y.; Liu, J. QoE-Driven Multi-User Video Transmission Over SM-NOMA Integrated Systems. IEEE J. Sel. Areas Commun. 2019, 37, 2102–2116. [Google Scholar] [CrossRef]
  47. Basavarajaiah, M.; Sharma, P. Cross Layer Optimization for Wireless Video Transmission Using Machine Learning. In Proceedings of the 2018 7th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO), Noida, India, 29–31 August 2018; IEEE: Toulouse, France, 2018; pp. 308–312. [Google Scholar]
  48. Matsumoto, T.; Goto, K.; Yamamoto, M. On Fairness Issue of ABR and TCP Algorithms in Video Streaming. In Proceedings of the 2020 IEEE 17th Annual Consumer Communications & Networking Conference (CCNC), Las Vegas, NV, USA, 10–13 January 2020; IEEE: Toulouse, France, 2020; pp. 1–2. [Google Scholar]
  49. Huang, T.; Sun, L. Deepmpc: A Mixture Abr Approach Via Deep Learning And Mpc. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 1231–1235. [Google Scholar] [CrossRef]
  50. Deng, Z.; Liu, Y.; Liu, J.; Argyriou, A. Cross-Layer DASH-Based Multipath Video Streaming over LTE and 802.11ac Networks. Multimed. Tools Appl. 2021, 80, 16007–16026. [Google Scholar] [CrossRef]
  51. Yaqoob, A.; Bi, T.; Muntean, G.M. A Survey on Adaptive 360° Video Streaming: Solutions, Challenges and Opportunities. IEEE Commun. Surv. Tutor. 2020, 22, 2801–2838. [Google Scholar] [CrossRef]
  52. Polese, M.; Chiariotti, F.; Bonetto, E.; Rigotto, F.; Zanella, A.; Zorzi, M. A Survey on Recent Advances in Transport Layer Protocols. IEEE Commun. Surv. Tutor. 2019, 21, 3584–3608. [Google Scholar] [CrossRef]
  53. Sanhueza, A.; Méric, H.; Estevez, C. Efficient video streaming rate control based on a deadline-sensitive selection of SVC layers. In Proceedings of the 2017 Ninth International Conference on Ubiquitous and Future Networks (ICUFN), Milan, Italy, 4–7 July 2017; pp. 502–506. [Google Scholar]
  54. Mineo, C.; Lines, D.; Cerniglia, D. Generalised bisection method for optimum ultrasonic ray tracing and focusing in multi-layered structures. Ultrasonics 2021, 111, 106330. [Google Scholar] [CrossRef] [PubMed]
  55. Etesami, R.; Madadi, M.; Mashinchi, M.; Ganjoei, R. A new method for rooting nonlinear equations based on the Bisection method. MethodsX 2021, 8, 101502. [Google Scholar] [CrossRef] [PubMed]
  56. Xiph.org Derf’s Test Media Collection. November 2023. Available online: https://media.xiph.org/video/derf/ (accessed on 26 October 2023).
Figure 1. Hierarchy for (a) I P B dependencies and (b) I B dependencies.
Figure 1. Hierarchy for (a) I P B dependencies and (b) I B dependencies.
Electronics 12 04542 g001
Figure 2. Hierarchy structure to a GOP of 8 for SVC with two layers.
Figure 2. Hierarchy structure to a GOP of 8 for SVC with two layers.
Electronics 12 04542 g002
Figure 3. Effects of the loss of layers in the video quality for (a) a normal video, (b) no EL-B, and (c) Only BL-I.
Figure 3. Effects of the loss of layers in the video quality for (a) a normal video, (b) no EL-B, and (c) Only BL-I.
Electronics 12 04542 g003
Figure 4. Outline of VMAF algorithm.
Figure 4. Outline of VMAF algorithm.
Electronics 12 04542 g004
Figure 6. Results of the arrival time of frames for theoretical and experimental scenarios.
Figure 6. Results of the arrival time of frames for theoretical and experimental scenarios.
Electronics 12 04542 g006
Figure 7. Results of the arrival time of frames for τ G and no protocol scenarios.
Figure 7. Results of the arrival time of frames for τ G and no protocol scenarios.
Electronics 12 04542 g007
Figure 8. Y-PSNR performance for τ G and no protocol scenarios.
Figure 8. Y-PSNR performance for τ G and no protocol scenarios.
Electronics 12 04542 g008
Figure 9. Video buffer variations for all scenarios on playout percent function.
Figure 9. Video buffer variations for all scenarios on playout percent function.
Electronics 12 04542 g009
Figure 10. Video quality results of MM method for BB, EI, and EB layers for experiment 3.
Figure 10. Video quality results of MM method for BB, EI, and EB layers for experiment 3.
Electronics 12 04542 g010
Figure 11. Video quality results of MM method for BB, EI, and EB layers for L S C 2 of experiment 4.
Figure 11. Video quality results of MM method for BB, EI, and EB layers for L S C 2 of experiment 4.
Electronics 12 04542 g011
Figure 12. Comparison between normalized layer sizes and optimal τ ranges.
Figure 12. Comparison between normalized layer sizes and optimal τ ranges.
Electronics 12 04542 g012
Figure 13. VB size comparison through video playout for ASViS and MPC.
Figure 13. VB size comparison through video playout for ASViS and MPC.
Electronics 12 04542 g013
Figure 14. Estimated bitrate distribution for ASViS and MPC.
Figure 14. Estimated bitrate distribution for ASViS and MPC.
Electronics 12 04542 g014
Figure 15. Video quality behavior of Y-PSNR for ASViS and MPC.
Figure 15. Video quality behavior of Y-PSNR for ASViS and MPC.
Electronics 12 04542 g015
Figure 16. Video quality behavior of VMAF for ASViS and MPC.
Figure 16. Video quality behavior of VMAF for ASViS and MPC.
Electronics 12 04542 g016
Figure 17. Boxplot and histogram for Y-PSNR comparison of ASViS and MPC.
Figure 17. Boxplot and histogram for Y-PSNR comparison of ASViS and MPC.
Electronics 12 04542 g017
Figure 18. Boxplot and histogram for a VMAF comparison of ASViS and MPC.
Figure 18. Boxplot and histogram for a VMAF comparison of ASViS and MPC.
Electronics 12 04542 g018
Table 1. Video parameters for all experiments.
Table 1. Video parameters for all experiments.
ParameterDetails
Filebus_cif.y4m
Rate30 fps
Length150 frames
Number of layers2
Quantization Layer 1—BL40
Quantization Layer 2—EL20
GOP8
Resolutionwidth 352, height 288
Amount of BI/EI19
Amount of BBI/EB121
Y-PSNR (Orig. Encoded)40.62 [dB]
VMAF (Orig. Encoded)99.97 [%]
Table 2. EPS for each layer for ASViS theoretical experiment.
Table 2. EPS for each layer for ASViS theoretical experiment.
LayerEPS [kB]
B I 7154
B B 584
E I 36,500
E B 11,388
Table 3. 7 τ gap configurations to ASViS.
Table 3. 7 τ gap configurations to ASViS.
τ G τ 2 —BB [RTT] τ 3 —EI [RTT] τ 4 —EB [RTT]
τ G 1 123
τ G 2 246
τ G 3 369
τ G 4 4812
τ G 5 51015
τ G 6 61218
τ G 7 71421
Table 4. MAPE comparison from theoretical and experimental results of ASViS.
Table 4. MAPE comparison from theoretical and experimental results of ASViS.
τ G MAPE [%]
τ G 1 6.5
τ G 2 7.3
τ G 3 9.3
τ G 4 10.7
τ G 5 10.1
τ G 6 14.2
τ G 7 13.7
Table 5. Average layer size for both LSCs.
Table 5. Average layer size for both LSCs.
Layer LSC 1  [kB] LSC 2  [kB]
BI6.9719.31
BB36.1323.43
EI0.461.73
EB10.296.66
Table 6. Video quality results and τ G coordinates for different layer size configurations.
Table 6. Video quality results and τ G coordinates for different layer size configurations.
τ G [RTT]Sent Layers [%]
τ G τ 2 —BB τ 3 —EI τ 4 —EBBIEIBBEB
L S C 1 01210096.660.226.6
L S C 2 00.50.7598.794.5342.721.4
Table 7. Chunk properties of each bitrate level R.
Table 7. Chunk properties of each bitrate level R.
LayerMetrics
BIBBEIEBAVQ Y-PSNR [dB] d ( R )  [Mb]
R 1 XX--28.40.32
R 2 XXX-37.71.45
R 3 XX-X31.02.70
R 4 XXXX41.63.83
Table 8. Detailed results of experiment 5.
Table 8. Detailed results of experiment 5.
ASViSMPC
MeanStdMeanStd
Y-PSNR [dB]41.12.439.26.0
VMAF [%]98.54.792.89.0
Bitrate [Mb/s]3.11.22.61.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peña-Ancavil, E.; Estevez, C.; Sanhueza, A.; Orchard, M. Adaptive Scalable Video Streaming (ASViS): An Advanced ABR Transmission Protocol for Optimal Video Quality. Electronics 2023, 12, 4542. https://doi.org/10.3390/electronics12214542

AMA Style

Peña-Ancavil E, Estevez C, Sanhueza A, Orchard M. Adaptive Scalable Video Streaming (ASViS): An Advanced ABR Transmission Protocol for Optimal Video Quality. Electronics. 2023; 12(21):4542. https://doi.org/10.3390/electronics12214542

Chicago/Turabian Style

Peña-Ancavil, Eliecer, Claudio Estevez, Andrés Sanhueza, and Marcos Orchard. 2023. "Adaptive Scalable Video Streaming (ASViS): An Advanced ABR Transmission Protocol for Optimal Video Quality" Electronics 12, no. 21: 4542. https://doi.org/10.3390/electronics12214542

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop