|
![]() |
ISSN 1214-9675 Server vznikl za podpory Grantové agentury ČR. 21. ročník |
Témata
Doporučujeme
Kontakt
|
![]()
Vydáno dne 24. 01. 2013 (16910 přečtení) |
Input YUV files | Foreman, News, Coastguard |
---|---|
Resolution | CIF |
Frame Rate | 30 fps |
Number of frames | 300 |
Number of layers | 1 |
GOP size | 16 |
Search range | 32 |
Search mode | 4 |
MGSControl | 1 |
CgsnrRefinement | 1 |
Base layer mode | 0 |
Encode key pictures | 1 |
The implementation methodology is shown in Fig. 4 above. The three YUV video files were first encoded using the JSVM Software Manual [12], according to the configurations further summarized in Table 1. A set of different QP scenarios was designed to cover a wide range of quality levels. We encoded each video using 7 scenarios in which the QP values for the base layer are varied for 44, 38, 32, 26, 20, 15 and 10. The coding efficiency of H.264/SVC is dependent on the quantization parameters of each layer. Packet traces (Network Abstraction Layer Units) of the H.264 bit streams are generated using BitStreamExtractor.
Fig. 5 Simulation topology
Parameter | Value |
---|---|
MAC type | 802.11 |
Radio propagation | Propagation/TwoRayGround |
Interface queue | Queue/DropTail |
Routing | DSDV |
Antenna model | Antenna/Omni Antenna |
Data rate | 11 Mbit/s |
Basic rate | 1 Mbit/s |
Number of mobile modes | 3 |
Interface queue | 50 |
The NALUs are prepared for transmission over the IP network (hinting, packetization). The resulting H.264 video trace files are hinted using MP4Box [13] which emulates the streaming of the *.h264 video over the network based on RTP/UDP/IP protocol stack. Large NALUs are thus split through IP layer fragmentation. Real-time Transport Protocol (RTP) is used for transfer of real-time data like video streaming. Existing transport protocols like UDP (User Datagram Protocol) will run under RTP. RTP provides applications that occur in real-time with end-to-end delivery services, such as sequence numbers, types, sizes of the video frames and the number of UDP packets used to transmit each frame, and timestamps (for packet loss and reordering detection, and end-to-end delay).
We conduct the simulations of H.264/SVC video transmission over IEEE 802.11 [14] using NS-2 [15]. The wireless channel configuration is summarized in Table 2. The simulated scenario consists of three wireless nodes, one multimedia server and two clients, all within reasonable transmission range. The multimedia server transmits H.264/SVC video and CBR traffic to Client 1, while Client 2 receives FTP and CBR traffic from the server, all happening simultaneously. Packet sizes were set to 1500 Bytes. The network topology is depicted in Fig. 5. The background traffic generated at the server and accessed by the two clients, while streaming video traffic, increases the virtual collisions that occur at the server’s MAC layer. All the packets were assigned equal priority and scheduled from the same access point of the multimedia server. The experiment is designed to study the impacts of competing background traffic with different sending rates on the streamed video quality. In order to overload the wireless transmission, the CBR flows for the two clients are varied from 0.1, 0.5 to 1 Mbit/s each, while streaming the different video sequences of different contents and different encoding QP values.
10 different initial seeds for random number generation were chosen for simulation. Results generated were averaged over these 10 runs. After simulation, the received trace file is generated. The received and the original NALU trace files are further combined and processed to generate the received NALU trace. Maximum playout buffer delay at the video client is set to 5 seconds. After further processing, the received NALU trace is passed through BitStreamExtractor which generates H.264 video, which is in turn decoded with the JSVM H264Decoder, thus obtaining an uncompressed YUV file. The reconstructed YUV file and the original one are compared with objective video quality metric, to compute the overall video quality.
Objective video quality algorithms are based on mathematical models that can predict image multimedia quality by comparing a distorted signal against a reference, typically by modeling the human visual system. Some existing objective criteria are Mean Error Square (MSE), Peak Signal-to-Noise Ratio (PSNR), SSIM (structural similarity) and VIF (Visual Information Fidelity). In this experiment, PSNR [16] is adopted as our objective metric. PSNR has been selected because it is the most widely used metric.
PSNR can be computed for both luminance (Y-PSNR) and chrominance (U-PSNR and V-PSNR) components of the video. The human eye is considered more sensitive to luminance (brightness) than chrominance (colour), therfore the PSNR is usually evaluted only for the luminance (Y) component. The equation below shows the relationship between the PSNR of the luminance component Y of original image and degraded image D:
Where Vpeak = 2k-1; k denotes number of bits per pixel. Ncol represents the number of columns; Nrow the number of rows in an image. PSNR computes the error between a reconstructed image and the original one. A larger PSNR value denotes better image quality.
Fig. 6 – Fig. 14 depict the results obtained from this experiment. Fig. 6 depicts the quality comparison of the encoded only video sequences, for the three contents, encoded at different quantization parameter values. Results indicate that lower quantization parameters lead to better perceptual quality, depicted by higher PSNR values. Fig. 7 plots the PSNR curve vs. frame number for the Foreman sequence, encoded only at QP = 44 and 10, and Foreman encoded at QP = 10 and transmitted under 1 Mbit/s background traffic level.
Fig. 6: Impact of quantization parameter on video quality
Fig. 7: Quality comparison for encoded only and transmitted sequences
Fig. 8: Quality comparison for transmitted sequences, QP =44
Fig. 9: Quality comparison for transmitted sequences, QP =38
Fig. 10: Quality comparison for transmitted sequences, QP =32
Fig. 11: Quality comparison for transmitted sequences, QP =26
Fig. 12: Quality comparison for transmitted sequences, QP =20
Fig. 13: Quality comparison for transmitted sequences, QP =15
Fig. 14: Quality comparison for transmitted sequences, QP =10
Analysis of the generated bit streams (Table 3) shows that the lower the quantization parameter, the higher the generated file size and consequently higher bit rates (204 Kbit/s for Foreman at QP = 44, 6.53 Mbit/s at QP = 10; 122 Kbit/s for News at QP = 44, 2.63 Mbit/s at QP = 10; 296 Kbit/s for Coastguard at QP = 44, 8.30 Mbit/s at QP = 10), however. The QP value may however vary during the encoding process, depending on the position of each frame within the Group of Pictures.
QP | Foreman Bit rates [Kbit/s] |
News Bit rates [Kbit/s] |
Coastguard Bit rates [Kbit/s] |
---|---|---|---|
44 | 204 | 122 | 296 |
20 | 1440 | 645 | 2700 |
15 | 3240 | 1240 | 4990 |
10 | 6530 | 2630 | 8300 |
Fig. 8 to Fig. 14 plot the PSNR values for the three video sequences encoded at seven different quantization levels and transmitted from same multimedia server accessed at varying background traffic bit rate levels. Encoded only videos generally have higher PSNR values compared to their transmitted counterparts. At higher quantization levels (QP = 44 to 32) and lower background bit rate level, the PSNR value of the streamed video sequences remain same as their coded only counterparts, meaning that no video packets were lost during transmission. However, at lower quantization levels and higher background traffic thresholds, the PSNR values of the streamed video decline sharply. Content-based analysis reveals that the video sequences can react differently to competition for channel bandwidth arising from background traffic of different bit rates. This could be attributed to different spatio-temporal complexities of the sequences. Given no packet pritotization, contents with high bit rates (e.g. Coastguard) suffer higher PSNR degradation, caused by collision- induced video packet loss at the MAC layer of the streaming server, even at same encoding quantization level.
This paper has presented a detailed video quality evaluation in the transmission of H.264/SVC video over IEEE 802.11 networks in the presence of background traffic. We considered a scenario where a wireless multimedia server is transmitting single-layer encoded H.264/SVC and background traffic to one client and two sets of background traffic to another client. We objectively evaluated the quality of the streamed video given background traffic with varying bit rates, contents with different spatio-temporal information encoded at different quantization parameter level. Results indicate that received video quality deteriorates with increasing background traffic and high content bit rate, given no packet differentiation at the MAC (Media Access Control) layer. Also, contents may be affected differently, depending on the scene complexity and coding efficiency. For future work, we intend to expand the studies to tradeoffs in video quality optimization in the presence of background traffic, which includes packet prioritization and QoS mapping, and the use of IEEE 802.11e for SVC content streaming in IEEE 802.11 networks.
This work was supported by the COST IC1003 European Network on Quality of Experience in Multimedia Systems and Services – QUALINET; by the COST CZ LD12018 Modeling and verification of methods for Quality of Experience (QoE) assessment in multimedia systems – MOVERIQ; by the grant No. P102/10/1320 Research and modeling of advanced methods of image quality evaluation of the Grant Agency of the Czech Republic; and by the project of the Student grant agency of the Czech Technical University in Prague SGS12/077/OHK3/1T/13, “Cross-Layer Quality Optimization in New Generation Heterogeneous Wireless Mobile Networks.”
[1] Z. He, J. Cai, C.W. Chen, “Joint source channel rate-distortion analysis for adaptive mode selection and rate control in wireless video coding,” IEEE Trans. Circuits Syst. Video Technol. 12 (6), 2002.
[2] C.-M. Chen, C.-W. Lin, H.-C. Wei, Y.-C. Chen, “Robust video streaming over wireless lans using multiple description transcoding and prioritized retransmission,” Visual Commun. Image Represent. 18 (3) 2007.
[3] C.H. Foh, Y. Zhang, Z. Ni, J. Cai, K.N. Ngan, “Optimized cross-layer design for scalable video transmission over the IEEE 802.11e networks,” IEEE Trans. Circuits Syst. Video Technol. 17 (12), 2007.
[4] A. Fiandrotti, D. Gallucci, E. Masala, E. Magli, “Traffic prioritization of H.264/SVC video over 802.11e ad hoc wireless networks, ” Proceedings of 17th International Conference on Computer Communications and Networks, Virgin Islands, USA, 2008.
[5] H. Schwarz, D. Marpe, and T. Wiegand, “Overview of the scalable video coding extension of the H.264/AVC standard,” IEEE Trans. Circuits Syst. Video Technol., vol. 17, no. 9, pp. 1103–1120, 2007.
[6] J. Lee, F. De Simone, and E. Ebrahimi, "Subjective quality assessment of scalable video coding: A survey,” 2011 Third International Workshop on Quality of Multimedia Experience (QoMEX), pp.25-30, 7-9 Sept. 2011.
[7] T. Schierl, T. Stockhammer, and T. Wiegand, “Mobile Video Transmission Using Scalable Video Coding,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 17, no. 9, pp. 1204-1217, Sept. 2007.
[8] S. Wenger, Y. K. Wang, T. Schierl, and A. Eleftheriadis, “RTP payload format for SVC video," Internet Engineering Task Force (IETF), September 2009.
[9] W. Ye-Kui, M. Hannuksela, S. Pateux, A. Eleftheriadis, and S. Wenger, “IEEE Transactions on System and transport interface of svc,” Circuits and Systems for Video Technology, vol. 17, no. 9, pp. 1149 –1163, Sept. 2007.
[10] Video Trace Library, http://dbq.multimediatech.cz/
[online]
[11] ITU T Rec. P.910, "Subjective video quality assessment methods for multimedia applications", Geneva, Sep. 1999.
[12] JSVM Software Manual,
http://evalsvc.googlecode.com/files/SoftwareManual.doc
[online]
[13] MP4BOX, http://www.videohelp.com/tools/mp4box
[online]
[14] IEEE Standard 802.11-2007, “Local and metropolitan area networks-Specific requirements Part 11: Wireless LAN Medium Access Control (MAC) and Physical Layer (PHY) Specifications,” June 2007.
[15] The Network Simulator-NS2, http://www.isi.edu/nsnam/ns/
[online]
[16] Z. Wang, L. Lu, and A. C. Bovik, “Video quality assessment based on structural distortion measurement,” Signal Processing: Image Communication, vol. 19, no. 2, pp.
121-132, 2004.
Tento web site byl vytvořen prostřednictvím phpRS - redakčního systému napsaného v PHP jazyce.
Na této stránce použité názvy programových produktů, firem apod. mohou být ochrannými známkami
nebo registrovanými ochrannými známkami příslušných vlastníků.