Change search
Link to record
Permanent link

Direct link
BETA
Rossholm, Andreas
Alternative names
Publications (8 of 8) Show all publications
Rossholm, A. & Lövström, B. (2016). A robust method for estimating synchronization and delay of audio and video for communication services. Multimedia tools and applications, 75(1), 527-545
Open this publication in new window or tab >>A robust method for estimating synchronization and delay of audio and video for communication services
2016 (English)In: Multimedia tools and applications, ISSN 1380-7501, E-ISSN 1573-7721, Vol. 75, no 1, p. 527-545Article in journal (Refereed) Published
Abstract [en]

One of the main contributions to the quality of experience in streaming services or in two-way communication of audio and video applications is synchronization. This has been shown in several studies and experiments but methods to measure synchronization are less frequent, especially for situations without internal access to the application and independent of platform and device. In this paper we present a method for measuring synchronization skewness as well as delay for audio and video. The solution incorporates audio and video reference streams, where audio and video frames are marked with frame numbers which are decoded on the receiver side to enable calculation of synchronization and delay. The method has been verified in a two-way communication application in a transparent network with and without inserting known delays, as well as in a network with 5 and 10 % packet loss levels. The method can be used for both streaming and two-way communication services, both with and without access to the internal structures, and enables measurements of applications running on e.g. smartphones, tablets, and laptops under various conditions.

Place, publisher, year, edition, pages
Springer US, 2016
Keywords
Lip sync · Synchronization, Delay, QoE, Video streaming, Video conferencing
National Category
Signal Processing
Identifiers
urn:nbn:se:bth-6407 (URN)10.1007/s11042-014-2306-6 (DOI)000367856500024 ()
Available from: 2015-02-27 Created: 2015-02-26 Last updated: 2017-12-04Bibliographically approved
Shahid, M., Pandremmenou,, K., Lisimachos P., K., Andreas, R. & Lövström, B. (2016). Perceptual quality estimation of H.264/AVC videos using reduced-reference and no-reference models. Journal of Electronic Imaging (JEI), 25(5)
Open this publication in new window or tab >>Perceptual quality estimation of H.264/AVC videos using reduced-reference and no-reference models
Show others...
2016 (English)In: Journal of Electronic Imaging (JEI), ISSN 1017-9909, E-ISSN 1560-229X, Vol. 25, no 5Article in journal (Refereed) Published
Abstract [en]

Reduced-reference (RR) and no-reference (NR) models for video quality estimation, using featuresthat account for the impact of coding artifacts, spatio-temporal complexity, and packet losses, are proposed. Thepurpose of this study is to analyze a number of potentially quality-relevant features in order to select the mostsuitable set of features for building the desired models. The proposed sets of features have not been used in theliterature and some of the features are used for the first time in this study. The features are employed by the leastabsolute shrinkage and selection operator (LASSO), which selects only the most influential of them toward per-ceptual quality. For comparison, we apply feature selection in the complete feature sets and ridge regression onthe reduced sets. The models are validated using a database of H.264/AVC encoded videos that were subjec-tively assessed for quality in an ITU-T compliant laboratory. We infer that just two features selected by RRLASSO and two bitstream-based features selected by NR LASSO are able to estimate perceptual qualitywith high accuracy, higher than that of ridge, which uses more features. The comparisons with competingworks and two full-reference metrics also verify the superiority of our models.

Place, publisher, year, edition, pages
SPIE - International Society for Optical Engineering, 2016
Keywords
no-reference; packet loss; perceptual quality estimation; reduced-reference; video q
National Category
Other Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:bth-13256 (URN)10.1117/1.JEI.25.5.053012 (DOI)000388216900023 ()
Available from: 2016-10-14 Created: 2016-10-14 Last updated: 2018-05-23Bibliographically approved
Minhas, T. N., Shahid, M., Lövström, B., Rossholm, A., Zepernick, H.-J. & Fiedler, M. (2016). QoE rating performance evaluation of ITU-T recommended video quality metrics in the context of video freezes. Australian Journal of Electrical and Electronics Engineering, 13(2), 122-131
Open this publication in new window or tab >>QoE rating performance evaluation of ITU-T recommended video quality metrics in the context of video freezes
Show others...
2016 (English)In: Australian Journal of Electrical and Electronics Engineering, ISSN 1448-837X, Vol. 13, no 2, p. 122-131Article in journal (Refereed) Published
Abstract [en]

In real-time video streaming, video quality can be degraded due to network performance issues. Among other artefacts, video freezing and video jumping are factors that influence user experience. Service providers, operators and manufacturers are interested in evaluating the quality of experience (QoE) objectively because subjective assessment of QoE is expensive and, in many user cases, subjective assessment is not possible to perform. Different algorithms have been proposed and implemented in this regard. Some of them are in the recommendation list of the ITU Telecommunication Standardization Sector (ITU-T). In this paper, we study the effect of the freezing artefact on user experience and compare the mean opinion score of these videos with the results of two algorithms, the perceptual evaluation of video quality (PEVQ) and temporal quality metric (TQM). Both metrics are part of the ITU-T Recommendation J.247 Annex B and C. PEVQ is a full-reference video quality metric, whereas TQM is a no-reference quality metric. Another contribution of this paper is the study of the impact of different resolutions and frame rates on user experience and how accurately PEVQ and TQM measure varying frame rates.

Place, publisher, year, edition, pages
Taylor & Francis, 2016
Keywords
Freezing; Image quality; Quality of service; Video signal processing; Video streaming, Objective video quality; Quality of experience (QoE); Subjective video quality; Temporal quality; Video quality, Quality control
National Category
Signal Processing Communication Systems
Identifiers
urn:nbn:se:bth-13135 (URN)10.1080/1448837X.2015.1094855 (DOI)2-s2.0-84965031048 (Scopus ID)
Available from: 2016-10-04 Created: 2016-10-03 Last updated: 2016-12-19Bibliographically approved
Rossholm, A., Shahid, M. & Lövström, B. (2014). Analysis of the impact of temporal, spatial, and quantization variations on perceptual video quality. In: : . Paper presented at IEEE/IFIP Network Operations and Management Symposium: Management in a Software Defined World, Krakow. IEEE
Open this publication in new window or tab >>Analysis of the impact of temporal, spatial, and quantization variations on perceptual video quality
2014 (English)Conference paper, Published paper (Refereed)
Abstract [en]

The growing consumer interest in video communication has increased the users' awareness in the visual quality of the delivered media. This in turn increases, at the service provider end, the need for intelligent methodologies of optimal techniques for adapting to varying network conditions. Recent studies show that constraints on the bandwidth of transmission media should not always be translated to an increase in compression ratio to lower the bitrate of the video. Instead, a suitable option for adaptive streaming is to scale down the video temporally or spatially before encoding to maintain a desirable level of perceptual quality, while the viewing resolution is constant. Most of the existing studies to examine these scenarios are either limited to low resolution videos or lack in provisioning of subjective assessment of quality. We present here the results of our campaign of subjective quality assessment experiments done on a range of spatial and temporal resolutions, up to VGA and 30 frames per second respectively, under a number of bitrate conditions. The analysis shows, among other things, that keeping the spatial resolution is perceptually preferred among the three parameters that have impact on the video quality, even in the case with high temporal activity.

Place, publisher, year, edition, pages
IEEE, 2014
Series
IEEE IFIP Network Operations and Management Symposium, ISSN 1542-1201
Keywords
Compression ratio (machinery), Consumer interests, Low resolution video, Perceptual quality, Perceptual video quality, Spatial and temporal resolutions, Subjective assessments, Subjective quality assessments, Video communications
National Category
Signal Processing
Identifiers
urn:nbn:se:bth-6406 (URN)10.1109/NOMS.2014.6838397 (DOI)000356862300166 ()9781479909131 (ISBN)
Conference
IEEE/IFIP Network Operations and Management Symposium: Management in a Software Defined World, Krakow
Available from: 2015-02-27 Created: 2014-11-24 Last updated: 2017-03-17Bibliographically approved
Shahid, M., Rossholm, A., Lövström, B. & Zepernick, H.-J. (2014). No-reference image and video quality assessment: a classification and review of recent approaches. EURASIP Journal on Image and Video Processing, 2014(40)
Open this publication in new window or tab >>No-reference image and video quality assessment: a classification and review of recent approaches
2014 (English)In: EURASIP Journal on Image and Video Processing, ISSN 1687-5176, E-ISSN 1687-5281, Vol. 2014, no 40Article, review/survey (Refereed) Published
Abstract [en]

The field of perceptual quality assessment has gone through a wide range of developments and it is still growing. In particular, the area of no-reference (NR) image and video quality assessment has progressed rapidly during the last decade. In this article, we present a classification and review of latest published research work in the area of NR image and video quality assessment. The NR methods of visual quality assessment considered for review are structured into categories and subcategories based on the types of methodologies used for the underlying processing employed for quality estimation. Overall, the classification has been done into three categories, namely, pixel-based methods, bitstream-based methods, and hybrid methods of the aforementioned two categories. We believe that the review presented in this article will be helpful for practitioners as well as for researchers to keep abreast of the recent developments in the area of NR image and video quality assessment. This article can be used for various purposes such as gaining a structured overview of the field and to carry out performance comparisons for the state-of-the-art methods.

Place, publisher, year, edition, pages
Springer, 2014
Keywords
No-reference, Image quality assessment, Video quality assessment, Perceptual quality
National Category
Signal Processing
Identifiers
urn:nbn:se:bth-6339 (URN)10.1186/1687-5281-2014-40 (DOI)000347551600001 ()oai:bth.se:forskinfoD7B57209B4F14304C1257D49004E907F (Local ID)oai:bth.se:forskinfoD7B57209B4F14304C1257D49004E907F (Archive number)oai:bth.se:forskinfoD7B57209B4F14304C1257D49004E907F (OAI)
Note

Open access journal

Available from: 2015-05-26 Created: 2014-09-04 Last updated: 2017-12-04Bibliographically approved
Rossholm, A. (2014). On Enhancement and Quality Assessment of Audio and Video in Communication Systems. (Doctoral dissertation). Karlskrona: Blekinge Institute of Technology
Open this publication in new window or tab >>On Enhancement and Quality Assessment of Audio and Video in Communication Systems
2014 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

The use of audio and video communication has increased exponentially over the last decade and has gone from speech over GSM to HD resolution video conference between continents on mobile devices. As the use becomes more widespread the interest in delivering high quality media increases even on devices with limited resources. This includes both development and enhancement of the communication chain but also the topic of objective measurements of the perceived quality. The focus of this thesis work has been to perform enhancement within speech encoding and video decoding, to measure influence factors of audio and video performance, and to build methods to predict the perceived video quality. The audio enhancement part of this thesis addresses the well known problem in the GSM system with an interfering signal generated by the switching nature of TDMA cellular telephony. Two different solutions are given to suppress such interference internally in the mobile handset. The first method involves the use of subtractive noise cancellation employing correlators, the second uses a structure of IIR notch filters. Both solutions use control algorithms based on the state of the communication between the mobile handset and the base station. The video enhancement part presents two post-filters. These two filters are designed to improve visual quality of highly compressed video streams from standard, block-based video codecs by combating both blocking and ringing artifacts. The second post-filter also performs sharpening. The third part addresses the problem of measuring audio and video delay as well as skewness between these, also known as synchronization. This method is a black box technique which enables it to be applied on any audiovisual application, proprietary as well as open standards, and can be run on any platform and over any network connectivity. The last part addresses no-reference (NR) bitstream video quality prediction using features extracted from the coded video stream. Several methods have been used and evaluated: Multiple Linear Regression (MLR), Artificial Neural Network (ANN), and Least Square Support Vector Machines (LS-SVM), showing high correlation with both MOS and objective video assessment methods as PSNR and PEVQ. The impact from temporal, spatial and quantization variations on perceptual video quality has also been addressed, together with the trade off between these, and for this purpose a set of locally conducted subjective experiments were performed.

Place, publisher, year, edition, pages
Karlskrona: Blekinge Institute of Technology, 2014
Series
Blekinge Institute of Technology Doctoral Dissertation Series, ISSN 1653-2090 ; 16
Keywords
QoE, video quality assessment, video quality metric, multi-linear regression, artificial neural network, support vector machine, quality predictor, machine learning, temporal scaling, spatial scaling, video compression, deblocking filter, noise cancelling, synchronization, audio delay, video delay, GSM interference signal, noise cancellation, notch filtering
National Category
Signal Processing
Identifiers
urn:nbn:se:bth-00604 (URN)oai:bth.se:forskinfoDE7E8BB7B60A3B4EC1257D820035F0B6 (Local ID)978-91-7295-295-9 (ISBN)oai:bth.se:forskinfoDE7E8BB7B60A3B4EC1257D820035F0B6 (Archive number)oai:bth.se:forskinfoDE7E8BB7B60A3B4EC1257D820035F0B6 (OAI)
Available from: 2014-12-11 Created: 2014-10-31 Last updated: 2017-03-14Bibliographically approved
Shahid, M., Rossholm, A. & Lövström, B. (2013). A No-Reference Machine Learning Based Video Quality Predictor. In: : . Paper presented at Fifth International Workshop on Quality of Multimedia Experience (QoMEX). Klagenfurt am Wörthersee: IEEE
Open this publication in new window or tab >>A No-Reference Machine Learning Based Video Quality Predictor
2013 (English)Conference paper, Published paper (Refereed)
Abstract [en]

The growing need of quick and online estimation of video quality necessitates the study of new frontiers in the area of no-reference visual quality assessment. Bitstream-layer model based video quality predictors use certain visual quality relevant features from the encoded video bitstream to estimate the quality. Contemporary techniques vary in the number and nature of features employed and the use of prediction model. This paper proposes a prediction model with a concise set of bitstream based features and a machine learning based quality predictor. Several full reference quality metrics are predicted using the proposed model with reasonably good levels of accuracy, monotonicity and consistency.

Place, publisher, year, edition, pages
Klagenfurt am Wörthersee: IEEE, 2013
Keywords
Video Quality, H.264/AVC, Bitstream Features, No-Reference, Support Vector Machine
National Category
Signal Processing
Identifiers
urn:nbn:se:bth-6686 (URN)10.1109/QoMEX.2013.6603233 (DOI)000331828000042 ()oai:bth.se:forskinfoF4C25327032FF159C1257BF3002DFA49 (Local ID)oai:bth.se:forskinfoF4C25327032FF159C1257BF3002DFA49 (Archive number)oai:bth.se:forskinfoF4C25327032FF159C1257BF3002DFA49 (OAI)
External cooperation:
Conference
Fifth International Workshop on Quality of Multimedia Experience (QoMEX)
Note

http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=6603233&isnumber=6603187

Available from: 2014-07-17 Created: 2013-09-27 Last updated: 2016-09-09Bibliographically approved
Shahid, M., Singam, A. K., Rossholm, A. & Lövstrom, B. (2012). Subjective Quality Assessment of H.264/AVC Encoded Low Resolution Videos. In: : . Paper presented at 5th International Congress on Image and Signal Processing (CISP). Chongqing, Sichuan, China: IEEE
Open this publication in new window or tab >>Subjective Quality Assessment of H.264/AVC Encoded Low Resolution Videos
2012 (English)Conference paper, Published paper (Refereed)
Abstract [en]

Advancements in the video processing area have been proliferated by services that require low delay. Such services involve applications being offered at various temporal and spatial resolutions. It necessitates to study the impacts of related video coding conditions upon perceptual quality. But most of studies concerned with quality assessment of videos affected by coding distortions lack in variety of spatio-temporal resolutions. This paper presents a work done on quality assessment of videos encoded by state-of-the-art H.264/AVC standard at different bitrates and frame rates. Overall, 120 test scenarios for video sequences having different spatial and temporal spectral information were studied. The used coded bistreams in this work and the corresponding subjective assessment scores have been made public for the research community to facilitate further studies

Place, publisher, year, edition, pages
Chongqing, Sichuan, China: IEEE, 2012
National Category
Signal Processing
Identifiers
urn:nbn:se:bth-7008 (URN)10.1109/CISP.2012.6470021 (DOI)oai:bth.se:forskinfo2610FB4434C4438CC1257B28005CD516 (Local ID)978-1-4673-0965-3 (ISBN)oai:bth.se:forskinfo2610FB4434C4438CC1257B28005CD516 (Archive number)oai:bth.se:forskinfo2610FB4434C4438CC1257B28005CD516 (OAI)
External cooperation:
Conference
5th International Congress on Image and Signal Processing (CISP)
Available from: 2013-03-11 Created: 2013-03-08 Last updated: 2016-09-08Bibliographically approved
Organisations

Search in DiVA

Show all publications