next up previous contents index
Next: Simulator Description Up: Measuring Video Quality Previous: Measuring Video Quality   Contents   Index


Introduction

Many real-time video transmission applications over the Internet have appeared in the last few years. We find today: video phones, video conferencing, video streaming, tele-medical applications, distance learning, telepresence, video on demand, etc., with a different number of requirements in bandwidth and perceived quality. This gives rise to a need for assessing the quality of the transmitted video in real time. Traditionally, the quality is measured as a function of the encoding algorithms, without taking into account the effect of packet networks' parameters. It is now easy to transfer audio in real time over the Internet with reasonable quality. On the other hand, it is not easy to do the same for video. This is because real-time video demands greater bandwidth and, in general, dedicated hardware encoders due to the high complexity of video compression algorithms. In addition, in the case of networks with differentiated services, when audio and video streams are to be transmitted over the same network, audio streams are privileged over video streams when there is congestion. Thus, to guarantee reasonable video quality a dedicated network with sufficiently high bandwidth and QoS levels must be used. This solution is not always feasible due to the high costs of these equipments and the per-service prices. Another factor that makes transferring video more difficult than audio over public networks is the sensitivity to errors. Public networks are usually error prone. However, encoded video streams do not tolerate errors as audio, as video can tolerate only up to 4 % loss without error concealment. While, in speech, with up to 20 %, we get reasonable quality and still understand the talk. Due to these problems, users tend to tradeoff the factor of quality in some cases to deliver real-time video with small resolutions (SQCIF or QCIF) over the low-cost public Internet. The high complexity of video encoding and decoding processes with respect to those of speech and audio is proportional to the complexity of pure objective video quality measures. Due to the diversity of the encoding algorithms and their complexity, it is difficult to have objective quality measures that take into account all the possible coding and network impairments. In the literature, there are some objective quality measures that work well in some restricted situations, and fail in many others. In addition, as they operate pixel-by-pixel-wise and involve too much computations, the resulting latency and CPU uses limit their use in real-time applications, knowing that video encoding also demands very high computational power. Another weak point is that most of the available objective video quality measures require the access to both the original signal and the impaired one. This is impractical in the case of real-time transmission over packet networks. Now, if we consider the parameters that affect the quality (quality-affecting parameters) of video transmission over packet networks, we can classify them as follows: Since we concentrate here only on pure video applications, we do not take into account parameters such as lip synchronization or other audio aspects. It is clear that quality is not linearly proportional to the variation of these parameters. The determination of the quality is a very complex problem, and there is no mathematical model that can take into account the effects of all these parameters. Hence, it is normal that, so far, there is no pure objective measure that can quantify the quality of all the effects of all these parameters altogether that can operate in all the situations. In Chapter 4, we outlined a new methodology to evaluate in real time the quality of any media type transmitted over packet networks in real time. We validated the result for speech in Chapter 5. We would like to apply our method in this Chapter to evaluate its applicability to the case of video. Thus, we show how Neural Networks (NN) can be used to mimic the way by which a group of human subjects assess video quality when this video is distorted by certain quality-affecting parameters (e.g. packet loss rate, loss distribution, bit rate, frame rate, encoded frame type, etc.). In order to illustrate its applicability, we chose to assess the quality of video flows transmitted over IP networks and we carried out subjective quality tests for video distorted by variations of those parameters. In this Section, we focus on how our approach can be applied to the case of pure video. (For the case when video is associated with other media types, speech or audio, see the end of Section 4.2 in page [*].) We give here a summary of the general description for the case of video. The overall procedure is summarized as follows. First, we have to identify the quality-affecting parameters, their ranges and values. Then we have to choose the original video sequences and to produce the distorted video database. After that, the subjective quality test is to be carried out for the distorted video database and the MOS scores (together with their statistical analysis) are to be calculated in order to form the video quality database. Then, we select the NN architecture and learning algorithm. Then the NN is to be trained and tested using the video quality database. Finally, we implement the final tool that consists of the two modules (parameters collection and MOS evaluation), in order to obtain a real-time quality evaluator. In most cases, real-time video applications can be considered one-way sessions (i.e. they consist of a sender that produces the video and a receiver that consumes it). This behavior is different from that of audio applications. Indeed, in audio, the interactivity may imply some other parameters (e.g. echo, crosstalk effect, number of participating sources, etc.) that affect the overall quality [92]. The rest of this Chapter is organized as follows. In Section 6.2, we describe the simulator that we used to generate the video sequences required to identify the relation between the subjective quality measures and the parameters' values. A description of the quality-affecting parameters we considered in our experiment is given in Section 6.3. We outline the subjective quality test and the MOS calculation we performed in Section 6.4. The obtained results and the performance of the NN are given in Section 6.5. In Section 6.6, we analyze the performance of some of the existing objective video quality mesures. In Section 6.7, we present an application we implemented based on the obtained results to evaluate in real time the quality of any video sequence. Finally, we conclude our work in this Chapter in Section 6.8.
next up previous contents index
Next: Simulator Description Up: Measuring Video Quality Previous: Measuring Video Quality   Contents   Index
Samir Mohamed 2003-01-08