monitoring video quality inside a network

28
AT&T Labs - Research SPS Santa Clara 09 - Page 1 09/09 Reibman Monitoring video quality inside a network Amy R. Reibman AT&T Labs – Research Florham Park, NJ [email protected] AT&T Labs - Research SPS Santa Clara 09 - Page 2 09/09 Reibman Outline Measuring video quality (inside a network) Anatomy of packet loss impairments (PLI) Estimate MSE due to a PLI Predicting visibility of PLI Conclusions and challenges

Upload: others

Post on 12-Apr-2022

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 109/09 Reibman

Monitoring video quality inside a network

Amy R. Reibman

AT&T Labs – ResearchFlorham Park, NJ

[email protected]

AT&T Labs - Research SPS Santa Clara 09 - Page 209/09 Reibman

Outline

• Measuring video quality (inside a network)

• Anatomy of packet loss impairments (PLI)

• Estimate MSE due to a PLI

• Predicting visibility of PLI

• Conclusions and challenges

Page 2: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 309/09 Reibman

Applications of video quality estimators

• Algorithm optimization– Automated in-the-loop assessment

• Product benchmarks– Vendor comparison to decide what product to buy

– Product marketing to convince customer to give you $$

• System provisioning– Determine how many servers, how much bandwidth, etc.

• Content acquisition and delivery (and SLAs)– Enter into legal agreements with other parties

• Outage detection and troubleshooting

AT&T Labs - Research SPS Santa Clara 09 - Page 409/09 Reibman

Measuring video quality inside the network

A video quality monitor for inside the network that is1. Real-time,2. Per stream,3. Scalable to many streams in network,4. Measures only impact of network impairments,5. Uses human perceptual properties, and

6. Accurate enough to answer the question:

To what degree are specific network impairments affecting the quality of this specific video content?

Page 3: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 509/09 Reibman

Factors that affect video quality

• Video compression algorithm factors– Decoder concealment, packetization, GOP structure, …

• Network-specific factors– Delay, Delay variation, Bit-rate, Packet losses

• Network independent factors– Sequence

• Content, amount of motion, amount of texture, spatial and temporal resolution

– User• Eyesight, interest, experience, involvement, expectations

– Environmental viewing conditions• Background and room lighting; display sensitivity, contrast,

and characteristics; viewing distance

AT&T Labs - Research SPS Santa Clara 09 - Page 609/09 Reibman

Where to measure?

• In the network– If corporate network is managed by third party

– Network operator does not have access to end-systems

– For videos traversing multiple ISPs

• Between LAN and WAN, or at access/peering points between ISPs

Video encoder

Videodecoder

networka dcb

Page 4: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 709/09 Reibman

What to measure?

• Not average network performance– Different ISPs,

– Different bandwidth capacities,

– Different time-varying loads

• Not only network-level measurements– Not all impairments produce same impact

– Example: some packet losses are invisible, others are highly visible

AT&T Labs - Research SPS Santa Clara 09 - Page 809/09 Reibman

What information can you gather?

• Original video X• Encoding parameters E(.)• Complete encoded bitstream E(X)• Network impairments (losses, jitter) L(.)• Lossy bitstream L(E(X))• Decoder (concealment, buffer, jitter) D(.)• Decoded pixels D(L(E(X)))

Video encoder

Videodecoder

networka dcb

Page 5: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 909/09 Reibman

Constraints imposed by “inside the network”

• Complexity, Scalability– If processing too complicated, can’t do for all streams

• Security, Proprietary algorithms– If encrypted content, can only process packet headers

• Structural constraints – Some data is unknowable (ex: environmental conditions)

– Make reasonable assumptions about decoder (buffer handling, error concealment)

• Measurement point(s) location?– Miss impairments between measurement point and viewer

– Not all measurements may be accurate

AT&T Labs - Research SPS Santa Clara 09 - Page 1009/09 Reibman

Categorizing image and video estimators

• Full and Reduced Reference (FR and RR)– Most available info; requires original and decoded pixels

• No-Reference Pixel-based methods (NR-P)– Requires decoded pixels: a decoder for each video stream

• No-Reference Bitstream-based methods (NR-B)– Processes packets containing bitstream, without decoder

FRNR-P

NR-B

Page 6: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 1109/09 Reibman

Traditional FR video quality measurements

• Original video X• Encoding parameters E(.)• Complete encoded bitstream E(X)• Network impairments (losses, jitter) L(.)• Lossy bitstream L(E(X))• Decoder (concealment, buffer, jitter) D(.)• Decoded pixels D(L(E(X)))

Video encoder

Videodecoder

networka dcb

AT&T Labs - Research SPS Santa Clara 09 - Page 1209/09 Reibman

Why doesn’t this solve our problem?

• Full-Reference: uses original and decoded video– Needs original video

– Needs decoded video: a decoder for each stream in network

– Cannot isolate impact of network impairments

– Perceptual Full-Reference estimators are REALLY complicated!

– Lots of parameter settings to get right

Page 7: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 1309/09 Reibman

NR-Pixel methods for video quality

• No-Reference pixel QE: uses only decoded video– Still needs a decoder for each stream in network

– Still cannot isolate impact of network impairments

• Black-frame detection• Video freezes• Blockiness (Wu ‘97, Wang ‘00, …)• Blurriness (Marziliano ‘02)• Jerkiness (Pastrana-Vidal ‘05, Huynh-Thu ‘06)• Ineffectiveness of error concealment (Yamada ‘07)• Spatial Aliasing (Reibman ‘08)

AT&T Labs - Research SPS Santa Clara 09 - Page 1409/09 Reibman

No-reference Bitstream methods

• Original video X• Encoding parameters E(.)• Complete encoded bitstream E(X)• Network impairments (losses, jitter) L(.)• Lossy bitstream L(E(X))• Decoder (concealment, buffer, jitter) D(.)• Decoded pixels D(L(E(X)))

Video encoder

Videodecoder

networka dcb

Page 8: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 1509/09 Reibman

NR-Bitstream methods for video quality

• FullParse – No complete decoding, but VLD– Mean, variance, spatial correlation, motion vectors– Location, extent, duration of losses

• QuickParse – “Easy-to-find” information only– Header information– Frame-level (or slice-level) summary information

• NoParse – Network-level stats only

FullParseQuickParse

NoParse

AT&T Labs - Research SPS Santa Clara 09 - Page 1609/09 Reibman

ITU-T SG 12 standardization of QoS/QoE

• P.NAMS– Non-intrusive parametric model for quality assessment

– Only packet-header information (IP through MPEG-2 TS)

– Useful if payload is encrypted

– Useful when processing capability is very limited

• P.NBAMS– Non-intrusive bitstream model for quality assessment

– Allowed to use coded bitstream

Page 9: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 1709/09 Reibman

Traditional network-based monitoring

• Original video X• Encoding parameters E(.)• Complete encoded bitstream E(X)• Network impairments (losses, jitter) L(.)• Lossy bitstream L(E(X))• Decoder (concealment, buffer, jitter) D(.)• Decoded pixels D(L(E(X)))

Video encoder

Videodecoder

networka dcb

AT&T Labs - Research SPS Santa Clara 09 - Page 1809/09 Reibman

Why is PLR not enough?

• For MPEG-2, average MSE is linear with PLR

• What is the correct slope for a given bitstream?• Depends on sequence-specific factors

– Source content: motion, texture

• Depends on encoder-specific factors– Frequency of Intra information, bit-rate

• What is specific error for the given loss pattern?• Depends on location of specific losses

– Which frame type, individual spatial and temporal extent, scene change?

Page 10: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 1909/09 Reibman

Influence of different content

0 1 2 3 4 5 6

x 10-3

0

20

40

60

80

100

120

140

Packet Loss Ratio

Seq

uenc

e M

SE

Eight 10-second MPEG-2 sequences, similar bit-rate

AT&T Labs - Research SPS Santa Clara 09 - Page 2009/09 Reibman

Variation due to different losses

Sequence F2Sequence G4

0 1 2 3 4 5 6

x 10-3

0

20

40

60

80

100

120

140

Packet Loss Ratio

Seq

uenc

e M

SE

Page 11: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 2109/09 Reibman

Quality assessment for networked video

• Compression effects– NR Estimation of MSE due to compression (Turaga ‘02, Ichigaya

‘04)

– Motion-compensated edge artifacts (Leontaris ‘05)

• Packet loss effects– Estimate MSE (Reibman ‘02, Naccari ‘08)

– Compute Mean Opinion Score (MOS) (Winkler 03, Liu 07, Lin 08)

– Estimate visibility of individual packet losses (Kanumuri ‘04)

– Estimate Mean Time Between Failures (Suresh ‘05)

• Timing effects (jitter)– Understand delivered video content in streaming scenario

(Reibman ‘04, Gustafsson ‘08)

AT&T Labs - Research SPS Santa Clara 09 - Page 2209/09 Reibman

Estimating MSE due to packet loss

where is encoded value at pixel i frame nand is decoded value at pixel i frame nand is error for pixel i frame n

• What clues are in the bitstream to estimate MSE?

• Map unstructured problem into equivalent structured problem

∑∑∑∑ =−=n in i

ineN

infinfN

MSE 22 ),(1

)),(~

),(ˆ(1

),(ˆ inf

),(~

inf

),( ine

Page 12: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 2309/09 Reibman

Impact of network losses

M 0: set of macroblocks initially lost

AT&T Labs - Research SPS Santa Clara 09 - Page 2409/09 Reibman

Impact of network losses

M 0: set of macroblocks initially lost

e0(n,i) : initial magnitude of error

Page 13: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 2509/09 Reibman

Impact of network losses

M 0: set of macroblocks initially lost

e0(n,i) : initial magnitude of error

ψ: prediction process(propagation of error;macroblock type and motion)

AT&T Labs - Research SPS Santa Clara 09 - Page 2609/09 Reibman

Impact of network losses

M 0: set of macroblocks initially lost

e0(n,i) : initial magnitude of error

ψ: prediction process(propagation of error;macroblock type and motion)

Page 14: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 2709/09 Reibman

Characterization of the error

• Error is completely characterized by1. Which macroblocks are initially in error (M 0)

2. How large the initial error is in those macroblocks (e0(n,i) )

3. How the error propagates in space and time (ψ)

AT&T Labs - Research SPS Santa Clara 09 - Page 2809/09 Reibman

Characterization of the error

• Error is completely characterized by1. Which macroblocks are initially in error (M 0)

• Entire picture lost, 1 slice lost, 2 slices lost, …

2. How large the initial error is in those macroblocks (e0(n,i) )

• Depends on source activity (still/moving)

• Depends on encoder prediction

• Depends on decoder’s concealment strategy

3. How the error propagates in space and time (ψ)

• Losses in B-frames only impact one frame

• Received I-frame cleans out previous errors

Page 15: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 2909/09 Reibman

Characterization of the error: in the network

• Error is completely characterized by1. Which macroblocks are initially in error (M 0)

• Can be measured directly from lossy bitstream (NR-B)

• Depends on compression, not on video content

2. How large the initial error is in those macroblocks (e0(n,i) )

• Very hard to estimate accurately from lossy bitstream

• Can be computed exactly given complete bitstream

3. How the error propagates in space and time (ψ)

• Characterized by motion vectors, macroblock types

• Can be extracted exactly from the lossy bitstream (NR-B)

AT&T Labs - Research SPS Santa Clara 09 - Page 3009/09 Reibman

Calculating MSE due to packet loss

• Encoder-based estimation of MSE– Uncertainty of loss location, M 0

– Exact knowledge of propagation, ψ– Exact knowledge of initial error, e0(n,i)

• Bitstream-based estimation of MSE– Exact knowledge of location of losses, M 0

– Exact knowledge of propagation, ψ– Unknown initial error, e0(n,i)

Page 16: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 3109/09 Reibman

Estimating MSE from lossy bitstream, L(E(X))

ExtractBitstreamData

EstimateInitialError

NewLoss?

Yes

PropagatePastErrors

No

Lossy video bitstream: L(E(X))

MSEestimate(alarm)

AT&T Labs - Research SPS Santa Clara 09 - Page 3209/09 Reibman

Extracting bitstream data

• How deeply can you process the packets?

QuickParse: Extracts slice-level information only– Frame type, slice location, slice bit-rate, slice quantizer

– Knows which macroblocks; knows when errors stop

– Approximates spatial spread of the error propagation

FullParse: macroblock-level -- no complete decoding!– Mean, variance, spatial correlation, motion vectors

– Knows exactly which macroblocks and how errors propagate

Page 17: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 3309/09 Reibman

Performance comparison: data

• 225 sample packet loss traces• 9 different PLR ranging from 5*10-5 to 5*10-3

• 25 sample traces per PLR

• 16 10-second MPEG-2 sequences– Wide range of sensitivity to packet loss

– 8 sequences in training set; 8 sequences in test set

AT&T Labs - Research SPS Santa Clara 09 - Page 3409/09 Reibman

NoParse: Performance

F1F2F3F4

0 50 100 150 2000

20

40

60

80

100

120

140

160

180

200

Actual sequence MSE

Est

imat

ed s

eque

nce

MS

E

Correlation0.71

Probability of error: 9-14%

Assumes MSE linear with PLR

Page 18: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 3509/09 Reibman

QuickParse: Performance

F1F2F3F4

0 50 100 150 2000

20

40

60

80

100

120

140

160

180

200

Actual sequence MSE

Est

imat

ed s

eque

nce

MS

E

Original QuickParse

Probability of error: 8-13%

Correlation0.79

AT&T Labs - Research SPS Santa Clara 09 - Page 3609/09 Reibman

FullParse: Performance

F1F2F3F4

0 50 100 150 2000

20

40

60

80

100

120

140

160

180

200

Actual sequence MSE

Est

imat

ed s

eque

nce

MS

E

FullParse

Correlation0.95

Probability of error: 3-4%

Page 19: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 3709/09 Reibman

Bounds: Estimating MSE from E(X)

ExtractBitstreamData

Use exact initial error Loss?

Yes

PropagatePastErrors

No

Lossy video Bitstream, L(E(X))

MSEestimate

Info from E(X)

AT&T Labs - Research SPS Santa Clara 09 - Page 3809/09 Reibman

FullParse: Performance

F1F2F3F4

0 50 100 150 2000

20

40

60

80

100

120

140

160

180

200

Actual sequence MSE

Est

imat

ed s

eque

nce

MS

E

FullParse

Correlation0.95

Probability of error: 3-4%

Page 20: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 3909/09 Reibman

FullParse Bound: Performance

F1F2F3F4

0 50 100 150 2000

20

40

60

80

100

120

140

160

180

200

Actual sequence MSE

Est

imat

ed s

eque

nce

MS

E

bound

Correlation0.998

Probability of error: 2%

AT&T Labs - Research SPS Santa Clara 09 - Page 4009/09 Reibman

QuickParse bound: Performance

F1F2F3F4

0 50 100 150 2000

20

40

60

80

100

120

140

160

180

200

Actual sequence MSE

Est

imat

ed s

eque

nce

MS

E

QuickParse bound

Probability of error: 2%

Correlation0.995

Page 21: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 4109/09 Reibman

Performance and bounds (16 seqs)

0

0.5

1

1.5

2

2.5

Reg

ress

ion

Coe

ffici

ents

FullParse FP Bound QP Bound QuickParse

AT&T Labs - Research SPS Santa Clara 09 - Page 4209/09 Reibman

Observations: Broadcast MPEG-2 MSE

• QuickParse: Widely different slopes for different sequences

• FullParse: More accurate slopes, but room for improvement

• FullParse bound: Slopes consistently near one, but underestimated

• QuickParse bound: Nearly same as FullParse bound!

Inaccuracy of QuickParse is not due to simpler propagation, but to inaccurate estimate of initial error

• Reduce the complexity of FullParse– Estimate initial error with FP, propagate with QP

Page 22: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 4309/09 Reibman

Outline

• Measuring video quality (inside a network)

• Anatomy of packet loss impairments (PLI)

• Estimate MSE due to a PLI

• Predicting visibility of PLI– NOT interested in quality given an average packet loss rate

– Want to understand impact of each individual packet loss

• Conclusions and challenges

AT&T Labs - Research SPS Santa Clara 09 - Page 4409/09 Reibman

Visibility vs. quality

• Quality– How good is the video? How annoying are the artifacts?

– Viewers provide MOS on a scale of 1—5

• Visibility– Did you see an artifact? What fraction of viewers saw

artifact?

• Applications– High-quality video transport over a mostly reliable network

– Design system so that less than some fraction of viewers will notice an impairment in the delivered video stream less than every (time period)?

– Prioritization of packets to minimize visible impairments

Page 23: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 4509/09 Reibman

Three Subjective DataSets

• Similar strategy (3455 isolated packet losses)– Measure each individual packet loss, NOT average quality

• Testing methodology– One packet loss every 4 seconds

– Viewers are “immersed”, no audio, CRT display

– “Press the space bar when you see an artifact”

– 12 viewers for every PLI

• Wide range of parameters– Various compression standards (H.264, MPEG-2)

– Different encoding parameters (Group of Picture, etc)

– Different approaches for error concealment at decoder

AT&T Labs - Research SPS Santa Clara 09 - Page 4609/09 Reibman

Subjective test results: Ground truth

0 1 2 3 4 5 6 7 8 9 10 11 120

200

400

600

800

1000

1200

1400

1600

1800

2000

Number of viewers who saw each error

Num

ber

of e

rror

s

52% of errors seen by no one

79% seen by 3 or fewer

10% seen by 9 or more

Page 24: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 4709/09 Reibman

Visibility of packet loss impairments

• Depends on error itself– Size, spatial pattern, location, duration, amplitude

• Depends on decoded signal at location of error– New temporal edges (jerkiness), added horizontal edges,

broken-up vertical edges

• Depends on encoded signal at location of error– Texture masking, luminance masking, motion masking may

hide error

– Motion tracking may enhance visibility in smoothly moving areas

– This provides an implicit internal reference, even if not seen

AT&T Labs - Research SPS Santa Clara 09 - Page 4809/09 Reibman

Exploratory data analysis (EDA)

• Visibility as a function of one variable– Temporal duration: short one-frame errors are usually

invisible

• 1.5% of one-frame errors are seen by 75%+ of people

• 63% of one-frame errors are seen by NO ONE!

– Spatial extent: smaller errors more likely to be invisible

– Motion: small motion losses typically invisible

– Initial MSE: smaller errors more likely to be invisible

– Scene motion: losses more likely to invisible with still camera

Page 25: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 4909/09 Reibman

Initial MSE vs. visibility

-8 -6 -4 -2 0 2 4 6 8 100

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

prob

abili

ty

initial MSE (log) of error

Visible errorsInvisible errors

AT&T Labs - Research SPS Santa Clara 09 - Page 5009/09 Reibman

Visual Glitch Detector for packet losses

• Always extract some information for all videos– Information about encoded signal

• Local means and variances, motion, motion accuracy– Information about surrounding scene

• Camera motion; Near a scene change?

• When there is a packet loss, extract:– Information about decoded signal

• Extra edges possibly introduced– Information about error signal

• Size, duration, initial MSE, initial SSIM

• Estimate visibility using logistic regression– Trained using subjective tests; Humans create “ground truth”

Page 26: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 5109/09 Reibman

0 1 0 2 0 3 0 4 0 5 0 6 00

0 . 5

1

T i m e ( s e c o n d s )

PLD

0 1 0 2 0 3 0 4 0 5 0 6 00

0 . 5

1

T i m e ( s e c o n d s )

VG

D

Visual Glitch Detector

PacketLossesOnly----------VisualGlitchDetector

AT&T Labs - Research SPS Santa Clara 09 - Page 5209/09 Reibman

Conclusions

• Many open problems in measuring video quality

• Characterizing impact of packet loss using M 0 , ψ , and e0(n,i)useful in many contexts related to video transport over networks

• Perceptual quality estimators can be very easy to implement

• Lots of room for improvement: No-Reference quality estimators that are effective– Across different image content and good enough for a legal

contract

Page 27: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 5309/09 Reibman

Thanks

• To all my immediate collaborators• To E. Koutsofios for “lefty” graphics package• To the community at large• To all our subjective test participants• To a patient audience

AT&T Labs - Research SPS Santa Clara 09 - Page 5409/09 Reibman

Collaborators

• Broadcast MPEG-2 with losses: MSE– Vinay Vaishampayan (AT&T)

– Swamy Sermadevi (Cornell/Microsoft)

• Video streaming using Microsoft Media– Shubho Sen (AT&T)

– Kobus van der Merwe (AT&T)

• Broadcast MPEG-2 with losses: Visibility– Sandeep Kanumuri (UCSD/DocomoUSA)

– Vinay Vaishampayan (AT&T)

– David Poole (AT&T)

– Pamela Cosman (UCSD)

Page 28: Monitoring video quality inside a network

AT&T Labs - Research SPS Santa Clara 09 - Page 5509/09 Reibman

My journal papers on assessing quality• A. R. Reibman, Y. Sermadevi and V. Vaishampayan, “Quality

monitoring of video over a network", IEEE Transactions on Multimedia, vol. 6, no. 2, pp. 327-334, April 2004.

• S. Kanumuri, P. C. Cosman, A. R. Reibman, and V. Vaishampayan, “Modeling packet-loss visibility in MPEG-2 Video", IEEE Transactions on Multimedia, April 2006.

• A. Leontaris, P. C. Cosman, and A. R. Reibman, “Quality evaluation of motion-compensated edge artifacts in compressed video", IEEE Transactions on Image Processing, vol. 16, no. 4, pp. 943--956, April 2007.

AT&T Labs - Research SPS Santa Clara 09 - Page 5609/09 Reibman

My conference papers on assessing quality• A. R. Reibman, Y. Sermadevi and V. Vaishampayan, “Quality monitoring of video over the Internet", 36th Asilomar

Conference on Signals, Systems, and Computers, vol. 2, pp. 1320-1324, Nov. 2002.• A. R. Reibman and V. Vaishampayan, “Quality monitoring for compressed video subjected to packet loss", IEEE

International Conference on Multimedia and Expo (ICME'03), pp. I-17--20, vol. 1, July 2003.• A. R. Reibman and V. Vaishampayan, ̀ `Low-complexity quality monitoring of MPEG-2 video in a network", IEEE

International Conference on Image Processing (ICIP'03), pp. III-261--264, Sept. 2003.• A. R. Reibman, S. Kanumuri, V. Vaishampayan, and P. C. Cosman, “Visibility of individual packet losses in MPEG-2

video", IEEE International Conference on Image Processing (ICIP'04), pp. 171--174, Oct. 2004.• S. Kanumuri, P. C. Cosman, and A. R. Reibman, “A generalized linear model for MPEG-2 packet-loss visibility",

Proc. International Workshop on Packet Video, Dec. 2004.• A. R. Reibman, S. Sen, and J. van der Merwe, “Network monitoring for video quality over IP", Proc. Picture Coding

Symposium, Dec 2004.• A. R. Reibman, S. Sen, and J. van der Merwe, “Analyzing the spatial quality of Internet streaming video", First

International Workshop on Video Processing and Quality Metrics, Scottsdale, AZ, Jan. 2005. (http://enpub.fulton.asu.edu/resp/Upld/VPQM05/papers/226.pdf)

• A. R. Reibman, S. Sen, and J. van der Merwe, “Video quality estimation for Internet streaming", Fourteenth International World Wide Web Conference, Chiba Japan, May 2005.

• A. Leontaris and A. R. Reibman, “Comparison of blocking and blurring metrics for video compression", IEEE International Conference on Acoustics, Speech, and Signal Processing, pp. 585-588, March 2005.

• A. Leontaris, P. C. Cosman, and A. R. Reibman, “Measuring the added high frequency energy in compressed video", IEEE International Conference on Image Processing (ICIP'05), pp. 498-501, Sept. 2005.

• A. R. Reibman and T. Schaper, “Subjective performance evaluation for super-resolution image enhancement", Second International Workshop on Video Processing and Quality Metrics, Scottsdale, Arizona, January 2006. (http://enpub.fulton.asu.edu/resp/vpqm2006/papers06/303.pdf)

• A. R. Reibman, R. Bell, and S. Gray, “Quality assessment for super-resolution image enhancement", IEEE International Conference on Image Processing, October 2006.

• S. Kanumuri, S. G. Subramanian, P. C. Cosman, and A. R. Reibman, “Packet loss visibility in H.264 videos using a reduced reference method", IEEE International Conference on Image Processing, October 2006.

• A. R. Reibman and D. Poole, “Characterizing packet-loss impairments in compressed video", IEEE International Conference on Image Processing, Sept. 2007.

• A. R. Reibman and D. Poole, “Predicting packet-loss visibility using scene characteristics", Sixteenth International Packet Video Workshop, Nov. 2007.

• A. R. Reibman and S. Suthaharan, “A no-reference spatial aliasing measure for digital image resizing", IEEE International Conference on Image Processing, Oct. 2008.

• T.-L. Lin, P. C. Cosman, and A. R. Reibman, “Perceptual impact of bursty versus isolated packet losses in H.264 compressed video", IEEE International Conference on Image Processing, Oct. 2008.