1
Exploiting High-Level Semantics for No-Reference Image Quality Assessment of Realistic Blur Images Dingquan Li, Tingting Jiang and Ming Jiang National Engineering Lab for Video Technology, Peking University Mountain View, CA USA October 23-27 2017 Background If you can’t measure it, you can’t improve it. - Peter Drucker Image quality assessment (IQA) is a critical precondi- tion for providing a satisfying end-user experience. Subjective IQA is reliable, but it is often cumbersome, ex- pensive and hard to carry out in reality. Objective IQA: full-reference IQA (FR-IQA), reduced- reference IQA (RR-IQA) and no-reference IQA (NR-IQA). NR-IQA is preferable but also more challenging in most prac- tical applications. Image blur is one of the most common distortions in practice, which relates to image quality. Figure 1: Typical causes of image blur in practice. Motivation Traditional blur-specific NR-IQA methods are mainly based on the assumptions that blur leads to the spread of edges (e.g., MDWE [1]), the smoothing effect (e.g., ARISM [2]), the re- duction of high-frequency energy (e.g., FISH [3]) or the loss of local phase coherence (e.g., LPC [4]). Which has better visual quality? Figure 2: The clear blue sky or a blurry mouse? The above low-level features based methods may contravene human perception on predicting the relative quality of im- age pairs with various image content. To tackle this issue, we resort to the high-level semantics, which is derived from a pre-trained deep convolutional neural network model. The Proposed Semantic Feature Aggregation (SFA) based Method Multi-patch Representation Pre-trained DCNN (ResNet- 50, pool 5 ) Mean&Std Aggregation Quantile Aggregation Moment Aggregation High-level Semantic Features PLSR PLSR PLSR Average Quality Score Step 2: Feature Extraction Step 3: Feature Aggregation Step 4: Quality Prediction Step 1: Image Representation Figure 3: The overall framework of the proposed method, mainly includes four steps: image representation, feature extraction, feature aggregation, and quality prediction. The choices in all steps are determined by the performance on validation data of BID. How to represent an image? 0.55 0.6 0.65 median SROCC in 1000 runs 0.7 fc 6 0.75 Multiple fc 7 Scaling Padding fc 8 Cropping Figure 4: Comparison among different image representations. Which layer to extract features? conv 1 conv 2 conv 3 conv 4 conv 5 fc 6 fc 7 fc 8 0 0.2 0.4 0.6 0.8 SROCC 0.2 0.4 0.6 0.8 1 PLCC 0.511 0.571 0.626 0.647 0.660 0.696 0.686 0.680 0.463 0.551 0.615 0.632 0.635 0.680 0.682 0.672 Low-level High-level Figure 5: Comparison among different layers in AlexNet. Which pre-trained model to extract features? 0.7023 0.6865 0.7007 0.6910 0.7577 0.7673 SROCC PLCC 0.6 0.65 0.7 0.75 0.8 median value in 1000 runs AlexNet GoogleNet ResNet-50 Figure 6: Comparison among different pre-trained models. How to aggregate the features? Aggregated Feature SROCC PLCC RMSE mean (f mean ) 0.7577 0.7673 0.8283 mean&std (f 1 ) 0.8022 0.8174 0.7333 quantile (f 2 ) 0.8109 0.8254 0.7135 moment (f 3 ) 0.8100 0.8254 0.7171 average-quality (f 1 , f 2 , f 3 ) 0.8154 0.8305 0.7055 Table 1: Comparison among different aggregation structures. Experiments Method BID CLIVE TID2008 LIVE MDWE [1] 0.3067 0.4313 0.8556 0.9188 FISH [3] 0.4736 0.4865 0.8737 0.9008 ARISM [2] 0.0151 0.2427 0.8851 0.9585 LPC [4] 0.3150 0.1483 0.8805 0.9469 Proposed 0.8269 0.8130 0.9098 0.9523 Table 2: Performance comparison (in terms of SROCC) on four databases. In each column, the best performance value is marked in boldface and the second best performance value is underlined. Train Test SROCC BID CLIVE 0.5729 CLIVE BID 0.6838 TID2008 LIVE 0.9166 LIVE TID2008 0.9243 Table 3: SROCC values of the proposed method in cross dataset evaluation. Conclusion A novel NR-IQA framework is proposed based on high-level semantic feature aggregation, whose superiority and generalization capability are verified on four popular image blur databases. High-level semantic information is experimentally verified to be crucial in quality estimation among various image content. References [1] P. Marziliano, F. Dufaux, S. Winkler, and T. Ebrahimi, “Perceptual blur and ringing metrics: Application to JPEG2000,” Signal Processing: Image Communication, vol. 19, no. 2, pp. 163–172, 2004. [2] K. Gu, G. Zhai, W. Lin, X. Yang, and W. Zhang, “No-reference image sharpness assessment in autoregressive parameter space,” IEEE Transactions on Image Processing, vol. 24, no. 10, pp. 3218–3231, 2015. [3]P. V. Vu and D. M. Chandler, “A fast wavelet-based algorithm for global and local image sharpness estimation,” IEEE Signal Processing Letters, vol. 19, no. 7, pp. 423–426, 2012. [4] R. Hassen, Z. Wang, and M. M. A. Salama, “Image sharpness assessment based on local phase coherence,” IEEE Transactions on Image Processing, vol. 22, no. 7, pp. 2798–2810, 2013. Contact Information Email: [email protected]

Exploiting High-Level Semantics for No-Reference Image ...Exploiting High-Level Semantics for No-Reference Image Quality Assessment of Realistic Blur Images DingquanLi,TingtingJiangandMingJiang

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Exploiting High-Level Semantics for No-Reference Image ...Exploiting High-Level Semantics for No-Reference Image Quality Assessment of Realistic Blur Images DingquanLi,TingtingJiangandMingJiang

Exploiting High-Level Semantics for No-Reference Image Quality Assessmentof Realistic Blur ImagesDingquan Li, Tingting Jiang and Ming Jiang

National Engineering Lab for Video Technology, Peking UniversityMountain View, CA USAOctober 23-27 2017

Background

If you can’t measure it, you can’t improve it. - Peter DruckerImage quality assessment (IQA) is a critical precondi-tion for providing a satisfying end-user experience.Subjective IQA is reliable, but it is often cumbersome, ex-pensive and hard to carry out in reality.Objective IQA: full-reference IQA (FR-IQA), reduced-reference IQA (RR-IQA) and no-reference IQA (NR-IQA).NR-IQA is preferable but also more challenging in most prac-tical applications.Image blur is one of the most common distortions in practice,which relates to image quality.

Figure 1: Typical causes of image blur in practice.

Motivation

Traditional blur-specific NR-IQA methods are mainly based onthe assumptions that blur leads to the spread of edges (e.g.,MDWE [1]), the smoothing effect (e.g., ARISM [2]), the re-duction of high-frequency energy (e.g., FISH [3]) or the loss oflocal phase coherence (e.g., LPC [4]).

Which has better visual quality?

Figure 2: The clear blue sky or a blurry mouse?

The above low-level features based methods may contravenehuman perception on predicting the relative quality of im-age pairs with various image content. To tackle this issue,we resort to the high-level semantics, which is derived froma pre-trained deep convolutional neural network model.

The Proposed Semantic Feature Aggregation (SFA) based Method

Multi-patch

Representation

Pre-trained

DCNN

(ResNet-

50, pool5)

Mean&Std

Aggregation

Quantile

Aggregation

Moment

Aggregation

High-level

Semantic

Features

PLSR

PLSR

PLSR

Average

Quality

Score

Step 2:

Feature Extraction

Step 3:

Feature Aggregation

Step 4:

Quality Prediction

Step 1:

Image Representation

Figure 3: The overall framework of the proposed method, mainly includes four steps: image representation, feature extraction,feature aggregation, and quality prediction. The choices in all steps are determined by the performance on validation data of BID.

How to represent an image?

0.55

0.6

0.65

med

ian

SRO

CC

in 1

000

runs

0.7

fc6

0.75

Multiplefc

7 ScalingPaddingfc

8Cropping

Figure 4: Comparison among different image representations.

Which layer to extract features?

conv1

conv2

conv3

conv4

conv5

fc6

fc7

fc8

0

0.2

0.4

0.6

0.8

SRO

CC

0.2

0.4

0.6

0.8

1

PLC

C

0.5110.571 0.626 0.647 0.660 0.696 0.686 0.680

0.4630.551

0.615 0.632 0.635 0.680 0.682 0.672

Low-level High-level

Figure 5: Comparison among different layers in AlexNet.

Which pre-trained model toextract features?

0.70

23

0.68

65

0.70

07

0.69

10

0.75

77

0.76

73

SROCC PLCC0.6

0.65

0.7

0.75

0.8

med

ian

valu

e in

100

0 ru

ns

AlexNet GoogleNet ResNet-50

Figure 6: Comparison among different pre-trained models.

How to aggregate the features?

Aggregated Feature SROCC PLCC RMSEmean (fmean) 0.7577 0.7673 0.8283mean&std (f1) 0.8022 0.8174 0.7333quantile (f2) 0.8109 0.8254 0.7135moment (f3) 0.8100 0.8254 0.7171average-quality (f1, f2, f3) 0.8154 0.8305 0.7055

Table 1: Comparison among different aggregation structures.

Experiments

Method BID CLIVE TID2008 LIVEMDWE [1] 0.3067 0.4313 0.8556 0.9188FISH [3] 0.4736 0.4865 0.8737 0.9008ARISM [2] 0.0151 0.2427 0.8851 0.9585LPC [4] 0.3150 0.1483 0.8805 0.9469Proposed 0.8269 0.8130 0.9098 0.9523

Table 2: Performance comparison (in terms of SROCC) onfour databases. In each column, the best performance valueis marked in boldface and the second best performance value isunderlined.

Train → Test SROCCBID → CLIVE 0.5729CLIVE → BID 0.6838

TID2008 → LIVE 0.9166LIVE → TID2008 0.9243

Table 3: SROCC values of the proposed method in cross datasetevaluation.

Conclusion

•A novel NR-IQA framework is proposed based onhigh-level semantic feature aggregation, whosesuperiority and generalization capability areverified on four popular image blur databases.

•High-level semantic information is experimentallyverified to be crucial in quality estimation amongvarious image content.

References

[1] P. Marziliano, F. Dufaux, S. Winkler, and T. Ebrahimi,“Perceptual blur and ringing metrics: Application toJPEG2000,” Signal Processing: Image Communication,vol. 19, no. 2, pp. 163–172, 2004.

[2] K. Gu, G. Zhai, W. Lin, X. Yang, and W. Zhang,“No-reference image sharpness assessment inautoregressive parameter space,” IEEE Transactions onImage Processing, vol. 24, no. 10, pp. 3218–3231, 2015.

[3] P. V. Vu and D. M. Chandler, “A fast wavelet-basedalgorithm for global and local image sharpnessestimation,” IEEE Signal Processing Letters, vol. 19,no. 7, pp. 423–426, 2012.

[4] R. Hassen, Z. Wang, and M. M. A. Salama, “Imagesharpness assessment based on local phase coherence,”IEEE Transactions on Image Processing, vol. 22, no. 7,pp. 2798–2810, 2013.

Contact Information

•Email: [email protected]