<?xml version="1.0" encoding="UTF-8"?>
<collection xmlns="http://www.loc.gov/MARC21/slim">
 <record>
  <leader>     naa a22        4500</leader>
  <controlfield tag="001">510729215</controlfield>
  <controlfield tag="003">CHVBK</controlfield>
  <controlfield tag="005">20180411082946.0</controlfield>
  <controlfield tag="007">cr unu---uuuuu</controlfield>
  <controlfield tag="008">180411e20130901xx      s     000 0 eng  </controlfield>
  <datafield tag="024" ind1="7" ind2="0">
   <subfield code="a">10.1007/s12200-013-0341-y</subfield>
   <subfield code="2">doi</subfield>
  </datafield>
  <datafield tag="035" ind1=" " ind2=" ">
   <subfield code="a">(NATIONALLICENCE)springer-10.1007/s12200-013-0341-y</subfield>
  </datafield>
  <datafield tag="245" ind1="0" ind2="0">
   <subfield code="a">Quality assessment method for geometrically distorted images</subfield>
   <subfield code="h">[Elektronische Daten]</subfield>
   <subfield code="c">[Binbing Liu, Ming Zhao, Haiqing Chen]</subfield>
  </datafield>
  <datafield tag="520" ind1="3" ind2=" ">
   <subfield code="a">The objective assessment of image quality is important for image processing, which has been paid much attention to in recent years. However, there were few reports about objective quality assessment methods for geometrically distorted images. Different from the routine image degradation processing (for example, noise addition, contrast change and lossy compression), the geometric distortion results in the changes of the spatial relationship of image pixels, which makes the traditional quality assessment algorithms, such as mean square error (MSE) and peak signal to noise ratio (PSNR) failure to obtain expected assessment results. In this paper, a full reference image quality assessment algorithm is proposed specifically for the quality evaluation of geometrically distorted images. This assessment algorithm takes into account three key factors, such as distortion intensity, distortion change rate and line feature index for perceptual quality assessment of images. Experimental results in this study show that the proposed assessment algorithm not only is significantly better than those of the traditional objective assessment methods such as PSNR and structural similarity index measurement (SSIM), but also has significant correlation with human subjective assessment.</subfield>
  </datafield>
  <datafield tag="540" ind1=" " ind2=" ">
   <subfield code="a">Higher Education Press and Springer-Verlag Berlin Heidelberg, 2013</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">image quality assessment</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">geometrical distortion</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">displacement vector field</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">line feature index</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">Hough transform</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="700" ind1="1" ind2=" ">
   <subfield code="a">Liu</subfield>
   <subfield code="D">Binbing</subfield>
   <subfield code="u">School of Optical and Electronics Information, Huazhong University of Science and Technology, 430074, Wuhan, China</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="700" ind1="1" ind2=" ">
   <subfield code="a">Zhao</subfield>
   <subfield code="D">Ming</subfield>
   <subfield code="u">School of Optical and Electronics Information, Huazhong University of Science and Technology, 430074, Wuhan, China</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="700" ind1="1" ind2=" ">
   <subfield code="a">Chen</subfield>
   <subfield code="D">Haiqing</subfield>
   <subfield code="u">School of Optical and Electronics Information, Huazhong University of Science and Technology, 430074, Wuhan, China</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="773" ind1="0" ind2=" ">
   <subfield code="t">Frontiers of Optoelectronics</subfield>
   <subfield code="d">Springer Berlin Heidelberg</subfield>
   <subfield code="g">6/3(2013-09-01), 275-281</subfield>
   <subfield code="x">2095-2759</subfield>
   <subfield code="q">6:3&lt;275</subfield>
   <subfield code="1">2013</subfield>
   <subfield code="2">6</subfield>
   <subfield code="o">12200</subfield>
  </datafield>
  <datafield tag="856" ind1="4" ind2="0">
   <subfield code="u">https://doi.org/10.1007/s12200-013-0341-y</subfield>
   <subfield code="q">text/html</subfield>
   <subfield code="z">Onlinezugriff via DOI</subfield>
  </datafield>
  <datafield tag="908" ind1=" " ind2=" ">
   <subfield code="D">1</subfield>
   <subfield code="a">research-article</subfield>
   <subfield code="2">jats</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">856</subfield>
   <subfield code="E">40</subfield>
   <subfield code="u">https://doi.org/10.1007/s12200-013-0341-y</subfield>
   <subfield code="q">text/html</subfield>
   <subfield code="z">Onlinezugriff via DOI</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">700</subfield>
   <subfield code="E">1-</subfield>
   <subfield code="a">Liu</subfield>
   <subfield code="D">Binbing</subfield>
   <subfield code="u">School of Optical and Electronics Information, Huazhong University of Science and Technology, 430074, Wuhan, China</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">700</subfield>
   <subfield code="E">1-</subfield>
   <subfield code="a">Zhao</subfield>
   <subfield code="D">Ming</subfield>
   <subfield code="u">School of Optical and Electronics Information, Huazhong University of Science and Technology, 430074, Wuhan, China</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">700</subfield>
   <subfield code="E">1-</subfield>
   <subfield code="a">Chen</subfield>
   <subfield code="D">Haiqing</subfield>
   <subfield code="u">School of Optical and Electronics Information, Huazhong University of Science and Technology, 430074, Wuhan, China</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">773</subfield>
   <subfield code="E">0-</subfield>
   <subfield code="t">Frontiers of Optoelectronics</subfield>
   <subfield code="d">Springer Berlin Heidelberg</subfield>
   <subfield code="g">6/3(2013-09-01), 275-281</subfield>
   <subfield code="x">2095-2759</subfield>
   <subfield code="q">6:3&lt;275</subfield>
   <subfield code="1">2013</subfield>
   <subfield code="2">6</subfield>
   <subfield code="o">12200</subfield>
  </datafield>
  <datafield tag="900" ind1=" " ind2="7">
   <subfield code="a">Metadata rights reserved</subfield>
   <subfield code="b">Springer special CC-BY-NC licence</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="898" ind1=" " ind2=" ">
   <subfield code="a">BK010053</subfield>
   <subfield code="b">XK010053</subfield>
   <subfield code="c">XK010000</subfield>
  </datafield>
  <datafield tag="949" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="F">NATIONALLICENCE</subfield>
   <subfield code="b">NL-springer</subfield>
  </datafield>
 </record>
</collection>
