<?xml version="1.0" encoding="UTF-8"?>
<collection xmlns="http://www.loc.gov/MARC21/slim">
 <record>
  <leader>     caa a22        4500</leader>
  <controlfield tag="001">605447055</controlfield>
  <controlfield tag="003">CHVBK</controlfield>
  <controlfield tag="005">20210128100130.0</controlfield>
  <controlfield tag="007">cr unu---uuuuu</controlfield>
  <controlfield tag="008">210128e20151101xx      s     000 0 eng  </controlfield>
  <datafield tag="024" ind1="7" ind2="0">
   <subfield code="a">10.1007/s11042-014-2079-y</subfield>
   <subfield code="2">doi</subfield>
  </datafield>
  <datafield tag="035" ind1=" " ind2=" ">
   <subfield code="a">(NATIONALLICENCE)springer-10.1007/s11042-014-2079-y</subfield>
  </datafield>
  <datafield tag="245" ind1="0" ind2="0">
   <subfield code="a">Noise-robust video super-resolution using an adaptive spatial-temporal filter</subfield>
   <subfield code="h">[Elektronische Daten]</subfield>
   <subfield code="c">[Jing Hu, Yupin Luo]</subfield>
  </datafield>
  <datafield tag="520" ind1="3" ind2=" ">
   <subfield code="a">In this paper, we introduce a new interpolation-based super-resolution scheme for super-resolving a low-resolution video that contains large-scale local motions and/or heavy noise. Our scheme leverages an efficient space-time descriptor to adapt the interpolation kernel to the video's spatial and temporal structures. Nevertheless, in the presence of large-scale local motions, the kernel suffers from tracking the motions incorrectly, leading to inaccurate temporal averaging. To address this problem, prior to computing the interpolation kernel, a mobile-neighborhood strategy that can identify the appropriate neighborhoods in adjacent frames is applied to neutralize the large-scale motions. Furthermore, we incorporate an adaptive sharpening technique into the kernel computation to remove the background noise and enhance the fine details simultaneously. Extensive experimental results on real-world videos show that the proposed method outperforms certain other state-of-the-art video super-resolution algorithms both visually and quantitatively, particularly in the presence of large-scale motions and/or heavy noise.</subfield>
  </datafield>
  <datafield tag="540" ind1=" " ind2=" ">
   <subfield code="a">Springer Science+Business Media New York, 2014</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">Video super-resolution</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">Interpolation-based</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">Mobile-neighborhood strategy</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">Adaptive sharpening</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="700" ind1="1" ind2=" ">
   <subfield code="a">Hu</subfield>
   <subfield code="D">Jing</subfield>
   <subfield code="u">Department of Automation, Tsinghua National Laboratory for Information Science and Technology (TNList), Tsinghua University, 100084, Beijing, China</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="700" ind1="1" ind2=" ">
   <subfield code="a">Luo</subfield>
   <subfield code="D">Yupin</subfield>
   <subfield code="u">Department of Automation, Tsinghua National Laboratory for Information Science and Technology (TNList), Tsinghua University, 100084, Beijing, China</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="773" ind1="0" ind2=" ">
   <subfield code="t">Multimedia Tools and Applications</subfield>
   <subfield code="d">Springer US; http://www.springer-ny.com</subfield>
   <subfield code="g">74/21(2015-11-01), 9259-9278</subfield>
   <subfield code="x">1380-7501</subfield>
   <subfield code="q">74:21&lt;9259</subfield>
   <subfield code="1">2015</subfield>
   <subfield code="2">74</subfield>
   <subfield code="o">11042</subfield>
  </datafield>
  <datafield tag="856" ind1="4" ind2="0">
   <subfield code="u">https://doi.org/10.1007/s11042-014-2079-y</subfield>
   <subfield code="q">text/html</subfield>
   <subfield code="z">Onlinezugriff via DOI</subfield>
  </datafield>
  <datafield tag="898" ind1=" " ind2=" ">
   <subfield code="a">BK010053</subfield>
   <subfield code="b">XK010053</subfield>
   <subfield code="c">XK010000</subfield>
  </datafield>
  <datafield tag="900" ind1=" " ind2="7">
   <subfield code="a">Metadata rights reserved</subfield>
   <subfield code="b">Springer special CC-BY-NC licence</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="908" ind1=" " ind2=" ">
   <subfield code="D">1</subfield>
   <subfield code="a">research-article</subfield>
   <subfield code="2">jats</subfield>
  </datafield>
  <datafield tag="949" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="F">NATIONALLICENCE</subfield>
   <subfield code="b">NL-springer</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">856</subfield>
   <subfield code="E">40</subfield>
   <subfield code="u">https://doi.org/10.1007/s11042-014-2079-y</subfield>
   <subfield code="q">text/html</subfield>
   <subfield code="z">Onlinezugriff via DOI</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">700</subfield>
   <subfield code="E">1-</subfield>
   <subfield code="a">Hu</subfield>
   <subfield code="D">Jing</subfield>
   <subfield code="u">Department of Automation, Tsinghua National Laboratory for Information Science and Technology (TNList), Tsinghua University, 100084, Beijing, China</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">700</subfield>
   <subfield code="E">1-</subfield>
   <subfield code="a">Luo</subfield>
   <subfield code="D">Yupin</subfield>
   <subfield code="u">Department of Automation, Tsinghua National Laboratory for Information Science and Technology (TNList), Tsinghua University, 100084, Beijing, China</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">773</subfield>
   <subfield code="E">0-</subfield>
   <subfield code="t">Multimedia Tools and Applications</subfield>
   <subfield code="d">Springer US; http://www.springer-ny.com</subfield>
   <subfield code="g">74/21(2015-11-01), 9259-9278</subfield>
   <subfield code="x">1380-7501</subfield>
   <subfield code="q">74:21&lt;9259</subfield>
   <subfield code="1">2015</subfield>
   <subfield code="2">74</subfield>
   <subfield code="o">11042</subfield>
  </datafield>
 </record>
</collection>
