Latex-Dateien der Masterarbeit
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

Masterarbeit_Navigation_Mobiler_Roboter_CV_File_BibLatex.bib 56KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742
  1. % Encoding: UTF-8
  2. % This file was created with Citavi 6.10.0.73
  3. @article{EncoderDecoder7803544,
  4. author={Badrinarayanan, Vijay and Kendall, Alex and Cipolla, Roberto},
  5. journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
  6. title={SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation},
  7. year={2017},
  8. volume={39},
  9. number={12},
  10. pages={2481-2495},
  11. doi={10.1109/TPAMI.2016.2644615}
  12. }
  13. @misc{Ranftl.3242021,
  14. author = {Ranftl, Ren{\'e} and Bochkovskiy, Alexey and Koltun, Vladlen},
  15. year = {2021},
  16. title = {Vision Transformers for Dense Prediction},
  17. archivePrefix = {arXiv},
  18. eprint = {arXiv:2103.13413v1},
  19. url = {https://arxiv.org/pdf/2103.13413},
  20. keywords = {Computer Science - Computer Vision and Pattern Recognition;Computer Vision and Pattern Recognition (cs.CV)},
  21. abstract = {We introduce dense vision transformers, an architecture that leverages vision transformers in place of convolutional networks as a backbone for dense prediction tasks. We assemble tokens from various stages of the vision transformer into image-like representations at various resolutions and progressively combine them into full-resolution predictions using a convolutional decoder. The transformer backbone processes representations at a constant and relatively high resolution and has a global receptive field at every stage. These properties allow the dense vision transformer to provide finer-grained and more globally coherent predictions when compared to fully-convolutional networks. Our experiments show that this architecture yields substantial improvements on dense prediction tasks, especially when a large amount of training data is available. For monocular depth estimation, we observe an improvement of up to 28{\%} in relative performance when compared to a state-of-the-art fully-convolutional network. When applied to semantic segmentation, dense vision transformers set a new state of the art on ADE20K with 49.02{\%} mIoU. We further show that the architecture can be fine-tuned on smaller datasets such as NYUv2, KITTI, and Pascal Context where it also sets the new state of the art. Our models are available at https://github.com/intel-isl/DPT.},
  22. pagetotal = {15},
  23. doi = {pages},
  24. note = {15 pages}
  25. }
  26. @online{NVIDIA.11262021,
  27. author = {NVIDIA},
  28. year = {2021},
  29. title = {NVIDIA Jetson Nano},
  30. url = {https://www.nvidia.com/de-de/autonomous-machines/embedded-systems/jetson-nano/product-development/},
  31. keywords = {AIoT;CUDA-X;Edge AI;Internet der Dinge;Jetson Nano;Jetson Nano Developer Kit;KI in der Peripherie;NVIDIA},
  32. urldate = {2021-12-12},
  33. abstract = {Das NVIDIA{\circledR} Jetson Nano{\texttrademark}-Entwicklerkit ist ein kleiner, leistungsstarker Computer, der trotz geringem Strombedarf moderne KI-Workloads stemmen kann. Statten Sie Millionen von Ger{\"a}ten in der Peripherie mit unglaublichen neuen Funktionen aus.}
  34. }
  35. @article{Sandler,
  36. author = {Sandler, Mark and Howard, Andrew and Zhu, Menglong and Zhmoginov, Andrey and Chen, Liang-Chieh},
  37. year = {2018},
  38. title = {MobileNetV2: Inverted Residuals and Linear Bottlenecks},
  39. archivePrefix = {arXiv},
  40. eprint = {arXiv:1801.04381v4},
  41. url = {https://arxiv.org/pdf/1801.04381},
  42. keywords = {Computer Vision and Pattern Recognition (cs.CV)},
  43. journaltitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
  44. abstract = {In this paper we describe a new mobile architecture, MobileNetV2, that improves the state of the art performance of mobile models on multiple tasks and benchmarks as well as across a spectrum of different model sizes. We also describe efficient ways of applying these mobile models to object detection in a novel framework we call SSDLite. Additionally, we demonstrate how to build mobile semantic segmentation models through a reduced form of DeepLabv3 which we call Mobile DeepLabv3.
  45. The MobileNetV2 architecture is based on an inverted residual structure where the input and output of the residual block are thin bottleneck layers opposite to traditional residual models which use expanded representations in the input an MobileNetV2 uses lightweight depthwise convolutions to filter features in the intermediate expansion layer. Additionally, we find that it is important to remove non-linearities in the narrow layers in order to maintain representational power. We demonstrate that this improves performance and provide an intuition that led to this design. Finally, our approach allows decoupling of the input/output domains from the expressiveness of the transformation, which provides a convenient framework for further analysis. We measure our performance on Imagenet classification, COCO object detection, VOC image segmentation. We evaluate the trade-offs between accuracy, and number of operations measured by multiply-adds (MAdd), as well as the number of parameters},
  46. file = {MobileNetV2 Inverted Residuals and Linear Bottlenecks:Attachments/MobileNetV2 Inverted Residuals and Linear Bottlenecks.pdf:application/pdf}
  47. }
  48. @online{FlatBuffer.852021,
  49. year = {2021},
  50. title = {FlatBuffers: FlatBuffers white paper},
  51. url = {https://google.github.io/flatbuffers/flatbuffers_white_paper.html},
  52. urldate = {2021-11-15},
  53. abstract = {}
  54. }
  55. @online{Colab.1072021,
  56. year = {2021},
  57. title = {Google Colaboratory},
  58. url = {https://colab.research.google.com},
  59. urldate = {2021-10-07},
  60. abstract = {}
  61. }
  62. @online{coco.8252021,
  63. year = {2021},
  64. title = {COCO - Common Objects in Context},
  65. url = {https://cocodataset.org},
  66. urldate = {2021-11-14},
  67. abstract = {}
  68. }
  69. @online{cocoEval.8252021,
  70. year = {2021},
  71. title = {COCO - Common Objects in Context},
  72. url = {https://cocodataset.org/#detection-eval},
  73. urldate = {2021-11-14},
  74. abstract = {}
  75. }
  76. @misc{Aleotti.2020,
  77. author = {Aleotti, Filippo and Zaccaroni, Giulio and Bartolomei, Luca and Poggi, Matteo and Tosi, Fabio and Mattoccia, Stefano},
  78. year = {2020},
  79. title = {Real-time single image depth perception in the wild with handheld devices},
  80. archivePrefix = {arXiv},
  81. eprint = {arXiv:2006.05724v1},
  82. url = {https://arxiv.org/pdf/2006.05724},
  83. keywords = {Computer Vision and Pattern Recognition (cs.CV);Graphics (cs.GR)},
  84. abstract = {Depth perception is paramount to tackle real-world problems, ranging from autonomous driving to consumer applications. For the latter, depth estimation from a single image represents the most versatile solution, since a standard camera is available on almost any handheld device. Nonetheless, two main issues limit its practical deployment: i) the low reliability when deployed in-the-wild and ii) the demanding resource requirements to achieve real-time performance, often not compatible with such devices. Therefore, in this paper, we deeply investigate these issues showing how they are both addressable adopting appropriate network design and training strategies -- also outlining how to map the resulting networks on handheld devices to achieve real-time performance. Our thorough evaluation highlights the ability of such fast networks to generalize well to new environments, a crucial feature required to tackle the extremely varied contexts faced in real applications. Indeed, to further support this evidence, we report experimental results concerning real-time depth-aware augmented reality and image blurring with smartphones in-the-wild.},
  85. file = {Real-time single image depth perception in the wild with:Attachments/Real-time single image depth perception in the wild with.pdf:application/pdf}
  86. }
  87. @article{Arulprakash.2021,
  88. author = {Arulprakash, Enoch and Aruldoss, Martin},
  89. year = {2021},
  90. title = {A study on generic object detection with emphasis on future research directions},
  91. issn = {13191578},
  92. journaltitle = {Journal of King Saud University - Computer and Information Sciences},
  93. doi = {10.1016/j.jksuci.2021.08.001},
  94. abstract = {},
  95. note = {PII: S1319157821002020}
  96. }
  97. @article{Everingham.2010,
  98. author = {Everingham, Mark and {van Gool}, Luc and Williams, Christopher K. I. and Winn, John and Zisserman, Andrew},
  99. year = {2010},
  100. title = {The Pascal Visual Object Classes (VOC) Challenge},
  101. pages = {303--338},
  102. pagination = {page},
  103. volume = {88},
  104. issn = {0920-5691},
  105. journaltitle = {International Journal of Computer Vision},
  106. shortjournal = {Int J Comput Vis},
  107. doi = {10.1007/s11263-009-0275-4},
  108. number = {2},
  109. abstract = {},
  110. note = {PII: 275}
  111. }
  112. @article{Everingham.2015,
  113. author = {Everingham, Mark and Eslami, S. M. Ali and {van Gool}, Luc and Williams, Christopher K. I. and Winn, John and Zisserman, Andrew},
  114. year = {2015},
  115. title = {The Pascal Visual Object Classes Challenge: A Retrospective},
  116. pages = {98--136},
  117. pagination = {page},
  118. volume = {111},
  119. issn = {0920-5691},
  120. journaltitle = {International Journal of Computer Vision},
  121. shortjournal = {Int J Comput Vis},
  122. doi = {10.1007/s11263-014-0733-5},
  123. number = {1},
  124. abstract = {},
  125. note = {PII: 733}
  126. }
  127. @misc{EveringhamDevelopmentKit.2012,
  128. title = {The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Development Kit},
  129. author = {Everingham, Mark and Winn, John},
  130. year = {2012},
  131. url = {http://host.robots.ox.ac.uk/pascal/VOC/voc2012/index.html},
  132. urldate = {2021-11-24},
  133. }
  134. @article{Hosang.2016,
  135. title={What Makes for Effective Detection Proposals?},
  136. volume={38},
  137. ISSN={2160-9292},
  138. url={http://dx.doi.org/10.1109/TPAMI.2015.2465908},
  139. DOI={10.1109/tpami.2015.2465908},
  140. number={4},
  141. journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
  142. publisher={Institute of Electrical and Electronics Engineers (IEEE)},
  143. author={Hosang, Jan and Benenson, Rodrigo and Dollar, Piotr and Schiele, Bernt},
  144. year={2016},
  145. month={4},
  146. pages={814–830}
  147. }
  148. @misc{Girshick.2013,
  149. author = {Girshick, Ross and Donahue, Jeff and Darrell, Trevor and Malik, Jitendra},
  150. year = {2013},
  151. title = {Rich feature hierarchies for accurate object detection and semantic segmentation},
  152. archivePrefix = {arXiv},
  153. eprint = {arXiv:1311.2524v5},
  154. url = {https://arxiv.org/pdf/1311.2524},
  155. keywords = {Computer Vision and Pattern Recognition (cs.CV)},
  156. abstract = {Object detection performance, as measured on the canonical PASCAL VOC dataset, has plateaued in the last few years. The best-performing methods are complex ensemble systems that typically combine multiple low-level image features with high-level context. In this paper, we propose a simple and scalable detection algorithm that improves mean average precision (mAP) by more than 30{\%} relative to the previous best result on VOC 2012---achieving a mAP of 53.3{\%}. Our approach combines two key insights: (1) one can apply high-capacity convolutional neural networks (CNNs) to bottom-up region proposals in order to localize and segment objects and (2) when labeled training data is scarce, supervised pre-training for an auxiliary task, followed by domain-specific fine-tuning, yields a significant performance boost. Since we combine region proposals with CNNs, we call our method R-CNN: Regions with CNN features. We also compare R-CNN to OverFeat, a recently proposed sliding-window detector based on a similar CNN architecture. We find that R-CNN outperforms OverFeat by a large margin on the 200-class ILSVRC2013 detection dataset. Source code for the complete system is available at this http URL.},
  157. file = {Rich feature hierarchies for accurate object detection a:Attachments/Rich feature hierarchies for accurate object detection a.pdf:application/pdf}
  158. }
  159. @online{.332021,
  160. year = {2021},
  161. title = {MNIST handwritten digit database, Yann LeCun, Corinna Cortes and Chris Burges},
  162. url = {http://yann.lecun.com/exdb/mnist/},
  163. urldate = {2021-10-20},
  164. abstract = {}
  165. }
  166. @online{Tensorflow.8202021,
  167. author = {TensorFlow},
  168. year = {2021},
  169. title = {TensorFlow 2 Schnellstart f{\"u}r Anf{\"a}nger ~|~ TensorFlow Core},
  170. url = {https://www.tensorflow.org/tutorials/quickstart/beginner?hl=de},
  171. urldate = {2021-10-20},
  172. abstract = {}
  173. }
  174. @online{GitHub.1062021,
  175. author = {GitHub},
  176. year = {2021},
  177. title = {models/research/object{\_}detection at master · tensorflow/models},
  178. url = {https://github.com/tensorflow/models/tree/master/research/object_detection},
  179. urldate = {2021-10-06},
  180. abstract = {Models and examples built with TensorFlow. Contribute to tensorflow/models development by creating an account on GitHub.}
  181. }
  182. @online{GitHub.1062021b,
  183. author = {GitHub},
  184. year = {2021},
  185. title = {GitHub - isl-org/MiDaS: Code for robust monocular depth estimation described in Ranftl et. al., Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer, TPAMI 2020},
  186. url = {https://github.com/isl-org/MiDaS},
  187. urldate = {2021-10-06},
  188. abstract = {Code for robust monocular depth estimation described in Ranftl et. al., Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer, TPAMI 2020 - GitHub - isl-org/MiDaS: Code for robust monocular depth estimation described in Ranftl et. al., Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer, TPAMI 2020}
  189. }
  190. @online{GitHub.1062021c,
  191. author = {GitHub},
  192. year = {2021},
  193. title = {GitHub - FilippoAleotti/mobilePydnet: Pydnet on mobile devices},
  194. url = {https://github.com/FilippoAleotti/mobilePydnet},
  195. urldate = {2021-10-06},
  196. abstract = {Pydnet on mobile devices. Contribute to FilippoAleotti/mobilePydnet development by creating an account on GitHub.}
  197. }
  198. @online{GitHubDetectionZoo.11152021,
  199. author = {GitHub},
  200. year = {2021},
  201. title = {models/tf2{\_}detection{\_}zoo.md at master · tensorflow/models},
  202. url = {https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf2_detection_zoo.md},
  203. urldate = {2021-11-15},
  204. abstract = {Models and examples built with TensorFlow. Contribute to tensorflow/models development by creating an account on GitHub.}
  205. }
  206. @online{GitHubTensorFlowAndroidExample.11222021,
  207. author = {GitHub},
  208. year = {2021},
  209. title = {examples/lite/examples/image{\_}classification/android at master · tensorflow/examples},
  210. url = {https://github.com/tensorflow/examples/tree/master/lite/examples/object_detection/android},
  211. urldate = {2021-11-22},
  212. abstract = {TensorFlow examples. Contribute to tensorflow/examples development by creating an account on GitHub.}
  213. }
  214. @online{GitHub.1062021d,
  215. author = {GitHub},
  216. year = {2021},
  217. title = {GitHub - tzutalin/labelImg: LabelImg is a graphical image annotation tool and label object bounding boxes in images},
  218. url = {https://github.com/tzutalin/labelImg},
  219. urldate = {2021-10-06},
  220. abstract = {LabelImg is a graphical image annotation tool and label object bounding boxes in images - GitHub - tzutalin/labelImg: LabelImg is a graphical image annotation tool and label object bounding boxes in images}
  221. }
  222. @online{GitHub.1282021,
  223. author = {GitHub},
  224. year = {2021},
  225. title = {GitHub - Genymobile/scrcpy: Display and control your Android device},
  226. url = {https://github.com/Genymobile/scrcpy},
  227. urldate = {2021-12-08},
  228. abstract = {Display and control your Android device. Contribute to Genymobile/scrcpy development by creating an account on GitHub.}
  229. }
  230. @online{GitHub.1282021b,
  231. author = {GitHub},
  232. year = {2021},
  233. title = {GitHub - lutzroeder/netron: Visualizer for neural network, deep learning, and machine learning models},
  234. url = {https://github.com/lutzroeder/netron},
  235. urldate = {2021-12-08},
  236. abstract = {Visualizer for neural network, deep learning, and machine learning models - GitHub - lutzroeder/netron: Visualizer for neural network, deep learning, and machine learning models}
  237. }
  238. @online{GitHub.12122021,
  239. author = {GitHub},
  240. year = {2021},
  241. title = {models/research/object{\_}detection at master · tensorflow/models},
  242. url = {https://github.com/tensorflow/models/blob/master/research/object_detection/protos/preprocessor.proto},
  243. urldate = {2021-12-12},
  244. abstract = {Models and examples built with TensorFlow. Contribute to tensorflow/models development by creating an account on GitHub.}
  245. }
  246. @book{Goodfellow.2018,
  247. author = {Goodfellow, Ian and Bengio, Yoshua and Courville, Aaron},
  248. year = {2018},
  249. title = {Deep Learning},
  250. url = {https://ebookcentral.proquest.com/lib/kxp/detail.action?docID=5598176},
  251. keywords = {Electronic books;Machine learning},
  252. edition = {1st edition},
  253. publisher = {mitp},
  254. isbn = {978-3-95845-700-3},
  255. subtitle = {Das umfassende Handbuch : Grundlagen, aktuelle Verfahren und Algorithmen, neue Forschungsans{\"a}tze},
  256. language = {ger},
  257. location = {Frechen},
  258. series = {mitp Professional},
  259. abstract = {Intro -- Impressum -- Website zum Buch -- Danksagung -- {\"U}ber die Fachkorrektoren zur deutschen Ausgabe -- Notation -- Einleitung -- F{\"u}r wen ist dieses Buch gedacht? -- Historische Entwicklungen im Deep Learning -- I Angewandte Mathematik und Grundlagen f{\"u}r das Machine Learning -- Lineare Algebra -- Skalare, Vektoren, Matrizen und Tensoren -- Multiplizieren von Matrizen und Vektoren -- Einheits- und Umkehrmatrizen -- Lineare Abh{\"a}ngigkeit und lineare H{\"u}lle -- Normen -- Spezielle Matrizen und Vektoren -- Eigenwertzerlegung -- Singul{\"a}rwertzerlegung -- Die Moore-Penrose-Pseudoinverse -- Der Spuroperator -- Die Determinante -- Beispiel: Hauptkomponentenanalyse -- Wahrscheinlichkeits- und Informationstheorie -- Warum Wahrscheinlichkeit? -- Zufallsvariablen -- Wahrscheinlichkeitsverteilungen -- Randwahrscheinlichkeit -- Bedingte Wahrscheinlichkeit -- Die Produktregel der bedingten Wahrscheinlichkeiten -- Unabh{\"a}ngigkeit und bedingte Unabh{\"a}ngigkeit -- Erwartungswert, Varianz und Kovarianz -- H{\"a}ufig genutzte Wahrscheinlichkeitsverteilungen -- N{\"u}tzliche Eigenschaften h{\"a}ufig verwendeter Funktionen -- Satz von Bayes -- Technische Einzelheiten stetiger Variablen -- Informationstheorie -- Strukturierte probabilistische Modelle -- Numerische Berechnung -- {\"U}berlauf und Unterlauf -- Schlechte Konditionierung -- Optimierung auf Gradientenbasis -- Optimierung unter Nebenbedingungen -- Beispiel: Lineare kleinste Quadrate -- Grundlagen f{\"u}r das Machine Learning -- Lernalgorithmen -- Kapazit{\"a}t, {\"U}beranpassung und Unteranpassung -- Hyperparameter und Validierungsdaten -- Sch{\"a}tzer, Verzerrung und Varianz -- Maximum-Likelihood-Sch{\"a}tzung -- Bayessche Statistik -- Algorithmen f{\"u}r {\"u}berwachtes Lernen -- Algorithmen f{\"u}r un{\"u}berwachtes Lernen -- Stochastisches Gradientenabstiegsverfahren -- Entwickeln eines Machine-Learning-Algorithmus -- Probleme, an denen Deep Learning w{\"a}chst.},
  260. pagetotal = {883},
  261. note = {Goodfellow, Ian (VerfasserIn)
  262. Bengio, Yoshua (VerfasserIn)
  263. Courville, Aaron (VerfasserIn)
  264. Lenz, Guido ({\"U}bersetzerIn)}
  265. }
  266. @book{Grus.2020,
  267. author = {Grus, Joel},
  268. year = {2020},
  269. title = {Einf{\"u}hrung in Data Science},
  270. url = {https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=2298669},
  271. edition = {2. Auflage},
  272. publisher = {O'Reilly and dpunkt.verlag},
  273. isbn = {978-3-96009-123-3},
  274. subtitle = {Grundprinzipien der Datenanalyse mit Python},
  275. language = {ger},
  276. location = {Heidelberg},
  277. abstract = {},
  278. note = {Grus, Joel (VerfasserIn)
  279. Rother, Kristian ({\"U}bersetzerIn)
  280. Demmig, Thomas ({\"U}bersetzerIn)}
  281. }
  282. @misc{He.2015,
  283. author = {He, Kaiming and Zhang, Xiangyu and Ren, Shaoqing and Sun, Jian},
  284. year = {2015},
  285. title = {Deep Residual Learning for Image Recognition},
  286. archivePrefix = {arXiv},
  287. eprint = {arXiv:1512.03385v1},
  288. url = {https://arxiv.org/pdf/1512.03385},
  289. keywords = {Computer Vision and Pattern Recognition (cs.CV)},
  290. abstract = {Deeper neural networks are more difficult to train. We present a residual learning framework to ease the training of networks that are substantially deeper than those used previously. We explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. We provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset we evaluate residual nets with a depth of up to 152 layers---8x deeper than VGG nets but still having lower complexity. An ensemble of these residual nets achieves 3.57{\%} error on the ImageNet test set. This result won the 1st place on the ILSVRC 2015 classification task. We also present analysis on CIFAR-10 with 100 and 1000 layers.
  291. The depth of representations is of central importance for many visual recognition tasks. Solely due to our extremely deep representations, we obtain a 28{\%} relative improvement on the COCO object detection dataset. Deep residual nets are foundations of our submissions to ILSVRC {\&} COCO 2015 competitions, where we also won the 1st places on the tasks of ImageNet detection, ImageNet localization, COCO detection, and COCO segmentation.},
  292. file = {Deep Residual Learning for Image Recognition:Attachments/Deep Residual Learning for Image Recognition.pdf:application/pdf}
  293. }
  294. @misc{HenriqueF.DeArruda.2019,
  295. author = {{Henrique F. De Arruda} and {Alexandre Benatti} and {C{\'e}sar Henrique Comin} and {Luciano Da F. Costa}},
  296. year = {2019},
  297. title = {Learning Deep Learning (CDT-15)},
  298. publisher = {Unpublished},
  299. abstract = {},
  300. doi = {10.13140/RG.2.2.29866.57283},
  301. language = {en}
  302. }
  303. @misc{Hosang.2017,
  304. author = {Hosang, Jan and Benenson, Rodrigo and Schiele, Bernt},
  305. year = {2017},
  306. title = {Learning non-maximum suppression},
  307. archivePrefix = {arXiv},
  308. eprint = {arXiv:1705.02950v2},
  309. url = {https://arxiv.org/pdf/1705.02950},
  310. keywords = {Computer Vision and Pattern Recognition (cs.CV)},
  311. abstract = {Object detectors have hugely profited from moving towards an end-to-end learning paradigm: proposals, features, and the classifier becoming one neural network improved results two-fold on general object detection. One indispensable component is non-maximum suppression (NMS), a post-processing algorithm responsible for merging all detections that belong to the same object. The de facto standard NMS algorithm is still fully hand-crafted, suspiciously simple, and -- being based on greedy clustering with a fixed distance threshold -- forces a trade-off between recall and precision. We propose a new network architecture designed to perform NMS, using only boxes and their score. We report experiments for person detection on PETS and for general object categories on the COCO dataset. Our approach shows promise providing improved localization and occlusion handling.},
  312. file = {Learning non-maximum suppression:Attachments/Learning non-maximum suppression.pdf:application/pdf}
  313. }
  314. @misc{Jacob.12152017,
  315. author = {Jacob, Benoit and Kligys, Skirmantas and Chen, Bo and Zhu, Menglong and Tang, Matthew and Howard, Andrew and Adam, Hartwig and Kalenichenko, Dmitry},
  316. year = {2017},
  317. title = {Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference},
  318. archivePrefix = {arXiv},
  319. eprint = {arXiv:1712.05877v1},
  320. url = {https://arxiv.org/pdf/1712.05877},
  321. keywords = {Computer Science - Learning;Machine Learning (cs.LG);Machine Learning (stat.ML);Statistics - Machine Learning},
  322. abstract = {The rising popularity of intelligent mobile devices and the daunting computational cost of deep learning-based models call for efficient and accurate on-device inference schemes. We propose a quantization scheme that allows inference to be carried out using integer-only arithmetic, which can be implemented more efficiently than floating point inference on commonly available integer-only hardware. We also co-design a training procedure to preserve end-to-end model accuracy post quantization. As a result, the proposed quantization scheme improves the tradeoff between accuracy and on-device latency. The improvements are significant even on MobileNets, a model family known for run-time efficiency, and are demonstrated in ImageNet classification and COCO detection on popular CPUs.},
  323. pagetotal = {14},
  324. note = {14 pages, 12 figures}
  325. }
  326. @misc{Howard.2017,
  327. author = {Howard, Andrew G. and Zhu, Menglong and Chen, Bo and Kalenichenko, Dmitry and Wang, Weijun and Weyand, Tobias and Andreetto, Marco and Adam, Hartwig},
  328. year = {2017},
  329. title = {MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications},
  330. archivePrefix = {arXiv},
  331. eprint = {arXiv:1704.04861v1},
  332. url = {https://arxiv.org/pdf/1704.04861},
  333. keywords = {Computer Vision and Pattern Recognition (cs.CV)},
  334. abstract = {We present a class of efficient models called MobileNets for mobile and embedded vision applications. MobileNets are based on a streamlined architecture that uses depth-wise separable convolutions to build light weight deep neural networks. We introduce two simple global hyper-parameters that efficiently trade off between latency and accuracy. These hyper-parameters allow the model builder to choose the right sized model for their application based on the constraints of the problem. We present extensive experiments on resource and accuracy tradeoffs and show strong performance compared to other popular models on ImageNet classification. We then demonstrate the effectiveness of MobileNets across a wide range of applications and use cases including object detection, finegrain classification, face attributes and large scale geo-localization.},
  335. file = {MobileNets Efficient Convolutional Neural Networks for M:Attachments/MobileNets Efficient Convolutional Neural Networks for M.pdf:application/pdf}
  336. }
  337. @misc{Ioffe.2015,
  338. author = {Ioffe, Sergey and Szegedy, Christian},
  339. year = {2015},
  340. title = {Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift},
  341. archivePrefix = {arXiv},
  342. eprint = {arXiv:1502.03167v3},
  343. url = {https://arxiv.org/pdf/1502.03167},
  344. keywords = {Machine Learning (cs.LG)},
  345. abstract = {Training Deep Neural Networks is complicated by the fact that the distribution of each layer's inputs changes during training, as the parameters of the previous layers change. This slows down the training by requiring lower learning rates and careful parameter initialization, and makes it notoriously hard to train models with saturating nonlinearities. We refer to this phenomenon as internal covariate shift, and address the problem by normalizing layer inputs. Our method draws its strength from making normalization a part of the model architecture and performing the normalization for each training mini-batch. Batch Normalization allows us to use much higher learning rates and be less careful about initialization. It also acts as a regularizer, in some cases eliminating the need for Dropout. Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin. Using an ensemble of batch-normalized networks, we improve upon the best published result on ImageNet classification: reaching 4.9{\%} top-5 validation error (and 4.8{\%} test error), exceeding the accuracy of human raters.}
  346. }
  347. @misc{Krishnamoorthi.2018,
  348. author = {Krishnamoorthi, Raghuraman},
  349. year = {2018},
  350. title = {Quantizing deep convolutional networks for efficient inference: A whitepaper},
  351. archivePrefix = {arXiv},
  352. eprint = {arXiv:1806.08342v1},
  353. url = {https://arxiv.org/pdf/1806.08342},
  354. keywords = {Computer Vision and Pattern Recognition (cs.CV);Machine Learning (cs.LG);Machine Learning (stat.ML)},
  355. abstract = {We present an overview of techniques for quantizing convolutional neural networks for inference with integer weights and activations. Per-channel quantization of weights and per-layer quantization of activations to 8-bits of precision post-training produces classification accuracies within 2{\%} of floating point networks for a wide variety of CNN architectures. Model sizes can be reduced by a factor of 4 by quantizing weights to 8-bits, even when 8-bit arithmetic is not supported. This can be achieved with simple, post training quantization of weights.We benchmark latencies of quantized networks on CPUs and DSPs and observe a speedup of 2x-3x for quantized implementations compared to floating point on CPUs. Speedups of up to 10x are observed on specialized processors with fixed point SIMD capabilities, like the Qualcomm QDSPs with HVX.
  356. Quantization-aware training can provide further improvements, reducing the gap to floating point to 1{\%} at 8-bit precision. Quantization-aware training also allows for reducing the precision of weights to four bits with accuracy losses ranging from 2{\%} to 10{\%}, with higher accuracy drop for smaller networks.We introduce tools in TensorFlow and TensorFlowLite for quantizing convolutional networks and review best practices for quantization-aware training to obtain high accuracy with quantized weights and activations. We recommend that per-channel quantization of weights and per-layer quantization of activations be the preferred quantization scheme for hardware acceleration and kernel optimization. We also propose that future processors and hardware accelerators for optimized inference support precisions of 4, 8 and 16 bits.},
  357. file = {Quantizing deep convolutional networks for efficient inf:Attachments/Quantizing deep convolutional networks for efficient inf.pdf:application/pdf}
  358. }
  359. @misc{Lin.1292016,
  360. author = {Lin, Tsung-Yi and Doll{\'a}r, Piotr and Girshick, Ross and He, Kaiming and Hariharan, Bharath and Belongie, Serge},
  361. year = {2016},
  362. title = {Feature Pyramid Networks for Object Detection},
  363. archivePrefix = {arXiv},
  364. eprint = {arXiv:1612.03144v2},
  365. url = {https://arxiv.org/pdf/1612.03144},
  366. keywords = {Computer Science - Computer Vision and Pattern Recognition;Computer Vision and Pattern Recognition (cs.CV)},
  367. abstract = {Feature pyramids are a basic component in recognition systems for detecting objects at different scales. But recent deep learning object detectors have avoided pyramid representations, in part because they are compute and memory intensive. In this paper, we exploit the inherent multi-scale, pyramidal hierarchy of deep convolutional networks to construct feature pyramids with marginal extra cost. A top-down architecture with lateral connections is developed for building high-level semantic feature maps at all scales. This architecture, called a Feature Pyramid Network (FPN), shows significant improvement as a generic feature extractor in several applications. Using FPN in a basic Faster R-CNN system, our method achieves state-of-the-art single-model results on the COCO detection benchmark without bells and whistles, surpassing all existing single-model entries including those from the COCO 2016 challenge winners. In addition, our method can run at 5 FPS on a GPU and thus is a practical and accurate solution to multi-scale object detection. Code will be made publicly available.},
  368. file = {Feature Pyramid Networks for Object Detection:Attachments/Feature Pyramid Networks for Object Detection.pdf:application/pdf}
  369. }
  370. @misc{Liu.2015,
  371. author = {Liu, Wei and Anguelov, Dragomir and Erhan, Dumitru and Szegedy, Christian and Reed, Scott and Fu, Cheng-Yang and Berg, Alexander C.},
  372. year = {2015},
  373. title = {SSD: Single Shot MultiBox Detector},
  374. archivePrefix = {arXiv},
  375. eprint = {arXiv:1512.02325v5},
  376. url = {https://arxiv.org/pdf/1512.02325},
  377. keywords = {Computer Vision and Pattern Recognition (cs.CV)},
  378. abstract = {We present a method for detecting objects in images using a single deep neural network. Our approach, named SSD, discretizes the output space of bounding boxes into a set of default boxes over different aspect ratios and scales per feature map location. At prediction time, the network generates scores for the presence of each object category in each default box and produces adjustments to the box to better match the object shape. Additionally, the network combines predictions from multiple feature maps with different resolutions to naturally handle objects of various sizes. Our SSD model is simple relative to methods that require object proposals because it completely eliminates proposal generation and subsequent pixel or feature resampling stage and encapsulates all computation in a single network. This makes SSD easy to train and straightforward to integrate into systems that require a detection component. Experimental results on the PASCAL VOC, MS COCO, and ILSVRC datasets confirm that SSD has comparable accuracy to methods that utilize an additional object proposal step and is much faster, while providing a unified framework for both training and inference. Compared to other single stage methods, SSD has much better accuracy, even with a smaller input image size. For
  379. 300$\times$300
  380. input, SSD achieves 72.1{\%} mAP on VOC2007 test at 58 FPS on a Nvidia Titan X and for
  381. 500$\times$500
  382. input, SSD achieves 75.1{\%} mAP, outperforming a comparable state of the art Faster R-CNN model. Code is available at this https URL .},
  383. file = {SSD Single Shot MultiBox Detector:Attachments/SSD Single Shot MultiBox Detector.pdf:application/pdf}
  384. }
  385. @misc{Poggi.2018,
  386. author = {Poggi, Matteo and Aleotti, Filippo and Tosi, Fabio and Mattoccia, Stefano},
  387. year = {2018},
  388. title = {Towards real-time unsupervised monocular depth estimation on CPU},
  389. archivePrefix = {arXiv},
  390. eprint = {arXiv:1806.11430v3},
  391. url = {https://arxiv.org/pdf/1806.11430},
  392. keywords = {Computer Vision and Pattern Recognition (cs.CV);Robotics (cs.RO)},
  393. abstract = {Unsupervised depth estimation from a single image is a very attractive technique with several implications in robotic, autonomous navigation, augmented reality and so on. This topic represents a very challenging task and the advent of deep learning enabled to tackle this problem with excellent results. However, these architectures are extremely deep and complex. Thus, real-time performance can be achieved only by leveraging power-hungry GPUs that do not allow to infer depth maps in application fields characterized by low-power constraints. To tackle this issue, in this paper we propose a novel architecture capable to quickly infer an accurate depth map on a CPU, even of an embedded system, using a pyramid of features extracted from a single input image. Similarly to state-of-the-art, we train our network in an unsupervised manner casting depth estimation as an image reconstruction problem. Extensive experimental results on the KITTI dataset show that compared to the top performing approach our network has similar accuracy but a much lower complexity (about 6{\%} of parameters) enabling to infer a depth map for a KITTI image in about 1.7 s on the Raspberry Pi 3 and at more than 8 Hz on a standard CPU. Moreover, by trading accuracy for efficiency, our network allows to infer maps at about 2 Hz and 40 Hz respectively, still being more accurate than most state-of-the-art slower methods. To the best of our knowledge, it is the first method enabling such performance on CPUs paving the way for effective deployment of unsupervised monocular depth estimation even on embedded systems.},
  394. file = {Towards real-time unsupervised monocular depth estimatio:Attachments/Towards real-time unsupervised monocular depth estimatio.pdf:application/pdf}
  395. }
  396. @misc{Ranftl.2019,
  397. author = {Ranftl, Ren{\'e} and Lasinger, Katrin and Hafner, David and Schindler, Konrad and Koltun, Vladlen},
  398. year = {2019},
  399. title = {Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer},
  400. archivePrefix = {arXiv},
  401. eprint = {arXiv:1907.01341v3},
  402. url = {https://arxiv.org/pdf/1907.01341},
  403. keywords = {Computer Vision and Pattern Recognition (cs.CV)},
  404. abstract = {The success of monocular depth estimation relies on large and diverse training sets. Due to the challenges associated with acquiring dense ground-truth depth across different environments at scale, a number of datasets with distinct characteristics and biases have emerged. We develop tools that enable mixing multiple datasets during training, even if their annotations are incompatible. In particular, we propose a robust training objective that is invariant to changes in depth range and scale, advocate the use of principled multi-objective learning to combine data from different sources, and highlight the importance of pretraining encoders on auxiliary tasks. Armed with these tools, we experiment with five diverse training datasets, including a new, massive data source: 3D films. To demonstrate the generalization power of our approach we use zero-shot cross-dataset transfer, i.e. we evaluate on datasets that were not seen during training. The experiments confirm that mixing data from complementary sources greatly improves monocular depth estimation. Our approach clearly outperforms competing methods across diverse datasets, setting a new state of the art for monocular depth estimation. Some results are shown in the supplementary video at this https URL},
  405. file = {Towards Robust Monocular Depth Estimation Mixing Dataset:Attachments/Towards Robust Monocular Depth Estimation Mixing Dataset.pdf:application/pdf}
  406. }
  407. @book{Raschka.2018,
  408. author = {Raschka, Sebastian},
  409. year = {2018},
  410. title = {Machine Learning mit Python und Scikit-learn und TensorFlow},
  411. url = {http://www.content-select.com/index.php?id=bib_view&ean=9783958457348},
  412. edition = {2., aktualisierte und erweiterte Auflage},
  413. publisher = {mitp},
  414. isbn = {978-3-95845-733-1},
  415. subtitle = {Das umfassende Praxis-Handbuch f{\"u}r Data Science, Deep Learning und Predictive Analytics},
  416. language = {ger},
  417. location = {Frechen},
  418. series = {mitp Professional},
  419. abstract = {},
  420. pagetotal = {577}
  421. }
  422. @misc{Redmon.2015,
  423. author = {Redmon, Joseph and Divvala, Santosh and Girshick, Ross and Farhadi, Ali},
  424. year = {2015},
  425. title = {You Only Look Once: Unified, Real-Time Object Detection},
  426. archivePrefix = {arXiv},
  427. eprint = {arXiv:1506.02640v5},
  428. url = {https://arxiv.org/pdf/1506.02640},
  429. keywords = {Computer Vision and Pattern Recognition (cs.CV)},
  430. abstract = {We present YOLO, a new approach to object detection. Prior work on object detection repurposes classifiers to perform detection. Instead, we frame object detection as a regression problem to spatially separated bounding boxes and associated class probabilities. A single neural network predicts bounding boxes and class probabilities directly from full images in one evaluation. Since the whole detection pipeline is a single network, it can be optimized end-to-end directly on detection performance.
  431. Our unified architecture is extremely fast. Our base YOLO model processes images in real-time at 45 frames per second. A smaller version of the network, Fast YOLO, processes an astounding 155 frames per second while still achieving double the mAP of other real-time detectors. Compared to state-of-the-art detection systems, YOLO makes more localization errors but is far less likely to predict false detections where nothing exists. Finally, YOLO learns very general representations of objects. It outperforms all other detection methods, including DPM and R-CNN, by a wide margin when generalizing from natural images to artwork on both the Picasso Dataset and the People-Art Dataset.},
  432. file = {You Only Look Once Unified Real-Time Object Detection:Attachments/You Only Look Once Unified Real-Time Object Detection.pdf:application/pdf}
  433. }
  434. @misc{Ren.2015,
  435. author = {Ren, Shaoqing and He, Kaiming and Girshick, Ross and Sun, Jian},
  436. year = {2015},
  437. title = {Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks},
  438. archivePrefix = {arXiv},
  439. eprint = {arXiv:1506.01497v3},
  440. url = {https://arxiv.org/pdf/1506.01497},
  441. keywords = {Computer Vision and Pattern Recognition (cs.CV)},
  442. abstract = {State-of-the-art object detection networks depend on region proposal algorithms to hypothesize object locations. Advances like SPPnet and Fast R-CNN have reduced the running time of these detection networks, exposing region proposal computation as a bottleneck. In this work, we introduce a Region Proposal Network (RPN) that shares full-image convolutional features with the detection network, thus enabling nearly cost-free region proposals. An RPN is a fully convolutional network that simultaneously predicts object bounds and objectness scores at each position. The RPN is trained end-to-end to generate high-quality region proposals, which are used by Fast R-CNN for detection. We further merge RPN and Fast R-CNN into a single network by sharing their convolutional features---using the recently popular terminology of neural networks with 'attention' mechanisms, the RPN component tells the unified network where to look. For the very deep VGG-16 model, our detection system has a frame rate of 5fps (including all steps) on a GPU, while achieving state-of-the-art object detection accuracy on PASCAL VOC 2007, 2012, and MS COCO datasets with only 300 proposals per image. In ILSVRC and COCO 2015 competitions, Faster R-CNN and RPN are the foundations of the 1st-place winning entries in several tracks. Code has been made publicly available.},
  443. file = {Faster R-CNN Towards Real-Time Object Detection with Reg:Attachments/Faster R-CNN Towards Real-Time Object Detection with Reg.pdf:application/pdf}
  444. }
  445. @misc{Rezatofighi.2252019,
  446. author = {Rezatofighi, Hamid and Tsoi, Nathan and Gwak, JunYoung and Sadeghian, Amir and Reid, Ian and Savarese, Silvio},
  447. year = {2019},
  448. title = {Generalized Intersection over Union: A Metric and A Loss for Bounding Box Regression},
  449. archivePrefix = {arXiv},
  450. eprint = {arXiv:1902.09630v2},
  451. url = {https://arxiv.org/pdf/1902.09630},
  452. keywords = {Artificial Intelligence (cs.AI);Computer Science - Artificial Intelligence;Computer Science - Computer Vision and Pattern Recognition;Computer Science - Learning;Computer Vision and Pattern Recognition (cs.CV);Machine Learning (cs.LG)},
  453. abstract = {Intersection over Union (IoU) is the most popular evaluation metric used in the object detection benchmarks. However, there is a gap between optimizing the commonly used distance losses for regressing the parameters of a bounding box and maximizing this metric value. The optimal objective for a metric is the metric itself. In the case of axis-aligned 2D bounding boxes, it can be shown that $IoU$ can be directly used as a regression loss. However, $IoU$ has a plateau making it infeasible to optimize in the case of non-overlapping bounding boxes. In this paper, we address the weaknesses of $IoU$ by introducing a generalized version as both a new loss and a new metric. By incorporating this generalized $IoU$ ($GIoU$) as a loss into the state-of-the art object detection frameworks, we show a consistent improvement on their performance using both the standard, $IoU$ based, and new, $GIoU$ based, performance measures on popular object detection benchmarks such as PASCAL VOC and MS COCO.},
  454. doi = {CVPR},
  455. file = {Generalized Intersection over Union A Metric and A Loss:Attachments/Generalized Intersection over Union A Metric and A Loss.pdf:application/pdf},
  456. note = {accepted in CVPR 2019}
  457. }
  458. @online{TensorFlow.8102021,
  459. author = {TensorFlow},
  460. year = {2021},
  461. title = {TensorFlow Lite-Konverter},
  462. url = {https://www.tensorflow.org/lite/convert},
  463. urldate = {2021-10-06},
  464. abstract = {}
  465. }
  466. @online{TensorFlow.8122021,
  467. author = {TensorFlow},
  468. year = {2021},
  469. title = {Quantisierung nach dem Training ~|~ TensorFlow Lite},
  470. url = {https://www.tensorflow.org/lite/performance/post_training_quantization},
  471. urldate = {2021-10-06},
  472. abstract = {}
  473. }
  474. @online{TensorFlow.8202021,
  475. author = {TensorFlow},
  476. year = {2021},
  477. title = {TensorFlow Lite | ML f{\"u}r Mobilger{\"a}te und Edge-Ger{\"a}te},
  478. url = {https://www.tensorflow.org/lite/?hl=de},
  479. urldate = {2021-10-06},
  480. abstract = {Ein Deep-Learning-Framework f{\"u}r ger{\"a}teinterne Inferenz. Trainieren und implementieren Sie Machine-Learning-Modelle auf mobilen und IoT-Ger{\"a}ten, Android, iOS, Edge TPU, Raspberry Pi.}
  481. }
  482. @online{TensorFlow.832021,
  483. author = {TensorFlow},
  484. year = {2021},
  485. title = {TensorFlow},
  486. url = {https://www.tensorflow.org/},
  487. urldate = {2021-10-06},
  488. abstract = {Eine End-to-End-Open-Source-Machine-Learning-Plattform f{\"u}r alle. Entdecken Sie das flexible {\"O}kosystem von TensorFlow aus Tools, Bibliotheken und Community-Ressourcen.}
  489. }
  490. @InProceedings{Xian_2018_CVPR,
  491. author = {Xian, Ke and Shen, Chunhua and Cao, Zhiguo and Lu, Hao and Xiao, Yang and Li, Ruibo and Luo, Zhenbo},
  492. title = {Monocular Relative Depth Perception With Web Stereo Data Supervision},
  493. booktitle = {Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
  494. month = {6},
  495. year = {2018}
  496. }
  497. @misc{Zhou.182017,
  498. author = {Zhou, Yiren and Song, Sibo and Cheung, Ngai-Man},
  499. year = {2017},
  500. title = {On Classification of Distorted Images with Deep Convolutional Neural Networks},
  501. archivePrefix = {arXiv},
  502. eprint = {arXiv:1701.01924v1},
  503. url = {http://arxiv.org/pdf/1701.01924v1},
  504. abstract = {Image blur and image noise are common distortions during image acquisition. In this paper, we systematically study the effect of image distortions on the deep neural network (DNN) image classifiers. First, we examine the DNN classifier performance under four types of distortions. Second, we propose two approaches to alleviate the effect of image distortion: re-training and fine-tuning with noisy images. Our results suggest that, under certain conditions, fine-tuning with noisy images can alleviate much effect due to distorted inputs, and is more practical than re-training.},
  505. pagetotal = {5},
  506. file = {On_Classification_of_Distorted_Images_with_Deep_Co:Attachments/On_Classification_of_Distorted_Images_with_Deep_Co.pdf:application/pdf},
  507. note = {5 pages, 8 figures, ICASSP 2017}
  508. }
  509. @book{Zhou.2021,
  510. author = {Zhou, Zhi-Hua},
  511. year = {2021},
  512. title = {Machine Learning},
  513. publisher = {{Springer Singapore}},
  514. isbn = {978-981-15-1966-6},
  515. location = {Singapore},
  516. abstract = {},
  517. doi = {10.1007/978-981-15-1967-3},
  518. file = {Machine Learning:Attachments/Machine Learning.pdf:application/pdf}
  519. }
  520. @InProceedings{PireStereoVision.2012,
  521. author = {Pire, Taihú and De Cristóforis, Pablo and Nitsche, Matias and Berlles, Julio},
  522. year = {2012},
  523. month = {12},
  524. pages = {},
  525. title = {Stereo vision obstacle avoidance using depth and elevation maps}
  526. }
  527. @InProceedings{Zhang.2019,
  528. author={Zhang, Zhenghong and Xiong, Mingkang and Xiong, Huilin},
  529. booktitle={2019 4th International Conference on Cloud Computing and Internet of Things (CCIOT)},
  530. title={Monocular Depth Estimation for UAV Obstacle Avoidance},
  531. year={2019},
  532. volume={},
  533. number={},
  534. pages={43-47},
  535. doi={10.1109/CCIOT48581.2019.8980350}}
  536. @InProceedings{Chakravarty.2017,
  537. author={Chakravarty, Punarjay and Kelchtermans, Klaas and Roussel, Tom and Wellens, Stijn and Tuytelaars, Tinne and Van Eycken, Luc},
  538. booktitle={2017 IEEE International Conference on Robotics and Automation (ICRA)},
  539. title={CNN-based single image obstacle avoidance on a quadrotor},
  540. year={2017},
  541. volume={},
  542. number={},
  543. pages={6369-6374},
  544. doi={10.1109/ICRA.2017.7989752}
  545. }
  546. @InProceedings{Macias-GarciaDetection.2020,
  547. author={Macias-Garcia, Edgar and Galeana-Perez, Deysy and Bayro-Corrochano, Eduardo},
  548. booktitle={2020 International Joint Conference on Neural Networks (IJCNN)},
  549. title={CNN Based Perception System for Collision Avoidance in Mobile Robots using Stereo Vision},
  550. year={2020},
  551. volume={},
  552. number={},
  553. pages={1-7},
  554. doi={10.1109/IJCNN48605.2020.9206747}
  555. }
  556. @InProceedings{Nalpantidis.2009,
  557. author = {Nalpantidis, Lazaros and Kostavelis, Ioannis and Gasteratos, Antonios},
  558. year = {2009},
  559. month = {12},
  560. pages = {195-204},
  561. title = {Stereovision-Based Algorithm for Obstacle Avoidance},
  562. isbn = {978-3-642-10816-7},
  563. doi = {10.1007/978-3-642-10817-4_19}
  564. }
  565. @inbook{Alvarez.2016,
  566. author = {Alvarez, H. and Paz, L.M. and Sturm, Jürgen and Cremers, D.},
  567. year = {2016},
  568. month = {11},
  569. pages = {195-209},
  570. title = {Collision Avoidance for Quadrotors with a Monocular Camera},
  571. volume = {109},
  572. isbn = {978-3-319-23777-0},
  573. doi = {10.1007/978-3-319-23778-7_14}
  574. }
  575. @misc{Hatch.2020,
  576. author = {Hatch, Kyle and Mern, John and Kochenderfer, Mykel},
  577. year = {2020},
  578. title = {Obstacle Avoidance Using a Monocular Camera},
  579. archivePrefix = {arXiv},
  580. eprint = {arXiv:2012.01608v2},
  581. url = {https://arxiv.org/pdf/2012.01608},
  582. keywords = {Artificial Intelligence (cs.AI);Robotics (cs.RO)},
  583. abstract = {A collision avoidance system based on simple digital cameras would help enable the safe integration of small UAVs into crowded, low-altitude environments. In this work, we present an obstacle avoidance system for small UAVs that uses a monocular camera with a hybrid neural network and path planner controller. The system is comprised of a vision network for estimating depth from camera images, a high-level control network, a collision prediction network, and a contingency policy. This system is evaluated on a simulated UAV navigating an obstacle course in a constrained flight pattern. Results show the proposed system achieves low collision rates while maintaining operationally relevant flight speeds.},
  584. file = {Obstacle Avoidance Using a Monocular Camera:Attachments/Obstacle Avoidance Using a Monocular Camera.pdf:application/pdf}
  585. }
  586. @misc{Jacob.12152017,
  587. author = {Jacob, Benoit and Kligys, Skirmantas and Chen, Bo and Zhu, Menglong and Tang, Matthew and Howard, Andrew and Adam, Hartwig and Kalenichenko, Dmitry},
  588. year = {2017},
  589. title = {Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference},
  590. archivePrefix = {arXiv},
  591. eprint = {arXiv:1712.05877v1},
  592. url = {https://arxiv.org/pdf/1712.05877},
  593. keywords = {Computer Science - Learning;Machine Learning (cs.LG);Machine Learning (stat.ML);Statistics - Machine Learning},
  594. abstract = {The rising popularity of intelligent mobile devices and the daunting computational cost of deep learning-based models call for efficient and accurate on-device inference schemes. We propose a quantization scheme that allows inference to be carried out using integer-only arithmetic, which can be implemented more efficiently than floating point inference on commonly available integer-only hardware. We also co-design a training procedure to preserve end-to-end model accuracy post quantization. As a result, the proposed quantization scheme improves the tradeoff between accuracy and on-device latency. The improvements are significant even on MobileNets, a model family known for run-time efficiency, and are demonstrated in ImageNet classification and COCO detection on popular CPUs.},
  595. pagetotal = {14},
  596. note = {14 pages, 12 figures}
  597. }
  598. @misc{Kim.11152015,
  599. author = {Kim, Dong Ki and Chen, Tsuhan},
  600. year = {2015},
  601. title = {Deep Neural Network for Real-Time Autonomous Indoor Navigation},
  602. archivePrefix = {arXiv},
  603. eprint = {arXiv:1511.04668v2},
  604. url = {https://arxiv.org/pdf/1511.04668},
  605. keywords = {Computer Science - Computer Vision and Pattern Recognition;Computer Vision and Pattern Recognition (cs.CV)},
  606. abstract = {Autonomous indoor navigation of Micro Aerial Vehicles (MAVs) possesses many challenges. One main reason is that GPS has limited precision in indoor environments. The additional fact that MAVs are not able to carry heavy weight or power consuming sensors, such as range finders, makes indoor autonomous navigation a challenging task. In this paper, we propose a practical system in which a quadcopter autonomously navigates indoors and finds a specific target, i.e., a book bag, by using a single camera. A deep learning model, Convolutional Neural Network (ConvNet), is used to learn a controller strategy that mimics an expert pilot's choice of action. We show our system's performance through real-time experiments in diverse indoor locations. To understand more about our trained network, we use several visualization techniques.},
  607. file = {Deep Neural Network for Real-Time Autonomous Indoor Navi:Attachments/Deep Neural Network for Real-Time Autonomous Indoor Navi.pdf:application/pdf}
  608. }
  609. @article{Liu.2017,
  610. author = {Liu, Canglong and Zheng, Bin and Wang, Chunyang and Zhao, Yongting and Fu, Shun and Li, Haochen},
  611. year = {2017},
  612. title = {CNN-Based Vision Model for Obstacle Avoidance of Mobile Robot},
  613. pages = {00007},
  614. pagination = {page},
  615. volume = {139},
  616. journaltitle = {MATEC Web of Conferences},
  617. shortjournal = {MATEC Web Conf.},
  618. doi = {10.1051/matecconf/201713900007},
  619. abstract = {},
  620. file = {CNN-Based Vision Model for Obstacle Avoidance of Mobile Robot:Attachments/CNN-Based Vision Model for Obstacle Avoidance of Mobile Robot.pdf:application/pdf},
  621. note = {Xu, Bing (Editor) Chen, Yinong (Editor) PII: matecconf{\_}icmite2017{\_}00007}
  622. }
  623. @misc{Yang.2017,
  624. author = {Yang, Shichao and Konam, Sandeep and Ma, Chen and Rosenthal, Stephanie and Veloso, Manuela and Scherer, Sebastian},
  625. year = {2017},
  626. title = {Obstacle Avoidance through Deep Networks based Intermediate Perception},
  627. archivePrefix = {arXiv},
  628. eprint = {arXiv:1704.08759v1},
  629. url = {https://arxiv.org/pdf/1704.08759},
  630. keywords = {Computer Vision and Pattern Recognition (cs.CV);Robotics (cs.RO)},
  631. abstract = {Obstacle avoidance from monocular images is a challenging problem for robots. Though multi-view structure-from-motion could build 3D maps, it is not robust in textureless environments. Some learning based methods exploit human demonstration to predict a steering command directly from a single image. However, this method is usually biased towards certain tasks or demonstration scenarios and also biased by human understanding. In this paper, we propose a new method to predict a trajectory from images. We train our system on more diverse NYUv2 dataset. The ground truth trajectory is computed from the designed cost functions automatically. The Convolutional Neural Network perception is divided into two stages: first, predict depth map and surface normal from RGB images, which are two important geometric properties related to 3D obstacle representation. Second, predict the trajectory from the depth and normal. Results show that our intermediate perception increases the accuracy by 20{\%} than the direct prediction. Our model generalizes well to other public indoor datasets and is also demonstrated for robot flights in simulation and experiments.},
  632. file = {Obstacle Avoidance through Deep Networks based Intermedi:Attachments/Obstacle Avoidance through Deep Networks based Intermedi.pdf:application/pdf}
  633. }
  634. @misc{Zhu.2016,
  635. author = {Zhu, Yuke and Mottaghi, Roozbeh and Kolve, Eric and Lim, Joseph J. and Gupta, Abhinav and Fei-Fei, Li and Farhadi, Ali},
  636. year = {2016},
  637. title = {Target-driven Visual Navigation in Indoor Scenes using Deep Reinforcement Learning},
  638. archivePrefix = {arXiv},
  639. eprint = {arXiv:1609.05143v1},
  640. url = {https://arxiv.org/pdf/1609.05143},
  641. keywords = {Computer Vision and Pattern Recognition (cs.CV)},
  642. abstract = {Two less addressed issues of deep reinforcement learning are (1) lack of generalization capability to new target goals, and (2) data inefficiency i.e., the model requires several (and often costly) episodes of trial and error to converge, which makes it impractical to be applied to real-world scenarios. In this paper, we address these two issues and apply our model to the task of target-driven visual navigation. To address the first issue, we propose an actor-critic model whose policy is a function of the goal as well as the current state, which allows to better generalize. To address the second issue, we propose AI2-THOR framework, which provides an environment with high-quality 3D scenes and physics engine. Our framework enables agents to take actions and interact with objects. Hence, we can collect a huge number of training samples efficiently.
  643. We show that our proposed method (1) converges faster than the state-of-the-art deep reinforcement learning methods, (2) generalizes across targets and across scenes, (3) generalizes to a real robot scenario with a small amount of fine-tuning (although the model is trained in simulation), (4) is end-to-end trainable and does not need feature engineering, feature matching between frames or 3D reconstruction of the environment.
  644. The supplementary video can be accessed at the following link: this https URL.}
  645. }
  646. @online{OpenImages.6252021,
  647. year = {2021},
  648. title = {Open Images V6},
  649. url = {https://storage.googleapis.com/openimages/web/index.html},
  650. urldate = {2021-12-12},
  651. abstract = {}
  652. }
  653. @Comment{jabref-meta: databaseType:biblatex;}