Latex-Dateien der Masterarbeit
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

abschlussarbeit.bbl 124KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024202520262027202820292030203120322033203420352036203720382039204020412042204320442045204620472048204920502051205220532054205520562057205820592060206120622063206420652066206720682069207020712072207320742075207620772078207920802081208220832084208520862087208820892090209120922093209420952096209720982099210021012102210321042105210621072108210921102111211221132114211521162117211821192120212121222123212421252126212721282129213021312132213321342135213621372138213921402141214221432144214521462147214821492150215121522153215421552156215721582159216021612162216321642165216621672168216921702171217221732174217521762177217821792180218121822183218421852186218721882189219021912192219321942195219621972198219922002201220222032204220522062207220822092210221122122213221422152216221722182219222022212222222322242225222622272228222922302231223222332234223522362237223822392240224122422243224422452246224722482249225022512252225322542255225622572258225922602261226222632264226522662267226822692270227122722273227422752276227722782279228022812282228322842285228622872288228922902291229222932294229522962297229822992300230123022303230423052306230723082309231023112312231323142315231623172318231923202321232223232324232523262327232823292330233123322333233423352336233723382339234023412342234323442345234623472348234923502351235223532354235523562357235823592360236123622363236423652366236723682369237023712372237323742375237623772378237923802381238223832384238523862387238823892390239123922393239423952396239723982399240024012402240324042405240624072408240924102411241224132414241524162417241824192420242124222423242424252426242724282429243024312432243324342435243624372438243924402441244224432444244524462447244824492450245124522453245424552456245724582459246024612462246324642465246624672468246924702471247224732474247524762477247824792480248124822483248424852486248724882489249024912492249324942495249624972498249925002501250225032504250525062507250825092510251125122513251425152516251725182519252025212522252325242525
  1. % $ biblatex auxiliary file $
  2. % $ biblatex bbl format version 3.1 $
  3. % Do not modify the above lines!
  4. %
  5. % This is an auxiliary file used by the 'biblatex' package.
  6. % This file may safely be deleted. It will be recreated by
  7. % biber as required.
  8. %
  9. \begingroup
  10. \makeatletter
  11. \@ifundefined{ver@biblatex.sty}
  12. {\@latex@error
  13. {Missing 'biblatex' package}
  14. {The bibliography requires the 'biblatex' package.}
  15. \aftergroup\endinput}
  16. {}
  17. \endgroup
  18. \refsection{0}
  19. \datalist[entry]{nty/global//global/global}
  20. \entry{Aleotti.2020}{misc}{}
  21. \name{author}{6}{}{%
  22. {{hash=0e141e967911dc1e9e3266ec98e4fc1d}{%
  23. family={Aleotti},
  24. familyi={A\bibinitperiod},
  25. given={Filippo},
  26. giveni={F\bibinitperiod}}}%
  27. {{hash=95a910f0b0e3c4905262ac1ab4cd3736}{%
  28. family={Zaccaroni},
  29. familyi={Z\bibinitperiod},
  30. given={Giulio},
  31. giveni={G\bibinitperiod}}}%
  32. {{hash=6d32bbe0bffb37cc564df8adb6c87702}{%
  33. family={Bartolomei},
  34. familyi={B\bibinitperiod},
  35. given={Luca},
  36. giveni={L\bibinitperiod}}}%
  37. {{hash=343dafb5548d904115e5782508eb8375}{%
  38. family={Poggi},
  39. familyi={P\bibinitperiod},
  40. given={Matteo},
  41. giveni={M\bibinitperiod}}}%
  42. {{hash=8d31b2206bee079ebdf6ff494cb9c6ee}{%
  43. family={Tosi},
  44. familyi={T\bibinitperiod},
  45. given={Fabio},
  46. giveni={F\bibinitperiod}}}%
  47. {{hash=97fde7508220953d74da9f2a50a93a4e}{%
  48. family={Mattoccia},
  49. familyi={M\bibinitperiod},
  50. given={Stefano},
  51. giveni={S\bibinitperiod}}}%
  52. }
  53. \strng{namehash}{7939deb38e016c99c47c9e866b3ade34}
  54. \strng{fullhash}{8314dffcc5cb6330e824ffeb8f7b00f2}
  55. \strng{bibnamehash}{7939deb38e016c99c47c9e866b3ade34}
  56. \strng{authorbibnamehash}{7939deb38e016c99c47c9e866b3ade34}
  57. \strng{authornamehash}{7939deb38e016c99c47c9e866b3ade34}
  58. \strng{authorfullhash}{8314dffcc5cb6330e824ffeb8f7b00f2}
  59. \field{sortinit}{A}
  60. \field{sortinithash}{2f401846e2029bad6b3ecc16d50031e2}
  61. \field{labelnamesource}{author}
  62. \field{labeltitlesource}{title}
  63. \field{abstract}{Depth perception is paramount to tackle real-world problems, ranging from autonomous driving to consumer applications. For the latter, depth estimation from a single image represents the most versatile solution, since a standard camera is available on almost any handheld device. Nonetheless, two main issues limit its practical deployment: i) the low reliability when deployed in-the-wild and ii) the demanding resource requirements to achieve real-time performance, often not compatible with such devices. Therefore, in this paper, we deeply investigate these issues showing how they are both addressable adopting appropriate network design and training strategies -- also outlining how to map the resulting networks on handheld devices to achieve real-time performance. Our thorough evaluation highlights the ability of such fast networks to generalize well to new environments, a crucial feature required to tackle the extremely varied contexts faced in real applications. Indeed, to further support this evidence, we report experimental results concerning real-time depth-aware augmented reality and image blurring with smartphones in-the-wild.}
  64. \field{eprinttype}{arXiv}
  65. \field{title}{Real-time single image depth perception in the wild with handheld devices}
  66. \field{year}{2020}
  67. \verb{eprint}
  68. \verb arXiv:2006.05724v1
  69. \endverb
  70. \verb{file}
  71. \verb Real-time single image depth perception in the wild with:Attachments/Real-time single image depth perception in the wild with.pdf:application/pdf
  72. \endverb
  73. \verb{urlraw}
  74. \verb https://arxiv.org/pdf/2006.05724
  75. \endverb
  76. \verb{url}
  77. \verb https://arxiv.org/pdf/2006.05724
  78. \endverb
  79. \keyw{Computer Vision and Pattern Recognition (cs.CV);Graphics (cs.GR)}
  80. \endentry
  81. \entry{Alvarez.2016}{inbook}{}
  82. \name{author}{4}{}{%
  83. {{hash=087415c33d15dad522378a03ed5ff0b5}{%
  84. family={Alvarez},
  85. familyi={A\bibinitperiod},
  86. given={H.},
  87. giveni={H\bibinitperiod}}}%
  88. {{hash=0dd1001c99b3c779d7ccd99a4efd66b3}{%
  89. family={Paz},
  90. familyi={P\bibinitperiod},
  91. given={L.M.},
  92. giveni={L\bibinitperiod}}}%
  93. {{hash=9e291584a3b29b27c88ecfe0a566274c}{%
  94. family={Sturm},
  95. familyi={S\bibinitperiod},
  96. given={Jürgen},
  97. giveni={J\bibinitperiod}}}%
  98. {{hash=b4995f3c4c5030db3a9e2fe4f3d61481}{%
  99. family={Cremers},
  100. familyi={C\bibinitperiod},
  101. given={D.},
  102. giveni={D\bibinitperiod}}}%
  103. }
  104. \strng{namehash}{d907d14b508c6dbd77cc25079a69a171}
  105. \strng{fullhash}{fc04147d83c649a9044a6fb80b221b9b}
  106. \strng{bibnamehash}{d907d14b508c6dbd77cc25079a69a171}
  107. \strng{authorbibnamehash}{d907d14b508c6dbd77cc25079a69a171}
  108. \strng{authornamehash}{d907d14b508c6dbd77cc25079a69a171}
  109. \strng{authorfullhash}{fc04147d83c649a9044a6fb80b221b9b}
  110. \field{sortinit}{A}
  111. \field{sortinithash}{2f401846e2029bad6b3ecc16d50031e2}
  112. \field{labelnamesource}{author}
  113. \field{labeltitlesource}{title}
  114. \field{isbn}{978-3-319-23777-0}
  115. \field{month}{11}
  116. \field{title}{Collision Avoidance for Quadrotors with a Monocular Camera}
  117. \field{volume}{109}
  118. \field{year}{2016}
  119. \field{pages}{195\bibrangedash 209}
  120. \range{pages}{15}
  121. \verb{doi}
  122. \verb 10.1007/978-3-319-23778-7_14
  123. \endverb
  124. \endentry
  125. \entry{Arulprakash.2021}{article}{}
  126. \name{author}{2}{}{%
  127. {{hash=4bbb0814acd56658521599c3a9cb4636}{%
  128. family={Arulprakash},
  129. familyi={A\bibinitperiod},
  130. given={Enoch},
  131. giveni={E\bibinitperiod}}}%
  132. {{hash=af8274fdf1755aa0001aecf866957874}{%
  133. family={Aruldoss},
  134. familyi={A\bibinitperiod},
  135. given={Martin},
  136. giveni={M\bibinitperiod}}}%
  137. }
  138. \strng{namehash}{595d5f35a0a949d2089c6f0c1195d885}
  139. \strng{fullhash}{595d5f35a0a949d2089c6f0c1195d885}
  140. \strng{bibnamehash}{595d5f35a0a949d2089c6f0c1195d885}
  141. \strng{authorbibnamehash}{595d5f35a0a949d2089c6f0c1195d885}
  142. \strng{authornamehash}{595d5f35a0a949d2089c6f0c1195d885}
  143. \strng{authorfullhash}{595d5f35a0a949d2089c6f0c1195d885}
  144. \field{sortinit}{A}
  145. \field{sortinithash}{2f401846e2029bad6b3ecc16d50031e2}
  146. \field{labelnamesource}{author}
  147. \field{labeltitlesource}{title}
  148. \field{issn}{13191578}
  149. \field{journaltitle}{Journal of King Saud University - Computer and Information Sciences}
  150. \field{note}{PII: S1319157821002020}
  151. \field{title}{A study on generic object detection with emphasis on future research directions}
  152. \field{year}{2021}
  153. \verb{doi}
  154. \verb 10.1016/j.jksuci.2021.08.001
  155. \endverb
  156. \endentry
  157. \entry{EncoderDecoder7803544}{article}{}
  158. \name{author}{3}{}{%
  159. {{hash=2b517df4f81c42dc3c5b32aede012cf9}{%
  160. family={Badrinarayanan},
  161. familyi={B\bibinitperiod},
  162. given={Vijay},
  163. giveni={V\bibinitperiod}}}%
  164. {{hash=4b1053f4ffb69cd61c1f122cc435c884}{%
  165. family={Kendall},
  166. familyi={K\bibinitperiod},
  167. given={Alex},
  168. giveni={A\bibinitperiod}}}%
  169. {{hash=95c5bd0a1d6e6649ed56043a63268bfa}{%
  170. family={Cipolla},
  171. familyi={C\bibinitperiod},
  172. given={Roberto},
  173. giveni={R\bibinitperiod}}}%
  174. }
  175. \strng{namehash}{fb26318591964ff2e23ee05578819393}
  176. \strng{fullhash}{fb26318591964ff2e23ee05578819393}
  177. \strng{bibnamehash}{fb26318591964ff2e23ee05578819393}
  178. \strng{authorbibnamehash}{fb26318591964ff2e23ee05578819393}
  179. \strng{authornamehash}{fb26318591964ff2e23ee05578819393}
  180. \strng{authorfullhash}{fb26318591964ff2e23ee05578819393}
  181. \field{sortinit}{B}
  182. \field{sortinithash}{d7095fff47cda75ca2589920aae98399}
  183. \field{labelnamesource}{author}
  184. \field{labeltitlesource}{title}
  185. \field{journaltitle}{IEEE Transactions on Pattern Analysis and Machine Intelligence}
  186. \field{number}{12}
  187. \field{title}{SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation}
  188. \field{volume}{39}
  189. \field{year}{2017}
  190. \field{pages}{2481\bibrangedash 2495}
  191. \range{pages}{15}
  192. \verb{doi}
  193. \verb 10.1109/TPAMI.2016.2644615
  194. \endverb
  195. \endentry
  196. \entry{Chakravarty.2017}{inproceedings}{}
  197. \name{author}{6}{}{%
  198. {{hash=b700b361eed5e85d647c6bbfef3eacd6}{%
  199. family={Chakravarty},
  200. familyi={C\bibinitperiod},
  201. given={Punarjay},
  202. giveni={P\bibinitperiod}}}%
  203. {{hash=abedc60b3acdfd251c34cf76e0aca39f}{%
  204. family={Kelchtermans},
  205. familyi={K\bibinitperiod},
  206. given={Klaas},
  207. giveni={K\bibinitperiod}}}%
  208. {{hash=6e8bb49e98d2fd65c0c240ceb60b922f}{%
  209. family={Roussel},
  210. familyi={R\bibinitperiod},
  211. given={Tom},
  212. giveni={T\bibinitperiod}}}%
  213. {{hash=f8bb13bef7b50a92cfa855b9c4c012c6}{%
  214. family={Wellens},
  215. familyi={W\bibinitperiod},
  216. given={Stijn},
  217. giveni={S\bibinitperiod}}}%
  218. {{hash=1fffde1e018446dc52c3da9c6e26ec11}{%
  219. family={Tuytelaars},
  220. familyi={T\bibinitperiod},
  221. given={Tinne},
  222. giveni={T\bibinitperiod}}}%
  223. {{hash=5868f82298c698db6e7249c1b16b1d6c}{%
  224. family={Van\bibnamedelima Eycken},
  225. familyi={V\bibinitperiod\bibinitdelim E\bibinitperiod},
  226. given={Luc},
  227. giveni={L\bibinitperiod}}}%
  228. }
  229. \strng{namehash}{71b3164dff5e2a7cf11049dd7e281ed1}
  230. \strng{fullhash}{f53b98e85eb07c9c6b8d6bb6edbc0493}
  231. \strng{bibnamehash}{71b3164dff5e2a7cf11049dd7e281ed1}
  232. \strng{authorbibnamehash}{71b3164dff5e2a7cf11049dd7e281ed1}
  233. \strng{authornamehash}{71b3164dff5e2a7cf11049dd7e281ed1}
  234. \strng{authorfullhash}{f53b98e85eb07c9c6b8d6bb6edbc0493}
  235. \field{sortinit}{C}
  236. \field{sortinithash}{4d103a86280481745c9c897c925753c0}
  237. \field{labelnamesource}{author}
  238. \field{labeltitlesource}{title}
  239. \field{booktitle}{2017 IEEE International Conference on Robotics and Automation (ICRA)}
  240. \field{title}{CNN-based single image obstacle avoidance on a quadrotor}
  241. \field{year}{2017}
  242. \field{pages}{6369\bibrangedash 6374}
  243. \range{pages}{6}
  244. \verb{doi}
  245. \verb 10.1109/ICRA.2017.7989752
  246. \endverb
  247. \endentry
  248. \entry{cocoEval.8252021}{online}{}
  249. \field{sortinit}{C}
  250. \field{sortinithash}{4d103a86280481745c9c897c925753c0}
  251. \field{labeltitlesource}{title}
  252. \field{title}{COCO - Common Objects in Context}
  253. \field{urlday}{14}
  254. \field{urlmonth}{11}
  255. \field{urlyear}{2021}
  256. \field{year}{2021}
  257. \field{urldateera}{ce}
  258. \verb{urlraw}
  259. \verb https://cocodataset.org/#detection-eval
  260. \endverb
  261. \verb{url}
  262. \verb https://cocodataset.org/#detection-eval
  263. \endverb
  264. \endentry
  265. \entry{coco.8252021}{online}{}
  266. \field{sortinit}{C}
  267. \field{sortinithash}{4d103a86280481745c9c897c925753c0}
  268. \field{labeltitlesource}{title}
  269. \field{title}{COCO - Common Objects in Context}
  270. \field{urlday}{14}
  271. \field{urlmonth}{11}
  272. \field{urlyear}{2021}
  273. \field{year}{2021}
  274. \field{urldateera}{ce}
  275. \verb{urlraw}
  276. \verb https://cocodataset.org
  277. \endverb
  278. \verb{url}
  279. \verb https://cocodataset.org
  280. \endverb
  281. \endentry
  282. \entry{EveringhamDevelopmentKit.2012}{misc}{}
  283. \name{author}{2}{}{%
  284. {{hash=655191b4f54a942147b118da2f50aeed}{%
  285. family={Everingham},
  286. familyi={E\bibinitperiod},
  287. given={Mark},
  288. giveni={M\bibinitperiod}}}%
  289. {{hash=19dcf449a33f27289a006b0acf819975}{%
  290. family={Winn},
  291. familyi={W\bibinitperiod},
  292. given={John},
  293. giveni={J\bibinitperiod}}}%
  294. }
  295. \strng{namehash}{71f075e5e08114d81693e397b204b1d8}
  296. \strng{fullhash}{71f075e5e08114d81693e397b204b1d8}
  297. \strng{bibnamehash}{71f075e5e08114d81693e397b204b1d8}
  298. \strng{authorbibnamehash}{71f075e5e08114d81693e397b204b1d8}
  299. \strng{authornamehash}{71f075e5e08114d81693e397b204b1d8}
  300. \strng{authorfullhash}{71f075e5e08114d81693e397b204b1d8}
  301. \field{sortinit}{E}
  302. \field{sortinithash}{8da8a182d344d5b9047633dfc0cc9131}
  303. \field{labelnamesource}{author}
  304. \field{labeltitlesource}{title}
  305. \field{title}{The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Development Kit}
  306. \field{urlday}{24}
  307. \field{urlmonth}{11}
  308. \field{urlyear}{2021}
  309. \field{year}{2012}
  310. \field{urldateera}{ce}
  311. \verb{urlraw}
  312. \verb http://host.robots.ox.ac.uk/pascal/VOC/voc2012/index.html
  313. \endverb
  314. \verb{url}
  315. \verb http://host.robots.ox.ac.uk/pascal/VOC/voc2012/index.html
  316. \endverb
  317. \endentry
  318. \entry{Everingham.2010}{article}{}
  319. \name{author}{5}{}{%
  320. {{hash=655191b4f54a942147b118da2f50aeed}{%
  321. family={Everingham},
  322. familyi={E\bibinitperiod},
  323. given={Mark},
  324. giveni={M\bibinitperiod}}}%
  325. {{hash=b296966d66d81277e4274fd491fdbf12}{%
  326. family={{van Gool}},
  327. familyi={v\bibinitperiod},
  328. given={Luc},
  329. giveni={L\bibinitperiod}}}%
  330. {{hash=d34fdfcbfcb5412397b946351370db22}{%
  331. family={Williams},
  332. familyi={W\bibinitperiod},
  333. given={Christopher\bibnamedelimb K.\bibnamedelimi I.},
  334. giveni={C\bibinitperiod\bibinitdelim K\bibinitperiod\bibinitdelim I\bibinitperiod}}}%
  335. {{hash=19dcf449a33f27289a006b0acf819975}{%
  336. family={Winn},
  337. familyi={W\bibinitperiod},
  338. given={John},
  339. giveni={J\bibinitperiod}}}%
  340. {{hash=c72fc39e94030f67717052309266a44d}{%
  341. family={Zisserman},
  342. familyi={Z\bibinitperiod},
  343. given={Andrew},
  344. giveni={A\bibinitperiod}}}%
  345. }
  346. \strng{namehash}{235de32ba219ebb749c8820ac16ff118}
  347. \strng{fullhash}{b62ea95ea4788089f092b405af5dc7a4}
  348. \strng{bibnamehash}{235de32ba219ebb749c8820ac16ff118}
  349. \strng{authorbibnamehash}{235de32ba219ebb749c8820ac16ff118}
  350. \strng{authornamehash}{235de32ba219ebb749c8820ac16ff118}
  351. \strng{authorfullhash}{b62ea95ea4788089f092b405af5dc7a4}
  352. \field{extraname}{1}
  353. \field{sortinit}{E}
  354. \field{sortinithash}{8da8a182d344d5b9047633dfc0cc9131}
  355. \field{labelnamesource}{author}
  356. \field{labeltitlesource}{title}
  357. \field{issn}{0920-5691}
  358. \field{journaltitle}{International Journal of Computer Vision}
  359. \field{note}{PII: 275}
  360. \field{number}{2}
  361. \field{pagination}{page}
  362. \field{shortjournal}{Int J Comput Vis}
  363. \field{title}{The Pascal Visual Object Classes (VOC) Challenge}
  364. \field{volume}{88}
  365. \field{year}{2010}
  366. \field{pages}{303\bibrangedash 338}
  367. \range{pages}{36}
  368. \verb{doi}
  369. \verb 10.1007/s11263-009-0275-4
  370. \endverb
  371. \endentry
  372. \entry{Everingham.2015}{article}{}
  373. \name{author}{6}{}{%
  374. {{hash=655191b4f54a942147b118da2f50aeed}{%
  375. family={Everingham},
  376. familyi={E\bibinitperiod},
  377. given={Mark},
  378. giveni={M\bibinitperiod}}}%
  379. {{hash=57a042c09397c84eaa88cad02f1a113d}{%
  380. family={Eslami},
  381. familyi={E\bibinitperiod},
  382. given={S.\bibnamedelimi M.\bibnamedelimi Ali},
  383. giveni={S\bibinitperiod\bibinitdelim M\bibinitperiod\bibinitdelim A\bibinitperiod}}}%
  384. {{hash=b296966d66d81277e4274fd491fdbf12}{%
  385. family={{van Gool}},
  386. familyi={v\bibinitperiod},
  387. given={Luc},
  388. giveni={L\bibinitperiod}}}%
  389. {{hash=d34fdfcbfcb5412397b946351370db22}{%
  390. family={Williams},
  391. familyi={W\bibinitperiod},
  392. given={Christopher\bibnamedelimb K.\bibnamedelimi I.},
  393. giveni={C\bibinitperiod\bibinitdelim K\bibinitperiod\bibinitdelim I\bibinitperiod}}}%
  394. {{hash=19dcf449a33f27289a006b0acf819975}{%
  395. family={Winn},
  396. familyi={W\bibinitperiod},
  397. given={John},
  398. giveni={J\bibinitperiod}}}%
  399. {{hash=c72fc39e94030f67717052309266a44d}{%
  400. family={Zisserman},
  401. familyi={Z\bibinitperiod},
  402. given={Andrew},
  403. giveni={A\bibinitperiod}}}%
  404. }
  405. \strng{namehash}{235de32ba219ebb749c8820ac16ff118}
  406. \strng{fullhash}{17d18af6db3a6b5860166567deed1fee}
  407. \strng{bibnamehash}{235de32ba219ebb749c8820ac16ff118}
  408. \strng{authorbibnamehash}{235de32ba219ebb749c8820ac16ff118}
  409. \strng{authornamehash}{235de32ba219ebb749c8820ac16ff118}
  410. \strng{authorfullhash}{17d18af6db3a6b5860166567deed1fee}
  411. \field{extraname}{2}
  412. \field{sortinit}{E}
  413. \field{sortinithash}{8da8a182d344d5b9047633dfc0cc9131}
  414. \field{labelnamesource}{author}
  415. \field{labeltitlesource}{title}
  416. \field{issn}{0920-5691}
  417. \field{journaltitle}{International Journal of Computer Vision}
  418. \field{note}{PII: 733}
  419. \field{number}{1}
  420. \field{pagination}{page}
  421. \field{shortjournal}{Int J Comput Vis}
  422. \field{title}{The Pascal Visual Object Classes Challenge: A Retrospective}
  423. \field{volume}{111}
  424. \field{year}{2015}
  425. \field{pages}{98\bibrangedash 136}
  426. \range{pages}{39}
  427. \verb{doi}
  428. \verb 10.1007/s11263-014-0733-5
  429. \endverb
  430. \endentry
  431. \entry{FlatBuffer.852021}{online}{}
  432. \field{sortinit}{F}
  433. \field{sortinithash}{2638baaa20439f1b5a8f80c6c08a13b4}
  434. \field{labeltitlesource}{title}
  435. \field{title}{FlatBuffers: FlatBuffers white paper}
  436. \field{urlday}{15}
  437. \field{urlmonth}{11}
  438. \field{urlyear}{2021}
  439. \field{year}{2021}
  440. \field{urldateera}{ce}
  441. \verb{urlraw}
  442. \verb https://google.github.io/flatbuffers/flatbuffers_white_paper.html
  443. \endverb
  444. \verb{url}
  445. \verb https://google.github.io/flatbuffers/flatbuffers_white_paper.html
  446. \endverb
  447. \endentry
  448. \entry{Girshick.2013}{misc}{}
  449. \name{author}{4}{}{%
  450. {{hash=bd5dadbe57bedc5957c19a3154c4d424}{%
  451. family={Girshick},
  452. familyi={G\bibinitperiod},
  453. given={Ross},
  454. giveni={R\bibinitperiod}}}%
  455. {{hash=913090a0d65c7a300c602f5a5a5a61e3}{%
  456. family={Donahue},
  457. familyi={D\bibinitperiod},
  458. given={Jeff},
  459. giveni={J\bibinitperiod}}}%
  460. {{hash=90180e1a30742e0d15328bfe637c2ef4}{%
  461. family={Darrell},
  462. familyi={D\bibinitperiod},
  463. given={Trevor},
  464. giveni={T\bibinitperiod}}}%
  465. {{hash=c75a5377b6dc5213831576e88ffe553c}{%
  466. family={Malik},
  467. familyi={M\bibinitperiod},
  468. given={Jitendra},
  469. giveni={J\bibinitperiod}}}%
  470. }
  471. \strng{namehash}{5aa9f085e8eccf547d1eb54a05cb1beb}
  472. \strng{fullhash}{9147daa2f6a058d5b8880ff5cf38abd6}
  473. \strng{bibnamehash}{5aa9f085e8eccf547d1eb54a05cb1beb}
  474. \strng{authorbibnamehash}{5aa9f085e8eccf547d1eb54a05cb1beb}
  475. \strng{authornamehash}{5aa9f085e8eccf547d1eb54a05cb1beb}
  476. \strng{authorfullhash}{9147daa2f6a058d5b8880ff5cf38abd6}
  477. \field{sortinit}{G}
  478. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  479. \field{labelnamesource}{author}
  480. \field{labeltitlesource}{title}
  481. \field{abstract}{Object detection performance, as measured on the canonical PASCAL VOC dataset, has plateaued in the last few years. The best-performing methods are complex ensemble systems that typically combine multiple low-level image features with high-level context. In this paper, we propose a simple and scalable detection algorithm that improves mean average precision (mAP) by more than 30{\%} relative to the previous best result on VOC 2012---achieving a mAP of 53.3{\%}. Our approach combines two key insights: (1) one can apply high-capacity convolutional neural networks (CNNs) to bottom-up region proposals in order to localize and segment objects and (2) when labeled training data is scarce, supervised pre-training for an auxiliary task, followed by domain-specific fine-tuning, yields a significant performance boost. Since we combine region proposals with CNNs, we call our method R-CNN: Regions with CNN features. We also compare R-CNN to OverFeat, a recently proposed sliding-window detector based on a similar CNN architecture. We find that R-CNN outperforms OverFeat by a large margin on the 200-class ILSVRC2013 detection dataset. Source code for the complete system is available at this http URL.}
  482. \field{eprinttype}{arXiv}
  483. \field{title}{Rich feature hierarchies for accurate object detection and semantic segmentation}
  484. \field{year}{2013}
  485. \verb{eprint}
  486. \verb arXiv:1311.2524v5
  487. \endverb
  488. \verb{file}
  489. \verb Rich feature hierarchies for accurate object detection a:Attachments/Rich feature hierarchies for accurate object detection a.pdf:application/pdf
  490. \endverb
  491. \verb{urlraw}
  492. \verb https://arxiv.org/pdf/1311.2524
  493. \endverb
  494. \verb{url}
  495. \verb https://arxiv.org/pdf/1311.2524
  496. \endverb
  497. \keyw{Computer Vision and Pattern Recognition (cs.CV)}
  498. \endentry
  499. \entry{GitHubTensorFlowAndroidExample.11222021}{online}{}
  500. \name{author}{1}{}{%
  501. {{hash=d3b7c913cd04ebfec0e9ec32cb6fd58c}{%
  502. family={GitHub},
  503. familyi={G\bibinitperiod}}}%
  504. }
  505. \strng{namehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  506. \strng{fullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  507. \strng{bibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  508. \strng{authorbibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  509. \strng{authornamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  510. \strng{authorfullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  511. \field{extraname}{1}
  512. \field{sortinit}{G}
  513. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  514. \field{labelnamesource}{author}
  515. \field{labeltitlesource}{title}
  516. \field{abstract}{TensorFlow examples. Contribute to tensorflow/examples development by creating an account on GitHub.}
  517. \field{title}{examples/lite/examples/image{\_}classification/android at master · tensorflow/examples}
  518. \field{urlday}{22}
  519. \field{urlmonth}{11}
  520. \field{urlyear}{2021}
  521. \field{year}{2021}
  522. \field{urldateera}{ce}
  523. \verb{urlraw}
  524. \verb https://github.com/tensorflow/examples/tree/master/lite/examples/object_detection/android
  525. \endverb
  526. \verb{url}
  527. \verb https://github.com/tensorflow/examples/tree/master/lite/examples/object_detection/android
  528. \endverb
  529. \endentry
  530. \entry{GitHub.1062021c}{online}{}
  531. \name{author}{1}{}{%
  532. {{hash=d3b7c913cd04ebfec0e9ec32cb6fd58c}{%
  533. family={GitHub},
  534. familyi={G\bibinitperiod}}}%
  535. }
  536. \strng{namehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  537. \strng{fullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  538. \strng{bibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  539. \strng{authorbibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  540. \strng{authornamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  541. \strng{authorfullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  542. \field{extraname}{2}
  543. \field{sortinit}{G}
  544. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  545. \field{labelnamesource}{author}
  546. \field{labeltitlesource}{title}
  547. \field{abstract}{Pydnet on mobile devices. Contribute to FilippoAleotti/mobilePydnet development by creating an account on GitHub.}
  548. \field{title}{GitHub - FilippoAleotti/mobilePydnet: Pydnet on mobile devices}
  549. \field{urlday}{6}
  550. \field{urlmonth}{10}
  551. \field{urlyear}{2021}
  552. \field{year}{2021}
  553. \field{urldateera}{ce}
  554. \verb{urlraw}
  555. \verb https://github.com/FilippoAleotti/mobilePydnet
  556. \endverb
  557. \verb{url}
  558. \verb https://github.com/FilippoAleotti/mobilePydnet
  559. \endverb
  560. \endentry
  561. \entry{GitHub.1282021}{online}{}
  562. \name{author}{1}{}{%
  563. {{hash=d3b7c913cd04ebfec0e9ec32cb6fd58c}{%
  564. family={GitHub},
  565. familyi={G\bibinitperiod}}}%
  566. }
  567. \strng{namehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  568. \strng{fullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  569. \strng{bibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  570. \strng{authorbibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  571. \strng{authornamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  572. \strng{authorfullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  573. \field{extraname}{3}
  574. \field{sortinit}{G}
  575. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  576. \field{labelnamesource}{author}
  577. \field{labeltitlesource}{title}
  578. \field{abstract}{Display and control your Android device. Contribute to Genymobile/scrcpy development by creating an account on GitHub.}
  579. \field{title}{GitHub - Genymobile/scrcpy: Display and control your Android device}
  580. \field{urlday}{8}
  581. \field{urlmonth}{12}
  582. \field{urlyear}{2021}
  583. \field{year}{2021}
  584. \field{urldateera}{ce}
  585. \verb{urlraw}
  586. \verb https://github.com/Genymobile/scrcpy
  587. \endverb
  588. \verb{url}
  589. \verb https://github.com/Genymobile/scrcpy
  590. \endverb
  591. \endentry
  592. \entry{GitHub.1062021b}{online}{}
  593. \name{author}{1}{}{%
  594. {{hash=d3b7c913cd04ebfec0e9ec32cb6fd58c}{%
  595. family={GitHub},
  596. familyi={G\bibinitperiod}}}%
  597. }
  598. \strng{namehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  599. \strng{fullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  600. \strng{bibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  601. \strng{authorbibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  602. \strng{authornamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  603. \strng{authorfullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  604. \field{extraname}{4}
  605. \field{sortinit}{G}
  606. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  607. \field{labelnamesource}{author}
  608. \field{labeltitlesource}{title}
  609. \field{abstract}{Code for robust monocular depth estimation described in Ranftl et. al., Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer, TPAMI 2020 - GitHub - isl-org/MiDaS: Code for robust monocular depth estimation described in Ranftl et. al., Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer, TPAMI 2020}
  610. \field{title}{GitHub - isl-org/MiDaS: Code for robust monocular depth estimation described in Ranftl et. al., Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer, TPAMI 2020}
  611. \field{urlday}{6}
  612. \field{urlmonth}{10}
  613. \field{urlyear}{2021}
  614. \field{year}{2021}
  615. \field{urldateera}{ce}
  616. \verb{urlraw}
  617. \verb https://github.com/isl-org/MiDaS
  618. \endverb
  619. \verb{url}
  620. \verb https://github.com/isl-org/MiDaS
  621. \endverb
  622. \endentry
  623. \entry{GitHub.1282021b}{online}{}
  624. \name{author}{1}{}{%
  625. {{hash=d3b7c913cd04ebfec0e9ec32cb6fd58c}{%
  626. family={GitHub},
  627. familyi={G\bibinitperiod}}}%
  628. }
  629. \strng{namehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  630. \strng{fullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  631. \strng{bibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  632. \strng{authorbibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  633. \strng{authornamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  634. \strng{authorfullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  635. \field{extraname}{5}
  636. \field{sortinit}{G}
  637. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  638. \field{labelnamesource}{author}
  639. \field{labeltitlesource}{title}
  640. \field{abstract}{Visualizer for neural network, deep learning, and machine learning models - GitHub - lutzroeder/netron: Visualizer for neural network, deep learning, and machine learning models}
  641. \field{title}{GitHub - lutzroeder/netron: Visualizer for neural network, deep learning, and machine learning models}
  642. \field{urlday}{8}
  643. \field{urlmonth}{12}
  644. \field{urlyear}{2021}
  645. \field{year}{2021}
  646. \field{urldateera}{ce}
  647. \verb{urlraw}
  648. \verb https://github.com/lutzroeder/netron
  649. \endverb
  650. \verb{url}
  651. \verb https://github.com/lutzroeder/netron
  652. \endverb
  653. \endentry
  654. \entry{GitHub.1062021d}{online}{}
  655. \name{author}{1}{}{%
  656. {{hash=d3b7c913cd04ebfec0e9ec32cb6fd58c}{%
  657. family={GitHub},
  658. familyi={G\bibinitperiod}}}%
  659. }
  660. \strng{namehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  661. \strng{fullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  662. \strng{bibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  663. \strng{authorbibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  664. \strng{authornamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  665. \strng{authorfullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  666. \field{extraname}{6}
  667. \field{sortinit}{G}
  668. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  669. \field{labelnamesource}{author}
  670. \field{labeltitlesource}{title}
  671. \field{abstract}{LabelImg is a graphical image annotation tool and label object bounding boxes in images - GitHub - tzutalin/labelImg: LabelImg is a graphical image annotation tool and label object bounding boxes in images}
  672. \field{title}{GitHub - tzutalin/labelImg: LabelImg is a graphical image annotation tool and label object bounding boxes in images}
  673. \field{urlday}{6}
  674. \field{urlmonth}{10}
  675. \field{urlyear}{2021}
  676. \field{year}{2021}
  677. \field{urldateera}{ce}
  678. \verb{urlraw}
  679. \verb https://github.com/tzutalin/labelImg
  680. \endverb
  681. \verb{url}
  682. \verb https://github.com/tzutalin/labelImg
  683. \endverb
  684. \endentry
  685. \entry{GitHub.1062021}{online}{}
  686. \name{author}{1}{}{%
  687. {{hash=d3b7c913cd04ebfec0e9ec32cb6fd58c}{%
  688. family={GitHub},
  689. familyi={G\bibinitperiod}}}%
  690. }
  691. \strng{namehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  692. \strng{fullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  693. \strng{bibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  694. \strng{authorbibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  695. \strng{authornamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  696. \strng{authorfullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  697. \field{extraname}{7}
  698. \field{sortinit}{G}
  699. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  700. \field{labelnamesource}{author}
  701. \field{labeltitlesource}{title}
  702. \field{abstract}{Models and examples built with TensorFlow. Contribute to tensorflow/models development by creating an account on GitHub.}
  703. \field{title}{models/research/object{\_}detection at master · tensorflow/models}
  704. \field{urlday}{6}
  705. \field{urlmonth}{10}
  706. \field{urlyear}{2021}
  707. \field{year}{2021}
  708. \field{urldateera}{ce}
  709. \verb{urlraw}
  710. \verb https://github.com/tensorflow/models/tree/master/research/object_detection
  711. \endverb
  712. \verb{url}
  713. \verb https://github.com/tensorflow/models/tree/master/research/object_detection
  714. \endverb
  715. \endentry
  716. \entry{GitHub.12122021}{online}{}
  717. \name{author}{1}{}{%
  718. {{hash=d3b7c913cd04ebfec0e9ec32cb6fd58c}{%
  719. family={GitHub},
  720. familyi={G\bibinitperiod}}}%
  721. }
  722. \strng{namehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  723. \strng{fullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  724. \strng{bibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  725. \strng{authorbibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  726. \strng{authornamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  727. \strng{authorfullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  728. \field{extraname}{8}
  729. \field{sortinit}{G}
  730. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  731. \field{labelnamesource}{author}
  732. \field{labeltitlesource}{title}
  733. \field{abstract}{Models and examples built with TensorFlow. Contribute to tensorflow/models development by creating an account on GitHub.}
  734. \field{title}{models/research/object{\_}detection at master · tensorflow/models}
  735. \field{urlday}{12}
  736. \field{urlmonth}{12}
  737. \field{urlyear}{2021}
  738. \field{year}{2021}
  739. \field{urldateera}{ce}
  740. \verb{urlraw}
  741. \verb https://github.com/tensorflow/models/blob/master/research/object_detection/protos/preprocessor.proto
  742. \endverb
  743. \verb{url}
  744. \verb https://github.com/tensorflow/models/blob/master/research/object_detection/protos/preprocessor.proto
  745. \endverb
  746. \endentry
  747. \entry{GitHubDetectionZoo.11152021}{online}{}
  748. \name{author}{1}{}{%
  749. {{hash=d3b7c913cd04ebfec0e9ec32cb6fd58c}{%
  750. family={GitHub},
  751. familyi={G\bibinitperiod}}}%
  752. }
  753. \strng{namehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  754. \strng{fullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  755. \strng{bibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  756. \strng{authorbibnamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  757. \strng{authornamehash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  758. \strng{authorfullhash}{d3b7c913cd04ebfec0e9ec32cb6fd58c}
  759. \field{extraname}{9}
  760. \field{sortinit}{G}
  761. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  762. \field{labelnamesource}{author}
  763. \field{labeltitlesource}{title}
  764. \field{abstract}{Models and examples built with TensorFlow. Contribute to tensorflow/models development by creating an account on GitHub.}
  765. \field{title}{models/tf2{\_}detection{\_}zoo.md at master · tensorflow/models}
  766. \field{urlday}{15}
  767. \field{urlmonth}{11}
  768. \field{urlyear}{2021}
  769. \field{year}{2021}
  770. \field{urldateera}{ce}
  771. \verb{urlraw}
  772. \verb https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf2_detection_zoo.md
  773. \endverb
  774. \verb{url}
  775. \verb https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf2_detection_zoo.md
  776. \endverb
  777. \endentry
  778. \entry{Goodfellow.2018}{book}{}
  779. \name{author}{3}{}{%
  780. {{hash=5d2585c11210cf1d4512e6e0a03ec315}{%
  781. family={Goodfellow},
  782. familyi={G\bibinitperiod},
  783. given={Ian},
  784. giveni={I\bibinitperiod}}}%
  785. {{hash=40a8e4774982146adc2688546f54efb2}{%
  786. family={Bengio},
  787. familyi={B\bibinitperiod},
  788. given={Yoshua},
  789. giveni={Y\bibinitperiod}}}%
  790. {{hash=ccec1ccd2e1aa86960eb2e872c6b7020}{%
  791. family={Courville},
  792. familyi={C\bibinitperiod},
  793. given={Aaron},
  794. giveni={A\bibinitperiod}}}%
  795. }
  796. \list{language}{1}{%
  797. {ger}%
  798. }
  799. \list{location}{1}{%
  800. {Frechen}%
  801. }
  802. \list{publisher}{1}{%
  803. {mitp}%
  804. }
  805. \strng{namehash}{3ae53fe582e8a815b118d26947eaa326}
  806. \strng{fullhash}{3ae53fe582e8a815b118d26947eaa326}
  807. \strng{bibnamehash}{3ae53fe582e8a815b118d26947eaa326}
  808. \strng{authorbibnamehash}{3ae53fe582e8a815b118d26947eaa326}
  809. \strng{authornamehash}{3ae53fe582e8a815b118d26947eaa326}
  810. \strng{authorfullhash}{3ae53fe582e8a815b118d26947eaa326}
  811. \field{sortinit}{G}
  812. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  813. \field{labelnamesource}{author}
  814. \field{labeltitlesource}{title}
  815. \field{abstract}{Intro -- Impressum -- Website zum Buch -- Danksagung -- Über die Fachkorrektoren zur deutschen Ausgabe -- Notation -- Einleitung -- Für wen ist dieses Buch gedacht? -- Historische Entwicklungen im Deep Learning -- I Angewandte Mathematik und Grundlagen für das Machine Learning -- Lineare Algebra -- Skalare, Vektoren, Matrizen und Tensoren -- Multiplizieren von Matrizen und Vektoren -- Einheits- und Umkehrmatrizen -- Lineare Abhängigkeit und lineare Hülle -- Normen -- Spezielle Matrizen und Vektoren -- Eigenwertzerlegung -- Singulärwertzerlegung -- Die Moore-Penrose-Pseudoinverse -- Der Spuroperator -- Die Determinante -- Beispiel: Hauptkomponentenanalyse -- Wahrscheinlichkeits- und Informationstheorie -- Warum Wahrscheinlichkeit? -- Zufallsvariablen -- Wahrscheinlichkeitsverteilungen -- Randwahrscheinlichkeit -- Bedingte Wahrscheinlichkeit -- Die Produktregel der bedingten Wahrscheinlichkeiten -- Unabhängigkeit und bedingte Unabhängigkeit -- Erwartungswert, Varianz und Kovarianz -- Häufig genutzte Wahrscheinlichkeitsverteilungen -- Nützliche Eigenschaften häufig verwendeter Funktionen -- Satz von Bayes -- Technische Einzelheiten stetiger Variablen -- Informationstheorie -- Strukturierte probabilistische Modelle -- Numerische Berechnung -- Überlauf und Unterlauf -- Schlechte Konditionierung -- Optimierung auf Gradientenbasis -- Optimierung unter Nebenbedingungen -- Beispiel: Lineare kleinste Quadrate -- Grundlagen für das Machine Learning -- Lernalgorithmen -- Kapazität, Überanpassung und Unteranpassung -- Hyperparameter und Validierungsdaten -- Schätzer, Verzerrung und Varianz -- Maximum-Likelihood-Schätzung -- Bayessche Statistik -- Algorithmen für überwachtes Lernen -- Algorithmen für unüberwachtes Lernen -- Stochastisches Gradientenabstiegsverfahren -- Entwickeln eines Machine-Learning-Algorithmus -- Probleme, an denen Deep Learning wächst.}
  816. \field{edition}{1st edition}
  817. \field{isbn}{978-3-95845-700-3}
  818. \field{note}{Goodfellow, Ian (VerfasserIn) Bengio, Yoshua (VerfasserIn) Courville, Aaron (VerfasserIn) Lenz, Guido (ÜbersetzerIn)}
  819. \field{pagetotal}{883}
  820. \field{series}{mitp Professional}
  821. \field{subtitle}{Das umfassende Handbuch : Grundlagen, aktuelle Verfahren und Algorithmen, neue Forschungsansätze}
  822. \field{title}{Deep Learning}
  823. \field{year}{2018}
  824. \verb{urlraw}
  825. \verb https://ebookcentral.proquest.com/lib/kxp/detail.action?docID=5598176
  826. \endverb
  827. \verb{url}
  828. \verb https://ebookcentral.proquest.com/lib/kxp/detail.action?docID=5598176
  829. \endverb
  830. \keyw{Electronic books;Machine learning}
  831. \endentry
  832. \entry{Colab.1072021}{online}{}
  833. \field{sortinit}{G}
  834. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  835. \field{labeltitlesource}{title}
  836. \field{title}{Google Colaboratory}
  837. \field{urlday}{7}
  838. \field{urlmonth}{10}
  839. \field{urlyear}{2021}
  840. \field{year}{2021}
  841. \field{urldateera}{ce}
  842. \verb{urlraw}
  843. \verb https://colab.research.google.com
  844. \endverb
  845. \verb{url}
  846. \verb https://colab.research.google.com
  847. \endverb
  848. \endentry
  849. \entry{Grus.2020}{book}{}
  850. \name{author}{1}{}{%
  851. {{hash=1ffc3eb84c3593aa01ba1d6000796064}{%
  852. family={Grus},
  853. familyi={G\bibinitperiod},
  854. given={Joel},
  855. giveni={J\bibinitperiod}}}%
  856. }
  857. \list{language}{1}{%
  858. {ger}%
  859. }
  860. \list{location}{1}{%
  861. {Heidelberg}%
  862. }
  863. \list{publisher}{2}{%
  864. {O'Reilly}%
  865. {dpunkt.verlag}%
  866. }
  867. \strng{namehash}{1ffc3eb84c3593aa01ba1d6000796064}
  868. \strng{fullhash}{1ffc3eb84c3593aa01ba1d6000796064}
  869. \strng{bibnamehash}{1ffc3eb84c3593aa01ba1d6000796064}
  870. \strng{authorbibnamehash}{1ffc3eb84c3593aa01ba1d6000796064}
  871. \strng{authornamehash}{1ffc3eb84c3593aa01ba1d6000796064}
  872. \strng{authorfullhash}{1ffc3eb84c3593aa01ba1d6000796064}
  873. \field{sortinit}{G}
  874. \field{sortinithash}{32d67eca0634bf53703493fb1090a2e8}
  875. \field{labelnamesource}{author}
  876. \field{labeltitlesource}{title}
  877. \field{edition}{2. Auflage}
  878. \field{isbn}{978-3-96009-123-3}
  879. \field{note}{Grus, Joel (VerfasserIn) Rother, Kristian (ÜbersetzerIn) Demmig, Thomas (ÜbersetzerIn)}
  880. \field{subtitle}{Grundprinzipien der Datenanalyse mit Python}
  881. \field{title}{Einführung in Data Science}
  882. \field{year}{2020}
  883. \verb{urlraw}
  884. \verb https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=2298669
  885. \endverb
  886. \verb{url}
  887. \verb https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=2298669
  888. \endverb
  889. \endentry
  890. \entry{Hatch.2020}{misc}{}
  891. \name{author}{3}{}{%
  892. {{hash=53f0f0debc0c887016671b82104af570}{%
  893. family={Hatch},
  894. familyi={H\bibinitperiod},
  895. given={Kyle},
  896. giveni={K\bibinitperiod}}}%
  897. {{hash=f0f6eec3d08415fb9191176c2693052d}{%
  898. family={Mern},
  899. familyi={M\bibinitperiod},
  900. given={John},
  901. giveni={J\bibinitperiod}}}%
  902. {{hash=e36ad022497396d3ca41ea1594ac09ec}{%
  903. family={Kochenderfer},
  904. familyi={K\bibinitperiod},
  905. given={Mykel},
  906. giveni={M\bibinitperiod}}}%
  907. }
  908. \strng{namehash}{d03c7a124089022fa1670fdc289053d9}
  909. \strng{fullhash}{d03c7a124089022fa1670fdc289053d9}
  910. \strng{bibnamehash}{d03c7a124089022fa1670fdc289053d9}
  911. \strng{authorbibnamehash}{d03c7a124089022fa1670fdc289053d9}
  912. \strng{authornamehash}{d03c7a124089022fa1670fdc289053d9}
  913. \strng{authorfullhash}{d03c7a124089022fa1670fdc289053d9}
  914. \field{sortinit}{H}
  915. \field{sortinithash}{23a3aa7c24e56cfa16945d55545109b5}
  916. \field{labelnamesource}{author}
  917. \field{labeltitlesource}{title}
  918. \field{abstract}{A collision avoidance system based on simple digital cameras would help enable the safe integration of small UAVs into crowded, low-altitude environments. In this work, we present an obstacle avoidance system for small UAVs that uses a monocular camera with a hybrid neural network and path planner controller. The system is comprised of a vision network for estimating depth from camera images, a high-level control network, a collision prediction network, and a contingency policy. This system is evaluated on a simulated UAV navigating an obstacle course in a constrained flight pattern. Results show the proposed system achieves low collision rates while maintaining operationally relevant flight speeds.}
  919. \field{eprinttype}{arXiv}
  920. \field{title}{Obstacle Avoidance Using a Monocular Camera}
  921. \field{year}{2020}
  922. \verb{eprint}
  923. \verb arXiv:2012.01608v2
  924. \endverb
  925. \verb{file}
  926. \verb Obstacle Avoidance Using a Monocular Camera:Attachments/Obstacle Avoidance Using a Monocular Camera.pdf:application/pdf
  927. \endverb
  928. \verb{urlraw}
  929. \verb https://arxiv.org/pdf/2012.01608
  930. \endverb
  931. \verb{url}
  932. \verb https://arxiv.org/pdf/2012.01608
  933. \endverb
  934. \keyw{Artificial Intelligence (cs.AI);Robotics (cs.RO)}
  935. \endentry
  936. \entry{He.2015}{misc}{}
  937. \name{author}{4}{}{%
  938. {{hash=6b4b60e909e78633945f3f9c9dc83e01}{%
  939. family={He},
  940. familyi={H\bibinitperiod},
  941. given={Kaiming},
  942. giveni={K\bibinitperiod}}}%
  943. {{hash=5e72bc22dbcf0984c6d113d280e36990}{%
  944. family={Zhang},
  945. familyi={Z\bibinitperiod},
  946. given={Xiangyu},
  947. giveni={X\bibinitperiod}}}%
  948. {{hash=bb295293acacd54387339079ebbe4ead}{%
  949. family={Ren},
  950. familyi={R\bibinitperiod},
  951. given={Shaoqing},
  952. giveni={S\bibinitperiod}}}%
  953. {{hash=f85751488058842b5777c7b4074077b5}{%
  954. family={Sun},
  955. familyi={S\bibinitperiod},
  956. given={Jian},
  957. giveni={J\bibinitperiod}}}%
  958. }
  959. \strng{namehash}{6edb98fe38401d2fe4a026f5ce6e8451}
  960. \strng{fullhash}{42c4b52dc3a62cebabbc11c73e1afb53}
  961. \strng{bibnamehash}{6edb98fe38401d2fe4a026f5ce6e8451}
  962. \strng{authorbibnamehash}{6edb98fe38401d2fe4a026f5ce6e8451}
  963. \strng{authornamehash}{6edb98fe38401d2fe4a026f5ce6e8451}
  964. \strng{authorfullhash}{42c4b52dc3a62cebabbc11c73e1afb53}
  965. \field{sortinit}{H}
  966. \field{sortinithash}{23a3aa7c24e56cfa16945d55545109b5}
  967. \field{labelnamesource}{author}
  968. \field{labeltitlesource}{title}
  969. \field{abstract}{Deeper neural networks are more difficult to train. We present a residual learning framework to ease the training of networks that are substantially deeper than those used previously. We explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. We provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset we evaluate residual nets with a depth of up to 152 layers---8x deeper than VGG nets but still having lower complexity. An ensemble of these residual nets achieves 3.57{\%} error on the ImageNet test set. This result won the 1st place on the ILSVRC 2015 classification task. We also present analysis on CIFAR-10 with 100 and 1000 layers. The depth of representations is of central importance for many visual recognition tasks. Solely due to our extremely deep representations, we obtain a 28{\%} relative improvement on the COCO object detection dataset. Deep residual nets are foundations of our submissions to ILSVRC {\&} COCO 2015 competitions, where we also won the 1st places on the tasks of ImageNet detection, ImageNet localization, COCO detection, and COCO segmentation.}
  970. \field{eprinttype}{arXiv}
  971. \field{title}{Deep Residual Learning for Image Recognition}
  972. \field{year}{2015}
  973. \verb{eprint}
  974. \verb arXiv:1512.03385v1
  975. \endverb
  976. \verb{file}
  977. \verb Deep Residual Learning for Image Recognition:Attachments/Deep Residual Learning for Image Recognition.pdf:application/pdf
  978. \endverb
  979. \verb{urlraw}
  980. \verb https://arxiv.org/pdf/1512.03385
  981. \endverb
  982. \verb{url}
  983. \verb https://arxiv.org/pdf/1512.03385
  984. \endverb
  985. \keyw{Computer Vision and Pattern Recognition (cs.CV)}
  986. \endentry
  987. \entry{HenriqueF.DeArruda.2019}{misc}{}
  988. \name{author}{4}{}{%
  989. {{hash=ee77957f6c8d1d0d6a9b57758027490b}{%
  990. family={{Henrique F. De Arruda}},
  991. familyi={H\bibinitperiod}}}%
  992. {{hash=1363d711e28a512ad7ec78e2b51689e5}{%
  993. family={{Alexandre Benatti}},
  994. familyi={A\bibinitperiod}}}%
  995. {{hash=056e0bc67b5bf46e849693b793631d1e}{%
  996. family={{César Henrique Comin}},
  997. familyi={C\bibinitperiod}}}%
  998. {{hash=df6db75b50bc38ba95ee9c4390dad80d}{%
  999. family={{Luciano Da F. Costa}},
  1000. familyi={L\bibinitperiod}}}%
  1001. }
  1002. \list{language}{1}{%
  1003. {en}%
  1004. }
  1005. \list{publisher}{1}{%
  1006. {Unpublished}%
  1007. }
  1008. \strng{namehash}{e503b9e1abe8168fc002d3288d0766aa}
  1009. \strng{fullhash}{b47eec58a7740fad75b1c18f6179c390}
  1010. \strng{bibnamehash}{e503b9e1abe8168fc002d3288d0766aa}
  1011. \strng{authorbibnamehash}{e503b9e1abe8168fc002d3288d0766aa}
  1012. \strng{authornamehash}{e503b9e1abe8168fc002d3288d0766aa}
  1013. \strng{authorfullhash}{b47eec58a7740fad75b1c18f6179c390}
  1014. \field{sortinit}{H}
  1015. \field{sortinithash}{23a3aa7c24e56cfa16945d55545109b5}
  1016. \field{labelnamesource}{author}
  1017. \field{labeltitlesource}{title}
  1018. \field{title}{Learning Deep Learning (CDT-15)}
  1019. \field{year}{2019}
  1020. \verb{doi}
  1021. \verb 10.13140/RG.2.2.29866.57283
  1022. \endverb
  1023. \endentry
  1024. \entry{Hosang.2017}{misc}{}
  1025. \name{author}{3}{}{%
  1026. {{hash=8192e2123af03bcb274c2e9444e0a153}{%
  1027. family={Hosang},
  1028. familyi={H\bibinitperiod},
  1029. given={Jan},
  1030. giveni={J\bibinitperiod}}}%
  1031. {{hash=b85b9efefe92bdad29985847fb978d46}{%
  1032. family={Benenson},
  1033. familyi={B\bibinitperiod},
  1034. given={Rodrigo},
  1035. giveni={R\bibinitperiod}}}%
  1036. {{hash=457dd3049a60059cf69e775a47f0c154}{%
  1037. family={Schiele},
  1038. familyi={S\bibinitperiod},
  1039. given={Bernt},
  1040. giveni={B\bibinitperiod}}}%
  1041. }
  1042. \strng{namehash}{a5b31c01e0efdee6397e46311b720a59}
  1043. \strng{fullhash}{a5b31c01e0efdee6397e46311b720a59}
  1044. \strng{bibnamehash}{a5b31c01e0efdee6397e46311b720a59}
  1045. \strng{authorbibnamehash}{a5b31c01e0efdee6397e46311b720a59}
  1046. \strng{authornamehash}{a5b31c01e0efdee6397e46311b720a59}
  1047. \strng{authorfullhash}{a5b31c01e0efdee6397e46311b720a59}
  1048. \field{sortinit}{H}
  1049. \field{sortinithash}{23a3aa7c24e56cfa16945d55545109b5}
  1050. \field{labelnamesource}{author}
  1051. \field{labeltitlesource}{title}
  1052. \field{abstract}{Object detectors have hugely profited from moving towards an end-to-end learning paradigm: proposals, features, and the classifier becoming one neural network improved results two-fold on general object detection. One indispensable component is non-maximum suppression (NMS), a post-processing algorithm responsible for merging all detections that belong to the same object. The de facto standard NMS algorithm is still fully hand-crafted, suspiciously simple, and -- being based on greedy clustering with a fixed distance threshold -- forces a trade-off between recall and precision. We propose a new network architecture designed to perform NMS, using only boxes and their score. We report experiments for person detection on PETS and for general object categories on the COCO dataset. Our approach shows promise providing improved localization and occlusion handling.}
  1053. \field{eprinttype}{arXiv}
  1054. \field{title}{Learning non-maximum suppression}
  1055. \field{year}{2017}
  1056. \verb{eprint}
  1057. \verb arXiv:1705.02950v2
  1058. \endverb
  1059. \verb{file}
  1060. \verb Learning non-maximum suppression:Attachments/Learning non-maximum suppression.pdf:application/pdf
  1061. \endverb
  1062. \verb{urlraw}
  1063. \verb https://arxiv.org/pdf/1705.02950
  1064. \endverb
  1065. \verb{url}
  1066. \verb https://arxiv.org/pdf/1705.02950
  1067. \endverb
  1068. \keyw{Computer Vision and Pattern Recognition (cs.CV)}
  1069. \endentry
  1070. \entry{Howard.2017}{misc}{}
  1071. \name{author}{8}{}{%
  1072. {{hash=0cedb03f907400fc304fdfaa1f7e2085}{%
  1073. family={Howard},
  1074. familyi={H\bibinitperiod},
  1075. given={Andrew\bibnamedelima G.},
  1076. giveni={A\bibinitperiod\bibinitdelim G\bibinitperiod}}}%
  1077. {{hash=d767e8e4d733bcf728bcdf2c193462f7}{%
  1078. family={Zhu},
  1079. familyi={Z\bibinitperiod},
  1080. given={Menglong},
  1081. giveni={M\bibinitperiod}}}%
  1082. {{hash=31960f03389184b7f052f5b197cc9fdf}{%
  1083. family={Chen},
  1084. familyi={C\bibinitperiod},
  1085. given={Bo},
  1086. giveni={B\bibinitperiod}}}%
  1087. {{hash=6cbb997a11c6922af719c32863261918}{%
  1088. family={Kalenichenko},
  1089. familyi={K\bibinitperiod},
  1090. given={Dmitry},
  1091. giveni={D\bibinitperiod}}}%
  1092. {{hash=47ad65c82b1de7d642988df185d7d8ea}{%
  1093. family={Wang},
  1094. familyi={W\bibinitperiod},
  1095. given={Weijun},
  1096. giveni={W\bibinitperiod}}}%
  1097. {{hash=7dcb9c6d4d4251a7e32b502d03c9354b}{%
  1098. family={Weyand},
  1099. familyi={W\bibinitperiod},
  1100. given={Tobias},
  1101. giveni={T\bibinitperiod}}}%
  1102. {{hash=8f221f2afb0b5a3d95b3e97101924922}{%
  1103. family={Andreetto},
  1104. familyi={A\bibinitperiod},
  1105. given={Marco},
  1106. giveni={M\bibinitperiod}}}%
  1107. {{hash=c707ec5b5997dc408a14a34a8380166c}{%
  1108. family={Adam},
  1109. familyi={A\bibinitperiod},
  1110. given={Hartwig},
  1111. giveni={H\bibinitperiod}}}%
  1112. }
  1113. \strng{namehash}{e1fc6cab9b6009340e110518e53868c4}
  1114. \strng{fullhash}{cffcf38c642164887a370768f5701b8e}
  1115. \strng{bibnamehash}{e1fc6cab9b6009340e110518e53868c4}
  1116. \strng{authorbibnamehash}{e1fc6cab9b6009340e110518e53868c4}
  1117. \strng{authornamehash}{e1fc6cab9b6009340e110518e53868c4}
  1118. \strng{authorfullhash}{cffcf38c642164887a370768f5701b8e}
  1119. \field{sortinit}{H}
  1120. \field{sortinithash}{23a3aa7c24e56cfa16945d55545109b5}
  1121. \field{labelnamesource}{author}
  1122. \field{labeltitlesource}{title}
  1123. \field{abstract}{We present a class of efficient models called MobileNets for mobile and embedded vision applications. MobileNets are based on a streamlined architecture that uses depth-wise separable convolutions to build light weight deep neural networks. We introduce two simple global hyper-parameters that efficiently trade off between latency and accuracy. These hyper-parameters allow the model builder to choose the right sized model for their application based on the constraints of the problem. We present extensive experiments on resource and accuracy tradeoffs and show strong performance compared to other popular models on ImageNet classification. We then demonstrate the effectiveness of MobileNets across a wide range of applications and use cases including object detection, finegrain classification, face attributes and large scale geo-localization.}
  1124. \field{eprinttype}{arXiv}
  1125. \field{title}{MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications}
  1126. \field{year}{2017}
  1127. \verb{eprint}
  1128. \verb arXiv:1704.04861v1
  1129. \endverb
  1130. \verb{file}
  1131. \verb MobileNets Efficient Convolutional Neural Networks for M:Attachments/MobileNets Efficient Convolutional Neural Networks for M.pdf:application/pdf
  1132. \endverb
  1133. \verb{urlraw}
  1134. \verb https://arxiv.org/pdf/1704.04861
  1135. \endverb
  1136. \verb{url}
  1137. \verb https://arxiv.org/pdf/1704.04861
  1138. \endverb
  1139. \keyw{Computer Vision and Pattern Recognition (cs.CV)}
  1140. \endentry
  1141. \entry{Jacob.12152017}{misc}{}
  1142. \name{author}{8}{}{%
  1143. {{hash=73f11c7544252601f7e8c5f614c660f2}{%
  1144. family={Jacob},
  1145. familyi={J\bibinitperiod},
  1146. given={Benoit},
  1147. giveni={B\bibinitperiod}}}%
  1148. {{hash=aa58f2ae95227046f2fbba114f0fb625}{%
  1149. family={Kligys},
  1150. familyi={K\bibinitperiod},
  1151. given={Skirmantas},
  1152. giveni={S\bibinitperiod}}}%
  1153. {{hash=31960f03389184b7f052f5b197cc9fdf}{%
  1154. family={Chen},
  1155. familyi={C\bibinitperiod},
  1156. given={Bo},
  1157. giveni={B\bibinitperiod}}}%
  1158. {{hash=d767e8e4d733bcf728bcdf2c193462f7}{%
  1159. family={Zhu},
  1160. familyi={Z\bibinitperiod},
  1161. given={Menglong},
  1162. giveni={M\bibinitperiod}}}%
  1163. {{hash=436cfd289182e9d6b0a46d79b8e91b6e}{%
  1164. family={Tang},
  1165. familyi={T\bibinitperiod},
  1166. given={Matthew},
  1167. giveni={M\bibinitperiod}}}%
  1168. {{hash=315c4166fc1f7cb66324a7f0d82827cd}{%
  1169. family={Howard},
  1170. familyi={H\bibinitperiod},
  1171. given={Andrew},
  1172. giveni={A\bibinitperiod}}}%
  1173. {{hash=c707ec5b5997dc408a14a34a8380166c}{%
  1174. family={Adam},
  1175. familyi={A\bibinitperiod},
  1176. given={Hartwig},
  1177. giveni={H\bibinitperiod}}}%
  1178. {{hash=6cbb997a11c6922af719c32863261918}{%
  1179. family={Kalenichenko},
  1180. familyi={K\bibinitperiod},
  1181. given={Dmitry},
  1182. giveni={D\bibinitperiod}}}%
  1183. }
  1184. \strng{namehash}{be1e8607db7344f1c8e395537c88d55a}
  1185. \strng{fullhash}{3ad38e526037fc692f33aef3160b0293}
  1186. \strng{bibnamehash}{be1e8607db7344f1c8e395537c88d55a}
  1187. \strng{authorbibnamehash}{be1e8607db7344f1c8e395537c88d55a}
  1188. \strng{authornamehash}{be1e8607db7344f1c8e395537c88d55a}
  1189. \strng{authorfullhash}{3ad38e526037fc692f33aef3160b0293}
  1190. \field{sortinit}{J}
  1191. \field{sortinithash}{b2f54a9081ace9966a7cb9413811edb4}
  1192. \field{labelnamesource}{author}
  1193. \field{labeltitlesource}{title}
  1194. \field{abstract}{The rising popularity of intelligent mobile devices and the daunting computational cost of deep learning-based models call for efficient and accurate on-device inference schemes. We propose a quantization scheme that allows inference to be carried out using integer-only arithmetic, which can be implemented more efficiently than floating point inference on commonly available integer-only hardware. We also co-design a training procedure to preserve end-to-end model accuracy post quantization. As a result, the proposed quantization scheme improves the tradeoff between accuracy and on-device latency. The improvements are significant even on MobileNets, a model family known for run-time efficiency, and are demonstrated in ImageNet classification and COCO detection on popular CPUs.}
  1195. \field{eprinttype}{arXiv}
  1196. \field{note}{14 pages, 12 figures}
  1197. \field{pagetotal}{14}
  1198. \field{title}{Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference}
  1199. \field{year}{2017}
  1200. \verb{eprint}
  1201. \verb arXiv:1712.05877v1
  1202. \endverb
  1203. \verb{urlraw}
  1204. \verb https://arxiv.org/pdf/1712.05877
  1205. \endverb
  1206. \verb{url}
  1207. \verb https://arxiv.org/pdf/1712.05877
  1208. \endverb
  1209. \keyw{Computer Science - Learning;Machine Learning (cs.LG);Machine Learning (stat.ML);Statistics - Machine Learning}
  1210. \endentry
  1211. \entry{Kim.11152015}{misc}{}
  1212. \name{author}{2}{}{%
  1213. {{hash=8095c82e8faeee6c563725c9a385ec5b}{%
  1214. family={Kim},
  1215. familyi={K\bibinitperiod},
  1216. given={Dong\bibnamedelima Ki},
  1217. giveni={D\bibinitperiod\bibinitdelim K\bibinitperiod}}}%
  1218. {{hash=770b8c88e0b93629ab1d35372bf0e3e4}{%
  1219. family={Chen},
  1220. familyi={C\bibinitperiod},
  1221. given={Tsuhan},
  1222. giveni={T\bibinitperiod}}}%
  1223. }
  1224. \strng{namehash}{ff2628bfa346d4778c90142904849c77}
  1225. \strng{fullhash}{ff2628bfa346d4778c90142904849c77}
  1226. \strng{bibnamehash}{ff2628bfa346d4778c90142904849c77}
  1227. \strng{authorbibnamehash}{ff2628bfa346d4778c90142904849c77}
  1228. \strng{authornamehash}{ff2628bfa346d4778c90142904849c77}
  1229. \strng{authorfullhash}{ff2628bfa346d4778c90142904849c77}
  1230. \field{sortinit}{K}
  1231. \field{sortinithash}{c02bf6bff1c488450c352b40f5d853ab}
  1232. \field{labelnamesource}{author}
  1233. \field{labeltitlesource}{title}
  1234. \field{abstract}{Autonomous indoor navigation of Micro Aerial Vehicles (MAVs) possesses many challenges. One main reason is that GPS has limited precision in indoor environments. The additional fact that MAVs are not able to carry heavy weight or power consuming sensors, such as range finders, makes indoor autonomous navigation a challenging task. In this paper, we propose a practical system in which a quadcopter autonomously navigates indoors and finds a specific target, i.e., a book bag, by using a single camera. A deep learning model, Convolutional Neural Network (ConvNet), is used to learn a controller strategy that mimics an expert pilot's choice of action. We show our system's performance through real-time experiments in diverse indoor locations. To understand more about our trained network, we use several visualization techniques.}
  1235. \field{eprinttype}{arXiv}
  1236. \field{title}{Deep Neural Network for Real-Time Autonomous Indoor Navigation}
  1237. \field{year}{2015}
  1238. \verb{eprint}
  1239. \verb arXiv:1511.04668v2
  1240. \endverb
  1241. \verb{file}
  1242. \verb Deep Neural Network for Real-Time Autonomous Indoor Navi:Attachments/Deep Neural Network for Real-Time Autonomous Indoor Navi.pdf:application/pdf
  1243. \endverb
  1244. \verb{urlraw}
  1245. \verb https://arxiv.org/pdf/1511.04668
  1246. \endverb
  1247. \verb{url}
  1248. \verb https://arxiv.org/pdf/1511.04668
  1249. \endverb
  1250. \keyw{Computer Science - Computer Vision and Pattern Recognition;Computer Vision and Pattern Recognition (cs.CV)}
  1251. \endentry
  1252. \entry{Krishnamoorthi.2018}{misc}{}
  1253. \name{author}{1}{}{%
  1254. {{hash=4a691df37485846988162c4f0639d2f0}{%
  1255. family={Krishnamoorthi},
  1256. familyi={K\bibinitperiod},
  1257. given={Raghuraman},
  1258. giveni={R\bibinitperiod}}}%
  1259. }
  1260. \strng{namehash}{4a691df37485846988162c4f0639d2f0}
  1261. \strng{fullhash}{4a691df37485846988162c4f0639d2f0}
  1262. \strng{bibnamehash}{4a691df37485846988162c4f0639d2f0}
  1263. \strng{authorbibnamehash}{4a691df37485846988162c4f0639d2f0}
  1264. \strng{authornamehash}{4a691df37485846988162c4f0639d2f0}
  1265. \strng{authorfullhash}{4a691df37485846988162c4f0639d2f0}
  1266. \field{sortinit}{K}
  1267. \field{sortinithash}{c02bf6bff1c488450c352b40f5d853ab}
  1268. \field{labelnamesource}{author}
  1269. \field{labeltitlesource}{title}
  1270. \field{abstract}{We present an overview of techniques for quantizing convolutional neural networks for inference with integer weights and activations. Per-channel quantization of weights and per-layer quantization of activations to 8-bits of precision post-training produces classification accuracies within 2{\%} of floating point networks for a wide variety of CNN architectures. Model sizes can be reduced by a factor of 4 by quantizing weights to 8-bits, even when 8-bit arithmetic is not supported. This can be achieved with simple, post training quantization of weights.We benchmark latencies of quantized networks on CPUs and DSPs and observe a speedup of 2x-3x for quantized implementations compared to floating point on CPUs. Speedups of up to 10x are observed on specialized processors with fixed point SIMD capabilities, like the Qualcomm QDSPs with HVX. Quantization-aware training can provide further improvements, reducing the gap to floating point to 1{\%} at 8-bit precision. Quantization-aware training also allows for reducing the precision of weights to four bits with accuracy losses ranging from 2{\%} to 10{\%}, with higher accuracy drop for smaller networks.We introduce tools in TensorFlow and TensorFlowLite for quantizing convolutional networks and review best practices for quantization-aware training to obtain high accuracy with quantized weights and activations. We recommend that per-channel quantization of weights and per-layer quantization of activations be the preferred quantization scheme for hardware acceleration and kernel optimization. We also propose that future processors and hardware accelerators for optimized inference support precisions of 4, 8 and 16 bits.}
  1271. \field{eprinttype}{arXiv}
  1272. \field{title}{Quantizing deep convolutional networks for efficient inference: A whitepaper}
  1273. \field{year}{2018}
  1274. \verb{eprint}
  1275. \verb arXiv:1806.08342v1
  1276. \endverb
  1277. \verb{file}
  1278. \verb Quantizing deep convolutional networks for efficient inf:Attachments/Quantizing deep convolutional networks for efficient inf.pdf:application/pdf
  1279. \endverb
  1280. \verb{urlraw}
  1281. \verb https://arxiv.org/pdf/1806.08342
  1282. \endverb
  1283. \verb{url}
  1284. \verb https://arxiv.org/pdf/1806.08342
  1285. \endverb
  1286. \keyw{Computer Vision and Pattern Recognition (cs.CV);Machine Learning (cs.LG);Machine Learning (stat.ML)}
  1287. \endentry
  1288. \entry{Lin.1292016}{misc}{}
  1289. \name{author}{6}{}{%
  1290. {{hash=08f925fe4692d130a1d7cb7d94483351}{%
  1291. family={Lin},
  1292. familyi={L\bibinitperiod},
  1293. given={Tsung-Yi},
  1294. giveni={T\bibinithyphendelim Y\bibinitperiod}}}%
  1295. {{hash=ecd149fdcb3e0503881d49e545744c3d}{%
  1296. family={Dollár},
  1297. familyi={D\bibinitperiod},
  1298. given={Piotr},
  1299. giveni={P\bibinitperiod}}}%
  1300. {{hash=bd5dadbe57bedc5957c19a3154c4d424}{%
  1301. family={Girshick},
  1302. familyi={G\bibinitperiod},
  1303. given={Ross},
  1304. giveni={R\bibinitperiod}}}%
  1305. {{hash=6b4b60e909e78633945f3f9c9dc83e01}{%
  1306. family={He},
  1307. familyi={H\bibinitperiod},
  1308. given={Kaiming},
  1309. giveni={K\bibinitperiod}}}%
  1310. {{hash=9cadf8ca1dd9b5f5a830ba2059b2517f}{%
  1311. family={Hariharan},
  1312. familyi={H\bibinitperiod},
  1313. given={Bharath},
  1314. giveni={B\bibinitperiod}}}%
  1315. {{hash=044d1db5259b74e4975282f599d8e767}{%
  1316. family={Belongie},
  1317. familyi={B\bibinitperiod},
  1318. given={Serge},
  1319. giveni={S\bibinitperiod}}}%
  1320. }
  1321. \strng{namehash}{3b7e00bc88d7aca4d4d5ddb8c252e36c}
  1322. \strng{fullhash}{407b643cc48d7ba78686be05e508bfa5}
  1323. \strng{bibnamehash}{3b7e00bc88d7aca4d4d5ddb8c252e36c}
  1324. \strng{authorbibnamehash}{3b7e00bc88d7aca4d4d5ddb8c252e36c}
  1325. \strng{authornamehash}{3b7e00bc88d7aca4d4d5ddb8c252e36c}
  1326. \strng{authorfullhash}{407b643cc48d7ba78686be05e508bfa5}
  1327. \field{sortinit}{L}
  1328. \field{sortinithash}{7c47d417cecb1f4bd38d1825c427a61a}
  1329. \field{labelnamesource}{author}
  1330. \field{labeltitlesource}{title}
  1331. \field{abstract}{Feature pyramids are a basic component in recognition systems for detecting objects at different scales. But recent deep learning object detectors have avoided pyramid representations, in part because they are compute and memory intensive. In this paper, we exploit the inherent multi-scale, pyramidal hierarchy of deep convolutional networks to construct feature pyramids with marginal extra cost. A top-down architecture with lateral connections is developed for building high-level semantic feature maps at all scales. This architecture, called a Feature Pyramid Network (FPN), shows significant improvement as a generic feature extractor in several applications. Using FPN in a basic Faster R-CNN system, our method achieves state-of-the-art single-model results on the COCO detection benchmark without bells and whistles, surpassing all existing single-model entries including those from the COCO 2016 challenge winners. In addition, our method can run at 5 FPS on a GPU and thus is a practical and accurate solution to multi-scale object detection. Code will be made publicly available.}
  1332. \field{eprinttype}{arXiv}
  1333. \field{title}{Feature Pyramid Networks for Object Detection}
  1334. \field{year}{2016}
  1335. \verb{eprint}
  1336. \verb arXiv:1612.03144v2
  1337. \endverb
  1338. \verb{file}
  1339. \verb Feature Pyramid Networks for Object Detection:Attachments/Feature Pyramid Networks for Object Detection.pdf:application/pdf
  1340. \endverb
  1341. \verb{urlraw}
  1342. \verb https://arxiv.org/pdf/1612.03144
  1343. \endverb
  1344. \verb{url}
  1345. \verb https://arxiv.org/pdf/1612.03144
  1346. \endverb
  1347. \keyw{Computer Science - Computer Vision and Pattern Recognition;Computer Vision and Pattern Recognition (cs.CV)}
  1348. \endentry
  1349. \entry{Liu.2017}{article}{}
  1350. \name{author}{6}{}{%
  1351. {{hash=10693ec37452f38e6c7e7ce45d8c429d}{%
  1352. family={Liu},
  1353. familyi={L\bibinitperiod},
  1354. given={Canglong},
  1355. giveni={C\bibinitperiod}}}%
  1356. {{hash=5b93b2c78389baaa078de347beede455}{%
  1357. family={Zheng},
  1358. familyi={Z\bibinitperiod},
  1359. given={Bin},
  1360. giveni={B\bibinitperiod}}}%
  1361. {{hash=3682934996fca823c7636639de0dcadb}{%
  1362. family={Wang},
  1363. familyi={W\bibinitperiod},
  1364. given={Chunyang},
  1365. giveni={C\bibinitperiod}}}%
  1366. {{hash=23cde46539de12057c4db04a3b4f0dbd}{%
  1367. family={Zhao},
  1368. familyi={Z\bibinitperiod},
  1369. given={Yongting},
  1370. giveni={Y\bibinitperiod}}}%
  1371. {{hash=1ad0a7a54bb4b8674aa795fef3f0a895}{%
  1372. family={Fu},
  1373. familyi={F\bibinitperiod},
  1374. given={Shun},
  1375. giveni={S\bibinitperiod}}}%
  1376. {{hash=37f7c16bc002b2fe9220ed6c15d35cdf}{%
  1377. family={Li},
  1378. familyi={L\bibinitperiod},
  1379. given={Haochen},
  1380. giveni={H\bibinitperiod}}}%
  1381. }
  1382. \strng{namehash}{a4bd7c4ed3a987468fe9d36a58d4e4f3}
  1383. \strng{fullhash}{1902e101876f8b68bcd0eb255993e0fe}
  1384. \strng{bibnamehash}{a4bd7c4ed3a987468fe9d36a58d4e4f3}
  1385. \strng{authorbibnamehash}{a4bd7c4ed3a987468fe9d36a58d4e4f3}
  1386. \strng{authornamehash}{a4bd7c4ed3a987468fe9d36a58d4e4f3}
  1387. \strng{authorfullhash}{1902e101876f8b68bcd0eb255993e0fe}
  1388. \field{extraname}{1}
  1389. \field{sortinit}{L}
  1390. \field{sortinithash}{7c47d417cecb1f4bd38d1825c427a61a}
  1391. \field{labelnamesource}{author}
  1392. \field{labeltitlesource}{title}
  1393. \field{journaltitle}{MATEC Web of Conferences}
  1394. \field{note}{Xu, Bing (Editor) Chen, Yinong (Editor) PII: matecconf{\_}icmite2017{\_}00007}
  1395. \field{pagination}{page}
  1396. \field{shortjournal}{MATEC Web Conf.}
  1397. \field{title}{CNN-Based Vision Model for Obstacle Avoidance of Mobile Robot}
  1398. \field{volume}{139}
  1399. \field{year}{2017}
  1400. \field{pages}{00007}
  1401. \range{pages}{1}
  1402. \verb{doi}
  1403. \verb 10.1051/matecconf/201713900007
  1404. \endverb
  1405. \verb{file}
  1406. \verb CNN-Based Vision Model for Obstacle Avoidance of Mobile Robot:Attachments/CNN-Based Vision Model for Obstacle Avoidance of Mobile Robot.pdf:application/pdf
  1407. \endverb
  1408. \endentry
  1409. \entry{Liu.2015}{misc}{}
  1410. \name{author}{7}{}{%
  1411. {{hash=c0e0d23e2d09e45e6f51cc2bcea6d9f9}{%
  1412. family={Liu},
  1413. familyi={L\bibinitperiod},
  1414. given={Wei},
  1415. giveni={W\bibinitperiod}}}%
  1416. {{hash=c1826f3465579186aff299a9b0e16ed7}{%
  1417. family={Anguelov},
  1418. familyi={A\bibinitperiod},
  1419. given={Dragomir},
  1420. giveni={D\bibinitperiod}}}%
  1421. {{hash=8bbc4c5d96f205bada839e74e0202146}{%
  1422. family={Erhan},
  1423. familyi={E\bibinitperiod},
  1424. given={Dumitru},
  1425. giveni={D\bibinitperiod}}}%
  1426. {{hash=ed568d9c3bb059e6bf22899fbf170f86}{%
  1427. family={Szegedy},
  1428. familyi={S\bibinitperiod},
  1429. given={Christian},
  1430. giveni={C\bibinitperiod}}}%
  1431. {{hash=698ee61a2f3fa29734204496d2d36aef}{%
  1432. family={Reed},
  1433. familyi={R\bibinitperiod},
  1434. given={Scott},
  1435. giveni={S\bibinitperiod}}}%
  1436. {{hash=ec820780d594e36d11c6e30c7a2614e0}{%
  1437. family={Fu},
  1438. familyi={F\bibinitperiod},
  1439. given={Cheng-Yang},
  1440. giveni={C\bibinithyphendelim Y\bibinitperiod}}}%
  1441. {{hash=963e9b2526a7150c418b4e9e9d19a82f}{%
  1442. family={Berg},
  1443. familyi={B\bibinitperiod},
  1444. given={Alexander\bibnamedelima C.},
  1445. giveni={A\bibinitperiod\bibinitdelim C\bibinitperiod}}}%
  1446. }
  1447. \strng{namehash}{623b50f45b666c1e9b84e5228e255810}
  1448. \strng{fullhash}{d5e1ac3dfe6687980f91e701611520ad}
  1449. \strng{bibnamehash}{623b50f45b666c1e9b84e5228e255810}
  1450. \strng{authorbibnamehash}{623b50f45b666c1e9b84e5228e255810}
  1451. \strng{authornamehash}{623b50f45b666c1e9b84e5228e255810}
  1452. \strng{authorfullhash}{d5e1ac3dfe6687980f91e701611520ad}
  1453. \field{extraname}{2}
  1454. \field{sortinit}{L}
  1455. \field{sortinithash}{7c47d417cecb1f4bd38d1825c427a61a}
  1456. \field{labelnamesource}{author}
  1457. \field{labeltitlesource}{title}
  1458. \field{abstract}{We present a method for detecting objects in images using a single deep neural network. Our approach, named SSD, discretizes the output space of bounding boxes into a set of default boxes over different aspect ratios and scales per feature map location. At prediction time, the network generates scores for the presence of each object category in each default box and produces adjustments to the box to better match the object shape. Additionally, the network combines predictions from multiple feature maps with different resolutions to naturally handle objects of various sizes. Our SSD model is simple relative to methods that require object proposals because it completely eliminates proposal generation and subsequent pixel or feature resampling stage and encapsulates all computation in a single network. This makes SSD easy to train and straightforward to integrate into systems that require a detection component. Experimental results on the PASCAL VOC, MS COCO, and ILSVRC datasets confirm that SSD has comparable accuracy to methods that utilize an additional object proposal step and is much faster, while providing a unified framework for both training and inference. Compared to other single stage methods, SSD has much better accuracy, even with a smaller input image size. For 300$\times$300 input, SSD achieves 72.1{\%} mAP on VOC2007 test at 58 FPS on a Nvidia Titan X and for 500$\times$500 input, SSD achieves 75.1{\%} mAP, outperforming a comparable state of the art Faster R-CNN model. Code is available at this https URL .}
  1459. \field{eprinttype}{arXiv}
  1460. \field{title}{SSD: Single Shot MultiBox Detector}
  1461. \field{year}{2015}
  1462. \verb{eprint}
  1463. \verb arXiv:1512.02325v5
  1464. \endverb
  1465. \verb{file}
  1466. \verb SSD Single Shot MultiBox Detector:Attachments/SSD Single Shot MultiBox Detector.pdf:application/pdf
  1467. \endverb
  1468. \verb{urlraw}
  1469. \verb https://arxiv.org/pdf/1512.02325
  1470. \endverb
  1471. \verb{url}
  1472. \verb https://arxiv.org/pdf/1512.02325
  1473. \endverb
  1474. \keyw{Computer Vision and Pattern Recognition (cs.CV)}
  1475. \endentry
  1476. \entry{Macias-GarciaDetection.2020}{inproceedings}{}
  1477. \name{author}{3}{}{%
  1478. {{hash=4174207e9e6591876e802651a493e3e5}{%
  1479. family={Macias-Garcia},
  1480. familyi={M\bibinithyphendelim G\bibinitperiod},
  1481. given={Edgar},
  1482. giveni={E\bibinitperiod}}}%
  1483. {{hash=90085c0512ccbf9d06c3fee7a592a160}{%
  1484. family={Galeana-Perez},
  1485. familyi={G\bibinithyphendelim P\bibinitperiod},
  1486. given={Deysy},
  1487. giveni={D\bibinitperiod}}}%
  1488. {{hash=27756fbb2b1f6458ee3602cf72fad839}{%
  1489. family={Bayro-Corrochano},
  1490. familyi={B\bibinithyphendelim C\bibinitperiod},
  1491. given={Eduardo},
  1492. giveni={E\bibinitperiod}}}%
  1493. }
  1494. \strng{namehash}{51215b919d0a7fbf4f95766cda42e590}
  1495. \strng{fullhash}{51215b919d0a7fbf4f95766cda42e590}
  1496. \strng{bibnamehash}{51215b919d0a7fbf4f95766cda42e590}
  1497. \strng{authorbibnamehash}{51215b919d0a7fbf4f95766cda42e590}
  1498. \strng{authornamehash}{51215b919d0a7fbf4f95766cda42e590}
  1499. \strng{authorfullhash}{51215b919d0a7fbf4f95766cda42e590}
  1500. \field{sortinit}{M}
  1501. \field{sortinithash}{4625c616857f13d17ce56f7d4f97d451}
  1502. \field{labelnamesource}{author}
  1503. \field{labeltitlesource}{title}
  1504. \field{booktitle}{2020 International Joint Conference on Neural Networks (IJCNN)}
  1505. \field{title}{CNN Based Perception System for Collision Avoidance in Mobile Robots using Stereo Vision}
  1506. \field{year}{2020}
  1507. \field{pages}{1\bibrangedash 7}
  1508. \range{pages}{7}
  1509. \verb{doi}
  1510. \verb 10.1109/IJCNN48605.2020.9206747
  1511. \endverb
  1512. \endentry
  1513. \entry{.332021}{online}{}
  1514. \field{sortinit}{M}
  1515. \field{sortinithash}{4625c616857f13d17ce56f7d4f97d451}
  1516. \field{labeltitlesource}{title}
  1517. \field{title}{MNIST handwritten digit database, Yann LeCun, Corinna Cortes and Chris Burges}
  1518. \field{urlday}{20}
  1519. \field{urlmonth}{10}
  1520. \field{urlyear}{2021}
  1521. \field{year}{2021}
  1522. \field{urldateera}{ce}
  1523. \verb{urlraw}
  1524. \verb http://yann.lecun.com/exdb/mnist/
  1525. \endverb
  1526. \verb{url}
  1527. \verb http://yann.lecun.com/exdb/mnist/
  1528. \endverb
  1529. \endentry
  1530. \entry{Nalpantidis.2009}{inproceedings}{}
  1531. \name{author}{3}{}{%
  1532. {{hash=8378a2bedf74d1ff3effafeb0146799c}{%
  1533. family={Nalpantidis},
  1534. familyi={N\bibinitperiod},
  1535. given={Lazaros},
  1536. giveni={L\bibinitperiod}}}%
  1537. {{hash=b9115d66032a5520d69886a7c7dcc101}{%
  1538. family={Kostavelis},
  1539. familyi={K\bibinitperiod},
  1540. given={Ioannis},
  1541. giveni={I\bibinitperiod}}}%
  1542. {{hash=12d89ed7954806488bac04f320f416ec}{%
  1543. family={Gasteratos},
  1544. familyi={G\bibinitperiod},
  1545. given={Antonios},
  1546. giveni={A\bibinitperiod}}}%
  1547. }
  1548. \strng{namehash}{e7c75a0dd778a461b79b06a080decaa1}
  1549. \strng{fullhash}{e7c75a0dd778a461b79b06a080decaa1}
  1550. \strng{bibnamehash}{e7c75a0dd778a461b79b06a080decaa1}
  1551. \strng{authorbibnamehash}{e7c75a0dd778a461b79b06a080decaa1}
  1552. \strng{authornamehash}{e7c75a0dd778a461b79b06a080decaa1}
  1553. \strng{authorfullhash}{e7c75a0dd778a461b79b06a080decaa1}
  1554. \field{sortinit}{N}
  1555. \field{sortinithash}{22369a73d5f88983a108b63f07f37084}
  1556. \field{labelnamesource}{author}
  1557. \field{labeltitlesource}{title}
  1558. \field{isbn}{978-3-642-10816-7}
  1559. \field{month}{12}
  1560. \field{title}{Stereovision-Based Algorithm for Obstacle Avoidance}
  1561. \field{year}{2009}
  1562. \field{pages}{195\bibrangedash 204}
  1563. \range{pages}{10}
  1564. \verb{doi}
  1565. \verb 10.1007/978-3-642-10817-4_19
  1566. \endverb
  1567. \endentry
  1568. \entry{NVIDIA.11262021}{online}{}
  1569. \name{author}{1}{}{%
  1570. {{hash=a2fe9051fe051f819bd739ef3e340377}{%
  1571. family={NVIDIA},
  1572. familyi={N\bibinitperiod}}}%
  1573. }
  1574. \strng{namehash}{a2fe9051fe051f819bd739ef3e340377}
  1575. \strng{fullhash}{a2fe9051fe051f819bd739ef3e340377}
  1576. \strng{bibnamehash}{a2fe9051fe051f819bd739ef3e340377}
  1577. \strng{authorbibnamehash}{a2fe9051fe051f819bd739ef3e340377}
  1578. \strng{authornamehash}{a2fe9051fe051f819bd739ef3e340377}
  1579. \strng{authorfullhash}{a2fe9051fe051f819bd739ef3e340377}
  1580. \field{sortinit}{N}
  1581. \field{sortinithash}{22369a73d5f88983a108b63f07f37084}
  1582. \field{labelnamesource}{author}
  1583. \field{labeltitlesource}{title}
  1584. \field{abstract}{Das NVIDIA{\circledR} Jetson Nano{\texttrademark}-Entwicklerkit ist ein kleiner, leistungsstarker Computer, der trotz geringem Strombedarf moderne KI-Workloads stemmen kann. Statten Sie Millionen von Geräten in der Peripherie mit unglaublichen neuen Funktionen aus.}
  1585. \field{title}{NVIDIA Jetson Nano}
  1586. \field{urlday}{12}
  1587. \field{urlmonth}{12}
  1588. \field{urlyear}{2021}
  1589. \field{year}{2021}
  1590. \field{urldateera}{ce}
  1591. \verb{urlraw}
  1592. \verb https://www.nvidia.com/de-de/autonomous-machines/embedded-systems/jetson-nano/product-development/
  1593. \endverb
  1594. \verb{url}
  1595. \verb https://www.nvidia.com/de-de/autonomous-machines/embedded-systems/jetson-nano/product-development/
  1596. \endverb
  1597. \keyw{AIoT;CUDA-X;Edge AI;Internet der Dinge;Jetson Nano;Jetson Nano Developer Kit;KI in der Peripherie;NVIDIA}
  1598. \endentry
  1599. \entry{OpenImages.6252021}{online}{}
  1600. \field{sortinit}{O}
  1601. \field{sortinithash}{2cd7140a07aea5341f9e2771efe90aae}
  1602. \field{labeltitlesource}{title}
  1603. \field{title}{Open Images V6}
  1604. \field{urlday}{12}
  1605. \field{urlmonth}{12}
  1606. \field{urlyear}{2021}
  1607. \field{year}{2021}
  1608. \field{urldateera}{ce}
  1609. \verb{urlraw}
  1610. \verb https://storage.googleapis.com/openimages/web/index.html
  1611. \endverb
  1612. \verb{url}
  1613. \verb https://storage.googleapis.com/openimages/web/index.html
  1614. \endverb
  1615. \endentry
  1616. \entry{PireStereoVision.2012}{inproceedings}{}
  1617. \name{author}{4}{}{%
  1618. {{hash=f498d360ad385c023730b43bd9812b69}{%
  1619. family={Pire},
  1620. familyi={P\bibinitperiod},
  1621. given={Taihú},
  1622. giveni={T\bibinitperiod}}}%
  1623. {{hash=dcc36a08c12e54445e394e423cbb7637}{%
  1624. family={De\bibnamedelima Cristóforis},
  1625. familyi={D\bibinitperiod\bibinitdelim C\bibinitperiod},
  1626. given={Pablo},
  1627. giveni={P\bibinitperiod}}}%
  1628. {{hash=7f69f8174769790ab71606007cffc744}{%
  1629. family={Nitsche},
  1630. familyi={N\bibinitperiod},
  1631. given={Matias},
  1632. giveni={M\bibinitperiod}}}%
  1633. {{hash=3e2ea478bcb80c478f83444955dc0c5a}{%
  1634. family={Berlles},
  1635. familyi={B\bibinitperiod},
  1636. given={Julio},
  1637. giveni={J\bibinitperiod}}}%
  1638. }
  1639. \strng{namehash}{f1bd8d752fd4388421c278800cc4be02}
  1640. \strng{fullhash}{5562cd4c7ba0183bc8b353b41d50f7f7}
  1641. \strng{bibnamehash}{f1bd8d752fd4388421c278800cc4be02}
  1642. \strng{authorbibnamehash}{f1bd8d752fd4388421c278800cc4be02}
  1643. \strng{authornamehash}{f1bd8d752fd4388421c278800cc4be02}
  1644. \strng{authorfullhash}{5562cd4c7ba0183bc8b353b41d50f7f7}
  1645. \field{sortinit}{P}
  1646. \field{sortinithash}{ff3bcf24f47321b42cb156c2cc8a8422}
  1647. \field{labelnamesource}{author}
  1648. \field{labeltitlesource}{title}
  1649. \field{month}{12}
  1650. \field{title}{Stereo vision obstacle avoidance using depth and elevation maps}
  1651. \field{year}{2012}
  1652. \endentry
  1653. \entry{Poggi.2018}{misc}{}
  1654. \name{author}{4}{}{%
  1655. {{hash=343dafb5548d904115e5782508eb8375}{%
  1656. family={Poggi},
  1657. familyi={P\bibinitperiod},
  1658. given={Matteo},
  1659. giveni={M\bibinitperiod}}}%
  1660. {{hash=0e141e967911dc1e9e3266ec98e4fc1d}{%
  1661. family={Aleotti},
  1662. familyi={A\bibinitperiod},
  1663. given={Filippo},
  1664. giveni={F\bibinitperiod}}}%
  1665. {{hash=8d31b2206bee079ebdf6ff494cb9c6ee}{%
  1666. family={Tosi},
  1667. familyi={T\bibinitperiod},
  1668. given={Fabio},
  1669. giveni={F\bibinitperiod}}}%
  1670. {{hash=97fde7508220953d74da9f2a50a93a4e}{%
  1671. family={Mattoccia},
  1672. familyi={M\bibinitperiod},
  1673. given={Stefano},
  1674. giveni={S\bibinitperiod}}}%
  1675. }
  1676. \strng{namehash}{f547ed4e1ed4a9803f760f7c779dd8b7}
  1677. \strng{fullhash}{72edd1d58afb42b32eda560296ac12df}
  1678. \strng{bibnamehash}{f547ed4e1ed4a9803f760f7c779dd8b7}
  1679. \strng{authorbibnamehash}{f547ed4e1ed4a9803f760f7c779dd8b7}
  1680. \strng{authornamehash}{f547ed4e1ed4a9803f760f7c779dd8b7}
  1681. \strng{authorfullhash}{72edd1d58afb42b32eda560296ac12df}
  1682. \field{sortinit}{P}
  1683. \field{sortinithash}{ff3bcf24f47321b42cb156c2cc8a8422}
  1684. \field{labelnamesource}{author}
  1685. \field{labeltitlesource}{title}
  1686. \field{abstract}{Unsupervised depth estimation from a single image is a very attractive technique with several implications in robotic, autonomous navigation, augmented reality and so on. This topic represents a very challenging task and the advent of deep learning enabled to tackle this problem with excellent results. However, these architectures are extremely deep and complex. Thus, real-time performance can be achieved only by leveraging power-hungry GPUs that do not allow to infer depth maps in application fields characterized by low-power constraints. To tackle this issue, in this paper we propose a novel architecture capable to quickly infer an accurate depth map on a CPU, even of an embedded system, using a pyramid of features extracted from a single input image. Similarly to state-of-the-art, we train our network in an unsupervised manner casting depth estimation as an image reconstruction problem. Extensive experimental results on the KITTI dataset show that compared to the top performing approach our network has similar accuracy but a much lower complexity (about 6{\%} of parameters) enabling to infer a depth map for a KITTI image in about 1.7 s on the Raspberry Pi 3 and at more than 8 Hz on a standard CPU. Moreover, by trading accuracy for efficiency, our network allows to infer maps at about 2 Hz and 40 Hz respectively, still being more accurate than most state-of-the-art slower methods. To the best of our knowledge, it is the first method enabling such performance on CPUs paving the way for effective deployment of unsupervised monocular depth estimation even on embedded systems.}
  1687. \field{eprinttype}{arXiv}
  1688. \field{title}{Towards real-time unsupervised monocular depth estimation on CPU}
  1689. \field{year}{2018}
  1690. \verb{eprint}
  1691. \verb arXiv:1806.11430v3
  1692. \endverb
  1693. \verb{file}
  1694. \verb Towards real-time unsupervised monocular depth estimatio:Attachments/Towards real-time unsupervised monocular depth estimatio.pdf:application/pdf
  1695. \endverb
  1696. \verb{urlraw}
  1697. \verb https://arxiv.org/pdf/1806.11430
  1698. \endverb
  1699. \verb{url}
  1700. \verb https://arxiv.org/pdf/1806.11430
  1701. \endverb
  1702. \keyw{Computer Vision and Pattern Recognition (cs.CV);Robotics (cs.RO)}
  1703. \endentry
  1704. \entry{Ranftl.3242021}{misc}{}
  1705. \name{author}{3}{}{%
  1706. {{hash=f13b87ed1a6410067504966eac03e9a2}{%
  1707. family={Ranftl},
  1708. familyi={R\bibinitperiod},
  1709. given={René},
  1710. giveni={R\bibinitperiod}}}%
  1711. {{hash=c0b2fdc20c37a4bd8d817822d374f809}{%
  1712. family={Bochkovskiy},
  1713. familyi={B\bibinitperiod},
  1714. given={Alexey},
  1715. giveni={A\bibinitperiod}}}%
  1716. {{hash=9025300f43a193cfed8e10e924f2a117}{%
  1717. family={Koltun},
  1718. familyi={K\bibinitperiod},
  1719. given={Vladlen},
  1720. giveni={V\bibinitperiod}}}%
  1721. }
  1722. \strng{namehash}{2593b6dcce725925f90629afdcead815}
  1723. \strng{fullhash}{2593b6dcce725925f90629afdcead815}
  1724. \strng{bibnamehash}{2593b6dcce725925f90629afdcead815}
  1725. \strng{authorbibnamehash}{2593b6dcce725925f90629afdcead815}
  1726. \strng{authornamehash}{2593b6dcce725925f90629afdcead815}
  1727. \strng{authorfullhash}{2593b6dcce725925f90629afdcead815}
  1728. \field{sortinit}{R}
  1729. \field{sortinithash}{5e1c39a9d46ffb6bebd8f801023a9486}
  1730. \field{labelnamesource}{author}
  1731. \field{labeltitlesource}{title}
  1732. \field{abstract}{We introduce dense vision transformers, an architecture that leverages vision transformers in place of convolutional networks as a backbone for dense prediction tasks. We assemble tokens from various stages of the vision transformer into image-like representations at various resolutions and progressively combine them into full-resolution predictions using a convolutional decoder. The transformer backbone processes representations at a constant and relatively high resolution and has a global receptive field at every stage. These properties allow the dense vision transformer to provide finer-grained and more globally coherent predictions when compared to fully-convolutional networks. Our experiments show that this architecture yields substantial improvements on dense prediction tasks, especially when a large amount of training data is available. For monocular depth estimation, we observe an improvement of up to 28{\%} in relative performance when compared to a state-of-the-art fully-convolutional network. When applied to semantic segmentation, dense vision transformers set a new state of the art on ADE20K with 49.02{\%} mIoU. We further show that the architecture can be fine-tuned on smaller datasets such as NYUv2, KITTI, and Pascal Context where it also sets the new state of the art. Our models are available at https://github.com/intel-isl/DPT.}
  1733. \field{eprinttype}{arXiv}
  1734. \field{note}{15 pages}
  1735. \field{pagetotal}{15}
  1736. \field{title}{Vision Transformers for Dense Prediction}
  1737. \field{year}{2021}
  1738. \verb{doi}
  1739. \verb pages
  1740. \endverb
  1741. \verb{eprint}
  1742. \verb arXiv:2103.13413v1
  1743. \endverb
  1744. \verb{urlraw}
  1745. \verb https://arxiv.org/pdf/2103.13413
  1746. \endverb
  1747. \verb{url}
  1748. \verb https://arxiv.org/pdf/2103.13413
  1749. \endverb
  1750. \keyw{Computer Science - Computer Vision and Pattern Recognition;Computer Vision and Pattern Recognition (cs.CV)}
  1751. \endentry
  1752. \entry{Ranftl.2019}{misc}{}
  1753. \name{author}{5}{}{%
  1754. {{hash=f13b87ed1a6410067504966eac03e9a2}{%
  1755. family={Ranftl},
  1756. familyi={R\bibinitperiod},
  1757. given={René},
  1758. giveni={R\bibinitperiod}}}%
  1759. {{hash=4ee0621ee5837a920708854b07fef7f7}{%
  1760. family={Lasinger},
  1761. familyi={L\bibinitperiod},
  1762. given={Katrin},
  1763. giveni={K\bibinitperiod}}}%
  1764. {{hash=89d0278ff885d57b7708ae120200f573}{%
  1765. family={Hafner},
  1766. familyi={H\bibinitperiod},
  1767. given={David},
  1768. giveni={D\bibinitperiod}}}%
  1769. {{hash=205714eb0cc6921c15493da21a574130}{%
  1770. family={Schindler},
  1771. familyi={S\bibinitperiod},
  1772. given={Konrad},
  1773. giveni={K\bibinitperiod}}}%
  1774. {{hash=9025300f43a193cfed8e10e924f2a117}{%
  1775. family={Koltun},
  1776. familyi={K\bibinitperiod},
  1777. given={Vladlen},
  1778. giveni={V\bibinitperiod}}}%
  1779. }
  1780. \strng{namehash}{9a396274beaaf111452927a46e80f547}
  1781. \strng{fullhash}{fa15ac494a832be56810827b8030d68a}
  1782. \strng{bibnamehash}{9a396274beaaf111452927a46e80f547}
  1783. \strng{authorbibnamehash}{9a396274beaaf111452927a46e80f547}
  1784. \strng{authornamehash}{9a396274beaaf111452927a46e80f547}
  1785. \strng{authorfullhash}{fa15ac494a832be56810827b8030d68a}
  1786. \field{sortinit}{R}
  1787. \field{sortinithash}{5e1c39a9d46ffb6bebd8f801023a9486}
  1788. \field{labelnamesource}{author}
  1789. \field{labeltitlesource}{title}
  1790. \field{abstract}{The success of monocular depth estimation relies on large and diverse training sets. Due to the challenges associated with acquiring dense ground-truth depth across different environments at scale, a number of datasets with distinct characteristics and biases have emerged. We develop tools that enable mixing multiple datasets during training, even if their annotations are incompatible. In particular, we propose a robust training objective that is invariant to changes in depth range and scale, advocate the use of principled multi-objective learning to combine data from different sources, and highlight the importance of pretraining encoders on auxiliary tasks. Armed with these tools, we experiment with five diverse training datasets, including a new, massive data source: 3D films. To demonstrate the generalization power of our approach we use zero-shot cross-dataset transfer, i.e. we evaluate on datasets that were not seen during training. The experiments confirm that mixing data from complementary sources greatly improves monocular depth estimation. Our approach clearly outperforms competing methods across diverse datasets, setting a new state of the art for monocular depth estimation. Some results are shown in the supplementary video at this https URL}
  1791. \field{eprinttype}{arXiv}
  1792. \field{title}{Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer}
  1793. \field{year}{2019}
  1794. \verb{eprint}
  1795. \verb arXiv:1907.01341v3
  1796. \endverb
  1797. \verb{file}
  1798. \verb Towards Robust Monocular Depth Estimation Mixing Dataset:Attachments/Towards Robust Monocular Depth Estimation Mixing Dataset.pdf:application/pdf
  1799. \endverb
  1800. \verb{urlraw}
  1801. \verb https://arxiv.org/pdf/1907.01341
  1802. \endverb
  1803. \verb{url}
  1804. \verb https://arxiv.org/pdf/1907.01341
  1805. \endverb
  1806. \keyw{Computer Vision and Pattern Recognition (cs.CV)}
  1807. \endentry
  1808. \entry{Raschka.2018}{book}{}
  1809. \name{author}{1}{}{%
  1810. {{hash=bb679961e3f3f4bba79fc1cf3aa7df1c}{%
  1811. family={Raschka},
  1812. familyi={R\bibinitperiod},
  1813. given={Sebastian},
  1814. giveni={S\bibinitperiod}}}%
  1815. }
  1816. \list{language}{1}{%
  1817. {ger}%
  1818. }
  1819. \list{location}{1}{%
  1820. {Frechen}%
  1821. }
  1822. \list{publisher}{1}{%
  1823. {mitp}%
  1824. }
  1825. \strng{namehash}{bb679961e3f3f4bba79fc1cf3aa7df1c}
  1826. \strng{fullhash}{bb679961e3f3f4bba79fc1cf3aa7df1c}
  1827. \strng{bibnamehash}{bb679961e3f3f4bba79fc1cf3aa7df1c}
  1828. \strng{authorbibnamehash}{bb679961e3f3f4bba79fc1cf3aa7df1c}
  1829. \strng{authornamehash}{bb679961e3f3f4bba79fc1cf3aa7df1c}
  1830. \strng{authorfullhash}{bb679961e3f3f4bba79fc1cf3aa7df1c}
  1831. \field{sortinit}{R}
  1832. \field{sortinithash}{5e1c39a9d46ffb6bebd8f801023a9486}
  1833. \field{labelnamesource}{author}
  1834. \field{labeltitlesource}{title}
  1835. \field{edition}{2., aktualisierte und erweiterte Auflage}
  1836. \field{isbn}{978-3-95845-733-1}
  1837. \field{pagetotal}{577}
  1838. \field{series}{mitp Professional}
  1839. \field{subtitle}{Das umfassende Praxis-Handbuch für Data Science, Deep Learning und Predictive Analytics}
  1840. \field{title}{Machine Learning mit Python und Scikit-learn und TensorFlow}
  1841. \field{year}{2018}
  1842. \verb{urlraw}
  1843. \verb http://www.content-select.com/index.php?id=bib_view&ean=9783958457348
  1844. \endverb
  1845. \verb{url}
  1846. \verb http://www.content-select.com/index.php?id=bib_view&ean=9783958457348
  1847. \endverb
  1848. \endentry
  1849. \entry{Redmon.2015}{misc}{}
  1850. \name{author}{4}{}{%
  1851. {{hash=99bced2e56a5253f3fe98a5f04e6d9b2}{%
  1852. family={Redmon},
  1853. familyi={R\bibinitperiod},
  1854. given={Joseph},
  1855. giveni={J\bibinitperiod}}}%
  1856. {{hash=05ca9f19da9ecbd2def4e5514f8043c8}{%
  1857. family={Divvala},
  1858. familyi={D\bibinitperiod},
  1859. given={Santosh},
  1860. giveni={S\bibinitperiod}}}%
  1861. {{hash=bd5dadbe57bedc5957c19a3154c4d424}{%
  1862. family={Girshick},
  1863. familyi={G\bibinitperiod},
  1864. given={Ross},
  1865. giveni={R\bibinitperiod}}}%
  1866. {{hash=396c6ddedb6f986906fc3e4994d19974}{%
  1867. family={Farhadi},
  1868. familyi={F\bibinitperiod},
  1869. given={Ali},
  1870. giveni={A\bibinitperiod}}}%
  1871. }
  1872. \strng{namehash}{e1203a0044715040adeb8c5079ee645a}
  1873. \strng{fullhash}{b5530443e433a4da53dbe3cf155225b4}
  1874. \strng{bibnamehash}{e1203a0044715040adeb8c5079ee645a}
  1875. \strng{authorbibnamehash}{e1203a0044715040adeb8c5079ee645a}
  1876. \strng{authornamehash}{e1203a0044715040adeb8c5079ee645a}
  1877. \strng{authorfullhash}{b5530443e433a4da53dbe3cf155225b4}
  1878. \field{sortinit}{R}
  1879. \field{sortinithash}{5e1c39a9d46ffb6bebd8f801023a9486}
  1880. \field{labelnamesource}{author}
  1881. \field{labeltitlesource}{title}
  1882. \field{abstract}{We present YOLO, a new approach to object detection. Prior work on object detection repurposes classifiers to perform detection. Instead, we frame object detection as a regression problem to spatially separated bounding boxes and associated class probabilities. A single neural network predicts bounding boxes and class probabilities directly from full images in one evaluation. Since the whole detection pipeline is a single network, it can be optimized end-to-end directly on detection performance. Our unified architecture is extremely fast. Our base YOLO model processes images in real-time at 45 frames per second. A smaller version of the network, Fast YOLO, processes an astounding 155 frames per second while still achieving double the mAP of other real-time detectors. Compared to state-of-the-art detection systems, YOLO makes more localization errors but is far less likely to predict false detections where nothing exists. Finally, YOLO learns very general representations of objects. It outperforms all other detection methods, including DPM and R-CNN, by a wide margin when generalizing from natural images to artwork on both the Picasso Dataset and the People-Art Dataset.}
  1883. \field{eprinttype}{arXiv}
  1884. \field{title}{You Only Look Once: Unified, Real-Time Object Detection}
  1885. \field{year}{2015}
  1886. \verb{eprint}
  1887. \verb arXiv:1506.02640v5
  1888. \endverb
  1889. \verb{file}
  1890. \verb You Only Look Once Unified Real-Time Object Detection:Attachments/You Only Look Once Unified Real-Time Object Detection.pdf:application/pdf
  1891. \endverb
  1892. \verb{urlraw}
  1893. \verb https://arxiv.org/pdf/1506.02640
  1894. \endverb
  1895. \verb{url}
  1896. \verb https://arxiv.org/pdf/1506.02640
  1897. \endverb
  1898. \keyw{Computer Vision and Pattern Recognition (cs.CV)}
  1899. \endentry
  1900. \entry{Ren.2015}{misc}{}
  1901. \name{author}{4}{}{%
  1902. {{hash=bb295293acacd54387339079ebbe4ead}{%
  1903. family={Ren},
  1904. familyi={R\bibinitperiod},
  1905. given={Shaoqing},
  1906. giveni={S\bibinitperiod}}}%
  1907. {{hash=6b4b60e909e78633945f3f9c9dc83e01}{%
  1908. family={He},
  1909. familyi={H\bibinitperiod},
  1910. given={Kaiming},
  1911. giveni={K\bibinitperiod}}}%
  1912. {{hash=bd5dadbe57bedc5957c19a3154c4d424}{%
  1913. family={Girshick},
  1914. familyi={G\bibinitperiod},
  1915. given={Ross},
  1916. giveni={R\bibinitperiod}}}%
  1917. {{hash=f85751488058842b5777c7b4074077b5}{%
  1918. family={Sun},
  1919. familyi={S\bibinitperiod},
  1920. given={Jian},
  1921. giveni={J\bibinitperiod}}}%
  1922. }
  1923. \strng{namehash}{f086ca4da3e532e8a41cb758ea461825}
  1924. \strng{fullhash}{008a132af3e2d4ff15eb01a8fb4b005c}
  1925. \strng{bibnamehash}{f086ca4da3e532e8a41cb758ea461825}
  1926. \strng{authorbibnamehash}{f086ca4da3e532e8a41cb758ea461825}
  1927. \strng{authornamehash}{f086ca4da3e532e8a41cb758ea461825}
  1928. \strng{authorfullhash}{008a132af3e2d4ff15eb01a8fb4b005c}
  1929. \field{sortinit}{R}
  1930. \field{sortinithash}{5e1c39a9d46ffb6bebd8f801023a9486}
  1931. \field{labelnamesource}{author}
  1932. \field{labeltitlesource}{title}
  1933. \field{abstract}{State-of-the-art object detection networks depend on region proposal algorithms to hypothesize object locations. Advances like SPPnet and Fast R-CNN have reduced the running time of these detection networks, exposing region proposal computation as a bottleneck. In this work, we introduce a Region Proposal Network (RPN) that shares full-image convolutional features with the detection network, thus enabling nearly cost-free region proposals. An RPN is a fully convolutional network that simultaneously predicts object bounds and objectness scores at each position. The RPN is trained end-to-end to generate high-quality region proposals, which are used by Fast R-CNN for detection. We further merge RPN and Fast R-CNN into a single network by sharing their convolutional features---using the recently popular terminology of neural networks with 'attention' mechanisms, the RPN component tells the unified network where to look. For the very deep VGG-16 model, our detection system has a frame rate of 5fps (including all steps) on a GPU, while achieving state-of-the-art object detection accuracy on PASCAL VOC 2007, 2012, and MS COCO datasets with only 300 proposals per image. In ILSVRC and COCO 2015 competitions, Faster R-CNN and RPN are the foundations of the 1st-place winning entries in several tracks. Code has been made publicly available.}
  1934. \field{eprinttype}{arXiv}
  1935. \field{title}{Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks}
  1936. \field{year}{2015}
  1937. \verb{eprint}
  1938. \verb arXiv:1506.01497v3
  1939. \endverb
  1940. \verb{file}
  1941. \verb Faster R-CNN Towards Real-Time Object Detection with Reg:Attachments/Faster R-CNN Towards Real-Time Object Detection with Reg.pdf:application/pdf
  1942. \endverb
  1943. \verb{urlraw}
  1944. \verb https://arxiv.org/pdf/1506.01497
  1945. \endverb
  1946. \verb{url}
  1947. \verb https://arxiv.org/pdf/1506.01497
  1948. \endverb
  1949. \keyw{Computer Vision and Pattern Recognition (cs.CV)}
  1950. \endentry
  1951. \entry{Rezatofighi.2252019}{misc}{}
  1952. \name{author}{6}{}{%
  1953. {{hash=c834f8dc7253578c3dfdf392681c186f}{%
  1954. family={Rezatofighi},
  1955. familyi={R\bibinitperiod},
  1956. given={Hamid},
  1957. giveni={H\bibinitperiod}}}%
  1958. {{hash=028a19dd53dc26b31bb9bb0f5aa212e7}{%
  1959. family={Tsoi},
  1960. familyi={T\bibinitperiod},
  1961. given={Nathan},
  1962. giveni={N\bibinitperiod}}}%
  1963. {{hash=d9806662b24ca486664faa2ddb093991}{%
  1964. family={Gwak},
  1965. familyi={G\bibinitperiod},
  1966. given={JunYoung},
  1967. giveni={J\bibinitperiod}}}%
  1968. {{hash=72c51c1b7ba21759de9bd158c8c74ad4}{%
  1969. family={Sadeghian},
  1970. familyi={S\bibinitperiod},
  1971. given={Amir},
  1972. giveni={A\bibinitperiod}}}%
  1973. {{hash=8d82839be6afd59c432fc7cf084f4326}{%
  1974. family={Reid},
  1975. familyi={R\bibinitperiod},
  1976. given={Ian},
  1977. giveni={I\bibinitperiod}}}%
  1978. {{hash=e29c20003a517abaaac28852301c03c2}{%
  1979. family={Savarese},
  1980. familyi={S\bibinitperiod},
  1981. given={Silvio},
  1982. giveni={S\bibinitperiod}}}%
  1983. }
  1984. \strng{namehash}{579d75aa87c34f094d34b3bc3670aa52}
  1985. \strng{fullhash}{d49fc6c7ff39001b1961fb11179318be}
  1986. \strng{bibnamehash}{579d75aa87c34f094d34b3bc3670aa52}
  1987. \strng{authorbibnamehash}{579d75aa87c34f094d34b3bc3670aa52}
  1988. \strng{authornamehash}{579d75aa87c34f094d34b3bc3670aa52}
  1989. \strng{authorfullhash}{d49fc6c7ff39001b1961fb11179318be}
  1990. \field{sortinit}{R}
  1991. \field{sortinithash}{5e1c39a9d46ffb6bebd8f801023a9486}
  1992. \field{labelnamesource}{author}
  1993. \field{labeltitlesource}{title}
  1994. \field{abstract}{Intersection over Union (IoU) is the most popular evaluation metric used in the object detection benchmarks. However, there is a gap between optimizing the commonly used distance losses for regressing the parameters of a bounding box and maximizing this metric value. The optimal objective for a metric is the metric itself. In the case of axis-aligned 2D bounding boxes, it can be shown that $IoU$ can be directly used as a regression loss. However, $IoU$ has a plateau making it infeasible to optimize in the case of non-overlapping bounding boxes. In this paper, we address the weaknesses of $IoU$ by introducing a generalized version as both a new loss and a new metric. By incorporating this generalized $IoU$ ($GIoU$) as a loss into the state-of-the art object detection frameworks, we show a consistent improvement on their performance using both the standard, $IoU$ based, and new, $GIoU$ based, performance measures on popular object detection benchmarks such as PASCAL VOC and MS COCO.}
  1995. \field{eprinttype}{arXiv}
  1996. \field{note}{accepted in CVPR 2019}
  1997. \field{title}{Generalized Intersection over Union: A Metric and A Loss for Bounding Box Regression}
  1998. \field{year}{2019}
  1999. \verb{doi}
  2000. \verb CVPR
  2001. \endverb
  2002. \verb{eprint}
  2003. \verb arXiv:1902.09630v2
  2004. \endverb
  2005. \verb{file}
  2006. \verb Generalized Intersection over Union A Metric and A Loss:Attachments/Generalized Intersection over Union A Metric and A Loss.pdf:application/pdf
  2007. \endverb
  2008. \verb{urlraw}
  2009. \verb https://arxiv.org/pdf/1902.09630
  2010. \endverb
  2011. \verb{url}
  2012. \verb https://arxiv.org/pdf/1902.09630
  2013. \endverb
  2014. \keyw{Artificial Intelligence (cs.AI);Computer Science - Artificial Intelligence;Computer Science - Computer Vision and Pattern Recognition;Computer Science - Learning;Computer Vision and Pattern Recognition (cs.CV);Machine Learning (cs.LG)}
  2015. \endentry
  2016. \entry{Sandler}{article}{}
  2017. \name{author}{5}{}{%
  2018. {{hash=8f90fd131c2bbfde4d0e9fdd7ed4ea8b}{%
  2019. family={Sandler},
  2020. familyi={S\bibinitperiod},
  2021. given={Mark},
  2022. giveni={M\bibinitperiod}}}%
  2023. {{hash=315c4166fc1f7cb66324a7f0d82827cd}{%
  2024. family={Howard},
  2025. familyi={H\bibinitperiod},
  2026. given={Andrew},
  2027. giveni={A\bibinitperiod}}}%
  2028. {{hash=d767e8e4d733bcf728bcdf2c193462f7}{%
  2029. family={Zhu},
  2030. familyi={Z\bibinitperiod},
  2031. given={Menglong},
  2032. giveni={M\bibinitperiod}}}%
  2033. {{hash=48f4090a93cf9f445057a9d6defe7973}{%
  2034. family={Zhmoginov},
  2035. familyi={Z\bibinitperiod},
  2036. given={Andrey},
  2037. giveni={A\bibinitperiod}}}%
  2038. {{hash=9dddfcfd529634a150ee38ee5c0203d7}{%
  2039. family={Chen},
  2040. familyi={C\bibinitperiod},
  2041. given={Liang-Chieh},
  2042. giveni={L\bibinithyphendelim C\bibinitperiod}}}%
  2043. }
  2044. \strng{namehash}{70179dd377b16c4cfcc0748ef0bd028a}
  2045. \strng{fullhash}{8d58d04315a5e1c06300772330b235d7}
  2046. \strng{bibnamehash}{70179dd377b16c4cfcc0748ef0bd028a}
  2047. \strng{authorbibnamehash}{70179dd377b16c4cfcc0748ef0bd028a}
  2048. \strng{authornamehash}{70179dd377b16c4cfcc0748ef0bd028a}
  2049. \strng{authorfullhash}{8d58d04315a5e1c06300772330b235d7}
  2050. \field{sortinit}{S}
  2051. \field{sortinithash}{b164b07b29984b41daf1e85279fbc5ab}
  2052. \field{labelnamesource}{author}
  2053. \field{labeltitlesource}{title}
  2054. \field{abstract}{In this paper we describe a new mobile architecture, MobileNetV2, that improves the state of the art performance of mobile models on multiple tasks and benchmarks as well as across a spectrum of different model sizes. We also describe efficient ways of applying these mobile models to object detection in a novel framework we call SSDLite. Additionally, we demonstrate how to build mobile semantic segmentation models through a reduced form of DeepLabv3 which we call Mobile DeepLabv3. The MobileNetV2 architecture is based on an inverted residual structure where the input and output of the residual block are thin bottleneck layers opposite to traditional residual models which use expanded representations in the input an MobileNetV2 uses lightweight depthwise convolutions to filter features in the intermediate expansion layer. Additionally, we find that it is important to remove non-linearities in the narrow layers in order to maintain representational power. We demonstrate that this improves performance and provide an intuition that led to this design. Finally, our approach allows decoupling of the input/output domains from the expressiveness of the transformation, which provides a convenient framework for further analysis. We measure our performance on Imagenet classification, COCO object detection, VOC image segmentation. We evaluate the trade-offs between accuracy, and number of operations measured by multiply-adds (MAdd), as well as the number of parameters}
  2055. \field{eprinttype}{arXiv}
  2056. \field{journaltitle}{The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}
  2057. \field{title}{MobileNetV2: Inverted Residuals and Linear Bottlenecks}
  2058. \field{year}{2018}
  2059. \verb{eprint}
  2060. \verb arXiv:1801.04381v4
  2061. \endverb
  2062. \verb{file}
  2063. \verb MobileNetV2 Inverted Residuals and Linear Bottlenecks:Attachments/MobileNetV2 Inverted Residuals and Linear Bottlenecks.pdf:application/pdf
  2064. \endverb
  2065. \verb{urlraw}
  2066. \verb https://arxiv.org/pdf/1801.04381
  2067. \endverb
  2068. \verb{url}
  2069. \verb https://arxiv.org/pdf/1801.04381
  2070. \endverb
  2071. \keyw{Computer Vision and Pattern Recognition (cs.CV)}
  2072. \endentry
  2073. \entry{TensorFlow.8122021}{online}{}
  2074. \name{author}{1}{}{%
  2075. {{hash=074dd699710da0ec1eb45f13b31788e3}{%
  2076. family={TensorFlow},
  2077. familyi={T\bibinitperiod}}}%
  2078. }
  2079. \strng{namehash}{074dd699710da0ec1eb45f13b31788e3}
  2080. \strng{fullhash}{074dd699710da0ec1eb45f13b31788e3}
  2081. \strng{bibnamehash}{074dd699710da0ec1eb45f13b31788e3}
  2082. \strng{authorbibnamehash}{074dd699710da0ec1eb45f13b31788e3}
  2083. \strng{authornamehash}{074dd699710da0ec1eb45f13b31788e3}
  2084. \strng{authorfullhash}{074dd699710da0ec1eb45f13b31788e3}
  2085. \field{extraname}{1}
  2086. \field{sortinit}{T}
  2087. \field{sortinithash}{9af77f0292593c26bde9a56e688eaee9}
  2088. \field{labelnamesource}{author}
  2089. \field{labeltitlesource}{title}
  2090. \field{title}{Quantisierung nach dem Training ~|~ TensorFlow Lite}
  2091. \field{urlday}{6}
  2092. \field{urlmonth}{10}
  2093. \field{urlyear}{2021}
  2094. \field{year}{2021}
  2095. \field{urldateera}{ce}
  2096. \verb{urlraw}
  2097. \verb https://www.tensorflow.org/lite/performance/post_training_quantization
  2098. \endverb
  2099. \verb{url}
  2100. \verb https://www.tensorflow.org/lite/performance/post_training_quantization
  2101. \endverb
  2102. \endentry
  2103. \entry{TensorFlow.832021}{online}{}
  2104. \name{author}{1}{}{%
  2105. {{hash=074dd699710da0ec1eb45f13b31788e3}{%
  2106. family={TensorFlow},
  2107. familyi={T\bibinitperiod}}}%
  2108. }
  2109. \strng{namehash}{074dd699710da0ec1eb45f13b31788e3}
  2110. \strng{fullhash}{074dd699710da0ec1eb45f13b31788e3}
  2111. \strng{bibnamehash}{074dd699710da0ec1eb45f13b31788e3}
  2112. \strng{authorbibnamehash}{074dd699710da0ec1eb45f13b31788e3}
  2113. \strng{authornamehash}{074dd699710da0ec1eb45f13b31788e3}
  2114. \strng{authorfullhash}{074dd699710da0ec1eb45f13b31788e3}
  2115. \field{extraname}{2}
  2116. \field{sortinit}{T}
  2117. \field{sortinithash}{9af77f0292593c26bde9a56e688eaee9}
  2118. \field{labelnamesource}{author}
  2119. \field{labeltitlesource}{title}
  2120. \field{abstract}{Eine End-to-End-Open-Source-Machine-Learning-Plattform für alle. Entdecken Sie das flexible Ökosystem von TensorFlow aus Tools, Bibliotheken und Community-Ressourcen.}
  2121. \field{title}{TensorFlow}
  2122. \field{urlday}{6}
  2123. \field{urlmonth}{10}
  2124. \field{urlyear}{2021}
  2125. \field{year}{2021}
  2126. \field{urldateera}{ce}
  2127. \verb{urlraw}
  2128. \verb https://www.tensorflow.org/
  2129. \endverb
  2130. \verb{url}
  2131. \verb https://www.tensorflow.org/
  2132. \endverb
  2133. \endentry
  2134. \entry{Tensorflow.8202021}{online}{}
  2135. \name{author}{1}{}{%
  2136. {{hash=074dd699710da0ec1eb45f13b31788e3}{%
  2137. family={TensorFlow},
  2138. familyi={T\bibinitperiod}}}%
  2139. }
  2140. \strng{namehash}{074dd699710da0ec1eb45f13b31788e3}
  2141. \strng{fullhash}{074dd699710da0ec1eb45f13b31788e3}
  2142. \strng{bibnamehash}{074dd699710da0ec1eb45f13b31788e3}
  2143. \strng{authorbibnamehash}{074dd699710da0ec1eb45f13b31788e3}
  2144. \strng{authornamehash}{074dd699710da0ec1eb45f13b31788e3}
  2145. \strng{authorfullhash}{074dd699710da0ec1eb45f13b31788e3}
  2146. \field{extraname}{3}
  2147. \field{sortinit}{T}
  2148. \field{sortinithash}{9af77f0292593c26bde9a56e688eaee9}
  2149. \field{labelnamesource}{author}
  2150. \field{labeltitlesource}{title}
  2151. \field{title}{TensorFlow 2 Schnellstart für Anfänger ~|~ TensorFlow Core}
  2152. \field{urlday}{20}
  2153. \field{urlmonth}{10}
  2154. \field{urlyear}{2021}
  2155. \field{year}{2021}
  2156. \field{urldateera}{ce}
  2157. \verb{urlraw}
  2158. \verb https://www.tensorflow.org/tutorials/quickstart/beginner?hl=de
  2159. \endverb
  2160. \verb{url}
  2161. \verb https://www.tensorflow.org/tutorials/quickstart/beginner?hl=de
  2162. \endverb
  2163. \endentry
  2164. \entry{TensorFlow.8202021}{online}{}
  2165. \name{author}{1}{}{%
  2166. {{hash=074dd699710da0ec1eb45f13b31788e3}{%
  2167. family={TensorFlow},
  2168. familyi={T\bibinitperiod}}}%
  2169. }
  2170. \strng{namehash}{074dd699710da0ec1eb45f13b31788e3}
  2171. \strng{fullhash}{074dd699710da0ec1eb45f13b31788e3}
  2172. \strng{bibnamehash}{074dd699710da0ec1eb45f13b31788e3}
  2173. \strng{authorbibnamehash}{074dd699710da0ec1eb45f13b31788e3}
  2174. \strng{authornamehash}{074dd699710da0ec1eb45f13b31788e3}
  2175. \strng{authorfullhash}{074dd699710da0ec1eb45f13b31788e3}
  2176. \field{extraname}{4}
  2177. \field{sortinit}{T}
  2178. \field{sortinithash}{9af77f0292593c26bde9a56e688eaee9}
  2179. \field{labelnamesource}{author}
  2180. \field{labeltitlesource}{title}
  2181. \field{abstract}{Ein Deep-Learning-Framework für geräteinterne Inferenz. Trainieren und implementieren Sie Machine-Learning-Modelle auf mobilen und IoT-Geräten, Android, iOS, Edge TPU, Raspberry Pi.}
  2182. \field{title}{TensorFlow Lite | ML für Mobilgeräte und Edge-Geräte}
  2183. \field{urlday}{6}
  2184. \field{urlmonth}{10}
  2185. \field{urlyear}{2021}
  2186. \field{year}{2021}
  2187. \field{urldateera}{ce}
  2188. \verb{urlraw}
  2189. \verb https://www.tensorflow.org/lite/?hl=de
  2190. \endverb
  2191. \verb{url}
  2192. \verb https://www.tensorflow.org/lite/?hl=de
  2193. \endverb
  2194. \endentry
  2195. \entry{TensorFlow.8102021}{online}{}
  2196. \name{author}{1}{}{%
  2197. {{hash=074dd699710da0ec1eb45f13b31788e3}{%
  2198. family={TensorFlow},
  2199. familyi={T\bibinitperiod}}}%
  2200. }
  2201. \strng{namehash}{074dd699710da0ec1eb45f13b31788e3}
  2202. \strng{fullhash}{074dd699710da0ec1eb45f13b31788e3}
  2203. \strng{bibnamehash}{074dd699710da0ec1eb45f13b31788e3}
  2204. \strng{authorbibnamehash}{074dd699710da0ec1eb45f13b31788e3}
  2205. \strng{authornamehash}{074dd699710da0ec1eb45f13b31788e3}
  2206. \strng{authorfullhash}{074dd699710da0ec1eb45f13b31788e3}
  2207. \field{extraname}{5}
  2208. \field{sortinit}{T}
  2209. \field{sortinithash}{9af77f0292593c26bde9a56e688eaee9}
  2210. \field{labelnamesource}{author}
  2211. \field{labeltitlesource}{title}
  2212. \field{title}{TensorFlow Lite-Konverter}
  2213. \field{urlday}{6}
  2214. \field{urlmonth}{10}
  2215. \field{urlyear}{2021}
  2216. \field{year}{2021}
  2217. \field{urldateera}{ce}
  2218. \verb{urlraw}
  2219. \verb https://www.tensorflow.org/lite/convert
  2220. \endverb
  2221. \verb{url}
  2222. \verb https://www.tensorflow.org/lite/convert
  2223. \endverb
  2224. \endentry
  2225. \entry{Xian_2018_CVPR}{inproceedings}{}
  2226. \name{author}{7}{}{%
  2227. {{hash=d34252df77136e3e43830c451bc16f4c}{%
  2228. family={Xian},
  2229. familyi={X\bibinitperiod},
  2230. given={Ke},
  2231. giveni={K\bibinitperiod}}}%
  2232. {{hash=9e49384c78f6861746fe787b12339aed}{%
  2233. family={Shen},
  2234. familyi={S\bibinitperiod},
  2235. given={Chunhua},
  2236. giveni={C\bibinitperiod}}}%
  2237. {{hash=5dd651375bfd096c12f9efedcbb23b26}{%
  2238. family={Cao},
  2239. familyi={C\bibinitperiod},
  2240. given={Zhiguo},
  2241. giveni={Z\bibinitperiod}}}%
  2242. {{hash=b61c7b6eeaa58572d109f9c87e5a61b0}{%
  2243. family={Lu},
  2244. familyi={L\bibinitperiod},
  2245. given={Hao},
  2246. giveni={H\bibinitperiod}}}%
  2247. {{hash=1f6352fb541fa4d76fa406a7a227e942}{%
  2248. family={Xiao},
  2249. familyi={X\bibinitperiod},
  2250. given={Yang},
  2251. giveni={Y\bibinitperiod}}}%
  2252. {{hash=7daa4ae161a7b30243906f8d8614d031}{%
  2253. family={Li},
  2254. familyi={L\bibinitperiod},
  2255. given={Ruibo},
  2256. giveni={R\bibinitperiod}}}%
  2257. {{hash=944b1099464311888b5869719fe97207}{%
  2258. family={Luo},
  2259. familyi={L\bibinitperiod},
  2260. given={Zhenbo},
  2261. giveni={Z\bibinitperiod}}}%
  2262. }
  2263. \strng{namehash}{d8bb47053eee030e5ea0df7c00d70552}
  2264. \strng{fullhash}{a8598030e163b0fcd7ad5a9a1d0e2305}
  2265. \strng{bibnamehash}{d8bb47053eee030e5ea0df7c00d70552}
  2266. \strng{authorbibnamehash}{d8bb47053eee030e5ea0df7c00d70552}
  2267. \strng{authornamehash}{d8bb47053eee030e5ea0df7c00d70552}
  2268. \strng{authorfullhash}{a8598030e163b0fcd7ad5a9a1d0e2305}
  2269. \field{sortinit}{X}
  2270. \field{sortinithash}{1965c258adceecf23ce3d67b05113442}
  2271. \field{labelnamesource}{author}
  2272. \field{labeltitlesource}{title}
  2273. \field{booktitle}{Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}
  2274. \field{month}{6}
  2275. \field{title}{Monocular Relative Depth Perception With Web Stereo Data Supervision}
  2276. \field{year}{2018}
  2277. \endentry
  2278. \entry{Yang.2017}{misc}{}
  2279. \name{author}{6}{}{%
  2280. {{hash=d063132f441798ef9f4d438c5d299f61}{%
  2281. family={Yang},
  2282. familyi={Y\bibinitperiod},
  2283. given={Shichao},
  2284. giveni={S\bibinitperiod}}}%
  2285. {{hash=be0784aa0265f37ffaf08730900a33ad}{%
  2286. family={Konam},
  2287. familyi={K\bibinitperiod},
  2288. given={Sandeep},
  2289. giveni={S\bibinitperiod}}}%
  2290. {{hash=0b13d29cc629ce6345c4d0376393515f}{%
  2291. family={Ma},
  2292. familyi={M\bibinitperiod},
  2293. given={Chen},
  2294. giveni={C\bibinitperiod}}}%
  2295. {{hash=059b8e68f832ebb69bef55f1fb50075b}{%
  2296. family={Rosenthal},
  2297. familyi={R\bibinitperiod},
  2298. given={Stephanie},
  2299. giveni={S\bibinitperiod}}}%
  2300. {{hash=108bc6e5f71a32699fcf07b7e8957310}{%
  2301. family={Veloso},
  2302. familyi={V\bibinitperiod},
  2303. given={Manuela},
  2304. giveni={M\bibinitperiod}}}%
  2305. {{hash=ad26bf16eeaee427e8e36bea35587822}{%
  2306. family={Scherer},
  2307. familyi={S\bibinitperiod},
  2308. given={Sebastian},
  2309. giveni={S\bibinitperiod}}}%
  2310. }
  2311. \strng{namehash}{6a59f26db512831317179bf5e8435bc2}
  2312. \strng{fullhash}{f79134b856bd1c1c1dc178e12c628a5a}
  2313. \strng{bibnamehash}{6a59f26db512831317179bf5e8435bc2}
  2314. \strng{authorbibnamehash}{6a59f26db512831317179bf5e8435bc2}
  2315. \strng{authornamehash}{6a59f26db512831317179bf5e8435bc2}
  2316. \strng{authorfullhash}{f79134b856bd1c1c1dc178e12c628a5a}
  2317. \field{sortinit}{Y}
  2318. \field{sortinithash}{fd67ad5a9ef0f7456bdd9aab10fe1495}
  2319. \field{labelnamesource}{author}
  2320. \field{labeltitlesource}{title}
  2321. \field{abstract}{Obstacle avoidance from monocular images is a challenging problem for robots. Though multi-view structure-from-motion could build 3D maps, it is not robust in textureless environments. Some learning based methods exploit human demonstration to predict a steering command directly from a single image. However, this method is usually biased towards certain tasks or demonstration scenarios and also biased by human understanding. In this paper, we propose a new method to predict a trajectory from images. We train our system on more diverse NYUv2 dataset. The ground truth trajectory is computed from the designed cost functions automatically. The Convolutional Neural Network perception is divided into two stages: first, predict depth map and surface normal from RGB images, which are two important geometric properties related to 3D obstacle representation. Second, predict the trajectory from the depth and normal. Results show that our intermediate perception increases the accuracy by 20{\%} than the direct prediction. Our model generalizes well to other public indoor datasets and is also demonstrated for robot flights in simulation and experiments.}
  2322. \field{eprinttype}{arXiv}
  2323. \field{title}{Obstacle Avoidance through Deep Networks based Intermediate Perception}
  2324. \field{year}{2017}
  2325. \verb{eprint}
  2326. \verb arXiv:1704.08759v1
  2327. \endverb
  2328. \verb{file}
  2329. \verb Obstacle Avoidance through Deep Networks based Intermedi:Attachments/Obstacle Avoidance through Deep Networks based Intermedi.pdf:application/pdf
  2330. \endverb
  2331. \verb{urlraw}
  2332. \verb https://arxiv.org/pdf/1704.08759
  2333. \endverb
  2334. \verb{url}
  2335. \verb https://arxiv.org/pdf/1704.08759
  2336. \endverb
  2337. \keyw{Computer Vision and Pattern Recognition (cs.CV);Robotics (cs.RO)}
  2338. \endentry
  2339. \entry{Zhang.2019}{inproceedings}{}
  2340. \name{author}{3}{}{%
  2341. {{hash=477b8da468a0e0969c47a18407e4f921}{%
  2342. family={Zhang},
  2343. familyi={Z\bibinitperiod},
  2344. given={Zhenghong},
  2345. giveni={Z\bibinitperiod}}}%
  2346. {{hash=c3cc50ba7a1919e3db6b64bc14843a9e}{%
  2347. family={Xiong},
  2348. familyi={X\bibinitperiod},
  2349. given={Mingkang},
  2350. giveni={M\bibinitperiod}}}%
  2351. {{hash=ad35b4d45aebfabe5d2d4be218eadfac}{%
  2352. family={Xiong},
  2353. familyi={X\bibinitperiod},
  2354. given={Huilin},
  2355. giveni={H\bibinitperiod}}}%
  2356. }
  2357. \strng{namehash}{53735448d457eaa108282eef5304063a}
  2358. \strng{fullhash}{53735448d457eaa108282eef5304063a}
  2359. \strng{bibnamehash}{53735448d457eaa108282eef5304063a}
  2360. \strng{authorbibnamehash}{53735448d457eaa108282eef5304063a}
  2361. \strng{authornamehash}{53735448d457eaa108282eef5304063a}
  2362. \strng{authorfullhash}{53735448d457eaa108282eef5304063a}
  2363. \field{sortinit}{Z}
  2364. \field{sortinithash}{96892c0b0a36bb8557c40c49813d48b3}
  2365. \field{labelnamesource}{author}
  2366. \field{labeltitlesource}{title}
  2367. \field{booktitle}{2019 4th International Conference on Cloud Computing and Internet of Things (CCIOT)}
  2368. \field{title}{Monocular Depth Estimation for UAV Obstacle Avoidance}
  2369. \field{year}{2019}
  2370. \field{pages}{43\bibrangedash 47}
  2371. \range{pages}{5}
  2372. \verb{doi}
  2373. \verb 10.1109/CCIOT48581.2019.8980350
  2374. \endverb
  2375. \endentry
  2376. \entry{Zhou.182017}{misc}{}
  2377. \name{author}{3}{}{%
  2378. {{hash=441742472ab4f4d6527bd49ddd3c6279}{%
  2379. family={Zhou},
  2380. familyi={Z\bibinitperiod},
  2381. given={Yiren},
  2382. giveni={Y\bibinitperiod}}}%
  2383. {{hash=0e339e2d98a016927dec3c34c9a12281}{%
  2384. family={Song},
  2385. familyi={S\bibinitperiod},
  2386. given={Sibo},
  2387. giveni={S\bibinitperiod}}}%
  2388. {{hash=54cff214a2905977051c8cbe018c5499}{%
  2389. family={Cheung},
  2390. familyi={C\bibinitperiod},
  2391. given={Ngai-Man},
  2392. giveni={N\bibinithyphendelim M\bibinitperiod}}}%
  2393. }
  2394. \strng{namehash}{6298ea7c9583f6fae23929e33e08d3d3}
  2395. \strng{fullhash}{6298ea7c9583f6fae23929e33e08d3d3}
  2396. \strng{bibnamehash}{6298ea7c9583f6fae23929e33e08d3d3}
  2397. \strng{authorbibnamehash}{6298ea7c9583f6fae23929e33e08d3d3}
  2398. \strng{authornamehash}{6298ea7c9583f6fae23929e33e08d3d3}
  2399. \strng{authorfullhash}{6298ea7c9583f6fae23929e33e08d3d3}
  2400. \field{sortinit}{Z}
  2401. \field{sortinithash}{96892c0b0a36bb8557c40c49813d48b3}
  2402. \field{labelnamesource}{author}
  2403. \field{labeltitlesource}{title}
  2404. \field{abstract}{Image blur and image noise are common distortions during image acquisition. In this paper, we systematically study the effect of image distortions on the deep neural network (DNN) image classifiers. First, we examine the DNN classifier performance under four types of distortions. Second, we propose two approaches to alleviate the effect of image distortion: re-training and fine-tuning with noisy images. Our results suggest that, under certain conditions, fine-tuning with noisy images can alleviate much effect due to distorted inputs, and is more practical than re-training.}
  2405. \field{eprinttype}{arXiv}
  2406. \field{note}{5 pages, 8 figures, ICASSP 2017}
  2407. \field{pagetotal}{5}
  2408. \field{title}{On Classification of Distorted Images with Deep Convolutional Neural Networks}
  2409. \field{year}{2017}
  2410. \verb{eprint}
  2411. \verb arXiv:1701.01924v1
  2412. \endverb
  2413. \verb{file}
  2414. \verb On_Classification_of_Distorted_Images_with_Deep_Co:Attachments/On_Classification_of_Distorted_Images_with_Deep_Co.pdf:application/pdf
  2415. \endverb
  2416. \verb{urlraw}
  2417. \verb http://arxiv.org/pdf/1701.01924v1
  2418. \endverb
  2419. \verb{url}
  2420. \verb http://arxiv.org/pdf/1701.01924v1
  2421. \endverb
  2422. \endentry
  2423. \entry{Zhou.2021}{book}{}
  2424. \name{author}{1}{}{%
  2425. {{hash=6b1a37e820a730061851a81dcca2d154}{%
  2426. family={Zhou},
  2427. familyi={Z\bibinitperiod},
  2428. given={Zhi-Hua},
  2429. giveni={Z\bibinithyphendelim H\bibinitperiod}}}%
  2430. }
  2431. \list{location}{1}{%
  2432. {Singapore}%
  2433. }
  2434. \list{publisher}{1}{%
  2435. {Springer Singapore}%
  2436. }
  2437. \strng{namehash}{6b1a37e820a730061851a81dcca2d154}
  2438. \strng{fullhash}{6b1a37e820a730061851a81dcca2d154}
  2439. \strng{bibnamehash}{6b1a37e820a730061851a81dcca2d154}
  2440. \strng{authorbibnamehash}{6b1a37e820a730061851a81dcca2d154}
  2441. \strng{authornamehash}{6b1a37e820a730061851a81dcca2d154}
  2442. \strng{authorfullhash}{6b1a37e820a730061851a81dcca2d154}
  2443. \field{sortinit}{Z}
  2444. \field{sortinithash}{96892c0b0a36bb8557c40c49813d48b3}
  2445. \field{labelnamesource}{author}
  2446. \field{labeltitlesource}{title}
  2447. \field{isbn}{978-981-15-1966-6}
  2448. \field{title}{Machine Learning}
  2449. \field{year}{2021}
  2450. \verb{doi}
  2451. \verb 10.1007/978-981-15-1967-3
  2452. \endverb
  2453. \verb{file}
  2454. \verb Machine Learning:Attachments/Machine Learning.pdf:application/pdf
  2455. \endverb
  2456. \endentry
  2457. \entry{Zhu.2016}{misc}{}
  2458. \name{author}{7}{}{%
  2459. {{hash=f86434cb0d86ba7eac5b0e3f3236abf0}{%
  2460. family={Zhu},
  2461. familyi={Z\bibinitperiod},
  2462. given={Yuke},
  2463. giveni={Y\bibinitperiod}}}%
  2464. {{hash=e1dbfa167aabb32138ad3d9df709e3dd}{%
  2465. family={Mottaghi},
  2466. familyi={M\bibinitperiod},
  2467. given={Roozbeh},
  2468. giveni={R\bibinitperiod}}}%
  2469. {{hash=9bfe220c99d5e3f19e82f842df626065}{%
  2470. family={Kolve},
  2471. familyi={K\bibinitperiod},
  2472. given={Eric},
  2473. giveni={E\bibinitperiod}}}%
  2474. {{hash=5556915c40a404e6fa83816576929682}{%
  2475. family={Lim},
  2476. familyi={L\bibinitperiod},
  2477. given={Joseph\bibnamedelima J.},
  2478. giveni={J\bibinitperiod\bibinitdelim J\bibinitperiod}}}%
  2479. {{hash=83dd9d464c9f0ea982254939a7f021d8}{%
  2480. family={Gupta},
  2481. familyi={G\bibinitperiod},
  2482. given={Abhinav},
  2483. giveni={A\bibinitperiod}}}%
  2484. {{hash=cd00ce5bc45f687c432e52e0fa1a7aa6}{%
  2485. family={Fei-Fei},
  2486. familyi={F\bibinithyphendelim F\bibinitperiod},
  2487. given={Li},
  2488. giveni={L\bibinitperiod}}}%
  2489. {{hash=396c6ddedb6f986906fc3e4994d19974}{%
  2490. family={Farhadi},
  2491. familyi={F\bibinitperiod},
  2492. given={Ali},
  2493. giveni={A\bibinitperiod}}}%
  2494. }
  2495. \strng{namehash}{abd154245b8f10c81399b395d5309db7}
  2496. \strng{fullhash}{6e77a10df1258517e9e9627a2a07a209}
  2497. \strng{bibnamehash}{abd154245b8f10c81399b395d5309db7}
  2498. \strng{authorbibnamehash}{abd154245b8f10c81399b395d5309db7}
  2499. \strng{authornamehash}{abd154245b8f10c81399b395d5309db7}
  2500. \strng{authorfullhash}{6e77a10df1258517e9e9627a2a07a209}
  2501. \field{sortinit}{Z}
  2502. \field{sortinithash}{96892c0b0a36bb8557c40c49813d48b3}
  2503. \field{labelnamesource}{author}
  2504. \field{labeltitlesource}{title}
  2505. \field{abstract}{Two less addressed issues of deep reinforcement learning are (1) lack of generalization capability to new target goals, and (2) data inefficiency i.e., the model requires several (and often costly) episodes of trial and error to converge, which makes it impractical to be applied to real-world scenarios. In this paper, we address these two issues and apply our model to the task of target-driven visual navigation. To address the first issue, we propose an actor-critic model whose policy is a function of the goal as well as the current state, which allows to better generalize. To address the second issue, we propose AI2-THOR framework, which provides an environment with high-quality 3D scenes and physics engine. Our framework enables agents to take actions and interact with objects. Hence, we can collect a huge number of training samples efficiently. We show that our proposed method (1) converges faster than the state-of-the-art deep reinforcement learning methods, (2) generalizes across targets and across scenes, (3) generalizes to a real robot scenario with a small amount of fine-tuning (although the model is trained in simulation), (4) is end-to-end trainable and does not need feature engineering, feature matching between frames or 3D reconstruction of the environment. The supplementary video can be accessed at the following link: this https URL.}
  2506. \field{eprinttype}{arXiv}
  2507. \field{title}{Target-driven Visual Navigation in Indoor Scenes using Deep Reinforcement Learning}
  2508. \field{year}{2016}
  2509. \verb{eprint}
  2510. \verb arXiv:1609.05143v1
  2511. \endverb
  2512. \verb{urlraw}
  2513. \verb https://arxiv.org/pdf/1609.05143
  2514. \endverb
  2515. \verb{url}
  2516. \verb https://arxiv.org/pdf/1609.05143
  2517. \endverb
  2518. \keyw{Computer Vision and Pattern Recognition (cs.CV)}
  2519. \endentry
  2520. \enddatalist
  2521. \endrefsection
  2522. \endinput