870 research outputs found

    Visual Representations: Defining Properties and Deep Approximations

    Full text link
    Visual representations are defined in terms of minimal sufficient statistics of visual data, for a class of tasks, that are also invariant to nuisance variability. Minimal sufficiency guarantees that we can store a representation in lieu of raw data with smallest complexity and no performance loss on the task at hand. Invariance guarantees that the statistic is constant with respect to uninformative transformations of the data. We derive analytical expressions for such representations and show they are related to feature descriptors commonly used in computer vision, as well as to convolutional neural networks. This link highlights the assumptions and approximations tacitly assumed by these methods and explains empirical practices such as clamping, pooling and joint normalization.Comment: UCLA CSD TR140023, Nov. 12, 2014, revised April 13, 2015, November 13, 2015, February 28, 201

    A Novel Hardware Architecture for Real Time Extraction of Local Features

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ)-- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ์ „๊ธฐยท์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2016. 8. ์ดํ˜์žฌ.์ปดํ“จํŒ… ์„ฑ๋Šฅ์˜ ๋น„์•ฝ์ ์ธ ๋ฐœ์ „๊ณผ ๋ณด๊ธ‰์€ ์ปดํ“จํ„ฐ ๊ธฐ์ˆ ์˜ ์ ์šฉ ๋ถ„์•ผ๋ฅผ ๋ฐ์Šคํฌํƒ‘์—์„œ ์Šค๋งˆํŠธํฐ, ์Šค๋งˆํŠธ TV, ์Šน์šฉ์ฐจ ๋“ฑ์— ์ด๋ฅด๊ธฐ๊นŒ์ง€ ํญ๋„“์€ ๋ฒ”์œ„๋กœ ๋„“ํžˆ๋Š” ๊ฒฐ๊ณผ๋ฅผ ์•ผ๊ธฐํ–ˆ๋‹ค. ๋ณ€ํ™”๋œ ํ™˜๊ฒฝ์—์„œ ๋Œ€์ค‘์€ ๊ธฐ์กด์— ์—†์—ˆ๋˜ ์ข€ ๋” ํ˜์‹ ์ ์ธ ๊ธฐ๋Šฅ์„ ๋ฐ›์•„๋“ค์ผ ์ค€๋น„๊ฐ€ ๋˜์—ˆ๊ณ , ์ด์— ๋ถ€ํ•ฉํ•˜๊ธฐ ์œ„ํ•ด Computer vision ๊ธฐ์ˆ ์€ ์ ์ฐจ ์ƒ์šฉํ™”์˜ ๊ธธ์„ ๊ฑท๊ฒŒ ๋˜์—ˆ๋‹ค. ๋ฌผ์ฒด ์ธ์‹ ๋ฐ ์ถ”์ , 3D reconstruction ๋“ฑ ํญ๋„“๊ฒŒ ์‘์šฉ๋  ์ˆ˜ ์žˆ๋Š” computer vision ๊ธฐ์ˆ ๋“ค์€ ์„œ๋กœ ๋‹ค๋ฅธ ์˜์ƒ ์‚ฌ์ด์—์„œ ๋™์ผํ•œ pixel์„ ์ฐพ๋Š” image matching ๊ธฐ์ˆ ์„ ํ•„์š”๋กœ ํ•œ๋‹ค. ๊ด€๋ จ ์—ฐ๊ตฌ๋“ค ์ค‘ ์˜์ƒ์˜ ํฌ๊ธฐ๊ฐ€ ๋ณ€ํ•˜๊ฑฐ๋‚˜ ํšŒ์ „ํ•˜์—ฌ๋„ ์•ˆ์ •์ ์œผ๋กœ matching์ด ๊ฐ€๋Šฅํ•œ Scale- Invariant Feature Transform (SIFT) ์•Œ๊ณ ๋ฆฌ์ฆ˜์ด ์ œ์•ˆ๋˜์—ˆ๊ณ , ์ดํ›„ ์นด๋ฉ”๋ผ์˜ viewpoint ๋ณ€ํ™”์—๋„ ๊ฐ•์ธํ•œ Affine Invariant Extension of SIFT (ASIFT) ์•Œ๊ณ ๋ฆฌ์ฆ˜์ด ์ œ์•ˆ๋˜์—ˆ๋‹ค. SIFT ๋ฐ ASIFT ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ image matching์˜ ์•ˆ์ •์„ฑ์ด ๋†’์€ ๋ฐ˜๋ฉด ๋งŽ์€ ์—ฐ์‚ฐ๋Ÿ‰์„ ์š”๊ตฌํ•œ๋‹ค. ์ด๋ฅผ ์‹ค์‹œ๊ฐ„ ์ฒ˜๋ฆฌํ•˜๊ธฐ ์œ„ํ•ด specifically designed hardware์„ ์ด์šฉํ•œ ์—ฐ์‚ฐ ๊ฐ€์† ์—ฐ๊ตฌ๊ฐ€ ์ง„ํ–‰๋˜๊ณ  ์žˆ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์‹ค์‹œ๊ฐ„ (30frames/sec)์œผ๋กœ ๋™์ž‘ ๊ฐ€๋Šฅํ•œ ASIFT ํ•˜๋“œ์›จ์–ด ๊ตฌ์กฐ๋ฅผ ์ œ์•ˆํ•œ๋‹ค. ์ด๋ฅผ ์œ„ํ•ด ์ฒซ๋ฒˆ์งธ๋กœ SIFT feature๋ฅผ ์‹ค์‹œ๊ฐ„์œผ๋กœ ์—ฐ์‚ฐ ํ•  ์ˆ˜ ์žˆ๋Š” SIFT ํ•˜๋“œ์›จ์–ด ๊ตฌ์กฐ๋ฅผ ์ œ์•ˆํ•œ๋‹ค. SIFT ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ๋„๋ฆฌ ์‚ฌ์šฉ๋˜๋Š” ๋งŒํผ ๋งŽ์€ ์ˆ˜์˜ ๊ฐ€์† ํ•˜๋“œ์›จ์–ด ๊ตฌ์กฐ๊ฐ€ ์—ฐ๊ตฌ๋˜์—ˆ๋‹ค. ๋Œ€๋ถ€๋ถ„์˜ ๊ธฐ์กด SIFT ํ•˜๋“œ์›จ์–ด๋Š” ์‹ค์‹œ๊ฐ„์„ฑ์„ ์ถฉ์กฑ์‹œํ‚ค์ง€๋งŒ, ์ด๋ฅผ ์œ„ํ•ด ๊ณผ๋„ํ•˜๊ฒŒ ๋งŽ์€ ๋‚ด๋ถ€ ๋ฉ”๋ชจ๋ฆฌ๋ฅผ ์‚ฌ์šฉํ•˜์—ฌ ํ•˜๋“œ์›จ์–ด ๋น„์šฉ์„ ํฌ๊ฒŒ ์ฆ๊ฐ€์‹œ์ผฐ๋‹ค. ์ด ์ด์Šˆ ์‚ฌํ•ญ์œผ๋กœ ์ธํ•ด ๋‚ด๋ถ€ ๋ฉ”๋ชจ๋ฆฌ์™€ ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ๋ฅผ ํ˜ผ์šฉํ•˜๋Š” ์ƒˆ๋กœ์šด SIFT ํ•˜๋“œ์›จ์–ด ๊ตฌ์กฐ๊ฐ€ ์ œ์•ˆ๋˜์—ˆ๋‹ค. ์ด ๊ฒฝ์šฐ ๋นˆ๋ฒˆํ•œ ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ ์‚ฌ์šฉ์€ ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ latency๋กœ ์ธํ•œ ๋™์ž‘ ์†๋„ ์ €ํ•˜ ๋ฌธ์ œ๋ฅผ ์ผ์œผํ‚จ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์€ ์ด ๋ฌธ์ œ๋ฅผ ํ•ด๊ฒฐํ•˜๊ธฐ ์œ„ํ•ด ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ์—์„œ ์ฝ์–ด์˜จ ๋ฐ์ดํ„ฐ ์žฌ์‚ฌ์šฉ ๋ฐฉ์•ˆ๊ณผ, ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ์— ์ €์žฅํ•˜๋Š” ๋ฐ์ดํ„ฐ์— ๋Œ€ํ•œ down-sampling ๋ฐ less significant bits ์ œ๊ฑฐ๋ฅผ ํ†ตํ•œ ๋ฐ์ดํ„ฐ๋Ÿ‰ ๊ฐ์†Œ ๋ฐฉ์•ˆ์„ ์ œ์•ˆํ•œ๋‹ค. ์ œ์•ˆํ•˜๋Š” SIFT ํ•˜๋“œ์›จ์–ด๋Š” Gaussian image๋ฅผ ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ์— ์ €์žฅํ•˜๋ฉฐ, ์ด ๊ฒฝ์šฐ descriptor ์ƒ์„ฑ์„ ์œ„ํ•ด local-patch๋ฅผ ์ฝ์–ด์˜ค๋Š”๋ฐ ๋งŽ์€ ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ ์ ‘๊ทผ์ด ๋ฐœ์ƒํ•œ๋‹ค. ์ด๋ฅผ ์ €๊ฐํ•˜๊ธฐ ์œ„ํ•ด, ์„œ๋กœ ๋‹ค๋ฅธ local-patch ์ƒ์˜ ์ค‘๋ณต ๋ฐ์ดํ„ฐ๋ฅผ ์žฌ์‚ฌ์šฉํ•˜๋Š” ๋ฐฉ์•ˆ๊ณผ ์ด๋ฅผ ์œ„ํ•œ ํ•˜๋“œ์›จ์–ด ๊ตฌ์กฐ๋ฅผ ์ œ์‹œํ•œ๋‹ค. ๋˜ํ•œ Gaussian image์˜ ๋ฐ์ดํ„ฐ๋Ÿ‰ ์ž์ฒด๋ฅผ ์ค„์ด๊ธฐ ์œ„ํ•ด down-sampling ๋ฐ less significant bits ์ œ๊ฑฐ ๋ฐฉ์•ˆ์„ ์ด์šฉํ•œ๋‹ค. ์ด๋•Œ SIFT ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ ์ •ํ™•๋„ ๊ฐ์†Œ๋ฅผ ์ตœ์†Œํ™”ํ•˜์˜€๋‹ค. ๊ฒฐ๊ณผ์ ์œผ๋กœ ๋ณธ ๋…ผ๋ฌธ์€ ๊ธฐ์กด state-of-the-art SIFT ํ•˜๋“œ์›จ์–ด์˜ 10.93% ํฌ๊ธฐ์˜ ๋‚ด๋ถ€ ๋ฉ”๋ชจ๋ฆฌ๋งŒ ์‚ฌ์šฉํ•˜๋ฉฐ, 3300๊ฐœ์˜ key-point์— ๋Œ€ํ•ด 30 frames/sec (fps)์˜ ์†๋„๋กœ ๋™์ž‘ ๊ฐ€๋Šฅํ•˜๋‹ค. ASIFT ์•Œ๊ณ ๋ฆฌ์ฆ˜ ์—ฐ์‚ฐ์„ ๊ณ ์†์œผ๋กœ ์ˆ˜ํ–‰ํ•˜๊ธฐ ์œ„ํ•ด์„œ๋Š” SIFT ํ•˜๋“œ์›จ์–ด์— affine transform๋œ ์˜์ƒ์„ ์ œ๊ณตํ•˜๋Š” affine transform ํ•˜๋“œ์›จ์–ด๊ฐ€ delay ์—†์ด ๋ฐ์ดํ„ฐ๋ฅผ ์ œ๊ณตํ•  ์ˆ˜ ์žˆ์–ด์•ผ ํ•œ๋‹ค. ํ•˜์ง€๋งŒ ์ผ๋ฐ˜์ ์ธ affine transform ์—ฐ์‚ฐ ๋ฐฉ์‹์„ ์ด์šฉํ•  ๊ฒฝ์šฐ affine transform ํ•˜๋“œ์›จ์–ด๋Š” ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ์—์„œ ์›๋ณธ ์˜์ƒ์„ ์ฝ์–ด ์˜ฌ ๋•Œ ๋ถˆ์—ฐ์†์ ์ธ ์ฃผ์†Œ๋กœ ์ ‘๊ทผํ•˜๊ฒŒ ๋œ๋‹ค. ์ด๋Š” ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ latency๋ฅผ ๋ฐœ์ƒ์‹œํ‚ค๋ฉฐ affine transform module์ด ์ถฉ๋ถ„ํ•œ ๋ฐ์ดํ„ฐ๋ฅผ SIFT ํ•˜๋“œ์›จ์–ด์— ๊ณต๊ธ‰ํ•ด์ฃผ์ง€ ๋ชปํ•˜๋Š” ๋ฌธ์ œ๋ฅผ ์•ผ๊ธฐํ•œ๋‹ค. ์ด ๋ฌธ์ œ๋ฅผ ํ•ด๊ฒฐํ•˜๊ธฐ ์œ„ํ•˜์—ฌ ๋ณธ ๋…ผ๋ฌธ์€ SIFT feature์˜ rotation-invariantํ•œ ํŠน์„ฑ์„ ์ด์šฉํ•˜์—ฌ, affine transform ์—ฐ์‚ฐ ๋ฐฉ์‹์„ ๋ณ€๊ฒฝํ•˜์˜€๋‹ค. ์ด ๋ฐฉ์‹์€ ASIFT ์•Œ๊ณ ๋ฆฌ์ฆ˜์ด ์ทจํ•˜๋Š” ๋ชจ๋“  affine transform ์—ฐ์‚ฐ์„ ์ˆ˜ํ–‰ํ•  ๋•Œ ์—ฐ์†๋œ ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ ์ฃผ์†Œ๋กœ ์ž…๋ ฅ ์˜์ƒ์„ ์ ‘๊ทผํ•  ์ˆ˜ ์žˆ๊ฒŒ ํ•ด์ค€๋‹ค. ์ด๋ฅผ ํ†ตํ•ด ๋ถˆํ•„์š”ํ•œ ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ latency๊ฐ€ ํฌ๊ฒŒ ๊ฐ์†Œ๋œ๋‹ค. ์ œ์•ˆ๋œ affine transform ์—ฐ์‚ฐ ๋ฐฉ์‹์€ ์›๋ณธ ์˜์ƒ์„ scalingํ•œ ๋’ค skewingํ•˜๋Š” ์—ฐ์‚ฐ ๊ณผ์ •์„ ๊ฑฐ์นœ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์€ ์ด ๊ณผ์ •์—์„œ scaling๋œ ์˜์ƒ ๋ฐ์ดํ„ฐ๋ฅผ ์„œ๋กœ ๋‹ค๋ฅธ affine transform ์—ฐ์‚ฐ์— ์žฌํ™œ์šฉํ•˜๋Š” ๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•œ๋‹ค. ์ด๋Š” scaling ์—ฐ์‚ฐ๋Ÿ‰์„ ๊ฐ์†Œ์‹œํ‚ฌ ๋ฟ ๋งŒ ์•„๋‹ˆ๋ผ ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ ์ ‘๊ทผ๋Ÿ‰๋„ ๊ฐ์†Œ์‹œํ‚จ๋‹ค. ์ œ์•ˆ๋œ ๋ฐฉ์•ˆ๋“ค๋กœ ์ธํ•œ affine transform ํ•˜๋“œ์›จ์–ด์˜ ์†๋„ ํ–ฅ์ƒ์€ SIFT ํ•˜๋“œ์›จ์–ด์— ๋Œ€๊ธฐ ์—†์ด ๋ฐ์ดํ„ฐ๋ฅผ ๊ณต๊ธ‰ํ•  ์ˆ˜ ์žˆ๊ฒŒ ํ•ด์ฃผ๊ณ , ์ตœ์ข…์ ์œผ๋กœ utilization ํ–ฅ์ƒ์„ ํ†ตํ•œ ASIFT ํ•˜๋“œ์›จ์–ด์˜ ๋™์ž‘ ์†๋„ ํ–ฅ์ƒ์— ๊ธฐ์—ฌํ•œ๋‹ค. ๊ฒฐ๊ณผ์ ์œผ๋กœ ๋ณธ ๋…ผ๋ฌธ์—์„œ ์ œ์•ˆํ•˜๋Š” ASIFT ํ•˜๋“œ์›จ์–ด๋Š” ๋†’์€ utilization์œผ๋กœ ๋™์ž‘์ด ๊ฐ€๋Šฅํ•˜๋ฉฐ, ์ด๋กœ ์ธํ•ด 2,500๊ฐœ์˜ key- point๊ฐ€ ๊ฒ€์ถœ๋˜๋Š” ์˜์ƒ์— ๋Œ€ํ•˜์—ฌ 30fps์˜ ๋™์ž‘ ์†๋„๋กœ ASIFT ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ˆ˜ํ–‰ํ•  ์ˆ˜ ์žˆ๋‹ค.์ œ 1 ์žฅ ์„œ๋ก  1 1.1 ์—ฐ๊ตฌ ๋ฐฐ๊ฒฝ 1 1.2 ์—ฐ๊ตฌ ๋‚ด์šฉ 4 1.3 ๋…ผ๋ฌธ ๊ตฌ์„ฑ 6 ์ œ 2 ์žฅ ์ด์ „ ์—ฐ๊ตฌ ์†Œ๊ฐœ ๋ฐ ๋ฌธ์ œ ์ œ์‹œ 7 2.1 SIFT ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๋ฐ ์—ฐ์‚ฐ ๊ฐ€์†ํ™” ๊ธฐ์ˆ  7 2.1.1 Scale-Invariant Feture Transform (SIFT) 7 2.1.2 ๊ธฐ์กด SIFT ์—ฐ์‚ฐ ๊ฐ€์†ํ™” ์—ฐ๊ตฌ ๋ฐ ๋ฌธ์ œ์  16 2.2 ASIFT ์•Œ๊ณ ๋ฆฌ์ฆ˜ ๋ฐ ์—ฐ์‚ฐ ๊ฐ€์†ํ™” ๊ธฐ์ˆ  19 2.2.1 Scale-Invariant Feture Transform (SIFT) 19 2.2.2 ๊ธฐ์กด SIFT ์—ฐ์‚ฐ ๊ฐ€์†ํ™” ์—ฐ๊ตฌ 23 2.3 ์‹ค์‹œ๊ฐ„ ASIFT ํ•˜๋“œ์›จ์–ด ๊ตฌํ˜„์„ ์œ„ํ•œ ์—ฐ๊ตฌ ๋ฐฉํ–ฅ 24 ์ œ 3 ์žฅ ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ bandwidth ์ €๊ฐ๋œ SIFT ํ•˜๋“œ์›จ์–ด ๊ตฌ์กฐ 26 3.1 ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ์— ์ €์žฅ๋  SIFT ์—ฐ์‚ฐ์˜ ์ค‘๊ฐ„ ๋ฐ์ดํ„ฐ ๊ณ ์ฐฐ 26 3.2 ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ bandwidth๋ฅผ ์ค„์ด๊ธฐ ์œ„ํ•œ ๋ฐฉ์•ˆ 31 3.2.1 Local-patch ์žฌ์‚ฌ์šฉ ๋ฐฉ์•ˆ 31 3.2.2 Local-patch down sampling ๋ฐฉ์•ˆ 44 3.2.3 Gaussian image์˜ less significant bit ์ œ๊ฑฐ 47 3.2.4 Bandwidth ์ตœ์ ํ™” ๋ฐฉ์•ˆ์ด ์ ์šฉ๋œ SIFT ํ•˜๋“œ์›จ์–ด ๊ตฌ์กฐ 50 3.3 SIFT ํ•˜๋“œ์›จ์–ด์— ๋Œ€ํ•œ ์‹คํ—˜ ๊ฒฐ๊ณผ 55 3.3.1 SIFT ํ•˜๋“œ์›จ์–ด์˜ ์ŠคํŽ™ 55 3.3.2 ์™ธ๋ถ€ ๋ฉ”๋ชจ๋ฆฌ bandwidth ์š”๊ตฌ๋Ÿ‰ ๋ถ„์„ 57 3.3.3 ๋™์ž‘ ์†๋„ 60 3.3.4 Feature matching ์ •ํ™•๋„ 64 ์ œ 4 ์žฅ ASIFT ํ•˜๋“œ์›จ์–ด ๊ตฌ์กฐ 68 4.1 ASIFT ํ•˜๋“œ์›จ์–ด์— ์ ํ•ฉํ•œ affine transform ๋ฐฉ์‹ 68 4.1.1 ์ƒˆ๋กœ์šด affine transform ๋ฐฉ์‹ 68 4.1.2 ๋‚ด๋ถ€ image buffer์˜ ๋ฉ”๋ชจ๋ฆฌ ๊ณต๊ฐ„ ์ตœ์ ํ™” 74 4.2 ASIFT ํ•˜๋“œ์›จ์–ด์˜ ๊ตฌ์กฐ 78 4.2.1 ๊ธฐ๋ณธ ํ•˜๋“œ์›จ์–ด ๊ตฌ์กฐ ๋ฐ scaling ์—ฐ์‚ฐ ์žฌ์‚ฌ์šฉ 78 4.2.2 Affine transform parameter์˜ ๊ตฌ์„ฑ 81 4.2.3 ASIFT ํ•˜๋“œ์›จ์–ด ๊ตฌ์กฐ ์„ค๋ช… 85 4.3 ASIFT ํ•˜๋“œ์›จ์–ด์— ๋Œ€ํ•œ ์‹คํ—˜ ๊ฒฐ๊ณผ 89 4.3.1 ์ƒˆ affine transform ๋ฐฉ์‹์— ์˜ํ•œ ๋ฉ”๋ชจ๋ฆฌ latency ๊ฐ์†Œ 89 4.3.2 Affine transform module์˜ ์ถœ๋ ฅ bandwidth ํ–ฅ์ƒ 91 4.3.3 ASIFT ํ•˜๋“œ์›จ์–ด์˜ ์ŠคํŽ™๊ณผ ๋™์ž‘ ์†๋„ 93 4.3.4 Feature matching ์ •ํ™•๋„ 95 ์ œ 5 ์žฅ ๊ฒฐ๋ก  104 ์ฐธ๊ณ  ๋ฌธํ—Œ 106 Abstract 109Docto

    Towards a Common Software/Hardware Methodology for Future Advanced Driver Assistance Systems

    Get PDF
    The European research project DESERVE (DEvelopment platform for Safe and Efficient dRiVE, 2012-2015) had the aim of designing and developing a platform tool to cope with the continuously increasing complexity and the simultaneous need to reduce cost for future embedded Advanced Driver Assistance Systems (ADAS). For this purpose, the DESERVE platform profits from cross-domain software reuse, standardization of automotive software component interfaces, and easy but safety-compliant integration of heterogeneous modules. This enables the development of a new generation of ADAS applications, which challengingly combine different functions, sensors, actuators, hardware platforms, and Human Machine Interfaces (HMI). This book presents the different results of the DESERVE project concerning the ADAS development platform, test case functions, and validation and evaluation of different approaches. The reader is invited to substantiate the content of this book with the deliverables published during the DESERVE project. Technical topics discussed in this book include:Modern ADAS development platforms;Design space exploration;Driving modelling;Video-based and Radar-based ADAS functions;HMI for ADAS;Vehicle-hardware-in-the-loop validation system

    GPU implementation of video analytics algorithms for aerial imaging

    Get PDF
    This work examines several algorithms that together make up parts of an image processing pipeline called Video Mosaicing and Summarization (VMZ). This pipeline takes as input geospatial or biomedical videos and produces large stitched-together frames (mosaics) of the video's subject. The content of these videos presents numerous challenges, such as poor lighting and a rapidly changing scene. The algorithms of VMZ were chosen carefully to address these challenges. With the output of VMZ, numerous tasks can be done. Stabilized imagery allows for easier object tracking, and the mosaics allow a quick understanding of the scene. These use-cases with aerial imagery are even more valuable when considered from the edge, where they can be applied as a drone is collecting the data. When executing video analytics algorithms, one of the most important metrics for real-life use is performance. All the accuracy in the world does not guarantee usefulness if the algorithms cannot provide that accuracy in a timely and actionable manner. Thus the goal of this work is to explore means and tools to implement video analytics algorithms, particularly the ones that make up the VMZ pipeline, on GPU devices{making them faster and more available for real-time use. This work presents four algorithms that have been converted to make use of the GPU in the GStreamer environment on NVIDIA GPUs. With GStreamer these algorithms are easily modular and lend themselves well to experimentation and real-life use even in pipelines beyond VMZ.Includes bibliographical references

    Medical Image Registration Using Artificial Neural Network

    Get PDF
    Image registration is the transformation of different sets of images into one coordinate system in order to align and overlay multiple images. Image registration is used in many fields such as medical imaging, remote sensing, and computer vision. It is very important in medical research, where multiple images are acquired from different sensors at various points in time. This allows doctors to monitor the effects of treatments on patients in a certain region of interest over time. In this thesis, artificial neural networks with curvelet keypoints are used to estimate the parameters of registration. Simulations show that the curvelet keypoints provide more accurate results than using the Discrete Cosine Transform (DCT) coefficients and Scale Invariant Feature Transform (SIFT) keypoints on rotation and scale parameter estimation

    Image Stitching for UAV remote sensing application

    Get PDF
    The objective of the project is to write an algorithm that is able to join top view images to create a big map. The project is done in the School of Castelldefels of UPC, within the research laboratory Icarus of EETAC Faculty. The goal of the project is to detect an area of this map, thanks to the analysis of this images. The images are taken by the two camera aboard on an Unmanned Aerial Vehicle (UAV) built by the Icarus group leaded by Enric Pastor. The implemented code is uploaded in Upc' svn at the adress: https://svn.fib.upc.es/svn/vincenzo.can

    Accelerated Object Tracking with Local Binary Features

    Get PDF
    Multi-object tracking is a problem with wide application in modern computing. Object tracking is leveraged in areas such as human computer interaction, autonomous vehicle navigation, panorama generation, as well as countless other robotic applications. Several trackers have demonstrated favorable results for tracking of single objects. However, modern object trackers must make significant tradeoffs in order to accommodate multiple objects while maintaining real-time performance. These tradeoffs include sacrifices in robustness and accuracy that adversely affect the results. This thesis details the design and multiple implementations of an object tracker that is focused on computational efficiency. The computational efficiency of the tracker is achieved through use of local binary descriptors in a template matching approach. Candidate templates are matched to a dictionary composed of both static and dynamic templates to allow for variation in the appearance of the object while minimizing the potential for drift in the tracker. Locality constraints have been used to reduce tracking jitter. Due to the significant promise for parallelization, the tracking algorithm was implemented on the Graphics Processing Unit (GPU) using the CUDA API. The tracker\u27s efficiency also led to its implantation on a mobile platform as one of the mobile trackers that can accurately track at faster than realtime speed. Benchmarks were performed to compare the proposed tracker to state of the art trackers on a wide range of standard test videos. The tracker implemented in this work has demonstrated a higher degree of accuracy while operating several orders of magnitude faster

    Towards a Common Software/Hardware Methodology for Future Advanced Driver Assistance Systems

    Get PDF
    The European research project DESERVE (DEvelopment platform for Safe and Efficient dRiVE, 2012-2015) had the aim of designing and developing a platform tool to cope with the continuously increasing complexity and the simultaneous need to reduce cost for future embedded Advanced Driver Assistance Systems (ADAS). For this purpose, the DESERVE platform profits from cross-domain software reuse, standardization of automotive software component interfaces, and easy but safety-compliant integration of heterogeneous modules. This enables the development of a new generation of ADAS applications, which challengingly combine different functions, sensors, actuators, hardware platforms, and Human Machine Interfaces (HMI). This book presents the different results of the DESERVE project concerning the ADAS development platform, test case functions, and validation and evaluation of different approaches. The reader is invited to substantiate the content of this book with the deliverables published during the DESERVE project. Technical topics discussed in this book include:Modern ADAS development platforms;Design space exploration;Driving modelling;Video-based and Radar-based ADAS functions;HMI for ADAS;Vehicle-hardware-in-the-loop validation system
    • โ€ฆ
    corecore