{"id":1680,"date":"2015-07-29T13:57:01","date_gmt":"2015-07-29T13:57:01","guid":{"rendered":"http:\/\/cvlab-dresden.de\/?page_id=1680"},"modified":"2015-10-19T20:27:24","modified_gmt":"2015-10-19T20:27:24","slug":"graphflow","status":"publish","type":"page","link":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/","title":{"rendered":"GraphFlow"},"content":{"rendered":"<p><span style=\"font-family: georgia, palatino, serif;\">H. Abu Alhaija, A. Sellent, D. Kondermann, C. Rother. &#8220;GraphFlow &#8211; 6D \u00a0Large Displacement Scene Flow via Graph Matching&#8221;, German Conference on Pattern Recognition (GCPR, a.k.a. DAGM), 2015. [<a href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/HTML\/people\/hassan_abualhaija\/publications\/GraphFlow.pdf\">pdf<\/a>]<\/span><\/p>\n<p>We present an approach for computing dense scene flow from two large displacement RGB-D images. When dealing with large displacements the crucial step is to estimate the overall motion correctly. While state-of-the-art approaches focus on RGB information to establish guiding correspondences, we explore the power of depth edges. To achieve this, we present a new graph matching technique that brings sparse depth edges into correspondence. An additional contribution is the formulation of a continuous-label energy which is used to densify the sparse graph matching output. We present results on challenging Kinect images, for which we outperform state-of-the-art techniques.<\/p>\n<h2>Graph Matching<\/h2>\n<p>In our two stage scene flow approach, instead of using sparse texture matches only, we additionally utilize depth edges extracted from the RGB-D images that describe object boundaries well.However, in the presence of large motion, they are actually not trivially described and matched.While exact edge description suffers from occlusion and distortion effects, more robust edge descriptors often lead to ambiguous matches.\u00a0To disambiguate edge matches with robust descriptors, we use a structured matching approach in the form of graph matching that profits from non-local information to assign edge matches.<\/p>\n<hr \/>\n<p><a href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_080315_030041_PM.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\" wp-image-1733 size-large aligncenter\" src=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_080315_030041_PM-1024x334.jpg\" alt=\"Screenshot_080315_030041_PM\" width=\"720\" height=\"235\" srcset=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_080315_030041_PM-1024x334.jpg 1024w, https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_080315_030041_PM-300x98.jpg 300w, https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_080315_030041_PM.jpg 1581w\" sizes=\"auto, (max-width: 720px) 100vw, 720px\" \/><\/a><\/p>\n<p style=\"text-align: center;\"><em>Fig. 1: Details of graph matching. (a) Our edge description segments are\u00a0represented by their center point, average appearance descriptor that describe\u00a0the foreground region and normalized depth gradient vector. In order to compute\u00a0a description segment we accumulate neighboring edge pixels whose descriptor\u00a0variance is lower than a threshold \u001bt and whose count is between rmin = 20\u00a0and rmax = 30 pixels. (b) For graph matching, the description segments centers\u00a0are connected to form a graph. In particular, each description segment center is\u00a0connected to its N = 3 nearest neighbors with respect to the geodesic distance of\u00a0the depth map to avoid connections across large depth changes. (c) Illustration\u00a0of the geometry term \u0001used for graph matching.<\/em><\/p>\n<hr \/>\n<p>&nbsp;<\/p>\n<p>Given ground truth matches, we can assign three class labels to each description segment matched by the graph matching algorithm:<br \/>\na) correct match; b) almost correct match and c) wrong match.\u00a0We compare our graph matching with conventional nearest neighbor (SIFT) matching and Hungarian matching of the description segments that admits at most one-to-one matching between description segments.<\/p>\n<p>&nbsp;<\/p>\n<hr \/>\n<p><a href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_072915_033041_PM.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\" wp-image-1685 size-large aligncenter\" src=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_072915_033041_PM-1024x559.jpg\" alt=\"Screenshot_072915_033041_PM\" width=\"720\" height=\"393\" srcset=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_072915_033041_PM-1024x559.jpg 1024w, https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_072915_033041_PM-300x164.jpg 300w, https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_072915_033041_PM.jpg 1565w\" sizes=\"auto, (max-width: 720px) 100vw, 720px\" \/><\/a><\/p>\n<p style=\"text-align: center;\"><em>Figure 02: Visual comparison of graph matching results. For illustration pur-\u00a0pose both RGB images are super-imposed. Green means a correct match (or\u00a0occlusion), blue is an almost correct match (de\fnition in text) and red is a\u00a0wrong match. Our result is clearly superior to the other techniques.<\/em><\/p>\n<hr \/>\n<h2><\/h2>\n<h2>Scene Flow<\/h2>\n<p style=\"text-align: center;\"><a href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_072915_033705_PM.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\" wp-image-1688 size-large aligncenter\" src=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_072915_033705_PM-1024x455.jpg\" alt=\"Screenshot_072915_033705_PM\" width=\"720\" height=\"320\" srcset=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_072915_033705_PM-1024x455.jpg 1024w, https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_072915_033705_PM-300x133.jpg 300w, https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_072915_033705_PM.jpg 1527w\" sizes=\"auto, (max-width: 720px) 100vw, 720px\" \/><\/a><\/p>\n<p style=\"text-align: center;\">Figure 03 : results of the Scene Flow: (a) Original image pairs (b) using SphereFlow [Hornacek et at] (c) using SphereFlow initialized with\u00a0our Graph Matching (d) Our results using\u00a0GraphFlow.<\/p>\n<p><iframe loading=\"lazy\" title=\"GraphFlow - 6D Large Displacement Scene Flow via Graph Matching (GCPR 2015)\" width=\"720\" height=\"405\" src=\"https:\/\/www.youtube.com\/embed\/rsMY3v31iRE?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture\" allowfullscreen><\/iframe><\/p>\n<h2>Dataset<\/h2>\n<p>The RGBD sequences can be downloaded <a href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/dresden_dataset_images.zip\">here<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>H. Abu Alhaija, A. Sellent, D. Kondermann, C. Rother. &#8220;GraphFlow &#8211; 6D \u00a0Large Displacement Scene Flow via Graph Matching&#8221;, German Conference on Pattern Recognition (GCPR, a.k.a. DAGM), 2015. [pdf] We present an approach for computing dense scene flow from two large displacement RGB-D images. When dealing with large displacements the crucial step is to estimate [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"parent":700,"menu_order":2,"comment_status":"open","ping_status":"open","template":"","meta":{"inline_featured_image":false,"footnotes":""},"class_list":["post-1680","page","type-page","status-publish","hentry","post"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>GraphFlow - Computer Vision and Learning Lab Heidelberg<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"GraphFlow - Computer Vision and Learning Lab Heidelberg\" \/>\n<meta property=\"og:description\" content=\"H. Abu Alhaija, A. Sellent, D. Kondermann, C. Rother. &#8220;GraphFlow &#8211; 6D \u00a0Large Displacement Scene Flow via Graph Matching&#8221;, German Conference on Pattern Recognition (GCPR, a.k.a. DAGM), 2015. [pdf] We present an approach for computing dense scene flow from two large displacement RGB-D images. When dealing with large displacements the crucial step is to estimate [&hellip;]\" \/>\n<meta property=\"og:url\" content=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/\" \/>\n<meta property=\"og:site_name\" content=\"Computer Vision and Learning Lab Heidelberg\" \/>\n<meta property=\"article:modified_time\" content=\"2015-10-19T20:27:24+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_080315_030041_PM-1024x334.jpg\" \/>\n<meta name=\"twitter:label1\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data1\" content=\"2 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/research\\\/image-matching\\\/graphflow\\\/\",\"url\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/research\\\/image-matching\\\/graphflow\\\/\",\"name\":\"GraphFlow - Computer Vision and Learning Lab Heidelberg\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/research\\\/image-matching\\\/graphflow\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/research\\\/image-matching\\\/graphflow\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/wp-content\\\/uploads\\\/2015\\\/07\\\/Screenshot_080315_030041_PM-1024x334.jpg\",\"datePublished\":\"2015-07-29T13:57:01+00:00\",\"dateModified\":\"2015-10-19T20:27:24+00:00\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/research\\\/image-matching\\\/graphflow\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/research\\\/image-matching\\\/graphflow\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/research\\\/image-matching\\\/graphflow\\\/#primaryimage\",\"url\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/wp-content\\\/uploads\\\/2015\\\/07\\\/Screenshot_080315_030041_PM.jpg\",\"contentUrl\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/wp-content\\\/uploads\\\/2015\\\/07\\\/Screenshot_080315_030041_PM.jpg\",\"width\":1581,\"height\":515},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/research\\\/image-matching\\\/graphflow\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Research\",\"item\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/research\\\/\"},{\"@type\":\"ListItem\",\"position\":3,\"name\":\"Image Matching\",\"item\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/research\\\/image-matching\\\/\"},{\"@type\":\"ListItem\",\"position\":4,\"name\":\"GraphFlow\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/#website\",\"url\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/\",\"name\":\"Computer Vision and Learning Lab Heidelberg\",\"description\":\"Computer Vision and Learning Lab Heidelberg\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"GraphFlow - Computer Vision and Learning Lab Heidelberg","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/","og_locale":"en_US","og_type":"article","og_title":"GraphFlow - Computer Vision and Learning Lab Heidelberg","og_description":"H. Abu Alhaija, A. Sellent, D. Kondermann, C. Rother. &#8220;GraphFlow &#8211; 6D \u00a0Large Displacement Scene Flow via Graph Matching&#8221;, German Conference on Pattern Recognition (GCPR, a.k.a. DAGM), 2015. [pdf] We present an approach for computing dense scene flow from two large displacement RGB-D images. When dealing with large displacements the crucial step is to estimate [&hellip;]","og_url":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/","og_site_name":"Computer Vision and Learning Lab Heidelberg","article_modified_time":"2015-10-19T20:27:24+00:00","og_image":[{"url":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_080315_030041_PM-1024x334.jpg","type":"","width":"","height":""}],"twitter_misc":{"Est. reading time":"2 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/","url":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/","name":"GraphFlow - Computer Vision and Learning Lab Heidelberg","isPartOf":{"@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/#website"},"primaryImageOfPage":{"@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/#primaryimage"},"image":{"@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/#primaryimage"},"thumbnailUrl":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_080315_030041_PM-1024x334.jpg","datePublished":"2015-07-29T13:57:01+00:00","dateModified":"2015-10-19T20:27:24+00:00","breadcrumb":{"@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/#primaryimage","url":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_080315_030041_PM.jpg","contentUrl":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-content\/uploads\/2015\/07\/Screenshot_080315_030041_PM.jpg","width":1581,"height":515},{"@type":"BreadcrumbList","@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/graphflow\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/"},{"@type":"ListItem","position":2,"name":"Research","item":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/"},{"@type":"ListItem","position":3,"name":"Image Matching","item":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/image-matching\/"},{"@type":"ListItem","position":4,"name":"GraphFlow"}]},{"@type":"WebSite","@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/#website","url":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/","name":"Computer Vision and Learning Lab Heidelberg","description":"Computer Vision and Learning Lab Heidelberg","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"}]}},"_links":{"self":[{"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/pages\/1680","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/comments?post=1680"}],"version-history":[{"count":22,"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/pages\/1680\/revisions"}],"predecessor-version":[{"id":2099,"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/pages\/1680\/revisions\/2099"}],"up":[{"embeddable":true,"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/pages\/700"}],"wp:attachment":[{"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/media?parent=1680"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}