{"id":1669,"date":"2015-07-27T14:47:39","date_gmt":"2015-07-27T14:47:39","guid":{"rendered":"http:\/\/cvlab-dresden.de\/?page_id=1669"},"modified":"2020-08-19T10:49:35","modified_gmt":"2020-08-19T10:49:35","slug":"iccv2015-occlusion-challenge","status":"publish","type":"page","link":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/iccv2015-occlusion-challenge\/","title":{"rendered":"ICCV2015 Occluded Object Challenge"},"content":{"rendered":"<h2><a name=\"Challenge\"><\/a>Challenge:<\/h2>\n<p>The purpose of this challenge is to compare different methods for object pose estimation in a realistic setting featuring heavy occlusion. Our <a href=\"#Dataset\">dataset<\/a> includes eight objects in a cluttered scene. Given a RGB-D image the method has to estimate the position and orientation (a total of six degrees of freedom) of each object. You can participate by applying your method to our data and <a href=\"#Submit\">submitting<\/a> your results. We will <a href=\"#Evaluate\">evaluate<\/a> submitted results according to multiple metrics and display the scores for <a href=\"#Scores\">comparison<\/a>.<\/p>\n<h2><a name=\"Scores\"><\/a>Scores:<\/h2>\n<table id=\"myTable\" class=\"tablesorter\" style=\"border-style: solid;\">\n<thead>\n<tr>\n<th>Method<\/th>\n<th><strong>AD<\/strong><\/th>\n<th><strong>5cm, 5deg<\/strong><\/th>\n<th><strong>IOU<br \/>\n<\/strong><\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td><a href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/research\/scene-understanding\/pose-estimation\/#ECCV14\">Learning 6D Object Pose Estimation<\/a><\/td>\n<td>56.58%<\/td>\n<td>22.74%<\/td>\n<td>62.02%<\/td>\n<\/tr>\n<tr>\n<td><\/td>\n<td><\/td>\n<td><\/td>\n<td><\/td>\n<\/tr>\n<tr>\n<td><\/td>\n<td><\/td>\n<td><\/td>\n<td><\/td>\n<\/tr>\n<tr>\n<td><\/td>\n<td><\/td>\n<td><\/td>\n<td><\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<h2><a name=\"Dataset\"><\/a> Dataset:<\/h2>\n<p><strong>NOTE<\/strong>: Below\u00a0you find the version of the occlusion dataset as it was used in our ICCV15 challenge. However, we released a reworked version of the dataset as part of the <a href=\"https:\/\/bop.felk.cvut.cz\/datasets\/\">BOP Challenge<\/a>. The reworked version contains all data (images, poses, 3D models of objects) and some annotation errors have been corrected. We advise to use the reworked version of the dataset.<\/p>\n<p>You can find our <a href=\"https:\/\/cloudstore.zih.tu-dresden.de\/index.php\/s\/a65ec05fedd4890ae8ced82dfcf92ad8\/download\">dataset here<\/a>. Please cite <a href=\"#Brachmann2014\">[Brachmann2014]<\/a> and <a href=\"#Hinterstoisser2012\">[Hinterstoisser2012]<\/a> when using it. Brachmann <em>et al.<\/em> provided additional annotations for a RGB-D sequence originally published By Hinterstoisser <em>et al.<\/em> Description of file formats and folder structure can be <a href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/HTML\/projects\/scene_understanding\/OcclusionDataset_ReadMe.pdf\">found here<\/a>. We provide the dataset under the\u00a0<a href=\"https:\/\/creativecommons.org\/licenses\/by-sa\/4.0\/\">CC BY-SA 4.0<\/a>\u00a0license.<\/p>\n<h2><a name=\"Evaluate\"><\/a>Evaluation:<\/h2>\n<p>We calculate the percentage of correctly estimated poses. We use three different criteria:<\/p>\n<ol>\n<li>The <strong>AD<\/strong> criterion <a href=\"#Hinterstoisser2012\">[Hinterstoisser2012]<\/a>: We calculate the Average Distance (AD) between all vertices in the 3D model of the part in the estimated pose and the ground truth pose. A pose is considered correct, when this average distance is below 10% of the object diameter.<\/li>\n<li><strong>5cm, 5deg <\/strong><a href=\"#Shotton2013\">[Shotton2013]<\/a>: A pose is considered correct when the translational error is below 5cm and the rotational error is below 5deg.<\/li>\n<li>The<strong> IOU<\/strong> criterion<strong>: <\/strong>We calculate the 2D axis aligned bounding boxes of the part in the estimated pose and ground truth pose. We calculate the IOU (Intersection Over Union) of the bounding boxes. A pose is considered correct, when this value is above a threshold of 0.5.<\/li>\n<\/ol>\n<h2><a name=\"Submit\"><\/a>How to participate?<\/h2>\n<p>In order to participate you have to:<\/p>\n<ol>\n<li>Download the <a href=\"https:\/\/cloudstore.zih.tu-dresden.de\/index.php\/s\/a65ec05fedd4890ae8ced82dfcf92ad8\/download\">dataset<\/a>.<\/li>\n<li>Apply your method. You can use anything as training data except the test sequences provided in the dataset.<\/li>\n<li>Write your pose estimates to .info text files using the exact format as in the dataset. <a href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/HTML\/projects\/scene_understanding\/OcclusionDataset_ReadMe.pdf\">It is described here in Section 2.2<\/a>.<\/li>\n<li>Compress your results to a single tar.gz file. Use the same folder structure as in this <a href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/HTML\/projects\/scene_understanding\/sample_occlusionChallenge.tgz\">sample.tar.gz<\/a>.<\/li>\n<li>Email your results to <strong>occlusion-challenge&lt;at&gt;cvlab-dresden&lt;dot&gt;de<\/strong>. Please provide the name of the method as well as a URL pointing to your project page or publication.<\/li>\n<\/ol>\n<h2>References<\/h2>\n<p><a name=\"Hinterstoisser2012\"><\/a>[Hinterstoisser2012]: Stefan Hinterstoisser, Vincent Lepetit, Slobodan Ilic, Stefan Holzer, Gary R. Bradski,<br \/>\nKurt Konolige, Nassir Navab:<br \/>\n<em>Model Based Training, Detection and Pose Estimation of Texture-Less 3D Objects in Heavily Cluttered Scenes.<\/em> ACCV 2012<\/p>\n<p><a name=\"Shotton2013\"><\/a>[Shotton2013]: Jamie Shotton, Ben Glocker, Christopher Zach, Shahram Izadi, Antonio Criminisi, Andrew Fitzgibbon:<br \/>\n<em>Scene Coordinate Regression Forests for Camera Relocalization in RGB-D Images.<\/em> CVPR 2013<\/p>\n<p><a name=\"Brachmann2014\"><\/a>[Brachmann2014]: Eric Brachmann, Alexander Krull, Frank Michel, Stefan Gumhold, Jamie Shotton, Carsten Rother:<br \/>\n<em>Learning 6D Object Pose Estimation using 3D Object Coordinates.<\/em> ECCV 2014<\/p>\n<p><a name=\"Michel2015\"><\/a>[Michel2015]: Frank Michel, Alexander Krull, Eric Brachmann, Michael. Y. Yang, Stefan Gumhold, Carsten Rother:<br \/>\n<em>Pose Estimation of Kinematic Chain Instances via Object Coordinate Regression.<\/em> BMVC 2015.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Challenge: The purpose of this challenge is to compare different methods for object pose estimation in a realistic setting featuring heavy occlusion. Our dataset includes eight objects in a cluttered scene. Given a RGB-D image the method has to estimate the position and orientation (a total of six degrees of freedom) of each object. You [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"open","ping_status":"open","template":"","meta":{"inline_featured_image":false,"footnotes":""},"class_list":["post-1669","page","type-page","status-publish","hentry","post"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.3 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>ICCV2015 Occluded Object Challenge - Computer Vision and Learning Lab Heidelberg<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/iccv2015-occlusion-challenge\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"ICCV2015 Occluded Object Challenge - Computer Vision and Learning Lab Heidelberg\" \/>\n<meta property=\"og:description\" content=\"Challenge: The purpose of this challenge is to compare different methods for object pose estimation in a realistic setting featuring heavy occlusion. Our dataset includes eight objects in a cluttered scene. Given a RGB-D image the method has to estimate the position and orientation (a total of six degrees of freedom) of each object. You [&hellip;]\" \/>\n<meta property=\"og:url\" content=\"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/iccv2015-occlusion-challenge\/\" \/>\n<meta property=\"og:site_name\" content=\"Computer Vision and Learning Lab Heidelberg\" \/>\n<meta property=\"article:modified_time\" content=\"2020-08-19T10:49:35+00:00\" \/>\n<meta name=\"twitter:label1\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data1\" content=\"3 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/iccv2015-occlusion-challenge\\\/\",\"url\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/iccv2015-occlusion-challenge\\\/\",\"name\":\"ICCV2015 Occluded Object Challenge - Computer Vision and Learning Lab Heidelberg\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/#website\"},\"datePublished\":\"2015-07-27T14:47:39+00:00\",\"dateModified\":\"2020-08-19T10:49:35+00:00\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/iccv2015-occlusion-challenge\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/iccv2015-occlusion-challenge\\\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/iccv2015-occlusion-challenge\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"ICCV2015 Occluded Object Challenge\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/#website\",\"url\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/\",\"name\":\"Computer Vision and Learning Lab Heidelberg\",\"description\":\"Computer Vision and Learning Lab Heidelberg\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/hci.iwr.uni-heidelberg.de\\\/vislearn\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"ICCV2015 Occluded Object Challenge - Computer Vision and Learning Lab Heidelberg","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/iccv2015-occlusion-challenge\/","og_locale":"en_US","og_type":"article","og_title":"ICCV2015 Occluded Object Challenge - Computer Vision and Learning Lab Heidelberg","og_description":"Challenge: The purpose of this challenge is to compare different methods for object pose estimation in a realistic setting featuring heavy occlusion. Our dataset includes eight objects in a cluttered scene. Given a RGB-D image the method has to estimate the position and orientation (a total of six degrees of freedom) of each object. You [&hellip;]","og_url":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/iccv2015-occlusion-challenge\/","og_site_name":"Computer Vision and Learning Lab Heidelberg","article_modified_time":"2020-08-19T10:49:35+00:00","twitter_misc":{"Est. reading time":"3 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/iccv2015-occlusion-challenge\/","url":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/iccv2015-occlusion-challenge\/","name":"ICCV2015 Occluded Object Challenge - Computer Vision and Learning Lab Heidelberg","isPartOf":{"@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/#website"},"datePublished":"2015-07-27T14:47:39+00:00","dateModified":"2020-08-19T10:49:35+00:00","breadcrumb":{"@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/iccv2015-occlusion-challenge\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/iccv2015-occlusion-challenge\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/iccv2015-occlusion-challenge\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/"},{"@type":"ListItem","position":2,"name":"ICCV2015 Occluded Object Challenge"}]},{"@type":"WebSite","@id":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/#website","url":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/","name":"Computer Vision and Learning Lab Heidelberg","description":"Computer Vision and Learning Lab Heidelberg","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"}]}},"_links":{"self":[{"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/pages\/1669","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/comments?post=1669"}],"version-history":[{"count":26,"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/pages\/1669\/revisions"}],"predecessor-version":[{"id":4930,"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/pages\/1669\/revisions\/4930"}],"wp:attachment":[{"href":"https:\/\/hci.iwr.uni-heidelberg.de\/vislearn\/wp-json\/wp\/v2\/media?parent=1669"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}