About Me
AboutMe.Main History
Hide minor edits - Show changes to output
Changed line 28 from:
* Did not get to work on medical apps myself, but ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
to:
* Did not get to work on medical apps myself, but `ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
Changed lines 9-10 from:
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. After some initial R&D projects we got selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. Watch our [[http://www.youtube.com/watch?v=9URIR-dEWBM|first prototype]] in action in 2011!
to:
The success of [[http://rgbdemo.org|RGBDemo]] led me to cofound `ManCTL in 2011. After some initial R&D projects we got selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. Watch our [[http://www.youtube.com/watch?v=9URIR-dEWBM|first prototype]] in action in 2011!
Deleted lines 16-18:
(:youtube YPLJsYYzFA4:)
Added lines 18-20:
* Kept improving our RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
(:youtube YPLJsYYzFA4:)
Changed line 9 from:
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. After some initial R&D projects we got selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. Watch our [[http://www.youtube.com/watch?v=9URIR-dEWBM|first prototype]] in action in 2011.
to:
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. After some initial R&D projects we got selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. Watch our [[http://www.youtube.com/watch?v=9URIR-dEWBM|first prototype]] in action in 2011!
Changed lines 15-17 from:
* Launch of Structure Sensor, wrote the first version of the Objective C API for [[http://structure.io/|Structure SDK]].
* RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
*
to:
* Launch of Structure Sensor, worked with our small team on the 3D reconstruction stack and wrote the first version of the Objective C API for [[http://structure.io/|Structure SDK]] to expose it to developers.
* Improved our RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
* Improved our RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
Changed line 28 from:
* Did not work on actual medical apps myself, but ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
to:
* Did not get to work on medical apps myself, but ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
Changed line 37 from:
* Positional tracking for AR/VR with a single camera and an IMU. Depth sensor not always required anymore :)
to:
* Positional tracking for AR/VR with a single camera and an IMU. Port from mobile to Windows/PC. Depth sensor not always required anymore :)
Changed line 15 from:
* Launch of Structure Sensor, wrote the first version of the API (Objective C) for [[http://structure.io/|Structure SDK]].
to:
* Launch of Structure Sensor, wrote the first version of the Objective C API for [[http://structure.io/|Structure SDK]].
Changed line 7 from:
I am a PhD in computer vision with a strong interest in real-time 3D tracking and mapping using RGB-D cameras. Between 2009 and 2012 I was a postdoc at Carlos III University of Madrid working on object 3D reconstruction and tracking for robotics.
to:
I am a PhD in computer vision with a strong interest in real-time 3D tracking and mapping. Between 2009 and 2012 I was a postdoc at Carlos III University of Madrid working on object 3D reconstruction and tracking for robotics.
Changed lines 9-12 from:
In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. Our [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, became the #6 most funded project on Kickstarter.
<br>
to:
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. After some initial R&D projects we got selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. Watch our [[http://www.youtube.com/watch?v=9URIR-dEWBM|first prototype]] in action in 2011.
In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. We participated to the launch of [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, and it became the #6 most funded project on Kickstarter.
Here are some of the projects I've participated to at Occipital:
* Launch of Structure Sensor, wrote the first version of the API (Objective C) for [[http://structure.io/|Structure SDK]].
* RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
(:youtube YPLJsYYzFA4:)
* [[https://apps.apple.com/us/app/structure-sensor-calibrator/id914275485|Calibrator]] iOS app to calibrate the iOS color camera with Structure Sensor, using feature matching between the IR camera in the sensor and color.
* Unbounded positional tracking using RGBD for CES 2015.
(:youtube UfQgkzfDwHw:)
* Mixed reality demo for iOS at CES 2016. Combining RGBD tracking, 3D reconstruction, and physics via an integration to the SceneKit (and later on Unity) game engines.
(:youtube cEnnbCSbijo:)
* Did not work on actual medical apps myself, but ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
(:youtube 9LgmQYkRiSY:)
* We adapted Bridge Engine to launch a VR headset for iPhone. Optimized for latency, and leveraged visual-inertial sensor fusion for pose prediction.
(:youtube qbkwew3bfWU:)
* Launch of Canvas (scan your home), that leveraged our work on real-time unbounded large-scale SLAM for mobile.
(:youtube XA7FMoNAK9M:)
* Positional tracking for AR/VR with a single camera and an IMU. Depth sensor not always required anymore :)
(:youtube aVdWED6kfKc:)
* [[http://tapmeasure.io|TapMeasure]] led a small team to build that iOS app in a very short time to leverage `ARKit and take 3D measurements.
* Positional tracking for AR/VR extended to stereo and including more room perception for CES 2018.
(:youtube ra4u5np4HXk:)
<br>
In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. We participated to the launch of [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, and it became the #6 most funded project on Kickstarter.
Here are some of the projects I've participated to at Occipital:
* Launch of Structure Sensor, wrote the first version of the API (Objective C) for [[http://structure.io/|Structure SDK]].
* RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
(:youtube YPLJsYYzFA4:)
* [[https://apps.apple.com/us/app/structure-sensor-calibrator/id914275485|Calibrator]] iOS app to calibrate the iOS color camera with Structure Sensor, using feature matching between the IR camera in the sensor and color.
* Unbounded positional tracking using RGBD for CES 2015.
(:youtube UfQgkzfDwHw:)
* Mixed reality demo for iOS at CES 2016. Combining RGBD tracking, 3D reconstruction, and physics via an integration to the SceneKit (and later on Unity) game engines.
(:youtube cEnnbCSbijo:)
* Did not work on actual medical apps myself, but ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
(:youtube 9LgmQYkRiSY:)
* We adapted Bridge Engine to launch a VR headset for iPhone. Optimized for latency, and leveraged visual-inertial sensor fusion for pose prediction.
(:youtube qbkwew3bfWU:)
* Launch of Canvas (scan your home), that leveraged our work on real-time unbounded large-scale SLAM for mobile.
(:youtube XA7FMoNAK9M:)
* Positional tracking for AR/VR with a single camera and an IMU. Depth sensor not always required anymore :)
(:youtube aVdWED6kfKc:)
* [[http://tapmeasure.io|TapMeasure]] led a small team to build that iOS app in a very short time to leverage `ARKit and take 3D measurements.
* Positional tracking for AR/VR extended to stereo and including more room perception for CES 2018.
(:youtube ra4u5np4HXk:)
<br>
Changed lines 7-11 from:
I am a PhD in computer vision with a strong interest in real-time 3D tracking and mapping using RGB-D cameras. Between 2009 and 2012 I was a postdoc at Carlos III University of Madrid working on object 3D reconstruction and tracking for robotics. The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. We were selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. Our [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, became the #6 most funded project on Kickstarter.
to:
I am a PhD in computer vision with a strong interest in real-time 3D tracking and mapping using RGB-D cameras. Between 2009 and 2012 I was a postdoc at Carlos III University of Madrid working on object 3D reconstruction and tracking for robotics.
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. We were selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras.
In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. Our [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, became the #6 most funded project on Kickstarter.
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. We were selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras.
In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. Our [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, became the #6 most funded project on Kickstarter.
Changed lines 7-9 from:
I am
to:
I am a PhD in computer vision with a strong interest in real-time 3D tracking and mapping using RGB-D cameras. Between 2009 and 2012 I was a postdoc at Carlos III University of Madrid working on object 3D reconstruction and tracking for robotics. The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. We were selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. Our [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, became the #6 most funded project on Kickstarter.
Changed line 9 from:
I am a PhD in computer vision and since 2009 I am a postdoctoral scholar at Carlos III University of Madrid. There, I am working on the perception aspects of the [[http://handle-project.eu|Handle]] FP7 european project on dextrous robotic manipulation. I am also involved in the Prosave project to study the suitability of Time-of-Flight cameras for airplane applications, in collaboration with Airbus Military. My main interests are computer vision for robotics, RGB-D cameras, object recognition and statistical approaches to visual event detection.
to:
I am a PhD in computer vision and since 2009 I am a postdoctoral scholar at Carlos III University of Madrid. There, I am working on the perception aspects of the [[http://handle-project.eu|Handle]] FP7 european project on dextrous robotic manipulation. I am also involved in the Prosave project to study the suitability of Time-of-Flight cameras for airplane applications, in collaboration with Airbus Military. My main interests are computer vision for robotics, RGB-D cameras, object recognition and statistical approaches to visual event detection.
Changed lines 7-121 from:
!! [[#CV]] CV
<!--HEVEA command line is: /usr/bin/hevea cv.tex -->
<h1 align="center">Docteur en Informatique<br><br>
<font size="h3"><i><b>Intelligence Artificielle et Imagerie</b></i></font></h1><br>
<h3>F O R M A T I O N</h3>
<hr>
<table cellspacing="10" width="100%">
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2005 - 2008</b></td>
<td>Doctorat à l'<b>UPMC</b> (Université Pierre et Marie Curie, Paris 6), laboratoire d'acceuil à l'<b>ENSTA</b> (École Nationale Supérieure de Techniques Avancées) sous la direction de [[http://www.ensta.fr/~tbernard|Thierry Bernard]] (ENSTA) et de [[http://rfv.insa-lyon.fr/~jolion/|Jean-Michel Jolion]] (<b>LIRIS - INSA Lyon</b>). <i>Apprentissage a-contrario et architecture efficace pour la détection d'évènements visuels significatifs ([[Research.Main#PhD|plus de détails]]).</i></td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2004 - 2005</b></td>
<td>Master M2 de recherche IAD (Intelligence Artificielle et Décision) à l'université Pierre et Marie Curie (Paris VI). Spécialisation Image et Son. Mention très bien.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2002 - 2004</b></td>
<td>Formation par la recherche au <b>LRDE</b> (Laboratoire de Recherche et de Développement d'EPITA), dirigé par Dr. Akim Demaille.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>1999 - 2004</b></td>
<td>Cycle préparatoire et cycle ingénierie à l'<b>EPITA</b> (Ecole Pour l'Informatique et les Techniques Avancées). Mention très bien.</td>
</tr>
</table><br>
<h3>E N S E I G N E M E N T</h3>
<hr>
<table cellspacing="10" width="100%">
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2007</b></td>
<td><b>ENSTA</b> : encadrement de projets en C (20h)</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2007</b></td>
<td><b>UPMC</b> (Université Pierre et Marie Curie) : TD et TP de <b>programmation par objets, design patterns</b> (50h)</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2006</b></td>
<td><b>UPMC</b> : TP de C et de shell Unix (75h).</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2005</b></td>
<td><b>ESILV</b> (École Supérieure d'Ingénieurs Léonard de Vinci) : TP et TD de <b>Java</b> (15h), TP de gestion de projet (10h).</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2005</b></td>
<td><b>CESI</b> (Centre d'Etudes Supérieures Industrielles) : mise en place d'un module d'introduction à Linux (20h).</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2005</b></td>
<td><b>Acadomia</b> : cours de soutien en maths et anglais.</td>
</tr>
</table><br>
<h3>E X P É R I E N C E P R O F E S S I O N N E L L E</h3>
<hr>
<table cellspacing="10" width="100%">
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2009 - 6 mois</b></td>
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour l'estimation monoculaire de pose du corps humain.<br> Application à l'analyse visuelle du langage des signes.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2005 - 6 mois</b></td>
<td><b>ENSTA</b> : stage de Master.<br> Détection statistique d'objets saillants sur rétine artificielle programmable.<br> Application à la détection de segments significatifs.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2004 - 7 mois</b></td>
<td><b>Siemens Corporate Research</b> (Princeton, États-Unis) : stage de recherche au sein du département imagerie.<br> Recherche d'un modèle d'application des filtres de particules au suivi de fibres cérébrales.<br> Recherche et implémentation (<b>C++</b>, <b>Windows/Linux</b>) d'algorithmes de visualisation d'images de tenseur de diffusion (DT-MRI).<br>[[http://www.freshpatents.com/Systems-devices-and-methods-for-diffusion-tractography-dt20061012ptan20060229856.php | Brevet]]</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2002 - 2003</b></td>
<td><b>LRDE</b> : recherche et développement.<br> Développement d'Evidenz, un moteur de raisonnement générique à base de théorie de l'évidence (Dempster-Shafer). Application au traitement d'image.<br> Participation à la conception et au développement d'Olena, une bibliothèque générique de traitement d'image en <b>C++ (metaprogramming)</b>.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2002 - 3 mois</b></td>
<td><b>Radio France</b> : stage au département de recherche et développement. Conception et réalisation d'applications de surveillance de flux audio en <b>C++</b> sous environnement <b>multiplate-forme</b> (Linux, `NetBSD, `OpenBSD, `FreeBSD).</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2001 - 2 mois</b></td>
<td>Snecma Services : génération de statistiques automatisées (VBA, Access).</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2000 - 2 mois</b></td>
<td>`WebValley : administration réseau sous <b>Linux</b> (serveur web / ftp, monitoring, sauvegardes, ...)</td>
</tr>
</table><br>
<h3>P U B L I C A T I O N S</h3>
<hr>
<br>
[[Research.Main|Check the research page]]
<br>
<h3>D I V E R S</h3>
<hr>
<table cellspacing="10" width="100%">
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>Langues</b></td>
<td>Anglais: courant (7 mois de vie aux États-Unis).<br>Espagnol: scolaire.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;">
<b>Activités</b></td>
<td>Contributions à divers logiciels libres.<br>Sports : volley, natation, roller.<br>Voyages.</td>
</tr>
</table><br>
<hr size="2">
<blockquote>
<em>This document was translated from L<sup>A</sup>T<sub>E</sub>X by</em> [[http://pauillac.inria.fr/~maranget/hevea/index.html | <em>H<font size="2"><sup>E</sup></font>V<font size="2"><sup>E</sup></font>A</em>]]<em>.</em>
</blockquote>
to:
'''Resume in english''' [[{$PubDirUrl}/files/aboutme/resume-burrus-march-2011.pdf | PDF]]
I am a PhD in computer vision and since 2009 I am a postdoctoral scholar at Carlos III University of Madrid. There, I am working on the perception aspects of the [[http://handle-project.eu|Handle]] FP7 european project on dextrous robotic manipulation. I am also involved in the Prosave project to study the suitability of Time-of-Flight cameras for airplane applications, in collaboration with Airbus Military. My main interests are computer vision for robotics, RGB-D cameras, object recognition and statistical approaches to visual event detection.
I am a PhD in computer vision and since 2009 I am a postdoctoral scholar at Carlos III University of Madrid. There, I am working on the perception aspects of the [[http://handle-project.eu|Handle]] FP7 european project on dextrous robotic manipulation. I am also involved in the Prosave project to study the suitability of Time-of-Flight cameras for airplane applications, in collaboration with Airbus Military. My main interests are computer vision for robotics, RGB-D cameras, object recognition and statistical approaches to visual event detection.
Changed lines 7-8 from:
Email: mailto:burrus@montefiore.ulg.ac.be
to:
Email: mailto:nicolas@burrus.name
Changed line 72 from:
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour l'estimation monoculaire de pose d'humains.<br> Application à l'analyse visuelle du langage des signes.</td>
to:
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour l'estimation monoculaire de pose du corps humain.<br> Application à l'analyse visuelle du langage des signes.</td>
Changed line 72 from:
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour le suivi de pose d'humains.<br> Application à l'analyse du langage des signes.</td>
to:
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour l'estimation monoculaire de pose d'humains.<br> Application à l'analyse visuelle du langage des signes.</td>
Added lines 71-74:
<td style="white-space: nowrap; vertical-align: top;"><b>2009 - 6 mois</b></td>
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour le suivi de pose d'humains.<br> Application à l'analyse du langage des signes.</td>
</tr>
<tr>
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour le suivi de pose d'humains.<br> Application à l'analyse du langage des signes.</td>
</tr>
<tr>
Changed line 76 from:
<td><b>Université de Liège</b> : postdoc.<br> \textbf{Approches statistiques} pour le suivi de pose d'humains.<br> Application à l'analyse du langage des signes.</td>
to:
<td><b>ENSTA</b> : stage de Master.<br> Détection statistique d'objets saillants sur rétine artificielle programmable.<br> Application à la détection de segments significatifs.</td>
Deleted lines 78-81:
<td><b>ENSTA</b> : stage de Master.<br> Détection statistique d'objets saillants sur rétine artificielle programmable.<br> Application à la détection de segments significatifs.</td>
</tr>
<tr>
Changed line 72 from:
<td><b>ENSTA</b> : stage de Master.<br> Détection statistique d'objets saillants sur rétine artificielle programmable.<br> Application à la détection de segments significatifs.</td>
to:
<td><b>Université de Liège</b> : postdoc.<br> \textbf{Approches statistiques} pour le suivi de pose d'humains.<br> Application à l'analyse du langage des signes.</td>
Added lines 75-78:
<td style="white-space: nowrap; vertical-align: top;"><b>2005 - 6 mois</b></td>
<td><b>ENSTA</b> : stage de Master.<br> Détection statistique d'objets saillants sur rétine artificielle programmable.<br> Application à la détection de segments significatifs.</td>
</tr>
<tr>
<td><b>ENSTA</b> : stage de Master.<br> Détection statistique d'objets saillants sur rétine artificielle programmable.<br> Application à la détection de segments significatifs.</td>
</tr>
<tr>
Changed lines 9-10 from:
'''Resume in english''' [[{$PubDirUrl}/files/aboutme/resume-burrus-nov-2008.pdf | PDF]]
to:
'''Resume in english''' [[{$PubDirUrl}/files/aboutme/resume-burrus-jul-2009.pdf | PDF]]
Changed line 22 from:
<td>Doctorat à l'UPMC (Université Pierre et Marie Curie, Paris 6), laboratoire d'acceuil à l'<b>ENSTA</b> (École Nationale Supérieure de Techniques Avancées) sous la direction de [[http://www.ensta.fr/~tbernard|Thierry Bernard]] (ENSTA) et de [[http://rfv.insa-lyon.fr/~jolion/|Jean-Michel Jolion]] (<b>LIRIS - INSA Lyon</b>). <i>Apprentissage a-contrario et architecture efficace pour la détection d'évènements visuels significatifs ([[Research.Main#PhD|plus de détails]]).</i></td>
to:
<td>Doctorat à l'<b>UPMC</b> (Université Pierre et Marie Curie, Paris 6), laboratoire d'acceuil à l'<b>ENSTA</b> (École Nationale Supérieure de Techniques Avancées) sous la direction de [[http://www.ensta.fr/~tbernard|Thierry Bernard]] (ENSTA) et de [[http://rfv.insa-lyon.fr/~jolion/|Jean-Michel Jolion]] (<b>LIRIS - INSA Lyon</b>). <i>Apprentissage a-contrario et architecture efficace pour la détection d'évènements visuels significatifs ([[Research.Main#PhD|plus de détails]]).</i></td>