About Me

AboutMe.Main History

Show minor edits - Show changes to output

June 07, 2020, at 02:40 PM by 81.164.27.8 -
Changed line 28 from:
* Did not get to work on medical apps myself, but ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
to:
* Did not get to work on medical apps myself, but `ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
June 07, 2020, at 06:25 AM by 81.164.27.8 -
Changed lines 9-10 from:
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. After some initial R&D projects we got selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. Watch our [[http://www.youtube.com/watch?v=9URIR-dEWBM|first prototype]] in action in 2011!
to:
The success of [[http://rgbdemo.org|RGBDemo]] led me to cofound `ManCTL in 2011. After some initial R&D projects we got selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. Watch our [[http://www.youtube.com/watch?v=9URIR-dEWBM|first prototype]] in action in 2011!
Deleted lines 16-18:
* Improved our RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
(:youtube YPLJsYYzFA4:)

Added lines 18-20:

* Kept improving our RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
(:youtube YPLJsYYzFA4:)
June 07, 2020, at 06:24 AM by 81.164.27.8 -
Changed line 9 from:
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. After some initial R&D projects we got selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. Watch our [[http://www.youtube.com/watch?v=9URIR-dEWBM|first prototype]] in action in 2011.
to:
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. After some initial R&D projects we got selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. Watch our [[http://www.youtube.com/watch?v=9URIR-dEWBM|first prototype]] in action in 2011!
June 07, 2020, at 06:18 AM by 81.164.27.8 -
June 07, 2020, at 06:18 AM by 81.164.27.8 -
Changed lines 15-17 from:
* Launch of Structure Sensor, wrote the first version of the Objective C API for [[http://structure.io/|Structure SDK]].

*
RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
to:
* Launch of Structure Sensor, worked with our small team on the 3D reconstruction stack and wrote the first version of the Objective C API for [[http://structure.io/|Structure SDK]] to expose it to developers.

* Improved our
RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
Changed line 28 from:
* Did not work on actual medical apps myself, but ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
to:
* Did not get to work on medical apps myself, but ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
Changed line 37 from:
* Positional tracking for AR/VR with a single camera and an IMU. Depth sensor not always required anymore :)
to:
* Positional tracking for AR/VR with a single camera and an IMU. Port from mobile to Windows/PC. Depth sensor not always required anymore :)
June 07, 2020, at 06:14 AM by 81.164.27.8 -
Changed line 15 from:
* Launch of Structure Sensor, wrote the first version of the API (Objective C) for [[http://structure.io/|Structure SDK]].
to:
* Launch of Structure Sensor, wrote the first version of the Objective C API for [[http://structure.io/|Structure SDK]].
June 07, 2020, at 06:12 AM by 81.164.27.8 -
Changed line 7 from:
I am a PhD in computer vision with a strong interest in real-time 3D tracking and mapping using RGB-D cameras. Between 2009 and 2012 I was a postdoc at Carlos III University of Madrid working on object 3D reconstruction and tracking for robotics.
to:
I am a PhD in computer vision with a strong interest in real-time 3D tracking and mapping. Between 2009 and 2012 I was a postdoc at Carlos III University of Madrid working on object 3D reconstruction and tracking for robotics.
June 07, 2020, at 06:12 AM by 81.164.27.8 -
Changed lines 9-12 from:
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. We were selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras.

In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. Our [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, became the #6 most funded project on Kickstarter.

<br>
to:
The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. After some initial R&D projects we got selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. Watch our [[http://www.youtube.com/watch?v=9URIR-dEWBM|first prototype]] in action in 2011.

In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. We participated to the launch of [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, and it became the #6 most funded project on Kickstarter.

Here are some of the projects I've participated to at Occipital:

* Launch of Structure Sensor, wrote the first version of the API (Objective C) for [[http://structure.io/|Structure SDK]].

* RGBD tracking, 3D reconstruction and texturing in real-time on iOS.
(:youtube YPLJsYYzFA4:)

* [[https://apps.apple.com/us/app/structure-sensor-calibrator/id914275485|Calibrator]] iOS app to calibrate the iOS color camera with Structure Sensor, using feature matching between the IR camera in the sensor and color.

* Unbounded positional tracking using RGBD for CES 2015.
(:youtube UfQgkzfDwHw:)

* Mixed reality demo for iOS at CES 2016. Combining RGBD tracking, 3D reconstruction, and physics via an integration to the SceneKit (and later on Unity) game engines.
(:youtube cEnnbCSbijo:)

* Did not work on actual medical apps myself, but ManCTL started with an R&D project for foot orthotics, but I've been very proud to see the many medical use of the Structure Sensor SDK and our live 3d reconstruction.
(:youtube 9LgmQYkRiSY:)

* We adapted Bridge Engine to launch a VR headset for iPhone. Optimized for latency, and leveraged visual-inertial sensor fusion for pose prediction.
(:youtube qbkwew3bfWU:)

* Launch of Canvas (scan your home), that leveraged our work on real-time unbounded large-scale SLAM for mobile.
(:youtube XA7FMoNAK9M:)

* Positional tracking for AR/VR with a single camera and an IMU. Depth sensor not always required anymore :)
(:youtube aVdWED6kfKc:)

* [[http://tapmeasure.io|TapMeasure]] led a small team to build that iOS app in a very short time to leverage `ARKit and take 3D measurements.

* Positional tracking for AR/VR extended to stereo and including more room perception for CES 2018.
(:youtube ra4u5np4HXk:)

<br>
May 26, 2014, at 01:59 AM by 208.66.25.130 -
Changed lines 7-11 from:
I am a PhD in computer vision with a strong interest in real-time 3D tracking and mapping using RGB-D cameras. Between 2009 and 2012 I was a postdoc at Carlos III University of Madrid working on object 3D reconstruction and tracking for robotics. The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. We were selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. Our [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, became the #6 most funded project on Kickstarter.
to:
I am a PhD in computer vision with a strong interest in real-time 3D tracking and mapping using RGB-D cameras. Between 2009 and 2012 I was a postdoc at Carlos III University of Madrid working on object 3D reconstruction and tracking for robotics.

The
success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. We were selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras.

In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. Our [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, became the #6 most funded project on Kickstarter.
May 26, 2014, at 01:58 AM by 208.66.25.130 -
Changed lines 7-9 from:
'''Resume in english''' [[{$PubDirUrl}/files/aboutme/resume-burrus-march-2011.pdf | PDF]]

I am
a PhD in computer vision and since 2009 I am a postdoctoral scholar at Carlos III University of Madrid. There, I am working on the perception aspects of the [[http://handle-project.eu|Handle]] FP7 european project on dextrous robotic manipulation. I am also involved in the Prosave project to study the suitability of Time-of-Flight cameras for airplane applications, in collaboration with Airbus Military. My main interests are computer vision for robotics, RGB-D cameras, object recognition and statistical approaches to visual event detection.
to:
I am a PhD in computer vision with a strong interest in real-time 3D tracking and mapping using RGB-D cameras. Between 2009 and 2012 I was a postdoc at Carlos III University of Madrid working on object 3D reconstruction and tracking for robotics. The success of [[http://labs.manctl.com/RGBDemo|RGBDemo]], my opensource software showcasing the possibilities of the Microsoft Kinect, led me to cofound `ManCTL in 2011. We were selected to become a member of the Microsoft Kinect Accelerator powered by Techstars in 2012, and developed [[http://skanect.com|Skanect]], a real-time 3D scanning software compatible with low-cost RGB-D cameras. In 2013 we joined forces with [[http://occipital.com|Occipital]] to better explore the possibilities of depth sensing on mobile devices. Our [[http://structure.io|Structure Sensor]], the first depth sensor for mobile, became the #6 most funded project on Kickstarter.
April 04, 2011, at 10:01 PM by 92.151.173.114 -
Added line 10:
<br>
April 04, 2011, at 10:00 PM by 92.151.173.114 -
Changed line 9 from:
I am a PhD in computer vision and since 2009 I am a postdoctoral scholar at Carlos III University of Madrid. There, I am working on the perception aspects of the [[http://handle-project.eu|Handle]] FP7 european project on dextrous robotic manipulation. I am also involved in the Prosave project to study the suitability of Time-of-Flight cameras for airplane applications, in collaboration with Airbus Military. My main interests are computer vision for robotics, RGB-D cameras, object recognition and statistical approaches to visual event detection.
to:
I am a PhD in computer vision and since 2009 I am a postdoctoral scholar at Carlos III University of Madrid. There, I am working on the perception aspects of the [[http://handle-project.eu|Handle]] FP7 european project on dextrous robotic manipulation. I am also involved in the Prosave project to study the suitability of Time-of-Flight cameras for airplane applications, in collaboration with Airbus Military. My main interests are computer vision for robotics, RGB-D cameras, object recognition and statistical approaches to visual event detection.
April 04, 2011, at 10:00 PM by 92.151.173.114 -
Changed lines 7-121 from:
'''Resume in english''' [[{$PubDirUrl}/files/aboutme/resume-burrus-jul-2009.pdf | PDF]]

!! [[#CV]] CV

<!--HEVEA command line is: /usr/bin/hevea cv.tex -->
<h1 align="center">Docteur en Informatique<br><br>
<font size="h3"><i><b>Intelligence Artificielle et Imagerie</b></i></font></h1><br>

<h3>F O R M A T I O N</h3>
<hr>
<table cellspacing="10" width="100%">
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2005 - 2008</b></td>
<td>Doctorat &agrave; l'<b>UPMC</b> (Université Pierre et Marie Curie, Paris 6), laboratoire d'acceuil &agrave; l'<b>ENSTA</b> (&Eacute;cole Nationale Sup&eacute;rieure de Techniques Avanc&eacute;es) sous la direction de [[http://www.ensta.fr/~tbernard|Thierry Bernard]] (ENSTA) et de [[http://rfv.insa-lyon.fr/~jolion/|Jean-Michel Jolion]] (<b>LIRIS - INSA Lyon</b>). <i>Apprentissage a-contrario et architecture efficace pour la détection d'évènements visuels significatifs ([[Research.Main#PhD|plus de détails]]).</i></td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2004 - 2005</b></td>
<td>Master M2 de recherche IAD (Intelligence Artificielle et D&eacute;cision) &agrave; l'universit&eacute; Pierre et Marie Curie (Paris VI). Sp&eacute;cialisation Image et Son. Mention tr&egrave;s bien.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2002 - 2004</b></td>
<td>Formation par la recherche au <b>LRDE</b> (Laboratoire de Recherche et de D&eacute;veloppement d'EPITA), dirig&eacute; par Dr. Akim Demaille.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>1999 - 2004</b></td>
<td>Cycle pr&eacute;paratoire et cycle ing&eacute;nierie &agrave; l'<b>EPITA</b> (Ecole Pour l'Informatique et les Techniques Avanc&eacute;es). Mention tr&egrave;s bien.</td>
</tr>
</table><br>

<h3>E N S E I G N E M E N T</h3>
<hr>
<table cellspacing="10" width="100%">
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2007</b></td>
<td><b>ENSTA</b> : encadrement de projets en C (20h)</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2007</b></td>
<td><b>UPMC</b> (Universit&eacute; Pierre et Marie Curie) : TD et TP de <b>programmation par objets, design patterns</b> (50h)</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2006</b></td>
<td><b>UPMC</b> : TP de C et de shell Unix (75h).</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2005</b></td>
<td><b>ESILV</b> (&Eacute;cole Sup&eacute;rieure d'Ing&eacute;nieurs L&eacute;onard de Vinci) : TP et TD de <b>Java</b> (15h), TP de gestion de projet (10h).</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2005</b></td>
<td><b>CESI</b> (Centre d'Etudes Sup&eacute;rieures Industrielles) : mise en place d'un module d'introduction &agrave; Linux (20h).</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2005</b></td>
<td><b>Acadomia</b> : cours de soutien en maths et anglais.</td>
</tr>
</table><br>

<h3>E X P &Eacute; R I E N C E &nbsp; P R O F E S S I O N N E L L E</h3>
<hr>
<table cellspacing="10" width="100%">
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2009 - 6 mois</b></td>
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour l'estimation monoculaire de pose du corps humain.<br> Application &agrave; l'analyse visuelle du langage des signes.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2005 - 6 mois</b></td>
<td><b>ENSTA</b> : stage de Master.<br> D&eacute;tection statistique d'objets saillants sur r&eacute;tine artificielle programmable.<br> Application &agrave; la d&eacute;tection de segments significatifs.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2004 - 7 mois</b></td>
<td><b>Siemens Corporate Research</b> (Princeton, &Eacute;tats-Unis) : stage de recherche au sein du d&eacute;partement imagerie.<br> Recherche d'un mod&egrave;le d'application des filtres de particules au suivi de fibres c&eacute;r&eacute;brales.<br> Recherche et impl&eacute;mentation (<b>C++</b>, <b>Windows/Linux</b>) d'algorithmes de visualisation d'images de tenseur de diffusion (DT-MRI).<br>[[http://www.freshpatents.com/Systems-devices-and-methods-for-diffusion-tractography-dt20061012ptan20060229856.php | Brevet]]</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2002 - 2003</b></td>
<td><b>LRDE</b> : recherche et d&eacute;veloppement.<br> D&eacute;veloppement d'Evidenz, un moteur de raisonnement g&eacute;n&eacute;rique &agrave; base de th&eacute;orie de l'&eacute;vidence (Dempster-Shafer). Application au traitement d'image.<br> Participation &agrave; la conception et au d&eacute;veloppement d'Olena, une biblioth&egrave;que g&eacute;n&eacute;rique de traitement d'image en <b>C++ (metaprogramming)</b>.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2002 - 3 mois</b></td>
<td><b>Radio France</b> : stage au d&eacute;partement de recherche et d&eacute;veloppement. Conception et r&eacute;alisation d'applications de surveillance de flux audio en <b>C++</b> sous environnement <b>multiplate-forme</b> (Linux, `NetBSD, `OpenBSD, `FreeBSD).</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2001 - 2 mois</b></td>
<td>Snecma Services : g&eacute;n&eacute;ration de statistiques automatis&eacute;es (VBA, Access).</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>2000 - 2 mois</b></td>
<td>`WebValley : administration r&eacute;seau sous <b>Linux</b> (serveur web / ftp, monitoring, sauvegardes, ...)</td>
</tr>
</table><br>

<h3>P U B L I C A T I O N S</h3>
<hr>
<br>
[[Research.Main|Check the research page]]
<br>

<h3>D I V E R S</h3>
<hr>
<table cellspacing="10" width="100%">
<tr>
<td style="white-space: nowrap; vertical-align: top;"><b>Langues</b></td>
<td>Anglais: courant (7 mois de vie aux &Eacute;tats-Unis).<br>Espagnol: scolaire.</td>
</tr>
<tr>
<td style="white-space: nowrap; vertical-align: top;">
<b>Activit&eacute;s</b></td>
<td>Contributions &agrave; divers logiciels libres.<br>Sports : volley, natation, roller.<br>Voyages.</td>
</tr>
</table><br>

<hr size="2">
<blockquote>
    <em>This document was translated from L<sup>A</sup>T<sub>E</sub>X by</em> [[http://pauillac.inria.fr/~maranget/hevea/index.html | <em>H<font size="2"><sup>E</sup></font>V<font size="2"><sup>E</sup></font>A</em>]]<em>.</em>
</blockquote>
to:
'''Resume in english''' [[{$PubDirUrl}/files/aboutme/resume-burrus-march-2011.pdf | PDF]]

I am a PhD in computer vision and since 2009 I am a postdoctoral scholar at Carlos III University of Madrid. There, I am working on the perception aspects of the [[http://handle-project.eu|Handle]] FP7 european project on dextrous robotic manipulation. I am also involved in the Prosave project to study the suitability of Time-of-Flight cameras for airplane applications, in collaboration with Airbus Military. My main interests are computer vision for robotics, RGB-D cameras, object recognition and statistical approaches to visual event detection.
March 31, 2011, at 07:26 PM by 87.217.160.151 -
Deleted lines 4-5:
27 ans
August 12, 2009, at 02:34 PM by 82.123.33.143 -
Changed lines 7-8 from:
Email: mailto:burrus@montefiore.ulg.ac.be
to:
Email: mailto:nicolas@burrus.name
August 12, 2009, at 02:28 PM by 82.123.33.143 -
Changed line 72 from:
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour l'estimation monoculaire de pose d'humains.<br> Application &agrave; l'analyse visuelle du langage des signes.</td>
to:
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour l'estimation monoculaire de pose du corps humain.<br> Application &agrave; l'analyse visuelle du langage des signes.</td>
August 12, 2009, at 02:27 PM by 82.123.33.143 -
Changed line 72 from:
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour le suivi de pose d'humains.<br> Application &agrave; l'analyse du langage des signes.</td>
to:
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour l'estimation monoculaire de pose d'humains.<br> Application &agrave; l'analyse visuelle du langage des signes.</td>
August 12, 2009, at 02:17 PM by 82.123.33.143 -
Added lines 71-74:
<td style="white-space: nowrap; vertical-align: top;"><b>2009 - 6 mois</b></td>
<td><b>Université de Liège</b> : postdoc.<br> <b>Approches statistiques</b> pour le suivi de pose d'humains.<br> Application &agrave; l'analyse du langage des signes.</td>
</tr>
<tr>
Changed line 76 from:
<td><b>Université de Liège</b> : postdoc.<br> \textbf{Approches statistiques} pour le suivi de pose d'humains.<br> Application &agrave; l'analyse du langage des signes.</td>
to:
<td><b>ENSTA</b> : stage de Master.<br> D&eacute;tection statistique d'objets saillants sur r&eacute;tine artificielle programmable.<br> Application &agrave; la d&eacute;tection de segments significatifs.</td>
Deleted lines 78-81:
<td style="white-space: nowrap; vertical-align: top;"><b>2005 - 6 mois</b></td>
<td><b>ENSTA</b> : stage de Master.<br> D&eacute;tection statistique d'objets saillants sur r&eacute;tine artificielle programmable.<br> Application &agrave; la d&eacute;tection de segments significatifs.</td>
</tr>
<tr>
August 12, 2009, at 02:17 PM by 82.123.33.143 -
Changed line 72 from:
<td><b>ENSTA</b> : stage de Master.<br> D&eacute;tection statistique d'objets saillants sur r&eacute;tine artificielle programmable.<br> Application &agrave; la d&eacute;tection de segments significatifs.</td>
to:
<td><b>Université de Liège</b> : postdoc.<br> \textbf{Approches statistiques} pour le suivi de pose d'humains.<br> Application &agrave; l'analyse du langage des signes.</td>
Added lines 75-78:
<td style="white-space: nowrap; vertical-align: top;"><b>2005 - 6 mois</b></td>
<td><b>ENSTA</b> : stage de Master.<br> D&eacute;tection statistique d'objets saillants sur r&eacute;tine artificielle programmable.<br> Application &agrave; la d&eacute;tection de segments significatifs.</td>
</tr>
<tr>
August 12, 2009, at 02:15 PM by 82.123.33.143 -
Changed lines 9-10 from:
'''Resume in english''' [[{$PubDirUrl}/files/aboutme/resume-burrus-nov-2008.pdf | PDF]]
to:
'''Resume in english''' [[{$PubDirUrl}/files/aboutme/resume-burrus-jul-2009.pdf | PDF]]
August 12, 2009, at 02:14 PM by 82.123.33.143 -
Changed line 22 from:
<td>Doctorat &agrave; l'UPMC (Université Pierre et Marie Curie, Paris 6), laboratoire d'acceuil &agrave; l'<b>ENSTA</b> (&Eacute;cole Nationale Sup&eacute;rieure de Techniques Avanc&eacute;es) sous la direction de [[http://www.ensta.fr/~tbernard|Thierry Bernard]] (ENSTA) et de [[http://rfv.insa-lyon.fr/~jolion/|Jean-Michel Jolion]] (<b>LIRIS - INSA Lyon</b>). <i>Apprentissage a-contrario et architecture efficace pour la détection d'évènements visuels significatifs ([[Research.Main#PhD|plus de détails]]).</i></td>
to:
<td>Doctorat &agrave; l'<b>UPMC</b> (Université Pierre et Marie Curie, Paris 6), laboratoire d'acceuil &agrave; l'<b>ENSTA</b> (&Eacute;cole Nationale Sup&eacute;rieure de Techniques Avanc&eacute;es) sous la direction de [[http://www.ensta.fr/~tbernard|Thierry Bernard]] (ENSTA) et de [[http://rfv.insa-lyon.fr/~jolion/|Jean-Michel Jolion]] (<b>LIRIS - INSA Lyon</b>). <i>Apprentissage a-contrario et architecture efficace pour la détection d'évènements visuels significatifs ([[Research.Main#PhD|plus de détails]]).</i></td>