Entry only available in french…
Contact: bogdan.stanciulescu at mines-paristech.fr
Position and duration: Postdoctorate – 12 Months full time contract
Starting date: 1st of April 2016
Qualifications and skills: Applicants must have a PhD in the field of computer science, electrical engineering, physics, or any other related field. The candidates need to have a strong background in scene interpretation, particularly in the following fields: 3D environment reconstruction, SLAM, feature extraction, scene recognition, visual object detection. The applicants must have good communication skills, be able to work in a team environment and have fluent English skills. French language knowledge is a plus, but not compulsory.
The application must contain information of research background and work experience, including:
Applications must be submitted by e-mail to bogdan.stanciulescu at mines-paristech.fr with the subject: POSTDOCTORAL POSITION.
The Robotics Laboratory of Mines-ParisTech (CAOR) has developed extensive competences and tools in the field of computer vision and pattern recognition for real-time object detection and classification (people, vehicles, faces, etc). One of the CAOR’s algorithms has been internationally recognised as the 2nd best Pascal VOC challenge 2006.
For its results in real-time object recognition and classification, the CAOR’s has been rewarded the Best Student Paper Award at the International Conference on Control, Automation, Robotics and Vision 2011, and again rewarded the International Joint Conference on Neural Networks 2011 object recognition challenge.
The postdoctoral associate could use the CAOR’s experience in real-time video processing, robust signature extraction from multiple images and machine learning.
Least but no last, the Robotics Lab has acquired a good experience in sensor data fusion for performing indoor SLAM. The Laboratory’s prototype « Corebots » has won 2 times out of 3 the DGA-ANR Carotte competition for mobile robots, by a precise 3D environment mapping and localisation.
SLAM laser mapping by Corebots prototype
This entry is only available in french…
Following on from the two previous successes of the International Workshop on Movement and Computing (MOCO’14) at IRCAM (Paris, France) in 2014, as well as MOCO’15 at Simon Fraser University (Vancouver, Canada) in 2015, we are pleased to announce MOCO’16, which will be hosted in Thessaloniki, Greece. MOCO’16 will be organized by MINES ParisTech, (France) in co-operation with the Paris 8 University (France), the University of Macedonia, Thessaloniki (Greece) and Aristotle University of Thessaloniki (Greece).
The vision of MOCO’16 is to bring together academics, researchers, engineers, designers, technologists, technocrats, creative artists, anthropologists, museologists, ergonomists and other practitioners interested in the phenomenon of the symbiosis between the human and the creative process, e.g. dancer-digital medias, musician-instrument, craftsman-object etc. This symbiosis takes the form of an interactional and gravitational relationship, where the human element is both a trigger and a transmitter, connecting perception (mind/environment interaction and cognition), knowledge (theoretical understanding of a process) and gesture (semantic motor skills). MOCO’16 invites researchers that have experiences of capturing the combined key elements of perception, knowledge and gesture/movement. MOCO’16 will be of interest to artists who work on the elucidation of the intersection between art, meaning cognition and technology by unlocking the hidden components in human creativity. The workshop also provides a forum for industrial partners, for whom the movement and gestures of the workers/operators consist of key elements in terms of ergonomics and health, to see and present state-of-the-art technologies.
A key feature of the MOCO’16 Workshop will be to open some of its demonstrations and artistic activities to the public-at-large in order to provide this extended audience with the opportunity to be informed about current scientific issues and topics by experts in an informal setting.
Professor at Center for Robotics, Philippe Fuchs talked about 360° immersion at “What’s Next”, 360° Immersion, in october, the 13th which took place at l’Usine – Saint-Denis. With the increasingly offer of HMD accessible to the general public (due to their low cost), it seems legitimate that the use of this type of device, quite intrusive, asks at least questions of public health.
Alexis Paljic has been invited to perform the inaugural presentation at the INC Day 2015 Conference. This year, the INC Day is dedicated to the many applications of Virtual Reality in Neurocience, Psychology and Psychiatry.
Snapshot of Alexis Paljic (MINES ParisTech) keynote during INC Day 2015 Conference
MINES ParisTech researcher and specialist in Virtual Reality, Alexis Paljic offered some elements of answers to the question “How far can we trust VR for the Simulation of Real Human Activities ?”.
Videos of the presentations are available on the website INC Day 2015.
The paper “Autonomous driving at intersections: Combining theoretical analysis with practical considerations” won a Best Paper Award at the ITS World Congress 2015 in Bordeaux.
The paper was written by Arnaud de La Fortelle (Director of the Robotics Centre and holder of the Chair Drive for You) and Xiangjun Qian (PhD student at the Robotics Centre) as part of the work on the European project Autonet 2030 and within the Chaire Drive for You.
The move towards automated driving is gaining impetus recently. This paper follows the approach of combining theoretical analysis with practical issues. It gives an insight of some practical problems that are encountered when running automated vehicles in real environments, using intersection crossing as a major example. The aim is not to try to be exhaustive but to show some criteria (safety, efficiency, reactivity, resilience, scalability…) for decision making in automated driving that have to be balanced before any mass deployment. In a second part we introduce mathematical tools that can help define algorithms and systems that improve current state of the art. We will also show some perspective for accommodating the hypotheses of these mathematical tools with real life constraints.
The mobile robot developed in the Center For Robotics at MINES Centre ParisTech was illustrated during the decade of the Carnot Institutes. The Carnot Institutes are composed of laboratory of excellence with lots of R&D activities.
The mobile robot his equipped by the CoreSLAM algorithm (developed in the Center for Robotics) to locate and mapping a previously unknown space, the robot can move autonomously in a room to perform various missions.
Tony Noël in charge of this mobile robot gave more information about the robot and its applications in a video made within the ten years of Carnot Institutes.
[Youtube id = “4uyef0byAQ0” width = “600” height = “350”]
The mobile robot also seen at “Usine Nouvelle“
The session SIS25 on “theoretical and technical challenges for automated” won the Best rated session award at ITS World Congress 2015 in Bordeaux.
This session has been the opportunity to present the theoretical and technical challenges for automated driving from different perspectives. Indeed the speakers were coming from all around the world, and with an academic or industrial focus.
The session has been organised and moderated by Arnaud de La Fortelle (MINES ParisTech, Center for robotics), in collaboration the following speakers:
The Chair Drive for You was well represented during this session (PSA Peugeot Citroën, Valeo, Shanghai Jiao Tong University, University of California at Berkeley and MINES ParisTech). This award is a strong encouragement for the research carried out in the Chair Drive for You.
Summary of the session:
We present a new dataset of annotations for aligning a Point Distribution Model onto the human body. The corresponding research is described in the paper O. Huynh and B.Stanciulescu, “Clustering and Classifying Deformations for Shape Regression applied to the Human Body”, in ISPA 2015. The InriaLHB uses the pedestrian images from the Inria Person Dataset. Annotated shapes correspond to a standing person in front or back view.