Vous êtes ici : Kiosque » Annonce

Identification

Identifiant: 
Mot de passe : 

Mot de passe oublié ?
Détails d'identification oubliés ?

Annonce

16 avril 2024

PhD position - Object Detection from Few Multispectral Examples


Catégorie : Doctorant


  • Academic lab: IRISA
  • Compagny: ATERMES
  • CIFRE PhD, European nationality required
  • Expected starting: September/October 2024
  • Application deadline: 15/05/2024

Context

ATERMES is an international mid-sized company, based in Montigny-le-Bretonneux with a strong expertise in high technology and system integration from the upstream design to the long-life maintenance cycle. It specializes in offering system solution for border surveillance. Its flagship product BARIER™ (“Beacon Autonomous Reconnaissance Identification and Evaluation Response”) provides ready application for temporary strategic site protection or ill-defined border regions in mountainous or remote terrain where fixed surveillance modes are impracticable or overly expensive to deploy. As another example, SURICATE is the first of its class optronic ground “RADAR” that covers very efficiently wide field with automatic classification of intruders thanks to multi-spectral deep learning detection.

The collaboration between ATERMES and IRISA was initiated through a first PhD thesis (Heng Zhang, defended December 2021, https://www.theses.fr/2021REN1S099/document). This successful collaboration led to multiple contributions on object detection in both mono-modal (RGB) and multi-modal (RGB+THERMAL) scenarios. Besides, this study allowed to identify remaining challenges that need to be solved to ensure multispectral object detection in the wild.

 

Objectives

The project aims at providing deep learning-based methods to detect objects in outdoor environments using multispectral data in a low supervision context, e.g., learning from few examples to detect scarcely-observed objects. The data consist of RGB and IR (Infra-red) images which are frames from calibrated and aligned multispectral videos.

Few-shot learning [1][2], semi-supervised learning [3][4] and continual learning [5][6] are among the most widely-used frameworks to tackle this task. For the first approach based on few-shot object detection (FSOD), the recent trend has relied on using meta learning or transfer learning approaches [1:1]. Yet, realistic settings including scarce objects may exist a domain shift that makes the task more challenging. The second approach based on semi-supervised learning considers a large amount of unlabeled data in the training process to foster the representation capacity of deep models, improving the peformance of object detection from a small amount of labeled samples. As the third approach, continual learning [5:1] aims to maintain the performance of the deep models on old categories and avoid the “catastrophic forgetting” phenomenon when learning new object categories. It has been also integrated into a FSOD task [7] to ensure that few-shot object detectors could learn new object concepts without forgetting previous object categories that still exist in prediction phase. Last but not least, with the dramastically rapid evolution of research in AI, another challenge to tackle is the investigation of modern AI models, and more specifically foundation models which involves multimodal transformers [8][9]. Indeed, these large machine learning models trained on a vast quantity of data at scale have been designed to be adapted to a wide range of downstream tasks (including object detection, see for instance UniDetector [10]) or CLIP2 [11]. These models leading to zero-shot object detection could very well be the ultimate answer for the task of having a true scene understanding.

 

Required background and skills

  • MSc or Engineering degree with excellent academic track and proven research experience in the following fields: computer science, applied maths, signal processing and computer vision;
  • Experience with machine learning, in particular deep learning;
  • Skills and proved experience in programming (Python is mandatory and knowledge about frameworks such as Pytorch is a real plus);
  • Excellent communication skills (spoken/written English) is required ;
  • Ambition to publish at the best level in the computer vision community (CVPR, ICCV, TPAMI, …) during the thesis.

 

Supervision team

The PhD will be co-supervised by Prof. Elisa Fromont (LACODAM team, IRISA/INRIA Rennes) and Prof. Sébastien Lefèvre (OBELIX team, IRISA Vannes). The supervision team will be completed by Dr. Minh-Tan Pham (Ass. Prof., OBELIX team) and Bruno Avignon (CSO, ATERMES).

Application Procedure

Your application (CV+cover letter+academic transcripts) should be sent before 15/05/2024 (but the sooner the better) to the 4 email addresses:
elisa.fromont@irisa.fr; sebastien.lefevre@irisa.fr; minh-tan.pham@irisa.fr; bavignon@atermes.fr

Applications will be treated and interviews will be conducted along the way.
The candidate (European nationality required) will be hired with a CIFRE contract by ATERMES.
The expected gross salary is around 3000€ per month for 3 years. Moreover, a financial package covering various scientific activities including participation in top national/international conferences, training at summer schools, etc. will be proposed for the recruited candidate.
The contract will start in September or October 2024. Atermes will hire the candidate (as an engineer, CDI) before the beginning of the CIFRE contract if necessary.

 

The full PhD topic can be found here: http://www-obelix.irisa.fr/files/2024/04/PhD_Cifre2024_IRISA_ATERMES.pdf

 

Dans cette rubrique

(c) GdR IASIS - CNRS - 2024.