APWeb-WAIM 2022 : The 6th APWeb-WAIM International Joint Conference on Web and Big Data

Date : 2022-08-11 => 2022-08-13
Lieu : Nanjing, China

The Asia Pacific Web (APWeb) and Web-Age Information Management (WAIM) Joint International Conference on Web and Big Data (APWeb-WAIM) is aiming at attracting professionals of different communities related to Web and Big Data, including Web technologies, database systems, information management, software engineering and big data.The year 2022 marks the 6th anniversary of APWeb-WAIM 2022, which will be held in Nanjing, China, 11-13 August, 2022. It is our great pleasure to invite you to contribute papers and participate in this premier annual event.With the increased focus on Big Data, the new joint conference is expected to attract more professionals from different industrial and academic communities, not only from the Asia Pacific countries but also from other continents.

Important Dates

• Abstract submission: March 15, 2022
• Full paper submission: March 22, 2022
• Acceptance Notification: May 23, 2022
• Camera Ready: June 6, 2022
• Conference Date: August 11-13, 2022

Topics of Interest, but not limited to

• Advanced database and Web applications
• Big data analytics
• Big data management
• Big data management and analytics
• Blockchain data management and applications
• Cloud computing
• Cloud computing
• Crowdsourcing
• Data and information quality
• Data management in edge computing
• Data management on new hardware
• Data mining
• Data provenance and workflow
• Data warehousing and OLAP
• Graph data management, Metadata, RDF, social networks
• Information extraction
• Information integration and heterogeneous systems
• Information management in Meta-verse
• Information retrieval
• Knowledge graph
• Machine Learning
• Multimedia information systems
• Parallel and distributed data management
• Query processing and optimization
• Recommender systems
• Security, privacy, and trust
• Semantic Web and ontology
• Service computing
• Spatial and multi-media data
• storage and indexing
• Streams, complex event processing
• Text database
• Uncertain data
• Web advertising and community analysis
• Web information quality and fusion
• Web search and meta-search
• Web service management

Authors should submit papers reporting original work that are currently not under review or published elsewhere. Accepted papers will be published in the conference proceedings, which will be published as a volume of Springer’s Lecture Notes in Computer Science (LNCS)series.

Paper Submission

All papers should be submitted through the Conference Management Tool at: https://cmt3.research.microsoft.com/APWEBWAIM2022
Submissions must be written in English and not exceed 15 pages in LNCS format, including references. All submissions must be in PDF format. Authors should avoid the use of non-English fonts to avoid problems with printing and viewing the submissions.

Submitted papers will undergo a double-blind reviewing process. The PC members and referees who review the paper will not know the identity of the authors. To ensure anonymity of authorship, authors must prepare their manuscript as follows:
Authors’ names and affiliations must not appear on the title page or elsewhere in the paper.
Funding sources must not be acknowledged on the title page or elsewhere in the paper.
Research group members, or other colleagues or collaborators, must not be acknowledged anywhere in the paper.
The paper’s file name must not identify the authors of the paper. It is strongly suggested that the submitted file be named with the assigned submission number. For example, if your assigned paper number is 386, then name your submission file 386.pdf.
Source file naming must also be done with care, to avoid identifying the authors’ name in the paper’s associated metadata. For example, if your name is Jane Smith and you submit a PDF file generated from a .dvi file called Jane-Smith.dvi, your authorship could be inferred by looking into the PDF file.
It is the responsibility of authors to do their very best to preserve anonymity. Papers that do not follow the above GUILDLINES, or otherwise potentially reveal the identity of the authors, are subject to desk rejection.

Recommendation to Journal

A number of best papers accepted at APWeb-WAIM 2022 will be recommended to a set of SCI indexed journals, including World Wide Web Journal (IF 2.716, JCR Q2), Knowledge-based Systems (IF 8.038, JCR Q1), Big Data Research (IF3.578, JCR Q1), and Journal Data Science Engineering.

Lien direct


Notre site web : www.madics.fr
Suivez-nous sur Tweeter : @GDR_MADICS
Pour vous désabonner de la liste, suivre ce lien.

INGÉNIEUR(E) R&D Réalité augmentée pour la téléopération sécurisée de robots

Offre en lien avec l’Action/le Réseau : – — –/– — –

Laboratoire/Entreprise : Awabot
Durée : 6 mois
Contact : ana-maria.roxin@ubfc.fr
Date limite de publication : 2022-04-10

Contexte :
Awabot Intelligence est spécialisée dans la prestation de conseils et de développement informatique pour des projets de conception et d’intégration robotique.

Awabot Intelligence développe et intègre des solutions de navigation autonome, d’intelligence artificielle et de téléprésence dans le domaine de la robotique de service.
Les robots sur lesquels ces solutions sont implémentées sont parfois amenés à être téléopérés par des utilisateurs au sein d’environnements complexes, conçus pour les êtres humains.

Cependant, en téléopération, les éléments perçus par l’utilisateur ne sont que ceux qui peuvent être captés par l’œil humain au travers d’un écran retranscrivant les données des capteurs embarqués sur le robot.

Sujet :
Afin d’améliorer l’expérience utilisateur en amplifiant ses capacités de perception du monde, Awabot Intelligence souhaite étudier l’utilisation de la réalité augmentée lors de la téléopération.

En particulier, l’équipe R&D souhaite pouvoir exploiter les capteurs visuels d’un robot pour détecter les éléments nuisibles à son fonctionnement (e.g., escaliers, câbles électriques, zones sécurisées, etc.) dans le but d’afficher des informations contextuelles à l’utilisateur afin de le guider dans ses manœuvres.

Le candidat retenu sera ainsi amené à intervenir sur le développement de nouveaux outils logiciels, que ce soit en se basant sur des technologies existantes ou en implémentant de nouveaux algorithmes de vision par ordinateur et de deep learning pour la réalité augmentée.

Le projet se déroulera idéalement sur 6 mois :
Etat de l’art sur la détection d’objets appliquée à la réalité augmentée. Analyse de l’existant.
Identification de solutions adaptées à la problématique. Définition de métriques pour l’analyse de performance.
Implémentation des solutions retenues pour différents cas d’usages.
Évaluation des solutions implémentées.
Documentation des avancées et des résultats.

Ce stage peut déboucher sur une proposition de thèse de doctorat.

Profil du candidat :
De formation Ingénieur ou Master 2 (Bac +5) en informatique/robotique/vision, vous recherchez un stage de fin d’études de six mois.

Formation et compétences requises :
Autonome, rigoureux et force de proposition, vous disposez des compétences suivantes :

bon niveau de compréhension des modèles mathématiques et de l’algorithmie,
aisance avec les langages de programmation (C++, Python, C#),
bonne capacité rédactionnelle.
Une première expérience avec le framework ROS/ROS2 et/ou un SDK de réalité augmentée est un plus.

Adresse d’emploi :
Awabot – 16 bis avenue de la République, 69200 Vénissieux

Assimilation of geodetic data for natural hazards forecasting

Offre en lien avec l’Action/le Réseau : – — –/– — –

Laboratoire/Entreprise : LISTIC
Durée : 3 ans
Contact : yajing.yan@univ-smb.fr
Date limite de publication : 2022-07-10

Contexte :
This Ph.D thesis is proposed along with the increasing and regular availability of the amount of remote sensing data and the response to the requirement of operational prediction of natural hazards. The main objective is to improve the near-real-time integration of remote sensing data and dynamical geophysical models for the mitigation of natural hazards. This thesis is partly (50%) funded by the national action plan in Artificial Intelligence. The relevance of the methodology developed in this thesis 1) compared to the actually emmerging data-driven methods, lies in the incorporation of geophysical knowledge (which helps increase the interpretability and the accountability of the results for operational purpose) and its near-real-time implementation ; 2) compared to previous attemps to improve the near-real-time integration of InSAR data based on
the Kalman Filter, lies in the capability in taking non-Gaussian error statistics (which can represent better the reality) into account. First application will be in volcanology, using InSAR & GNSS data, but the methodology can be easily utilized for other natural hazards (e.g. landslides, slow slip, etc.), as well as for anthropogenic hazards like forest fire.

Sujet :
In a perspective of volcanic hazard assessment, it is fundamental to be able to know, in advance, if magma that has started to propagate from a reservoir will reach the surface, where and when. The propagation phase is generally rapid, lasting a few hours to a few months but it induces
seismicity and deformation signals. These signals are recorded by continuous sensors (GNSS) and InSAR data whose temporal sampling frequency as well as latency have been greatly improved in recent years. In addition, we have dynamic magmatic intrusion propagation models that can be used to calculate, depending on the physical parameters of the Earth’s crust, the properties of the magma and the state of local stress, the trajectory followed by the magma and its propagation velocity (Pinel et al. 2017). Data assimilation, a method that combines a dynamic model with observations at present and in the past based on error statistics and predicts the future state of the observed system, is therefore an appropriate tool to respond to the need to be able to predict
the position and timing of an eruption in volcanology.

Among numerous data assimilation methods, the particle filter is distinguished from others by its great ability to deal with non-linear models and non-Gaussian error statistics (van Leeuwen P.J 2009, van Leeuwen P.J, 2010). The particle filter is based on a representation of the probability density of the dynamic model by a discrete set of states of the model (namely particles) and relies on the Bayes theorem so without worrying about the distribution of model errors (different from most other data assimilation methods which assume a Gaussian distribution of the errors). The evolution of the probability density of the model is realized through the propagation of all the particles (states of the model) over time following the model equation (the physics). When
observations of the system are available, the relative contributions of the particles are modified so that the information available in the observations is integrated into the particles swarm. The particle filter does not directly correct the values of particles, but their relative contributions, this is very important for estimating magmatic intrusion propagation trajectories. The particle filter is therefore the appropriate tool in the present specific context of estimating the position of a volcanic eruption.

In this thesis, we will develop an efficient data assimilation strategy using the particle filter allowing to use timely available geodedic data to predict the location and timing of eruptive vents induced by magmatic intrusion propagation. This work will be based on the Ph.D thesis of Mary Grace Bato who, under our supervision and for the first time, successfully applied sequential data assimilation techniques (i.e. Ensemble Kalman Filter) to volcanological problems focussing on the pressurization and rupture of magmatic reservoirs (Bato et al, 2017, Bato et al., 2018). Moreover, this thesis will benefit from the results obtained in the TOSCA AssimSAR project (2018-2019). It will be part of the Franco-German ANR MagmaPropagator (ANR-18-CE92-0037, 2019-2022) with an application to Piton de la Fournaise volcano. It will also be the oppotunity to collaborate with Andy Hooper from the University of Leeds, who is developping new methods to automatically extract a physical signal from InSAR time series (Gaddes et al, 2019).

Selected references :
1) Bato M.-G., Pinel V., Yan Y., Jouanne F., Vandemeulebrouck J., “Possible deep connection between volcanic systems evidenced by sequential assimilation of geodetic d”, Scientific Reports, Nature Publishing Group, 2018, https://doi.org/10.1038/s41598-018-29811-x
2) Bato M.-G., Pinel V., Yan Y., “Assimilation of Deformation Data for Eruption Forecasting: Potentiality Assessment Based on Syntheti”, Frontiers in Earth Science, Frontiers Media, 2017, pp.doi: 10.3389/feart.2017.00048
3) Dalaison, M., Jolivet, R., A Kalman filter time series analysis method for InSAR, Journal of Geophysical Research : Solid Earth , 2020, e2019JB019150. e2019JB019150 2019JB019150.
4) Gaddes, M. E., Hooper, A., Bagnardi, M. (2019), Using machine learning to automatically detect volcanic unrest in a time series of interferograms, Journal of Geophysical Research : Solid Earth , 124(11), 12304–12322.
5) Pinel V., Carrara A., Maccaferri F., Rivalta E., Corbi F., A two-step model for dynamical dike propagation in two dimensions: Application to the July 2001 Etna eruption, 2017, Journal of Geophysical Research, vol. 122, doi:10.1002/2016JB013630.
6) van Leeuwen P.J., Review Particle Filtering in Geophysical System, Mathematical Advances in Data Assimilation, 2009, pp. 4089-4114.
7) van Leeuwen P.J. Nonlinear data assimilation in geosciences : an extremely efficient particle filter, 2010, Quarterly Journal of the Royal Meteorological Society, vol. 136, pp. 1991-1999.

Profil du candidat :
The Ph.D candidate should have good skills in signal/image processing,
mathematics/statistics or geophysics.

Formation et compétences requises :
mathematics/statistics, image processing, remote sensing

Adresse d’emploi :
Laboratoire d’Informatique, Systèmes, Traitement de l’Information et de la Connaissance, Université Savoie Mont-Blanc, Annecy, France

Document attaché : 202203081350_sujet_these_DA_volcan_en2022.pdf

DOING@ADBIS’2020

Date : 2022-09-05 => 2022-09-08
Lieu : Torino, Italy

*******************
IMPORTANT DATES
******************

Paper submission: May 3, 2022 at 5 a.m. CET
Notification of acceptance: May 23, 2022
Camera-ready due: June 7, 2022
Workshop day: September 5, 2022

******************
SUBMISSIONS
******************

DOING workshop accepts short (limited to 6-8 pages) and long (limited to 12 pages) papers. DOING reserves the right to accept as short papers those submitted as long, describing interesting and innovative ideas but still requiring further technical development. Papers should be written in English, formatted in Latex and present substantially original results. We adopt a double blind review policy: the papers submitted for review MUST NOT contain the authors’ names, affiliations, or any information that may disclose the authors’ identity. Authors should consult Springer’s authors’ guidelines and use their proceedings templates (you can download the templates available on the bottom of that page).

ADBIS 2022 follows a Diversity and Inclusion policy that invites authors to adopt inclusive language in their papers and presentations (https://dbdni.github.io/pages/inclusivewriting.html and https://dbdni.github.io/pages/inclusivetalks.html). We also kindly ask all participants to adopt a proper code on conduct (https://dbdni.github.io/pages/codeofconduct.html).

Accepted papers will be published in the Springer CCIS series and the best papers will be invited to a special issue of the journal Computer Science and Information Systems.

Papers should be submitted in PDF format using the EasyChair online submission system:
https://easychair.org/conferences/?conf=adbis2022
Be careful to select the track of the WORKSHOP DOING: Intelligent Data – From Data to Knowledge.

*******************
AIMS AND SCOPE
******************

The workshop focuses on transforming data into information and then into knowledge. The idea is to gather researchers to discuss two main problems :

+ how to extract information from textual data and represent it in knowledge bases;
+ how to propose intelligent methods for handling and maintaining these databases with new forms of requests, including efficient, flexible, and secure analysis mechanisms, adapted to the user, and with quality and privacy preservation guarantees.

******************
TOPICS OF INTEREST
******************

We invite the submission of work-in-progress that address various aspects of information extraction from textual data, intelligent and efficient interrogation, and maintenance of (large) knowledge bases.
The workshop welcomes submissions of theoretical, technical, experimental, methodological papers, application papers, position papers and papers on experience reports addressing – though not limited to – the following topics:

Artificial intelligence in databases and information systems
Data curation, annotation, and provenance
Data management and analytics
Data mining and knowledge discovery
Data models and query languages
Data quality and data cleansing
Data science (theory and techniques)
Context-aware and adaptive information systems
Constraints extraction from text
Natural language processing
Indexing, query processing and optimization
Information and knowledge extraction
Information integration
Information quality
Graph databases
Knowledge bases (querying, management, evolution and dynamics)
Machine learning for knowledge graph construction, completion, refinement
Machine learning for knowledge and information extraction, for instance, named entity disambiguation, sentiment analysis, relation extraction, or the detection of claims, facts and stances from unstructured documents
Machine Learning in NLP
Management of large volumes of data
Methodologies, models, algorithms, and architectures for applied data science
NLP for Digital Humanities
NLP & Knowledge Graphs
Privacy, trust and security in databases
Query processing and optimization
Question answering over knowledge graphs
Text databases

Preferred Application Domains (but not limited to).

Bio-sciences and healthcare
Environmental issues

Lien direct


Notre site web : www.madics.fr
Suivez-nous sur Tweeter : @GDR_MADICS
Pour vous désabonner de la liste, suivre ce lien.

Applying Hybrid Evolutionary Machine Learning techniques on Environmental DNA data to predict biodiversity

Offre en lien avec l’Action/le Réseau : – — –/– — –

Laboratoire/Entreprise : LAMSADE
Durée : 3 years
Contact : sana.mrabet@dauphine.psl.eu
Date limite de publication : 2022-04-15

Contexte :
Evaluating the level of biodiversity in marine and terrestrial locations is currently a universal practice for estimating the impact of human activities and climate change on the ecosystems of these environments. An innovative and modern approach to perform this observation is the high-throughput sequencing of DNA taken from the environment (“environmental DNA)
Environmental DNA (eDNA) analysis consists mainly of the identification of species from the DNA they leave in their environment. Many studies show that the use of eDNA allows a good estimation of the different species (or taxa) present in different types of environments. eDNA analysis is based on classical molecular biology techniques (PCR, sequencing…). The identified taxa are assigned to ecological weights which are used to calculate the biotic indices (BI). These indices are used to determine the ecological quality of the site in question (generally classified in five categories from “very poor” to “very good”). However, these studies relied on reference sequence databases (Silva Greengenes LTP) for taxonomic assignment, in order to retrieve taxon-specific ecological weights and to calculate BI values. This phase may fail if some sequences in the environment to be analysed are incomplete or do not appear in the reference databases.
Recent works [1,2,5] has demonstrated that machine learning can be used to predict accurate values of biotic indices from eDNA metabarcoding, regardless of sequence affiliation. The idea is to generate a clustering model to group closely related sequences (high percentage of identity) belonging to a well-defined taxon. The approach then consists in using a clustering technique on the analysis data of the marine samples applicable even if the determined sequences are incomplete. These sequences are then affiliated automatically to the taxon of the central sequence of the cluster identified in a reference database. The validity of the model can be checked by searching for the exact affiliation of the cluster centres in reference databases.

Sujet :
The subject of this thesis has a double objective. The first objective is the implementation of conventional techniques for clustering eDNA data of a marine environment allowing to evaluate the BI of this environment. The generated models in this phase may be non-generalizable solutions and very dependent on the sequencing methods used in the analysed samples. Hence the second objective of the thesis which aims at the generalization of prediction models to data from other samples of the same marine environment using evolutionary methods. The idea is to develop a hybrid evolutionary algorithm that evolves clustering models as a function of sample parameters (i.e. sequencing markers) and clustering parameters (i.e. distance measurement indicators). The model scheme can be based on the “Genetic K-means algorithm” (GKA) of Krishna and Murty.
The first models are to be tested on marine samples studied by Cordier [1,2] and available online. Other samples will be studied after validation of the algorithm. The validation of the method and its application on other marine data may lead to a possible exchange with OFB (French Office of Biodiversity).

Profil du candidat :
Master’s degree in computer science or equivalent degree giving access to PhD studies.

Formation et compétences requises :
– Good knowledge in Machine Learning
– Good knowledge in Computer Science and Mathematics.
– Good programming skills, especially in Python programming.
– Some knowledge in Bio-informatic would be good but it is not mandatory.
– Good command of written and spoken English.

Adresse d’emploi :
LAMSADE CNRS UMR 7243, University of Paris Dauphine, France

Document attaché : 202203072123_PhD Application – Lamsade 2022.pdf

Poste MCF

Offre en lien avec l’Action/le Réseau : – — –/– — –

Laboratoire/Entreprise : Université de Strasbourg
Durée : 42ans
Contact : gancarski@unistra.fr
Date limite de publication : 2022-04-28

Contexte :
L’Université de Strasbourg (UFR Mathématique et Informatique) propose un poste de Maitre de conférences en Informatique avec un

Sujet :
Profil de recherche “Science des données et/ou des réseaux de communication” (voir pièce jointe)

Profil du candidat :
Doctorat en Informatique – Qualifié 27ieme section

Formation et compétences requises :
Compétences en SDIA et Réseaux

Adresse d’emploi :
Université de Strasbourg

Document attaché : 202203071131_MCF_Université_Strasbourg_0277.pdf

Statistical learning for satellite SAR image based Earth deformation observation

Offre en lien avec l’Action/le Réseau : MACLEAN/– — –

Laboratoire/Entreprise : LISTIC (Université Savoie Mont-Blanc)
Durée : 3 years
Contact : guillaume.ginolhac@univ-smb.fr
Date limite de publication : 2022-07-03

Contexte :
The thesis will be funded by the ANR REPED-SARIX project and will be start from the 1st September 2022 (or possibly october 1st).

Keywords : SAR interferometry, robust statistics, recursive estimation, missing data imputation, time series.

Sujet :
The systematic acquisition of and free access to Sentinel-1 A/B Synthetic Aperture Radar (SAR) images covering Europe every 6 days (every 12 days elsewhere) provide scientists with both opportunities and challenges for operational monitoring of Earth deformation by SAR image time series. For displacement estimation from SAR image time series, numerous multi-temporal Interferometry SAR (InSAR) methods, such as Small BAseline Subset, Permanent Scatterer Interferometry, SqueeSAR, Phase Linking methods, Multi-link InSAR, CAESAR, Least-Square estimator and EMI, have been extensively developed and implemented. Thanks to these methods, the accuracy of the displacement velocity estimation has been revolutionized to millimeters per year. However, these methods are mainly retrospective analysis tools and do not allow efficient gradual integration of new SAR images that arrive over time and it is necessary to restart part of or
the whole displacement estimation processing chain, which would be prohibitively expensive in practice and does not answer the need for operational monitoring. Therefore, it is still necessary to develop more elaborated recursive multi-temporal InSAR methods allowing for efficient gradual integration of new arriving SAR images and considering non Gaussianity of data statistics.

In this Ph.D thesis, we aim to develop a novel robust and recursive multi-temporal InSAR approach for operational displacement estimation from SAR image time series. We consider the state-of-theart Phase Linking approach as the baseline approach in which the sample covariance matrix of SAR image time series is fully exploited. First, we propose a sequential or recursive estimation of the covariance matrix of SAR images, taking into account the structure of the covariance matrix that is
directly related to the decorrelation properties of the targets under observation. Second, we integrate temporal decorrelation models (with possible unknown parameters) providing prior information on the structure of the covariance matrix in the sequential or recursive estimation process in order to improve the efficiency. We then deploy the Expectation – Maximization (EM) algorithm to estimate jointly the unknown model parameters and the covariance matrix in an
iterative way. The displacement time series can be later obtained from the properly estimated covariance matrix. This displacement times series is finally used to estimate physical parameters of the deformation source in depth. However, missing data can exist in the displacement time series, mainly due to the coherence loss that results in unreliable displacement estimations. Data gaps can hinder the full understanding of the phenomenon under observation. Therefore, the third objective
of this Ph.D consists of imputing missing data in displacement time series, with the missing data mechanism taken into account by assuming statistical laws and estimating the parameters that describe these statistical laws.

We consider the ‘’Piton de la Fournaise’’ and Merapi volcano test sites as proving ground for the developed approach in this Ph.D thesis. Both descending and ascending Sentinel-1 A/B acquisitions are available. GPS measurements from permanent GNSS stations are also available for results comparison and validation.

Profil du candidat :
The Ph.D candidate should have good skills in mathematics/statistics and/or signal/image processing. Knowledge in Interferometry SAR is appreciated.

Formation et compétences requises :
Statistics, Optimisation, Python, Remote Sensing.

Adresse d’emploi :
LISTIC, Annecy, FRANCE.

Document attaché : 202203071519_sujet_these_InSAR_en2022.pdf

Sujet de thèse en IA appliquée à l’hydroacoustique

Offre en lien avec l’Action/le Réseau : – — –/– — –

Laboratoire/Entreprise : Université de Brest
Durée : 3 ans
Contact : dorian.cazau@ensta-bretagne.fr
Date limite de publication : 2022-07-31

Contexte :
cf Présentation du projet du pdf joint

Sujet :
Application de l’intelligence artificielle à la détection automatique d’évènements pour les observat

Profil du candidat :
Le candidat doit avoir une solide expérience en programmation et en AI, et en éventuellement mathématiques et physique (traitement du signal et propagation des ondes).

Formation et compétences requises :
Formation universitaire ou école d’ingénieurs niveau master 2 en informatique, mathématiques ET/OU en océanographie, physique, sismologie

Adresse d’emploi :
Laboratoire Geo-Ocean (https://www.geo-ocean.fr/), Technopôle Brest-Iroise, Brest 29200

Document attaché : 202203041050_Bazin_LGO_IACOUSA_fr.pdf

Vers des données ouvertes de microscopie fonctionnelle en neurosciences

Offre en lien avec l’Action/le Réseau : – — –/– — –

Laboratoire/Entreprise : Institut de Neurosciences de la Timone, INT, Marse
Durée : 4-6 mois
Contact : sylvain.takerkart@univ-amu.fr
Date limite de publication : 2022-07-31

Contexte :
L’Institut de Neurosciences de la Timone (INT, http://www.int.univ-amu.fr ) est une unité mixte de recherche qui a pour objectif de développer des recherches interdisciplinaires en neuroscience. Situé sur le Campus de la Faculté de Médecine d’Aix Marseille Université, il est doté de plateformes technologiques de haut niveau au service d’équipes de recherche en neurosciences théoriques et expérimentales.

Sujet :
Dans le cadre du plan national pour la science ouverte (https://www.ouvrirlascience.fr/plan-national-pour-la-science-ouverte/), la mise en place de procédures d’ouverture des données scientifiques récoltées en neurosciences reste un challenge. En effet, ces données sont complexes et la mise en place de standards basés sur des formats de données ouverts sont des initiatives récentes. En particulier, les microscopes de dernières génération qui permettent d’enregistrer l’activité cérébrale en temps réel fournissent des gros volumes de données qu’il est important de gérer de manière efficace afin d’obtenir des données FAIR (Faciles à trouver, Accessibles, Intéropérables, Réutilisables : https://www.go-fair.org/fair-principles/). L’objectif de ce stage est de développer des composants logiciels open source qui permettront la production de données FAIR-by-design en partant des données brutes acquises sur les microscopes récemment acquis dans le laboratoire.

Profil du candidat :
Nous recherchons un.e candidat.e qui soit:
– curieux.se pour les applications en imagerie biomédicale;
– volontaire et sachant avancer de manière autonome;
– bon.ne communiquant.e et sachant partager ses progrès et les obstacles rencontrés;
– motivé.e pour coder dans un environnement “open source”.

Le stage peut se dérouler sur toute l’année 2022, suivant le calendrier des stages du cursus suivi par le.la candidat.e.

Formation et compétences requises :
Formation: bac + 4 ou bac + 5, cursus « Sciences des données » ou « Développement logiciel »

Compétences requises:
– bonne connaissance de l’écosystème des sciences des données
– maitrise des concepts avancés en développement logiciel (test unitaires, gestion de version avec des outils de type GIT, intégration continue, etc.)
– maitrise du langage python
– connaissances en mathématiques appliquées et en algorithmie
– intérêt pour la biologie et/ou les neurosciences et/ou l’imagerie médicale

Adresse d’emploi :
Institut de Neurosciences de la Timone, INT
27 boulevard Jean Moulin
13005 Marseille

Ingénieur•e de recherche en informatique pour le traitement du signal et le machine learning / Contractuel de recherche

Offre en lien avec l’Action/le Réseau : – — –/– — –

Laboratoire/Entreprise : CRIStAL UMR 9189 – Centrale Lille Institut
Durée : 12 mois renouvelable
Contact : pierre.chainais@centralelille.fr
Date limite de publication : 2022-07-31

Contexte :
Ce poste en CDD de 12 mois renouvelables au sein du laboratoire CRIStAL, Centre de Recherche en Informatique, Automatique et Signal de Lille (UMR 9189), est financé par l’ANR qui soutient la chaire IA Sherlock portée par Pierre Chainais (2021-2025), professeur à Centrale Lille et membre de l’équipe SigMA, Signaux, Modèles et Applications. L’équipe SigMA est composée de 16 membres permanents (enseignants-chercheurs et chercheurs CNRS) et de 18 doctorants et post-doctorants et offre un environnement scientifique stimulant et de haut niveau.

Sujet :
Missions du projet de recherche
Le projet de recherche de la chaire IA Sherlock porte sur « Fast inference with controlled uncertainty : application to astrophysical observations ». Il inclut le développement et l’implémentation de différents méthodes d’inférence en machine learning ou pour la résolution de problèmes inverses en traitement du signal et des images. La complexité et le coût de calcul de ces méthodes augmente très vite lorsque l’on travaille avec de grandes masses de données ou sur des problèmes en grande dimension.

Nous travaillons d’une part sur des méthodes d’optimisation, mais aussi et surtout sur des méthodes bayésiennes dans une logique de parallélisation et de distribution des calculs sur plusieurs nœuds de calculs. Au-delà de l’estimation ponctuelle de paramètres, nous nous intéressons à l’estimation de leur distribution en vue de la quantification des incertitudes. Il s’agit d’un enjeu crucial pour garantir la qualité des prédictions, notamment en l’absence de vérité terrain comme en astrophysique par exemple. L’inférence de distribution se traduit le plus souvent par le recours à des méthodes d’échantillonnage telles que les méthodes de Monte Carlo par chaînes de Markov. Ces méthodes sont réputées coûteuses en temps de calcul. Nous développons de nouvelles méthodes qui ouvrent la voie à la parallélisation/distribution des algorithmes associés.

L’ingénieur•e recruté•e sera chargé•e du développement de codes parallélisés/distribués permettant d’accélérer les calculs. Il/elle devra respecter une logique de recherche reproductible et de logiciel libre favorisant une diffusion publique large des codes produits. Il/elle sera chargé•e d’accompagner les membres de l’équipe via des actions de formation internes en vue de partager ses connaissances avec l’équipe SigMA. Il/elle contribuera activement aux activités de recherche dans une logique de collaboration.

Profil du candidat :
M2 ou ingénieur•e avec une spécialité en informatique, traitement du signal et des images, ou machine learning, idéalement titulaire d’un doctorat.

Formation et compétences requises :
Langages : Python, C, C++, OpenMP, MPI, Cuda, et techniques de parallélisation CPU et GPU.

Niveau d’Anglais B2 requis – Bonnes pratiques du développement collaboratif – Bonnes qualités de communication

Candidatures : envoyer CV détaillé, diplômes, relevés de notes des 2 dernières années Bac+4/+5 ; si titulaire d’un doctorat, joindre les rapports sur le manuscrit et la soutenance.
Indiquer 2 référents pour recommandation. Lettre de motivation

Candidature à transmettre à : pole.rh@centralelille.fr et pierre.chainais@centralelille.fr

Adresse d’emploi :
CRIStAL, Centre de Recherche en Informatique, Automatique et Signal de Lille (UMR 9189)
Employeur : Centrale Lille Institut, 59651 Villeneuve d’Ascq.

Contact Recherche: Pierre Chainais, pierre.chainais@centralelille.fr
Contact Administratif : Pôle des Ressources Humaines, pole.rh@centralelille.fr

Document attaché : 202203031604_annonce_ingénieur_informatique_2022_Sherlock.pdf