Offre en lien avec l’Action/le Réseau : – — –/Doctorants
Laboratoire/Entreprise : ETIS/CY Paris Cergy Université
Durée : 18 months
Contact : aikaterini.tzompanaki@cyu.fr
Date limite de publication : 2025-10-01
Contexte :
The position is funded by the prestigious EU Horizon PANDORA project, A Comprehensive Framework enabling the Delivery of Trustworthy Datasets for Efficient AIoT Operation. The main goal of the project is to contribute towards the creation of a dynamic AI pipeline in the context of IoT applications, focusing on i) the creation of synthetic but trustworthy data, and ii) on the development of AI algorithms for various tasks (from classic classification, regression and anomaly detection, to forecasting and predictive maintenance) taking into account the specific data characteristics of continuous data generation settings and leveraging the domain knowledge when available. Explainability is a fundamental property for rendering the systems reliable, as it can help i) enhance the acceptability of the system by the system experts and further aid in decision making, ii) optimize model performance (time and accuracy) by revealing the actual causes behind predictions, and iii) repair data acquisition processes or model training/updating by exposing errors and/or drifts that may arise through time. The project gathers over 20 academic and industrial partners, with real and challenging use-cases and thus provides a unique opportunity to contribute to cutting-edge research with significant real-world impact.
Sujet :
Subject: Predictive maintenance (PDM) in industrial settings spans from identifying anomalies and categorizing failures in already observed data, to prognostically predicting the Remaining Useful Life (RUL) and the Failure Time (FT) of machines, appliances, etc., in the future. Typically such predictive tasks are implemented using Deep Learning and/or statistical analysis techniques, which may be complex to interpret, while their performance is challenged by multiple sources of errors (sensors’ tuning, aging, failures, etc.) and the non-stationary nature of IoT data streams. In particular, concept drifts can originate from the changes in the underlying data generation mechanism that reflects different states of the monitored system. In this context, the predictive performance of already trained models f(X,Y) (e.g., classification, regression) may start degrading after a certain point in time and hence models need to be adapted at the right frequency. However, not all types of changes in the joint probability distribution P(X,Y) have the same impact on model performance. In this respect, we need to distinguish between changes in the posterior probability distribution P(Y|X) (i.e., Model drifts) between input features X and the target variable Y, from class-conditioned data distribution changes P(X|Y) (i.e., likelihood drifts) and changes in the distribution of input features P(X) (i.e., covariate drifts). Clearly, not all types of changes of the joint probability distribution $P(X, y)$ influence predictive models in the same way and hence, they require different mitigation actions. In this project we are particularly interested in how data quality and concept drifts affect the performance of PDM tasks such RUL and FT.
Moreover, besides developing performant, robust, and stable FT and RUL prediction algorithms through time, we are also interested in enhancing interpretability of their results. On the one hand, engineers need to know the root causes for a predicted machine failure at a time t in the future, so that they may take the best possible action towards preventing the failure to happen, or replace a machine in time before having to take the system down for replacement of the compromised machine. Such explanations should cover both the time parameter (why a failure will happen after a time interval) and the type of failure (why a specific type of error will happen). On the other hand, fine-grained explanations of the different types of concept drifts can guide data analysts to take timely, and informed actions for adapting the prediction algorithm to the observed concept drift. While explainability has been a major research interest in recent years, explanation methods for concept drift are still in their infancy. Some of the approaches aim for the detection and quantification of drift, its localization in space or its visualization, while others focus on feature-wise representations of drift. In this project, we aim to investigate actionable concept drift explanations, adding in the equation weak and strong signals for failure events. We believe that concept drift explanations constitute a form of actionable explanations responding to both aforementioned expert needs, and thus can be more valuable than standard feature importance explanations.
Profil du candidat :
Responsibilities/Opportunities
Conduct high quality research in Explainable AI in non-stationary settings.
Develop novel algorithms and methodologies for Predictive Maintenance using IoT streams.
Publish in top-tier conferences and journals.
Collaborate with an interdisciplinary team of researchers and industry partners.
Participate in project meetings and contribute to the project’s management.
Formation et compétences requises :
PhD in Machine Learning, AI, Data Science, Statistics, or a related field.
Strong background in at least one of the following: deep learning, continual learning, time series analysis, or predictive maintenance.
Experience with explainable AI (XAI) methods.
Proficiency in Python and relevant ML frameworks (PyTorch, TensorFlow, Scikit-learn).
Excellent publication record in Data management, Artificial Intelligence, Machine Learning, or IoT applications.
Adresse d’emploi :
Equipe DATA&AI – ETIS Laboratory, CY Cergy Paris Université
33 Bd du Port, 95000 Cergy
Document attaché : 202503241752_Post-Doc Position in Explainable AI in Industrial Settings_DATAIA_PARIS.pdf