01 Spatio-Temporal Data Analytics on Heterogeneous Data (Prof. Paolo Garza)
02 Ultra-low latency multimedia streaming over HTTP/3 (Prof. Antonio Servetti)
03 Media Quality Optimization using Machine Learning on Large Scale Datasets (Prof. Enrico Masala)
04 Security Analysis and Automation in Smart Systems (Prof. Fulvio Valenza)
05 Local energy markets in citizen-centered energy communities (Prof. Edoardo Patti)
06 Simulation and Modelling of V2X connectivity with traffic simulation (Prof. Edoardo Patti)
08 Multi-Device Programming for Artistic Expression (Prof. Luigi De Russis)
09 Digital Wellbeing By Desing (Prof. Alberto Monge Roffarello)
10 Management Solutions for Autonomous Networks (Prof. Guido Marchetto)
12 Digital Twin development for the enhancement of manufacturing systems (Prof. Sara Vinco)
13 State-of-Health diagnostic framework towards battery digital twins (Prof. Sara Vinco)
14 Modeling, simualtion and validation of modern electronic systems (Prof. Sara Vinco)
15 Robust AI systems for data-limited applications (Prof. Santa Di Cataldo)
16 Artificial Intelligence applications for advanced manufacturing systems (Prof. Santa Di Cataldo)
17 AI for Secured Networks: Language Models for Automated Security Log Analysis (Prof. Marco Mellia)
20 Cloud continuum machine learning (Prof. Daniele Apiletti)
21 Graph network models for Data Science (Prof. Daniele Apiletti)
26 Computational Intelligence for Computer-Aided Design (Prof. Giovanni Squillero)
28 Security of Software Networks (Prof. Cataldo Basile)
31 Monitoring systems and techniques for precision agriculture (Prof. Renato Ferrero)
34 AI-driven cybersecurity assessment for automotive (Prof. Luca Cagliero)
35 Applications of Large Language Models in time-evolving scenarios (Prof. Luca Cagliero)
36 Building Adaptive Embodied Agents in XR to Enhance Educational Activities (Prof. Andrea Bottino)
37 Real-Time Generative AI for Enhanced Extended Reality (Prof. Andrea Bottino)
41 Machine unlearning (Prof. Elena Maria Baralis)
42 Generative AI models for enhanced text-to-image synthesis (Prof. Lia Morra)
44 Cybersecurity for a quantum world (Prof. Antonio Lioy)
45 Bridging Human Expertise and Generative AI in Software Engineering (Prof. Luca Ardito)
46 Explaining AI (XAI) models for spatio-temporal data (Prof. Elena Maria Baralis)
48 Functional Safety Techniques for Automotive oriented Systems-on-Chip (Prof. Paolo Bernardi)
49 Human-aware robot behaviour learning for HRI (Prof. Giuseppe Bruno Averta)
Spatio-Temporal Data Analytics on Heterogeneous Data |
|
Proposer |
Paolo Garza |
Topics |
Data science, Computer vision and AI |
Group website |
https://dbdmg.polito.it/, https://linksfoundation.com/ |
Summary of the proposal |
Spatio-Temporal data are continuously increasing (e.g., remote sensing images, LiDAR acquisitions, and time series collected from IoT sensors). Although Spatio-Temporal data have been extensively studied, most current data analytics approaches do not effectively manage the heterogeneous nature of the data, especially considering the aforementioned domains, with most of the state-of-the-art approaches focusing on one modality at a time. Innovative AI approaches designed to solve practical tasks leveraging the multimodality and heterogeneity of information conveyed by multiscale and multitemporal geospatial data sources are the primary goals of this proposal. |
Research objectives and methods |
The main objective of this research activity is to design machine learning algorithms aimed at big data-driven applications to analyze heterogeneous Spatio-Temporal data (e.g., images from satellites, aerial vehicles or UAVs, 3D acquisitions from LiDAR sensors, or time series collected from IoT sensors), considering both descriptive and predictive problems. The main research questions that will be addressed are as follows. Heterogeneity. Several data sources are available, characterized by different data types and modalities. Each data source represents a facet of the analyzed phenomena and provides additional insights, especially when adequately integrated with other sources. Innovative integration techniques based, for instance, on latent spaces will be studied to leverage the opportunities provided by such diverse data sources. An effective integration of heterogeneous modalities could enable better performances of AI-based tasks. Scalability. Spatio-Temporal data are frequently big (e.g., vast collections of remote sensing data). Hence, big data solutions are needed to process and analyze them, mainly when historical data are analyzed. Timeliness. Timeliness is crucial in several domains (e.g., emergency management). Efficient machine learning algorithms shall be designed and implemented to deal with rapid and near real-time scenarios, with an eye towards practical and deployable solutions. The work plan for the three years is organized as follows. 1st year. Analysis of the state-of-the-art algorithms and ML frameworks for heterogeneous and multimodal Spatio-Temporal data. Based on the pros and cons of the current solutions, a preliminary common data representation based on latent spaces will be studied and designed to integrate heterogeneous data effectively. Based on the proposed data representation, novel algorithms will be designed, developed, and validated on historical data related to specific domains (e.g., emergency management). 2nd year. State-of-the-art representations of multimodal Spatio-Temporal data will be further analyzed and proposed, focusing on scalable algorithms. 3rd year. The timeliness aspect will be considered, especially during the last year. Specifically, the focus will be on near real-time Spatio-Temporal data analysis based on efficient ML-based algorithms. The activity is part of a well-established collaboration with the LINKS Foundation. The outcomes of the research activity are expected to be published at IEEE/ACM/CVF International Conferences and in any of the following journals:- ACM Transactions on Spatial Algorithms and Systems- IEEE Transactions on Knowledge and Data Engineering- IEEE Transactions on Geoscience and Remote Sensing- IEEE Transactions on Big Data- IEEE Transactions on Emerging Topics in Computing- IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing - Information Sciences (Elsevier)- Expert Systems with Applications (Elsevier)Machine Learning with Applications (Elsevier) |
Required skills |
Strong background in data science fundamentals and machine and deep learning algorithms. Strong programming skills. Knowledge of frameworks such as PyTorch or Spark is advisable but not mandatory. |
Ultra-low latency multimedia streaming over HTTP/3 |
|
Proposer |
Antonio Servetti |
Topics |
Computer graphics and Multimedia, Parallel and distributed systems, Quantum computing |
Group website |
https://media.polito.it/ |
Summary of the proposal |
The growing demand for interactive web services has also led to the need for interactive video applications, capable of accommodating a much larger audience than videoconferencing tools, but with almost the same, i.e., strict, requirements in end-to-end latency. This proposal aims to define and study new media coding and transmission techniques that will exploit new HTTP/3 features, such as QUIC and WebTransport, to improve the scalability and reduce the latency of current streaming solutions. |
Research objectives and methods |
Research objectives |
Required skills |
The candidate is expected to have a good background in multimedia coding/streaming, computer networking, and web development. A reasonable knowledge of network programming and software development in the Unix/Linux environment is appreciated. |
Media Quality Optimization using Machine Learning on Large Scale Datasets |
|
Proposer |
Enrico Masala |
Topics |
Computer graphics and Multimedia, Data science, Computer vision and AI |
Group website |
http://media.polito.it |
Summary of the proposal |
Machine learning (ML) significantly changed the way many optimization tasks are addressed. Here the focus is on optimizing the media compression and communication scenario, trying to predict users' quality of experience. Key objectives of the proposal are creation of tools to analyze and exploit large scale datasets using ML to identify media characteristics and features that most influence perceptual quality. Such new knowledge will be fundamental to improve existing measures and algorithms |
Research objectives and methods |
In recent years, ML has been successfully employed to develop video quality estimation algorithm (see, e.g., the Netflix VMAF proposal) to be integrated in media quality optimization frameworks. However, despite these improvements, no technique can currently be considered reliable, partly because the inner workings of ML models cannot be easily and fully understood especially when they are based on "black box" neural network models. We aim to improve the situation by developing more reliable and explainable quality prediction models. Starting from Internet Media Group's ongoing work on modeling the behavior of single human subjects in media quality experiments, the candidate will derive a systematic approach by employing several subjectively annotated datasets (i.e., with quality scores given by human subjects). With such an approach we expect to be able to identify meaningful media quality features useful to develop new reliable and explainable quality prediction models. The workplan of the activities is detailed in the following. In the first year the PhD candidate will first familiarize with the recently proposed ML and AI-based techniques for media quality optimization, as well as the characteristics of the publicly available datasets for research purposes. Possible targets for research publications, well known to the proposer, include IEEE Transactions on Multimedia, Elsevier Signal Processing: Image Communication, ACM Transactions on Multimedia Computing Communications and Applications, Elsevier Multimedia Tools and Applications, various IEEE/ACM international conferences (IEEE ICME, IEEE MMSP, QoMEX, ACM MM, ACM MMSys). The proposer is actively collaborating with the Video Quality Experts Group (VQEG), an international group of experts from academia and industry that aims to develop new standards in the context of video quality. In particular the tutor is co-chair of the JEG-Hybrid project which is very interested in the activity previously described. |
Required skills |
The PhD candidate is expected to have: strong analytical skills; some background on ML systems; good English writing and communication skills; reasonably good ability/willingness to learn how to work with large quantities of data on remote server systems, in particular by automating the procedures with scripts, pipelines, etc. |
Security Analysis and Automation in Smart Systems |
|
Proposer |
Fulvio Valenza |
Topics |
Cybersecurity |
Group website |
http://netgroup.polito.it |
Summary of the proposal |
Cyber-physical systems and their smart components are pervasive in our daily activities. Unfortunately, identifying the potential threats and issues in these systems and selecting and configuring enough protection is challenging, given that such environments combine human, physical, and cyber aspects to the system design and implementation. This research aims to fill this gap by defining a novel, highly automated system to analyze and enforce security formally and optimally. |
Research objectives and methods |
The main objective of the proposed research is to improve the state of the art of security analysis and automation in novel Cyber-Physical Systems (i.e., Smart Systems), mainly focusing on the automated implementation of threat analysis and access control and defense policies. |
Required skills |
In order to successfully develop the proposed activity, the candidate should have a good background in cybersecurity (especially in network security), and good programming skills. Some knowledge of formal methods can be useful, but it is not required: the candidate can acquire this knowledge and related skills as part of the PhD Program, by exploiting specialized courses. |
Local energy markets in citizen-centered energy communities |
|
Proposer |
Edoardo Patti |
Topics |
Software engineering and Mobile computing, Parallel and distributed systems, Quantum computing, Computer architectures and Computer aided design |
Group website |
www.eda.polito.it |
Summary of the proposal |
A smart citizen-centric energy system is at the centre of the energy transition. Energy communities will enable citizens to participate actively in local energy markets by exploiting new digital tools. Citizens will need to understand how to interact with smart energy systems, novel digital tools and local energy markets. Thus, new complex socio-techno-economic interactions will take place in such smart systems which need to be analyzed and simulated to evaluate possible future impacts. For this purpose, a novel co-simulation framework is needed, which combines agent-based modelling techniques with external simulators of the grid and energy sources. |
Research objectives and methods |
The diffusion of distributed (renewable) energy sources poses new challenges in the underlying energy infrastructure, e.g., distribution and transmission networks and/or within micro (private) electric grids. The optimal, efficient and safe management and dispatch of electricity flows among different actors (i.e., prosumers) is key to supporting the diffusion of the distributed energy sources paradigm. The goal of the project is to explore different corporate structures, billing and sharing mechanisms inside energy communities. For instance, the use of smart energy contracts based on Distributed Ledger Technology (blockchain) for energy management in local energy communities will be studied. A testbed comprising of physical hardware (e.g., smart meters) connected in the loop with a simulated energy community environment (e.g., a building or a cluster of buildings) exploiting different Renewable Energy Sources (RES) and energy storage technology will be developed and tested during the three-year program. Hence, the research will focus on the development of agents capable of describing:- the final customer/prosumer beliefs desires and intentions and opinions.- the local energy market where prosumers can trade their energy and or flexibility- the local system operator that has to provide the grid reliability |
Required skills |
Programming and Object-Oriented Programming (preferable in Python). Frameworks for Multi Agent Systems Development (preferable). Development in web environment (e.g. REST web services). Computer Networks |
Simulation and Modelling of V2X connectivity with traffic simulation |
|
Proposer |
Edoardo Patti |
Topics |
Data science, Computer vision and AI, Parallel and distributed systems, Quantum computing, Software engineering and Mobile computing |
Group website |
www.eda.polito.it |
Summary of the proposal |
The development of novel ICT solutions in smart-grids has opened new opportunities to foster novel services for energy management and savings in all end-use sectors, with particular emphasis on Electric Vehicle connectivity, such as demand flexibility. Thus, there will be a strong interaction among transportation, traffic trends and energy distribution systems. New simulation tools are needed to evaluate the impact of Electric Vehicles in the grid by considering citizens behaviors. |
Research objectives and methods |
This research aims at developing novel simulation tools for smart cities/smart grid scenarios that exploit the Agent-Based Modelling (ABM) approach to evaluate novel strategies to manage the V2X connectivity with traffic simulation. The candidate will develop an ABM simulator that will provide a realistic and virtual city where different scenarios will be executed. The ABM should be based on real data, demand profiles and traffic patterns. Furthermore, the simulation framework should be flexible and extendable so that i) It can be improved with new data from the field; ii) it can be interfaced with other simulation layers (i.e. physical grid simulators, communication simulators); iii) It can interact with external tools executing real policies (such as energy aggregation). This simulator will be a useful tool to analyse how V2X connectivity and the associated services impact both social behaviours and traffic. It will also help the understanding of the impact of new actors and companies (e.g., sharing companies) in both the marketplace and the society, again by analysing the social behaviours and the traffic conditions. In a nutshell, ABM simulator will simulate both traffic variation and the possible advantages of V2X connectivity strategies in a smart grid context. This ABM simulator will be designed and developed to span different spatial-temporal resolutions. All the software entities will be coupled with external simulators of grid and energy sources in a plug-and-play fashion to be ready for being integrated with external simulators and platforms. This will enhance the resulting AMB framework unlocking also hardware in the loop features. |
Required skills |
Programming and Object-Oriented Programming (preferable in Python), |
Machine Learning techniques for real-time State-of-Health estimation of Electric Vehicles batteries |
|
Proposer |
Edoardo Patti |
Topics |
Data science, Computer vision and AI, Software engineering and Mobile computing, Computer architectures and Computer aided design |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
This Ph.D. research proposal aims at studying novel software solutions based on Machine Learning (ML) techniques to estimate the State-of-Health (SoH) of batteries in Electric Vehicles (EV) in near-real-time. This research area is gaining a strong interest in the last years as the number of EVs is constantly rising. Knowing this SoH can unlock different possible strategies i) to reuse EVs' batteries in other contexts, e.g. stationary energy storage systems in Smart Grids, or ii) to recycle them. |
Research objectives and methods |
In the last years, the number of Electric Vehicles (EVs) increased significantly and it is expected to grow in the upcoming years. Due to the use of high-value materials, there is a strong economic, environmental and political interest in implementing solutions to recycle EV's batteries for example by reusing them in stationary applications to become useful energy storage systems in Smart Grids. To achieve it, novel tools are needed to estimate the battery State-of-Health (SoH), i.e. the battery capacity measurement, in near-real-time. Currently, SoH is determined by bench discharging tests taking several hours and making this process time-consuming and expensive. |
Required skills |
Programming and Object-Oriented Programming (preferable in Python). Knowledge of Machine Learning and Neural Networks. Knowledge of frameworks to develop models based on Machine Learning and Neural Networks. Knowledge of development of Internet of Things Applications |
Multi-Device Programming for Artistic Expression |
|
Proposer |
Luigi De Russis |
Topics |
Computer graphics and Multimedia, Software engineering and Mobile computing |
Group website |
https://elite.polito.it |
Summary of the proposal |
Media artists use smartphones and IoT devices as material for creative exploration. However, some do not code and have a low interest in learning. In addition, programming artworks enclose characteristics that differ from traditional coding. This Ph.D. proposal aims to extend our comprehension of the needs of artists for creative coding through the design, implementation, and evaluation of toolkits that serve them to realize code-based artworks across multiple devices and media effectively. |
Research objectives and methods |
The recent availability of smartphones, AR/VR headsets, IoT-enabled devices, and microcontroller kits creates new opportunities for creative explorations for media artists and designers. The field of creative coding emphasizes the goal of expression rather than function, and creative coders combine computational skills with creative insight. In some cases, artists and designers are interested in creative coding but lack the knowledge or programming skills to benefit from the offered possibilities. The main research objective of this Ph.D. proposal is to extend our comprehension of the needs of media artists and designers for creative coding across multiple devices and media. To reach this objective, the Ph.D. student will study, design, develop, and evaluate proper models and novel technical solutions (e.g., toolkits and tools) for supporting creative coders. The proposal envisions focusing on both creative coders and end-user programmers. The work will start from the needs of the stakeholders (i.e., artists and designers), complemented by the existing literature. Using a participatory approach, the Ph.D. student will keep the stakeholders involved in the various phases of the work. In particular, the Ph.D. research activity will focus on the following: - Study of the creative coding field, stakeholders' needs and current tools, and HCI techniques able to support the identification of suitable requirements and the creation of technical solutions to effectively support creative exploration and coding. - Creation of a theoretical framework to satisfy the identified needs and requirements, able to adapt to different media, devices, and skills. For instance, it can include end-user personalization as a way to allow end-users to create code-based artifacts and AI techniques to support the creation of programs. - Development of a toolkit and related tools to experiment with the theoretical framework's facets. The creation and evaluation of the tools will serve as the validation for the framework. The work plan will be organized according to the following four phases, partially overlapped:- Phase 1 (months 0-6): literature review about creative coders and coding; focus groups and interviews with designers and media artists of various skills; definitions and development of a set of use cases and promising strategies to be adopted.- Phase 2 (months 6-18): research, definition, and experimentation of an initial version of the theoretical framework and a first toolkit for creative coders, starting from the outcome of the previous phase. In this phase, the focus will be on the most common target devices, i.e., the smartphone and the PC, with the design, implementation, and evaluation of suitable tools.- Phase 3 (months 12-24): research, definition, and experimentation of a second toolkit (or an evolution of the previous one) for novice creative coders and end-users. Such a toolkit uses artificial intelligence and machine learning to help during the coding process. The subsequent design, implementation, evaluation of suitable tools, and extension of the framework.- Phase 4 (months 24-36): extension and generalization of the previous phases to include additional target devices and consolidate the theoretical framework; evaluation of the toolkit and the tools in real settings with a large number of artists. It is expected that the results of this research will be published in some of the top conferences in the Human-Computer Interaction field (e.g., ACM CHI, ACM CSCW, ACM C&C, and ACM IUI). One or more journal publications are expected on a subset of the following international journals: ACM Transactions on Computer-Human Interaction, ACM Transactions on Interactive Intelligent Systems, and International Journal of Human Computer Studies. |
Required skills |
A candidate interested in the proposal should (i) be able to critically analyze and evaluate existing research, as well as gather and interpret data from various sources; (ii) be able to communicate research findings through writing and presenting; (iii) have a solid foundation in computer science/engineering and possess the relevant technical skills; (iv) have a good understanding of HCI research methods, especially around needfinding. |
Digital Wellbeing By Desing |
|
Proposer |
Alberto Monge Roffarello |
Topics |
Computer graphics and Multimedia, Data science, Computer vision and AI, Software engineering and Mobile computing |
Group website |
https://elite.polito.it/ |
Summary of the proposal |
Tools for digital wellbeing allow users to self-control their habits with distractive apps and websites. Yet, they are ineffective in the long term, as tech companies still adopt attention-capture designs, e.g., infinite scroll, that compromise users' self-control. This PhD proposal investigates innovative strategies for designers and end users to consider digital wellbeing in user interface design, recognizing the need to foster healthy digital experiences without depending on external support. |
Research objectives and methods |
In today's attention economy, tech companies compete to capture users' attention, e.g., by introducing visual features and functionalities - from guilty-pleasures recommendations to content autoplay - that are purposely designed to maximize metrics such as daily visits and time spent. These Attention-Capture Damaging Patterns (ACDPs) [1] compromise users' sense of agency and self-control, ultimately undermining their digital wellbeing. |
Required skills |
A candidate interested in the proposal should ideally: be able to critically analyze and evaluate existing research, as well as gather and interpret data from various sources; be able to communicate research findings through writing and presenting; have a solid foundation in computer science/engineering and possess relevant technical skills; have a good understanding of HCI research methods, especially around needfinding. |
Management Solutions for Autonomous Networks |
|
Proposer |
Guido Marchetto |
Topics |
Parallel and distributed systems, Quantum computing |
Group website |
http://www.netgroup.polito.it |
Summary of the proposal |
Next-Generation (NextG) networks are expected to support advanced and critical services, incorporating computation, communication, and intelligent decision making. |
Research objectives and methods |
Two research questions (RQ) guide the proposed work: RQ1: How can we design and implement on local and larger-scale testbeds effective transport and routing network protocols that integrate the network stack at different scopes using recent advances in supervised and unsupervised learning? RQ2: To scale the use of machine learning-based solutions in network management, what are the most efficient distributed machine learning architectures that can be implemented at the network edge layer? The final target of the research work is to answer these questions, also by evaluating the proposed solutions on small-scale network emulators or large-scale virtual network testbeds, using a few applications, including virtual and augmented reality, precision agriculture, or haptic wearables. In essence, the main goals are to provide innovation in network monitoring, network adaptation, and network resilience, using centralized and distributed learning integrated with edge computing infrastructures. Both vertical and horizontal integration will be considered. By vertical integration, we mean considering learning problems that integrate states across network hardware and software, as well as states across the network stack across different scopes. For example, the candidate will design data-driven algorithms for congestion control problems to address the tussle between in-network and end-to-end congestion notifications. By horizontal learning, we mean using states from local (e.g., physical layer) and wide area (e.g., transport layer) as input for the learning-based algorithms. The data needed by these algorithms are carried to the learning actor by means of newly defined in-band network telemetry mechanisms. Aside from supporting resiliency with the vertical integration, solutions must offer resiliency across a wide (horizontal) range of network operations: from close-edge, i.e., near the device, to the far-edge, with the design of secure data-centric resource allocation (federated) algorithms. The research activity will be organized in three phases: Phase 1 (1st year): the candidate will analyze the state-of-the-art solutions for network management, with particular emphasis on knowledge-based network automation techniques. The candidate will then define detailed guidelines for the development of architectures and protocols that are suitable for automatic operation and configuration of NextG networks, with particular reference to edge infrastructures. Specific use-cases will also be defined during this phase (e.g., in virtual reality). Such use cases will help identifying ad-hoc requirements and will include peculiarities of specific environments. With these use cases in mind, the candidate will also design and implement novel solutions to deal with the partial availability of data within distributed edge infrastructures. Results of this work will likely result in conference publications. Phase 2 (2nd year): the candidate will consolidate the approaches proposed in the previous year, focusing on the design and implementation of mechanisms for vertical and horizontal integration of supervised and unsupervised learning with network virtualization. Network, and computational resources will be considered for the definition of proper allocation algorithms. All solutions will be implemented and tested. Results will be published, targeting at least one journal publication. Phase 3 (3rd year): the consolidation and the experimentation of the proposed approach will be completed. Particular emphasis will be given to the identified use cases, properly tuning the developed solutions to real scenarios. Major importance will be given to the quality offered to the service, with specific emphasis on the minimization of latencies in order to enable a real-time network automation for critical environments (e.g., telehealth systems, precision agriculture, or haptic wearables). Further conference and journal publications are expected. The research activity is in collaboration with Saint Louis University, MO, USA, also in the context of the NSF grant #2201536 ?Integration-Small: A Software-Defined Edge Infrastructure Testbed for Full-stack Data-Driven Wireless Network Applications?. Furthermore, it is related to active collaborations with Futurewei Inc. and Tiesse SpA, both interested in the covered topics. The contributions produced by the proposed research can be published in conferences and journals belonging to the areas of networking and machine learning (e.g. IEEE INFOCOM, ICML, ACM/IEEE Transactions on Networking, or IEEE Transactions on Network and Service Management) and cloud/fog computing (e.g. IEEE/ACM SEC, IEEE ICFEC, IEEE Transactions on Cloud Computing), as well as in publications related to the specific areas that could benefit from the proposed solutions (e.g., IEEE Transactions on Industrial Informatics, IEEE Transactions on Vehicular Technology) |
Required skills |
The ideal candidate has good knowledge and experience in networking and machine learning, or at least in one of the two topics. Availability for spending periods abroad (mainly but not only at Saint Louis University) is also important for a profitable development of the research topic. |
Preserving privacy and fairness with generative AI-based synthetic data production |
|
Proposer |
Antonio Vetro' |
Topics |
Software engineering and Mobile computing, Data science, Computer vision and AI |
Group website |
https://nexa.polito.it |
Summary of the proposal |
Synthetic data generation is fundamental in contexts of data scarcity or low economical resources to collect data. However, several challenges are still open in this research field, the most important being the trade-off between privacy, fairness and accuracy. The goal of this PhD proposal is to design, develop and test new generative models for synthetic data production able to preserve privacy, guarantee fairness and good levels of accuracy. |
Research objectives and methods |
Synthetic data generation enables the reproduction, diversification, and augmentation of real data in contexts where data is scarce and where preserving privacy is paramount. However, synthetic data may come at costs of unrealistic synthetic populations or limited accuracy in the downstream predictions and classifications. In addition, reliable techniques for reaching satisfactory trade-offs between contrasting requirements (e.g., privacy, fairness, and accuracy) are still object of research and experimentation, as well as how to produce suitable dataset and model documentation. RQ 3. Task 3.1) State of art of existing documentation suites for datasets (e.g., datasheets) and models (e.g., model cards); collect and organize online software systems that are suitable candidates for black box testing against discrimination (e.g., insurance, online advertising, etc.). RQ 3. - Task 3.4) Quantitative analysis of discrimination in selected existing software systems using the synthetic data generated. |
Required skills |
The candidate should have: ? Basic knowledge on software testing concepts, techniques, and methodologies. ? Basic knowledge of AI techniques. ? Good knowledge of statistical methods for analyzing experimental data. ? Proficiency in data analysis techniques and tools. ? Strong programming skills. ? Basic knowledge of the problem of algorithm bias. ? Research aptitude and curiosity to cross disciplinary boundaries. The candidate should also possess good communication and presentation skills. |
Digital Twin development for the enhancement of manufacturing systems |
|
Proposer |
Sara Vinco |
Topics |
Data science, Computer vision and AI, Controls and system engineering |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
Industry 4.0 has deeply changed manufacturing: enormous quantity of data allows to build data-based decision-support strategies and to reduce down time and defects. Many challenges are posed by the heterogeneity and variety of data and by the construction of effective data-based analytics. This program tackles such challenges to build a virtual replica of a manufacturing system (digital twin), e.g.. targeting production lines, tire production, semiconductor manufacturing, battery management. |
Research objectives and methods |
The main goal of this PhD program is the construction of a digital twin of a manufacturing system, to improve production effectiveness. A digital twin is a virtual replica of the system that exploits available technologies (Artificial Intelligence, data management and mining, Internet of Things, etc.) to enhance production automatically or through decision support systems. While the technologies per se are well established, their application in real life scenarios is still preliminary. Manufacturing systems indeed entail challenges such as: extreme data variety and variability, protocol heterogeneity, lack of data collection infrastructures, reduced data availability for the training of algorithms. This PhD program seeks solutions to these challenges, to allow e.g., anomaly detection, maintenance support, and automatic optimization of the production flow. Example of application scenarios are new generation manufacturing systems, such as tire production, line production, semiconductor manufacturing. All scenarios will be investigated with the support and with case studies provided by industrial and research partners, such as Michelin, STMicroelectornics, Technoprobe. |
Required skills |
The ideal candidate to this PhD program has: |
State-of-Health diagnostic framework towards battery digital twins |
|
Proposer |
Sara Vinco |
Topics |
Controls and system engineering, Data science, Computer vision and AI |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
The adoption of EVs is limited by their reliance on batteries with low energy and power densities compared to liquid fuels and subject to aging and performance deterioration. For this reason, monitoring the battery state-of-charge (SoC) and -health (SoH) is a very relevant problem. This PhD program focuses on the development of models for battery SoC and SoH, with the goal of enabling continuous monitoring of batteries and of improving their design and management throughout their lifetime. |
Research objectives and methods |
The main goal of this PhD program is the construction of a framework to simulate battery behavior over time, to create a virtual replica and allow the analysis of different management strategies and configurations. This will require: - The identification, analysis and adoption of datasets (both public and private) of batteries - The construction of models with different levels of accuracy and different flows, e.g., based on Artificial Intelligence techniques (e.g., Physically Informed Neural Networks, Machine Learning) and on top-down modeling techniques (e.g., circuit models)- The definition of the monitoring architecture to be installed at the level of the Battery Manage System (BMS) or in an IT infrastructure, to define decision-support solutions, digital twins to the customer, or other services |
Required skills |
The ideal candidate to this PhD program has: |
Modeling, simualtion and validation of modern electronic systems |
|
Proposer |
Sara Vinco |
Topics |
Computer architectures and Computer aided design, Controls and system engineering |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
The current international semiconductor scenario is extremely competitive and is pushing for strong innovation advancement. This PhD program focuses on the development of modeling, simulation and validation flows of innovative systems, including not only digital functionality but also thermal and power flows, mechanical components (e.g., accelerometers) and analog subsystems (e.g., gate drivers). Research is supported by international projects and partners. |
Research objectives and methods |
Modern electronic systems are tightly coupled to mechanical, thermal, power aspects that must be taken into account at design time to ensure the correct operation of the final system. Ignoring behaviors or potential faults of connected analog subsystems or mechanical actuators may indeed lead to unsafe or incorrect behaviors, that prevent the operation of the design after deployment. This requires to extend the traditional design, simulation and validation flows with a sensibility to extra-functional and non-digital aspects. The main goal of this PhD program is the definition of such flows, through the adoption of OpenHW, standard, technologies such as SystemC(-AMS), RISC-V, IP-XACT, and other technologies that fall under the ChipsAct umbrella of EU research. Example of application scenarios are smart systems such as drones, and automotive and robotics applications. All scenarios will be investigated with the support and with case studies provided by industrial and research partners, such as Infineon and STMicroelectornics. |
Required skills |
The ideal candidate to this PhD program has: |
Robust AI systems for data-limited applications |
|
Proposer |
Santa Di Cataldo |
Topics |
Data science, Computer vision and AI |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
Artificial Intelligence is driving a revolution in many important sectors in society. Deep learning networks, and especially supervised ones such as Convolutional Neural Networks, remain the go-to approach for many important tasks. Nonetheless, training these models typically requires massive amount of good-quality annotated data, which makes them impractical in many real-world applications. This PhD program seeks answers to such problems, targeting important use-cases in today's society (among the others: industry 4.0 and biomedical applications). |
Research objectives and methods |
The main goal of this PhD program is the investigation of robust AI-based decision making in data-limited situations. This includes three possible scenarios, which are typical of many important real-world applications:- the training data is difficult to obtain, or it is available in limited quantity.- obtaining the training data is not difficult. Nonetheless, it is either difficult or economically impractical to have human experts labelling the data.- the training data/annotations are available, but the quality of such data is very poor. |
Required skills |
The ideal candidate to this PhD program has: |
Artificial Intelligence applications for advanced manufacturing systems |
|
Proposer |
Santa Di Cataldo |
Topics |
Data science, Computer vision and AI |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
Industry 4.0 refers to digital technologies designed to sense, predict, and interact with production systems, to make decisions that support productivity, energy-efficiency, and sustainability. While Artificial Intelligence plays a crucial role in this paradigm, many challenges are still posed by the nature and dimensionality of the data, and by the immaturity and intrinsic complexity of some of the processes involved. The aim of this PhD program is to successfully tackle these challenges. |
Research objectives and methods |
The main goal of this PhD program is the investigation, design and deployment of state-of-the-art Artificial Intelligence approaches in the context of the smart factory, with special regards with new generation manufacturing systems. These tasks include:- quality assurance and inspection of manufactured product via heterogeneous sensors data (e.g., images from visible range or IR cameras, time-series, etc.)- process monitoring and forecasting- anomaly detection- failure prediction and maintenance planning support |
Required skills |
The ideal candidate to this PhD program has: |
AI for Secured Networks: Language Models for Automated Security Log Analysis |
|
Proposer |
Marco Mellia |
Topics |
Cybersecurity, Data science, Computer vision and AI |
Group website |
https://smartdata.polito.it/ |
Summary of the proposal |
Network security analysts are a key component of the defence infrastructure of an organization. They continuously and manually analyze security alarms and logs to make decisions against undesired intrusions. |
Research objectives and methods |
Research objectives: The candidate will perform research to determine whether, and to what extent, the recent advances in language models could be used to automate and assist security analysts in the process (i) of learning the security-device rules by example and (ii) autonomously investigating the challenging cases currently analyzed by humans. Outline of the research work plan: 1st year- Study of the state-of-the-art of security log analysis and state-of-the-art language models in ML.- Data collection and analysis of raw and structured data on security devices such as Firewall/Intrusion Prevention Systems (IPS), Endpoint Detection and Response (EDR) and Cloud security services. 2nd year- Adaptation and extension solutions to learn the security-device rules by example and autonomously investigate complex cases.- Propose and develop innovative solutions to the problems of cyber threats analysis with Language models.- Propose multi-modal embeddings for network raw data and security logs. 3rd year - Tune the developed techniques and highlight possible strategies to counteract the various threats.- Application of the strategies to new data for validation and testing. References: List of possible venues for publications: |
Required skills |
- Good programming skills (such as Python, Torch, Spark) - Excellent Machine Learning knowledge - Knowledge NLP and LM - Basics of Networking and security |
Leveraging Machine Learning Analytics for Intelligent Transport Systems Optimization in Smart Cities |
|
Proposer |
Marco Mellia |
Topics |
Data science, Computer vision and AI |
Group website |
https://smartdata.polito.it/ |
Summary of the proposal |
Electrification and big data are changing the design of transport systems. The availability of large amounts of data collected by black boxes for insurance/safety opens innovative challenges and opportunities to improve transport systems and reduce carbon footprint. |
Research objectives and methods |
Research objectives: This proposal outlines a comprehensive plan to leverage big data analytics for intelligent transport systems in smart cities. The impact of mobility based on electric vehicles and its comparison with previous habits will be a core part of the study. Outline of the research work plan: 1st year- Study of the state-of-the-art data analysis techniques for transportation and mobility.- Data collection, Exploration and Pre-processing: Extract and pre-process raw data from black boxes, ensuring data quality and compatibility for further analysis. Develop techniques to handle missing or incomplete data.- Investigate and implement privacy-preserving methods to ensure ethical use of mobility data while still deriving valuable insights. 2nd year- Apply machine learning algorithms to identify patterns in mobility data, extracting insights into traffic flows, congestion, and usage patterns.- Implement anomaly detection mechanisms to identify unusual events and improve system resilience- Develop predictive models to forecast traffic conditions, enabling proactive measures to alleviate congestion and enhance overall traffic management.- Explore adaptive algorithms for real-time adjustments based on dynamic traffic patterns. 3rd year - Integrate developed algorithms into a cohesive system for intelligent transport systems.- Validate the system using real-world scenarios and fine-tune algorithms for optimal performance. References:- Ciociola, A., Cocca, M., Giordano, D., Mellia, M., Morichetta, A., Putina, A., & Salutari, F. (2017, August). UMAP: Urban mobility analysis platform to harvest car-sharing data. In SmartWorld/(pp. 1-8). IEEE.- Cocca, M., Giordano, D., Mellia, M., & Vassio, L. (2019). Free-floating electric car sharing: A data-driven approach for system design. IEEE Transactions on Intelligent Transportation Systems, 20(12), 4691-4703.- Cocca, M., Giordano, D., Mellia, M., & Vassio, L. (2019). Free-floating electric car sharing design: Data-driven optimisation. Pervasive and Mobile Computing, 55, 59-75.
|
Required skills |
- Good programming and data analysis skills (such as Python, Pandas, Torch, Spark) |
Natural Language Processing e Large Language Models for source code generation |
|
Proposer |
Edoardo Patti |
Topics |
Data science, Computer vision and AI, Software engineering and Mobile computing |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
This Ph.D. research is focused on revolutionizing source code generation by harnessing the capabilities of Natural Language Processing by exploring novel methodologies to facilitate the creation of high-quality code through enhanced human-machine collaboration. By leveraging advanced language models, like Generative Pretrained Transformer models, the research seeks to optimize the process, leading to more efficient, expressive, and context-aware source code generation in software development. |
Research objectives and methods |
The integration of Artificial Intelligence, especially Machine/Deep Learning, in industrial processes promises swift changes. Companies stand to benefit in the short term with improved production quality, efficiency, and automated routine tasks, fostering positive impacts on work environments. In addition to Natural Language Processing, Large Language Models (LLMs) have already demonstrated significant progress in healthcare, education, software development, finance, journalism, scientific research, and customer support. The future entails optimizing LLMs for widespread use, enhancing the competitiveness of the industrial system and streamlining collaborative supply chain management. The objective of this Ph.D. proposal consists of the design and development of AI-assisted models based on Natural Language Processing (NLP) and Large Language Models (LLMs) to optimize the AI-assisted source code generation in the context of software development by enhancing the process, leading to more efficient, expressive, and context-aware. During the three years of the Ph.D., the research activity will be divided into five phases:- Survey existing literature on NLP applications in software engineering and analyze methodologies and challenges in source code generation using language models.- Design and develop Large Language Models for improved programming language understanding by investigating techniques for domain-specific customization of language models.- Develop algorithms and strategies for context-aware source code generation by implementing prototype systems for evaluation and refinement.- Design and implement a collaborative framework that seamlessly integrates developer input with language model suggestions.- Evaluate the effectiveness of the collaboration framework through user studies and real-world projects. Possible international scientific journals and conference:- IEEE Transactions on Audio, Speech, and Language Processing- IEEE Transactions on Software Engineering- IEEE Transaction on Industrial Informatics,- IEEE Transactions on Industry Applications,- Engineering Applications of Artificial Intelligence,- Expert Systems with Applications,- IEEE NLP-KE internat. conf.- IEEE ICNLP internat. conf.- IEEE Compsac internat. conf. |
Required skills |
Programming and Object-Oriented Programming (preferable in Python), |
Cloud continuum machine learning |
|
Proposer |
Daniele Apiletti |
Topics |
Data science, Computer vision and AI, Parallel and distributed systems, Quantum computing, Software engineering and Mobile computing |
Group website |
|
Summary of the proposal |
As the demand for novel distributed machine learning models operating at the edge continues to grow, so does the call for cloud continuum frameworks to support machine learning. |
Research objectives and methods |
Research Objectives |
Required skills |
Knowledge of the basic computer science concepts. |
Graph network models for Data Science |
|
Proposer |
Daniele Apiletti |
Topics |
Data science, Computer vision and AI, Parallel and distributed systems, Quantum computing, Software engineering and Mobile computing |
Group website |
|
Summary of the proposal |
Machine learning approaches extract information from data with generalized optimization methods. However, besides the knowledge brought by the data, extra a-priori knowledge of the modeled phenomena is often available. Hence an inductive bias can be introduced from domain knowledge and physical constraints, as proposed by the emerging field of Theory-Guided Data Science. |
Research objectives and methods |
Research Objectives |
Required skills |
- Knowledge of the basic computer science concepts. |
Automatic composability of Large Co-simulation Scenarios for smart energy communities |
|
Proposer |
Edoardo Patti |
Topics |
Parallel and distributed systems, Quantum computing, Data science, Computer vision and AI, Computer architectures and Computer aided design |
Group website |
www.eda.polito.it |
Summary of the proposal |
The emerging concept of multi-energy systems is linked to heterogeneous competencies spanning from energy systems to cyber-physical systems and active prosumers. Studying such complex systems needs the usage of co-simulation techniques. However, the setup of co-simulation scenarios requires a deep knowledge of the framework and a time-consuming setup of the distributed infrastructure. The research program aims to develop automatic composability of multi-energy system co-simulations to ease usage |
Research objectives and methods |
A complex system such as a multi-energy system requires the accurate modelling of the heterogeneous aspects that constitute the overall phenomena under study. To achieve this goal researchers in different fields have started using co-simulation and model coupling to build new models capable of describing the interactions and the overall complexity. Such approaches give the possibility of coupling different models, running on different simulators and/or simulation engines, by exchanging data via some standard protocols over the internet. Indeed, such models have been developed and validated following a methodology that can be compared to service-oriented architecture, thus, reducing the time and complexity of building new models from scratch. Moreover, such an approach disease the interconnection of the vertical knowledge coming from each discipline/domain that is involved in the complex system, eg. ICT or Energy experts. Examples of models can be software entities that replicate the realistic behaviour of a photovoltaic (PV) system, energy storage, heating distribution networks or, even, human beans. Nowadays, researchers have invested in the usage of co-simulation orchestrators to achieve the goal of interconnection and synchronization of different models and simulators, including real-time simulators. However, the setup of the co-simulation is not an easy and trivial task as it is time-consuming and it requires the involvement of domain and co-simulation experts. This research topic aims to develop a framework, that exploits existing co-simulation orchestrators, for the automatic composability of co-simulation scenarios in a distributed infrastructure to assess different aspects of Multi-Energy-Systems. The framework will integrate models in a plug-and-play fashion reducing as much as possible the coding phase and the presence of a co-simulation expert easing the work of multi-energy systems engineers. Moreover, the framework will ease the setup in terms of computational resources for the modelling of complex and large scenarios. The final purpose consists of simulating the impact and management of future energy systems to foster the energy transition. Thus, the resulting infrastructure will integrate with a semantic approach in a distributed environment heterogeneous i) data sources, ii) cyber-physical-systems, i.e. Internet-of-Things devices iii) models of energy systems and iv) real-time simulators. The starting point of this activity will be the already existing EC-L co-simulation platform, which will be enhanced by embedding all the aforementioned features. Hence the research will focus on developing:- a methodology based on semantic web technologies for linking and interconnecting simulators automatically in a co-simulation approach- a domain-specific ontology for describing the components and interconnection of multi-energy system models- a methodology for the automatic composability and setup of the distributed infrastructures of the energetic scenario to assess (e.g., the impact of PV systems and EVs in a city) The outcomes of this research will be a distributed co-simulation platform for:- planning the evolution of the future smart multi-energy system by taking into account the operational phase- evaluating the effect of different policies and related customer satisfaction- evaluating the performances of hardware components in a realistic test bench During the first year, the candidate will study the literature solutions of existing co-simulation platforms to identify the best available solution for i) large-scale smart energy system simulation in distributed environments and ii) semantic web solutions to describe complex systems with a focus on the multi-energy system domain. Finally, the student will design the overall framework starting from the requirements identification and definition. During the second year, the candidate will face the implementation of the visual and semantic framework for model coupling and scenario creation. Furthermore, the candidate will start developing software solutions to automatic composability and setup of the co-simulation environments in terms of simulator deployment in a cloud system. During the third year, the candidate will complete the overall framework development and test it in different case study scenarios to assess the capabilities of the platform in terms of automatic scenario composition and setup. Possible international scientific journals and conferences:- IEEE Transaction Smart Grid- IEEE Transaction on Industrial Informatics,- IEEE Transaction on sustainable computing,- IEEE EEEIC internat. conf.- IEEE SEST internat. conf.- IEEE Compsac internat. conf. |
Required skills |
Programming and Object-Oriented Programming (preferable in Python), |
Multivariate time series representation learning for vehicle telematics data analysis |
|
Proposer |
Luca Cagliero |
Topics |
Data science, Computer vision and AI |
Group website |
https://smartdata.polito.it/ |
Summary of the proposal |
This PhD proposal aims to study new techniques for embedding multivariate time series, apply them to solve established downstream tasks, and leverage these solutions in Data Science pipelines to analyze vehicles' telematics data such as CAN Bus signals. Embeddings will not only capture the series' temporal properties but also their multi-dimensional relations. These models will be used to classify, segment, and cluster signals and to detect anomalies and communities for industrial vehicle usage. |
Research objectives and methods |
Description: |
Required skills |
The PhD candidate is expected to - be proficient in Python and SQL programming- have a deep knowledge of statistics and probability fundamentals- have a solid background in data preprocessing and mining techniques- know the most established machine learning and deep learning techniques- have natural inclination for teamwork- be proficient in English speaking, reading, and writingWe seek motivated students who are willing to work at the intersection between academia and industry. |
Designing a cloud-based heterogeneous prototyping platform for the development of fog computing apps |
|
Proposer |
Gianvito Urgese |
Topics |
Parallel and distributed systems, Quantum computing, Computer architectures and Computer aided design, Software engineering and Mobile computing |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
The PhD project enables SW developers to prototype complex solutions on heterogeneous systems (CPU, GPU, FPGA, Neuromorphic HW) effectively. |
Research objectives and methods |
Research objectives |
Required skills |
MS degree in computer engineering, electronics engineering. |
Designing a Development Framework for Engineering Edge-Based AIoT Sensor Solutions |
|
Proposer |
Gianvito Urgese |
Topics |
Data science, Computer vision and AI, Life sciences, Parallel and distributed systems, Quantum computing |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
The transition to digitalization, driven by the Industry 4.0 paradigm, requires advanced frameworks and tools to effectively integrate System of Systems (SoS) within industrial use case scenarios. |
Research objectives and methods |
Research objectives |
Required skills |
MS degree in computer engineering, electronics engineering or physics of complex systems. |
Computational Intelligence for Computer-Aided Design |
|
Proposer |
Giovanni Squillero |
Topics |
Computer architectures and Computer aided design, Data science, Computer vision and AI |
Group website |
https://cad.polito.it |
Summary of the proposal |
The proposal focuses on the use and development of "Intelligent" algorithms specifically tweaked on the need and peculiarities of CAD industries. Generic techniques ascribable to Computational Intelligence have long been used in the CAD field: probabilistic methods for the analysis of failures or the classification of processes; bio-inspired algorithms for the generation of tests and optimization of parameters or the definition of surrogate models. |
Research objectives and methods |
The recent fortune of the term "Machine Learning" renewed the interests in many automatic processes; moreover, the publicized successes of (deep) neural networks smoothed down the bias against other non-explicable black-box approaches, such as Evolutionary Algorithms, or the use of complex kernels in linear models. The goal of the research is twofold: from an academic point of view, tweaking existing methodologies, as well as developing new ones, specifically able to tackle CAD problems; from an industrial point of view, creating a highly qualified expert able to bring the scientific know-how into a company, while being also able to understand the practical needs, such as how data are selected and possibly collected. The need to team the experts from industry with more mathematically minded researchers is apparent: frequently a great knowledge of the practicalities is not accompanied by an adequate understanding of the statistical models used for analysis and predictions. In the first year, the research will consider techniques less able to process large amount of information, but perhaps more able to exploit all problem-specific knowledge available. It will almost certainly include bio-inspired techniques for generating, optimizing, minimizing test programs; statistical methods for analyzing and predicting the outcome of industrial processes (e.g., predicting the maximum operating frequency of a programmable unit based on the frequencies measured by some ring oscillators; detecting dangerous elements in a circuit; predicting catastrophic events). The activity is also like to exploit (deep) neural networks, however developing novel, creative results in this area is not a priority. On the contrary, the research shall face problems related to dimensionality reduction, feature extraction and prototypes identification/creation. Then the research shall also focus on the study of surrogate measures, that is, the use of measures that can be easily and inexpensively gathered as a proxy for others, more industrially relevant but expensive. In this regard, the tutors are working with a semiconductor manufacturer for using in-situ sensors values as a proxy for the prediction of operating frequency. The work could then proceed by tackling problems related to "dimensionality reduction", useful to limit the number of input data of the model, and "feature selection", essential when each single feature is the result of a costly measurement. At the same time, the research is likely to help the introduction of more advanced optimization techniques in everyday tasks. From a practical standpoint, starting in the second year, the activity would continue by analyzing a current practical need, namely: "predictive maintenance". A significant amount of data is currently collected by many industries, although in a rather disorganized way. The student would start by analyzing the practical problems of data collection, storage, and transmission, while, at the same time, practicing with the principles of data profiling, classification, and regression (all topics that are currently considered part of "machine learning"). The analysis of sequences to predict the final event, or rather identify a trigger, is an open research topic, with implications far beyond CAD. Unfortunately, unlikely popular ML scenarios, the availability of data is a significant limitation, a situation where the amount of available data for training is insufficient and is sometimes labeled "small data". Expected target publications: Top journals with impact factors * ASOC -- Applied Soft Computing Top conferences * ITC -- International Test Conference Notes: * The CAD Group has a long record of successful applications of intelligent systems in several different domains. For the specific activities, the list of possibly involved companies include: SPEA, Infineon (through the Ph.D. student Niccol? Bellarmino), ST Microelectronics, Comau (through the Ph.D. student Eliana Giovannitti) |
Required skills |
Required skills: Proficiency in Python (including deep understanding of object-oriented principles and design patterns); Proficiency in using libraries such as NumPy and SciPy for data analysis and manipulation // Preferred: Knowledge of Electronic CAD |
Security of Software Networks |
|
Proposer |
Cataldo Basile |
Topics |
Cybersecurity, Parallel and distributed systems, Quantum computing |
Group website |
https://www.dauin.polito.it/research/research_groups/torsec_security_group |
Summary of the proposal |
The massive progress in software network complexity, flexibility, and manageability was only marginally used to increase the security of these networks: attacks may remain undiscovered for months, and human errors mainly cause them. |
Research objectives and methods |
Nowadays, attackers are always one or more steps behind the security defenders. When vulnerabilities are found, patches follow only days later, and anti-virus signature updates come after discovering new malware. Intrusion Prevention Systems provide simple reactions triggered by simplistic conditions often considered ineffective by large companies. Moreover, companies face risks of misconfiguration whenever security policies or network layouts need an update. Statistics are clear: attacks are discovered with unacceptable delays, and in most cases, attacks are caused by human errors. The solution is also clear: providing defensive tools with more intelligence and a higher level of automation. |
Required skills |
The candidate needs to have a solid background in cybersecurity (risk management), defensive controls (e.g., firewall technologies and VPNs), monitoring controls (e.g., IDS/IPS and threat intelligence) and incident response. Moreover, he should also possess a background in software network technologies (SDN, NFV, Kubernetes) and cloud computing. Having skills in formal modelling and logical systems is a plus. |
Emerging Topics in Evolutionary Computation: Diversity Promotion and Graph-GP |
|
Proposer |
Giovanni Squillero |
Topics |
Computer architectures and Computer aided design, Data science, Computer vision and AI |
Group website |
https://www.cad.polito.it/ |
Summary of the proposal |
Soft Computing, including evolutionary computation (EC), is currently experiencing a unique moment. While fewer scientific papers focus solely on EC, traditional EC techniques are frequently utilized in practical activities under different labels. The objective of this analysis is to examine both the new representations that scholars are currently exploring and the old, yet still pressing, problems that practitioners are facing. |
Research objectives and methods |
Although the classical approach to representing solutions in EC involves bit strings and expression trees, far more complex encodings have been recetly proposed. More specifically, graph-based representations have led to novel applications of EC in circuit design, cryptography, image analysis, and other fields. At the same time, divergence of character, or, more precisely, the lack of it, is widely recognized as the most impairing single problem in the field of EC. While divergence of character is a cornerstone of natural evolution, in EC all candidate solutions eventually crowd the very same areas in the search space, such a "lack of speciation" has been pointed out in the seminal work of Holland back in 1975. It is usually labeled with the oxymoron "premature convergence" to stress the tendency of an algorithm to convergence toward a point where it was not supposed to converge to in the first place. The research activity would tackle "diversity promotion", that is either "increasing" or "preserving" diversity in an EC population, both from a practical and theoretical point of view. It will also include the related problems of defining and measuring diversity. The research project shall include an extensive experimental study of existing diversity preservation methods across various global optimization problems. Open-source, general-purpose EA toolkits, inspyred and DEAP, will also be used to study the influence of various methodologies and modifications on the population dynamics. Solutions that do not require the analysis of the internal structure of the individual (e.g., Cellular EAs, Deterministic Crowding, Hierarchical Fair Competition, Island Models, or Segregation) shall be considered. This study should allow the development of a, possibly new, effective methodology, able to generalize and coalesce most of the cited techniques. During the first year, the candidate will take a course in Artificial Intelligence, and all Ph.D. courses of the educational path on Data Science. Additionally, the candidate is required to improve the knowledge of Python. Starting from the second year, the research activity shall include Turing-complete program generation. The candidate will work on an open-source Python project, currently under active development. The candidate will try to replicate the work of the first year on much more difficult genotype-level methodologies, such as Clearing, Diversifiers, Fitness Sharing, Restricted Tournament Selection, Sequential Niching, Standard Crowding, Tarpeian Method, and Two-level Diversity Selection. At some point, probably toward the end of the second year, the new methodologies will be integrated into the Grammatical Evolution framework developed at the Machine Learning Lab of University of Trieste ? GE allows a sharp distinction between phenotype, genotype and fitness, creating an unprecedented test bench (the research group is already collaborating with a group in UniTS on these topics, see "Multi-level diversity promotion strategies for Grammar-guided Genetic Programming" Applied Soft Computing, 2019). A remarkable goal of this research would be to link phenotype-level methodologies to genotype measures. Target Publications Journals with impact factors Top conferences - ACM GECCO - Genetic and Evolutionary Computation Conference Notes: The tutors regularly present tutorials on Diversity Preservation at top conferences in the field, such as GECCO, PPSN, and CEC. Additionally, they are involved in the organization of a workshops focused on graph-based representation for EA. Moreover, the research group is in contact with industries that actively consider exploiting evolutionary machine-learning for enhancing their biological models, for instance, KRD (Czech Republic), Teregroup (Italy), and BioVal Process (France). The research group has also a long record of successful applications of evolutionary algorithms in several different domains. For instance, the on-going collaboration with STMicroelectronics on test and validation of programmable devices, does exploit evolutionary algorithms and would benefit from the research. |
Required skills |
Proficiency in Python (including deep understanding of object-oriented principles and design patterns, and handling of parallelism); Preferred: Experience with metaheuristcs, Experience with optimization algorithms |
Advanced ICT solutions and AI-driven methodologies for Cultural Heritage resilience |
|
Proposer |
Edoardo Patti |
Topics |
Data science, Computer vision and AI, Software engineering and Mobile computing, Parallel and distributed systems, Quantum computing |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
This Ph.D. research leverages on cutting-edge technologies to preserve Cultural Heritage (e.g., monuments, historical sites, etc.) against natural disasters, climate change, and human-related threats. The interdisciplinary approach integrates ICT tools, Machine Learning, and Data Analytics to develop proactive strategies for risk assessment, monitoring, and preservation of cultural assets by addressing challenges through innovative solutions for sustainable conservation and resilience |
Research objectives and methods |
Recent crises and disasters have affected the European citizens' lives, livelihoods, and environment in unforeseen and unprecedented ways. They have transformed our very understanding of them by reshaping hitherto unchallenged notions of the ?local? and the ?global? and putting into question well-rehearsed conceptual distinctions of ?natural? and ?man-made? disasters. Modern and high-performance ICT solutions need to be deployed in order to prevent and mitigate the effects of disasters and climate change events by enabling critical thinking and framing a holistic approach for better understanding of catastrophic events. The objective of this Ph.D. proposal consists of the design and development of ICT-driven solutions to develop proactive strategies for risk assessment, monitoring, and preservation of Cultural Heritage. The candidate will adopt a comprehensive interdisciplinary approach, seamlessly integrating modern techniques rooted in IoT, Machine/Deep Learning, and Big Data paradigms within the realm of cultural heritage resilience. This approach transcends purely technical facets, encompassing social and cultural dimensions to provide a holistic understanding and effective solutions. During the three years of the Ph.D., the research activity will be divided into five phases:- Survey existing literature on modern Ai-driven ICT solutions and applications in software engineering and analyze methodologies and challenges Cultural Heritage resilience.- Design and develop a data-driven digital ecosystem - i.e., distributed IoT platform - for the collection and harmonization of heterogeneous data from the real world to enable on-top advanced visualization and analysis services (e.g., Digital Twins). A multidisciplinary approach ranging from IoT paradigms to the application of Machine/Deep Learning methodologies for Big Data analysis is required in order to allow the development of proactive strategies for risk assessment, monitoring, and preservation of Cultural Heritage.- Develop algorithms and strategies for a context-aware Cultural Heritage resilience by implementing prototype systems for evaluation and refinement.- Design and implement continuous improvement and fine-tuning strategies for the development of increasingly effective and high-performing prevention strategies.- Evaluate the effectiveness of the data-driven digital ecosystem and developed strategies through user studies and real-world projects. Possible international scientific journals and conference: - IEEE Transactions on Computational Social Systems- IEEE Transactions on Industrial Informatics- Journal on Computing and Cultural Heritage- Journal of Cultural Heritage- Engineering Applications of Artificial Intelligence,- Expert Systems with Applications,- IEEE CoStProgramming and Object-Oriented Programming (preferable in Python).- Knowledge of web application programming.- Knowledge of IoT paradigms.- Knowledge of Machine Learning and Deep Learning.- Knowledgeof frameworks to develop models based on Machine Learning and Deep Learning Model- internat. Conf.- IEEE SKIMA internat. Conf. |
Required skills |
Programming and Object-Oriented Programming (preferable in Python). |
Monitoring systems and techniques for precision agriculture |
|
Proposer |
Renato Ferrero |
Topics |
Data science, Computer vision and AI, Software engineering and Mobile computing |
Group website |
|
Summary of the proposal |
The most challenging current demand of the agricultural sector is the production of sufficient and safe food for a growing population without over-exploiting natural resources. This challenge is placed in a difficult context of unstable climate conditions, with competition for land, water, energy, and in an increasingly urbanized world. The research activity aims to increase the competitiveness of the agri-food system in terms of safety, quality, sustainability, and added value of food products. |
Research objectives and methods |
The research activity of the PhD candidate will investigate devices and techniques for monitoring the agricultural produce in a holistic vision, with the aim of limiting environmental pollution, preventing the misuse of pesticides and fertilizers, reducing water and energy request, and increasing net profit. |
Required skills |
As the research activity regards the design, development, and evaluation of digital technologies for the next generation agriculture in a holistic vision, the PhD candidate is required to own multidisciplinary skills: e.g., distributed computing, embedded systems, computer networks, security, computer graphics, programming, database management. |
Designing heterogeneous digital/neuromorphic fog computing systems and development framework |
|
Proposer |
Gianvito Urgese |
Topics |
Parallel and distributed systems, Quantum computing, Life sciences, Data science, Computer vision and AI |
Group website |
https://eda.polito.it/ |
Summary of the proposal |
The candidate will be involved in the development of:A Heterogeneous Prototyping Platform (HPP) for Spiking Neural Network (SNN) simulations and AI applications on digital/neuromorphic systems.A framework for end-to-end engineering of SNN simulations on neuromorphic devices.A SW library optimizing SNN on RISC-V-based edge devices. |
Research objectives and methods |
Research objectives |
Required skills |
MS degree in computer engineering, electronics engineering or physics of complex systems. |
Cloud at the edge: creating a seamless computing platform with opportunistic datacenters |
|
Proposer |
Fulvio Giovanni Ottavio Risso |
Topics |
Computer architectures and Computer aided design, Parallel and distributed systems, Quantum computing, Software engineering and Mobile computing |
Group website |
https://netgroup.polito.it Project website: https://liqo.io |
Summary of the proposal |
The idea is to aggregate the huge number of traditional computing/storage devices available in modern environments (such as desktop/laptop computers, embedded devices, etc.) into an opportunistic datacenter, hence transforming all the current devices into datacenter nodes. |
Research objectives and methods |
Cloud-native technologies are increasingly deployed at the edge of the network, usually through tiny datacenters made by a few servers that maintain the main characteristics (powerful CPUs, high-speed network) of the well-known cloud datacenters. However, most of current domestic environments and enterprises host a huge number of traditional computing/storage devices, such as desktop/laptop computers, embedded devices, and more, which run mostly underutilized. |
Required skills |
The ideal candidate has good knowledge and experience in computing architectures, cloud computing and networking. Availability for spending periods abroad would be preferred for a more profitable investigation of the research topic. |
AI-driven cybersecurity assessment for automotive |
|
Proposer |
Luca Cagliero |
Topics |
Data science, Computer vision and AI, Cybersecurity |
Group website |
https://www.dauin.polito.it/en/research/research_groups/dbdm_database_and_data_mining_group https://www.dauin.polito.it/research/research_groups/torsec_security_group https://www.drivesec.com/ |
Summary of the proposal |
This PhD proposal aims to investigate how to leverage Generative AI techniques for assessing the cybersecurity posture of vehicles and automotive infrastructures and evaluating the compliance with existing standards (e.g., ISO 21434). It will also propose innovative LLM-based approaches to retrieve, recommend, and generate penetration tests and vulnerability-related information. It will also study innovative methodologies based on Multimodal Learning and Retrieval Augmented Generation. |
Research objectives and methods |
Research objectives |
Required skills |
The PhD candidate is expected to |
Applications of Large Language Models in time-evolving scenarios |
|
Proposer |
Luca Cagliero |
Topics |
Data science, Computer vision and AI |
Group website |
https://dbdmg.polito.it/ https://smartdata.polito.it |
Summary of the proposal |
Large Language Models are Generative AI models pretrained on a huge mass of data. Since training examples are sampled at a fixed time interval, LLMs require specific interventions to deal with time-evolving scenarios. Furthermore, they are not designed to process timestamped data such as time series and temporal sequences. The PhD proposal aims to propose new LLM-based approaches to analyze textual and multimedia sources in time-evolving scenarios and to leverage LLMs in timestamped data mining. |
Research objectives and methods |
Context Research objectives Tentative work plan |
Required skills |
The PhD candidate is expected to |
Building Adaptive Embodied Agents in XR to Enhance Educational Activities |
|
Proposer |
Andrea Bottino |
Topics |
Computer graphics and Multimedia, Data science, Computer vision and AI |
Group website |
https://www.polito.it/cgvg |
Summary of the proposal |
This research explores the integration of Memory-Augmented Neural Networks (MANNs) in Embodied Conversational Agents (ECAs) to create interactive, personalized and engaging learning experiences in XR. Such ECAs can adapt to the learner's/learner group characteristics and progression to personalize education for more effective learning outcomes in both individual and collaborative learning. The challenge is to develop complex yet accessible ECAs for different educational environments. |
Research objectives and methods |
In the evolving landscape of AI and educational technology, the integration of MANNs and gamification in ECAs offers new opportunities to push the field of AI agents and create highly interactive, adaptive and engaging learning experiences in XR. The use of MANNs allows ECAs to store and recall previous interactions, resolving different problems of actual conversational agents and enabling unprecedented levels of personalized content and engagement. Ultimately, these ECAs can provide an enhanced learning experience that is both dynamic and responsive to learners' individual needs. In collaborative learning scenarios, these ECAs should be designed to act not just as facilitators but as active participants, encouraging group interaction and the overall learning process. The integration of these technologies offers the potential to explore new educational methodologies that align with the evolving digital competencies of today's learners. 1. Enhancing ECAs with MANNs: 2. Facilitate collaborative learning through ECAs: 4. Assessment: COLLABORATIONS |
Required skills |
The ideal candidate for this PhD project should possess the following skills and characteristics: |
Real-Time Generative AI for Enhanced Extended Reality |
|
Proposer |
Andrea Bottino |
Topics |
Computer graphics and Multimedia, Data science, Computer vision and AI |
Group website |
https://www.polito.it/cgvg |
Summary of the proposal |
The integration of generative AI (GenAI) in extended reality (XR) offers transformative potential for the creation of dynamic and immersive experiences in many fields. The project aims to develop optimized GenAI models for XR, with a focus on algorithms that efficiently generate realistic content within the computational limitations of XR hardware. |
Research objectives and methods |
In the rapidly evolving field of extended reality (XR), the integration of GenAI offers transformative opportunities. GenAI is at the forefront of creating realistic, dynamic and immersive XR experiences. Its ability to automatically generate complex data such as geometries, textures, animations and even emotional voice modulations has a significant impact on various sectors, including education, entertainment and professional training. However, the practical implementation of these advanced technologies in XR faces critical challenges. |
Required skills |
The ideal candidate should have a strong background in computer science and AI, with specific skills in generative algorithms and XR. Problem-solving abilities, creativity, and knowledge of model optimization for low-power devices are essential. Experience in GPU programming and immersive user interface development is also required. We also require good communication and collaboration skills and publication and scientific writing skills |
Transferable and efficient robot learning across tasks, environments, and embodiments |
|
Proposer |
Raffaello Camoriano |
Topics |
Data science, Computer vision and AI |
Group website |
http://vandal.polito.it/ |
Summary of the proposal |
The project's goal is the design of efficient methods for training, transfer, and inference of high-capacity models for embodied systems. Promising approaches include knowledge distillation, recent fine-tuning and approximation methods reducing the policy execution cost while retaining performance levels. Moreover, constraining model output space to low-dimensional manifold structures arising from the physics of the target problem also holds promise to improve policy efficiency and safety. |
Research objectives and methods |
Classical learning methods for robotic perception and control tend to target specific skills and embodiments, due to the difficulties in extracting transferable and actionable representations which are invariant to physical properties of the environment and of the robot. However, the performance of such specialized agents can be limited by low model capacity and training on relatively few examples. This can be particularly problematic when tackling complex and long-horizon tasks for which the cost of large-scale data collection on a single robot can be prohibitively high and the complexity of the policy to be learned might benefit from a more expressive function class (i.e., with a larger number of parameters). Conversely, recent high-capacity, highly flexible machine learning models, such as vision transformers and large multimodal models, proved their worth in less constrained domains such as computer vision and NLP. In such domains, pre-training on large and diverse datasets is possible due to web-scale data availability. This results in rich ?generalist? pre-trained models enabling model fine tuning and adaptation to specific target tasks with large savings in terms of target data collection and positive transfer to new tasks and visual appearances. A growing research line investigates the extension of high-capacity models to robotic tasks to enable complex skill learning across embodiments and modalities, thanks to the high flexibility of high-capacity architectures (e.g., GATO [1]). RobotCat [2] demonstrates how such models can be applied to solve complex robotic manipulation tasks with visually defined goals, while Open X Embodiment [3] demonstrates positive transfer for task goals specified in natural language. Octo further extends this concept by supporting multimodal goal definitions [4], while AutoRT [5] also supports multi-robot coordination. Large language models can also be employed to guide exploration and automate reward design for reinforcement learning [6]. However, these methods rely on very large numbers of parameters (i.e., in the order of billions), rendering model storage and real-time inference a challenge. This is a relevant roadblock when local execution on limited robotic hardware is required, as is often the case in open-world unstructured environments. Some of the most advanced multi-embodiment models (e.g., RT-2-X [3]) are so extensive that they cannot be stored locally and require communication with cloud environments to perform inference. Even more so when model fine-tuning or open-ended learning are required for tackling new tasks. Impractical computational and communication costs and catastrophic forgetting of previous tasks indeed represent a major challenge. The objective of this project is the development of efficient methods for training, transfer, and inference of generalist high-capacity models for embodied and robotic tasks. Several approaches will be investigated, including the use of knowledge distillation, recent fine-tuning methods which proved to reduce the cost of execution of robotic policies (i.e., RT-2-X) from quadratic to linear while retaining performance levels [7], and approximation methods to reduce the number of parameters while retaining approximation power [8]. Moreover, constraining model output space to low-dimensional manifolds arising from the physical constraints of the target problem also holds promise to improve policy efficiency and safety [9] [10]. Potential publication venues include major AI, ML, robotics, and computer vision venues (e.g., TRO, RAL, TPAMI, JMLR, ICRA, IROS, CoRL, NeurIPS, ICML, ICLR, etc.) Preliminary Main Activities Plan- M1-M4 Literature review on foundation models for robot learning- M3-M7 Empirical analysis of state-of-the-art methods for improving foundation model efficiency- M8-M15 Design and development of novel efficient methods focusing on robotic requirements and resource constraints- M16-M22 Experimental evaluation of the proposed methods- M23-M28 Development of novel methods incorporating output space constraints to enforce safety requirements while retaining efficiency and predictive capabilities- M28-M32 Experimental validation and dissemination of the results- M32-M36 Thesis writing References |
Required skills |
We seek candidates highly motivated to conduct methodological research in ML and robotics. |
Neural Network reliability assessment and hardening for safety-critical embedded systems |
|
Proposer |
Matteo Sonza Reorda |
Topics |
Computer architectures and Computer aided design, Data science, Computer vision and AI |
Group website |
https://cad.polito.it/ |
Summary of the proposal |
Neural Networks are increasingly used within embedded systems in many application domains, including cases where safety is crucial (e.g., automotive, space, robotics). Possible hardware faults affecting the underlying hardware (CPU, GPU, TCU) can severely impact the produced results. The goal of the proposed research activity is first to estimate the probability that critical failures are produced, and then to devise effective solutions for system hardening, playing mainly at the software level. |
Research objectives and methods |
NNs are increasingly adopted in the area of embedded systems, even for safety-critical applications (e.g., in the automotive, aerospace and robotics domains), where the probability of failures must be lower than well-defined (and extremely low) thresholds. This goal is particularly challenging, since the hw used to run the NN often corresponds to extremely advanced devices (e.g., GPUs, or dedicated AI accelerators), built with highly sophisticated (and hence less mature) semiconductor technologies. On the other side, NNs are known to own some intrinsic robustness, and can tolerate a given number of faults inside the hardware. Unfortunately, given the complexity of the NN algorithms and of the underlying architectures, an extensive analysis to understand which (and how many) faults are particularly critical is difficult to perform, at least when usual computational resources are available. The planned research activities aim first at exploring the effects of faults affecting the hardware of a GPU/AI accelerator supporting the NN execution. Experiments will study the effects of the considered faults on the results produced by the NN. This study will mainly be performed resorting to fault injection experiments. In order to keep the computational effort reasonable, different solutions will be considered, combining simulation- and emulation-based fault injection with multi-level one. The trade-off between the accuracy of the results and the required computational effort will also be evaluated. Based on the gathered results, hardening solutions acting on the hardware and/or the software will be devised, aimed at improving the resilience of the whole application with respect to faults, and thus matching the safety requirements of the target applications. The proposed plan of activities is organized in the following phases (for each phase, the indicative time span in months from the beginning of the PhD period is reported): Phases 2 to 4 will include dissemination activities, based on writing papers and presenting them at conferences (e.g., ETS, VTS, IOLTS, DATE). The most relevant proposed methods and results will be submitted for publication on the journals in the field, such as the IEEE Transactions on Computers, CAD, and VLSI, as well as Elsevier Microelectronics & Reliability. We also plan for a strong cooperation with the researchers of other universities and research centers working in the area, such as the University of Trento, the University of California at Irvine (US), the Federal University of Rio Grande do Sul (Brazil), NVIDIA. |
Required skills |
The candidate should own basic skills in |
Design of an integrated system for testing headlamp optical functionalities |
|
Proposer |
Bartolomeo Montrucchio |
Topics |
Computer graphics and Multimedia, Data science, Computer vision and AI |
Group website |
https://www.dauin.polito.it/it/la_ricerca/gruppi_di_ricerca/grains_graphics_and_intelligent_systems https://www.italdesign.it/services-electric-and-electronics/harness-and-lighting/ |
Summary of the proposal |
Automobile recent development are based on several sensors such as cameras, radars, and others. These sensors are used also for improving road illumination, both for human and autonomous drivers. |
Research objectives and methods |
Automobile evolution requires increasingly automatic systems for driving and traffic detection, for example of other cars, bicycles or other vehicles. Therefore, also lighting systems are in fast evolution. In particular future vehicles' headlamps will move towards several light sources independently driven, up to several thousand of different sources, each of them driven by means of a technology similar to the one used in digital micromirror projectors. The final purpose is to develop a headlamp able to move automatically the light on obstacles like pedestrian on bicycles suddenly appeared on the road. In order to find where to move the light all the sensors available in the car can be used, mainly cameras and radars. |
Required skills |
The ideal candidate should have an interest in optics, computer vision, and image processing. |
Machine unlearning |
|
Proposer |
Elena Maria Baralis |
Topics |
Data science, Computer vision and AI |
Group website |
https://dbdmg.polito.it |
Summary of the proposal |
Machine Unlearning is the task of selectively erasing or modifying previously acquired knowledge from machine learning models. This is particularly relevant nowadays due to the increasing concerns surrounding privacy (e.g. the Right To Be Forgotten required by GDPR) and copyright infringements, as highlighted by recent cases involving Large Language Models. The key goal of this proposal is to propose novel architectures, algorithms and evaluation metrics for Machine Unlearning. |
Research objectives and methods |
In recent years, the rapid advancement of machine learning models, particularly Large Language Models (LLMs), has raised significant concerns regarding privacy and intellectual property rights. The need for responsible AI practices has become increasingly evident, driven by legal frameworks such as the General Data Protection Regulation (GDPR) that mandates the Right To Be Forgotten. Additionally, high-profile cases involving LLMs have highlighted the need to address issues related to the unintentional retention of sensitive information and potential copyright infringements. The proposed research activity on Machine Unlearning (MU) aims to tackle these challenges by developing novel techniques to selectively erase or modify previously acquired knowledge from machine learning models. The primary objectives of this research are twofold: first, to explore the current state of the art in MU, and second, to propose innovative architectures, algorithms, and evaluation metrics to enhance the efficacy of the unlearning process. Through these goals, the aim is to contribute to the establishment of ethical and responsible AI practices, ensuring compliance with legal requirements and mitigating the risks associated with unintentional information retention by machine learning models. The workplan for this PhD is structured to comprehensively address the multifaceted challenges of MU. The research will focus on proposing novel architectures and algorithms that facilitate effective unlearning while preserving the model's overall performance. Given the current lack of definitive metrics for MU, part of the research efforts will be focused toward trying to identify more suitable and comprehensive metrics. The research activity progresses from foundational research to the practical implementation, validation and application of MU techniques. An outline of the possible research plan is as follows. - First year - Second year -Third year |
Required skills |
The candidate should have a strong computer and data science background, in particular for what concerns: |
Generative AI models for enhanced text-to-image synthesis |
|
Proposer |
Lia Morra |
Topics |
Computer graphics and Multimedia, Data science, Computer vision and AI |
Group website |
http://grains.polito.it - http://dbmg.polito.it |
Summary of the proposal |
This research proposal aims to overcome limitations in current generative text-to-image models. Despite advancements in visual fidelity, existing models struggle with precise control over generated images in response to detailed prompts. The candidate will research innovative strategies to improve spatial composition and alignment with user-defined specifications, including the application of neuro-symbolic AI to embed logical constraints and leveraging background ontological knowledge. |
Research objectives and methods |
While current generative text-to-image latent diffusion models have reached unprecedented results in terms of visual fidelity, there are still open issues to be addressed in exerting precise control over the generated images. On the one hand, generative models have difficulty in creating correct images when the textual prompt contains many details and often with object placement and spatial awareness. Recent text-to-image latent diffusion models have shown substantial improvements in prompt following, yet still struggle with the use of words such as ?left? or ?behind?. Increasing the size of the model has so far led to small improvements on these aspects, in the face of a significant increase in hardware requirements. Alternatively, other recent works have looked into improving captions at training time. Neither approach, so far, as successfully addressed spatial composition. One possible reason lies in the inherent limitations of the text embedding employed to condition the generation process, that fails to learn sufficiently detailed and disentangled representation; this issue would not be necessarily solved by increasing the amount or complexity of training data. One the other hand, there is an also an ongoing struggle in aligning the generated output with human values. Generative models may generate offending images, perpetuate societal biases and stereotypes embedded in the training data, or ?regurgitate? training samples potentially exposing the user to inadvertent copyright infringements. While vendors have generally responded by establishing safeguards for specific inputs or outputs, a more general, robust and reliable solution is called for. For instance, recent preliminary results have shown that neuro-symbolic AI techniques could be used, in toy datasets, to sample from an unconditioned model under user-defined logical constraints. Research objectives: Outline of the research plan: |
Required skills |
- Good knowledge of machine learning, deep learning, and generative models. |
Test, reliability, and safety of intelligent and dependable devices supporting sustainable mobility |
|
Proposer |
Riccardo Cantoro |
Topics |
Computer architectures and Computer aided design, Cybersecurity |
Group website |
https://cad.polito.it |
Summary of the proposal |
The research addresses the pressing need for dependable electronic systems in safety-critical domains, specifically focusing on sustainable mobility. The objective is to develop innovative hardware and software methodologies to qualify electronic systems against stringent reliability and safety requirements. The work will involve developing suitable hardening techniques on the hardware, software safety mechanisms, and a comprehensive assessment methodology supported by EDA partners. |
Research objectives and methods |
Research objectives The objectives of this research are summarized as follows:- Identify a suitable hardware platform for sustainable mobility applications with particular emphasis on RISC-V based systems.- Identify suitable software for mobility applications to be used as a representative benchmark for the qualification activities.- Assess dependability figures on the identified hardware/software infrastructure to identify critical parts of the design that require hardening.- Develop innovative hardening solutions to improve the reliability of critical areas in the design.- Focus on sustainable mobility as an emerging area of research with great potential for real-world impact.- Establish a comprehensive assessment methodology in collaboration with EDA partners. Outline of possible research plan First year: Second year: Third year: List of possible venues for publications Possible venues for publications could include:- IEEE Transactions on Computers - IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems- IEEE Transactions on Very Large Scale Integration (VLSI) Systems- International Conference on Computer-Aided Design (ICCAD)- International Test Conference (ITC)- IEEE European Test Symposium (ETS)- Design, Automation and Test in Europe Conference (DATE)- RISC-V Summit Projects |
Required skills |
Background in digital design and verification. |
Cybersecurity for a quantum world |
|
Proposer |
Antonio Lioy |
Topics |
Cybersecurity, Parallel and distributed systems, Quantum computing |
Group website |
https://security.polito.it/ |
Summary of the proposal |
Cybersecurity is typically based on cryptographic algorithms (e.g. RSA, ECDSA, ECDH) that are threatened by the advent of quantum computing. |
Research objectives and methods |
Hard security is typically based on mathematical cryptographic algorithms that support computation of symmetric and asymmetric encryption, key exchange, digital signature, and hash values. This research is part of the Horizon Europe QUBIP project (Quantum-oriented Update to Browsers and Infrastructures for the PQ Transition) https://quibip.eu/ |
Required skills |
REQUIRED SKILLS |
Bridging Human Expertise and Generative AI in Software Engineering |
|
Proposer |
Luca Ardito |
Topics |
Software engineering and Mobile computing |
Group website |
https://softeng.polito.it |
Summary of the proposal |
In collaboration with Vodafone Digital and the ZTO team in Network Operations, the PhD project aims to define a framework to generate code from functional requirements, fostering synergy between human developers and AI-based actors. The project will involve a systematization of metrics and methodologies to evaluate the correctness of the generated code and requirements performed by the generative AI components to increase the effectiveness and gain trust in the outputs of generative AI. |
Research objectives and methods |
Research objectives The main objectives of the PhD programme are the following:The identification of generative AI mechanisms that can aid in code generation from software requirements. The development and assessment of methods for the evaluation of the correctness and the dependability of the application of generative AI to code development;The conduction of formal experiments to evaluate how code generated by AI Compares to human-written code in both functional and non-functional terms. Outline of the research work plan Task 1: Preliminary evaluation of state-of-the-art solutions (M1-M3) The task involves a comprehensive assessment of current solutions in the domain. The objective is to evaluate existing methodologies, technologies, and frameworks relevant to the research context. This preliminary analysis will be conducted systematically by applying Kitchenham's guidelines for conducting Systematic Literature Reviews in the Software Engineering research field. The systematic literature review will also consider grey literature sources (i.e., non-peer-reviewed sources available on various internet sources) to cope with the high novelty of the generative AI research field. The systematic evaluation of the state of the art will be complemented with open and structured interviews with practitioners and developers to understand their main needs and most common practices. Task 2: Selection and Integration of Generative AIs for Code Generation (M4-M18) This task focuses on selecting, customizing, integrating, and training a Generative AI, specifically a Large Language or Foundation Model, to generate code from formal requirements. It includes understanding various use case and formal requirement languages and creating modules for translating natural language requirements into structured notations like Use Case Diagrams. The process involves preprocessing data -collecting, cleaning, and structuring use case language datasets- and training the AI to understand these scenarios. Ongoing evaluation and refinement of the AI are crucial for accuracy. The main goal is to develop a solution that translates use case specifications into high-quality code, with evaluations based on the development effort, error rates, and requirement alignment. The task will use Software Repository Mining (MSR) techniques for diverse dataset collection. The implementation phase of this research will follow the Agile Software Development practices, streamlining the entire software development lifecycle to assess the efficacy of AI-generated code against existing tools and manually written code for both front-end and back-end applications. Furthermore, the research is dedicated to pioneering methods for automatically generating synchronized documentation and unit testing. It will also investigate strategies for conducting code quality reviews, monitoring resource usage efficiently, and evaluating the software's business impact, thereby tailoring the development process to meet the demands of network operations. Task 3: Definition of assessment methods for Generative AI-based code development (M13-M24) The task focuses on defining robust methods to assess Generative AI-based code development. This entails the definition of structured procedures to assess the accuracy, reliability, and compliance with the requirements of the generated code. The goal is to establish a rigorous framework for ensuring the quality of code produced by Generative AI, thus advancing the state of the art in Generative AI code development. Task 4: Analysis of the non-functional implications of Generative AI-based code development (M22-36) The task focuses on a comprehensive analysis of the non-functional implications inherent to Generative AI-based code development. This includes scrutinizing factors such as scalability, performance, readability and maintainability of the generated code. The objective is to discern and mitigate any adverse effects of integrating Generative AI into the code development process. List of possible venues for publications The target for the PhD research includes a set of conferences in the general area of software engineering (ICSE, ESEM, EASE, ASE, ICSME) as well as in the specific area of testing (ICST, ISSTA). |
Required skills |
The main skills required by the candidate are the following: |
Explaining AI (XAI) models for spatio-temporal data |
|
Proposer |
Elena Maria Baralis |
Topics |
Data science, Computer vision and AI |
Group website |
https://dbdmg.polito.it |
Summary of the proposal |
Spatio-temporal data allow an effective representation of many interesting phenomena in application domains ranging from transportation to finance. Current state-of-the-art deep learning techniques (e.g., LM, CNN, RNN) provide black-box models, i.e., models that do not expose the motivations for their predictions. The main goal of this research activity is the study of methods to allow human-in-the-loop inspection of reasons behind classifier predictions for spatio-temporal data. |
Research objectives and methods |
Machine learning models are increasingly adopted to assist human experts in decision-making. Especially in critical tasks, understanding the reasons behind machine learning model predictions is essential for trusting the model itself. For example, experts can detect model wrong behaviors and actively work on model debugging and improvement. Unfortunately, most high-performance ML models lack interpretability. PhD years organization |
Required skills |
The candidate should have a strong computer and data science background, in particular for what concerns: |
Advanced data modeling and innovative data analytics solutions for complex application domains |
|
Proposer |
Silvia Anna Chiusano |
Topics |
Data science, Computer vision and AI |
Group website |
dbdmg.polito.it |
Summary of the proposal |
Data science projects entail the acquisition, modelling, integration, and analysis of big and heterogeneous data collections generated by a diversity of sources, to profile the different facets and issues of the considered application context. However, data analytics in many application domains is still a daunting task, because data collections are generally too big and heterogeneous to be processed through machine learning techniques currently available. Thus advanced data modeling and machine learning/artificial intelligence techniques needs to be devised to uneart meaningful insights and efficiently manage large volumes of data. |
Research objectives and methods |
The PhD student will work on the study, design and development of proper data models and novel solutions for the integration, storage, management and analysis of big volumes of heterogeneous data collections in complex application domains. The research activity involves multidisciplinary knowledge and skills including database, machine learning and artificial intelligence algorithms, and advanced programming. More in detail, the following challenges will be addressed during the PhD: |
Required skills |
The candidate should have good programming skills, and competencies in data modelling and techniques for data analysis. |
Functional Safety Techniques for Automotive oriented Systems-on-Chip |
|
Proposer |
Paolo Bernardi |
Topics |
Computer architectures and Computer aided design |
Group website |
|
Summary of the proposal |
The activities planned for this proposal include efforts toward Functional Safety Techniques for Automotive Systems-on-Chip (SoC): |
Research objectives and methods |
The phd student will pursue objectives in the broader research field of the Automotive Reliability and Testing. Techniques for grading and developing System-level Test (SLT) libraries The working plan for the PhD student is recalling the objectives drawn in the previous sections. The order is not fixed and may vary according to the advancement during the PhD program. |
Required skills |
C/C++, ASM, Simulation and Fault Simulation, VHDL, Firmware |
Human-aware robot behaviour learning for HRI |
|
Proposer |
Giuseppe Bruno Averta |
Topics |
Data science, Computer vision and AI, Controls and system engineering |
Group website |
vandal.polito.it |
Summary of the proposal |
Humans are naturally multi-task agents, with an innate capability to interact with objects and tools and plan complex sequences of actions to address a specific activity. Advanced robots, on the other side, are still far from such capabilities. The goal of this work is to investigate how to learn from humans the capability to quickly plan and execute complex procedures in unstructured scenarios and transfer such skills to intelligent robots for an effective human-robot cooperation. |
Research objectives and methods |
This proposed PhD thesis aims to explore the domain of learning human skills from egocentric videos and transferring them to robotic systems. The increasing integration of robots into various aspects of human life highlight the need of developing a more intuitive and adaptive approach to skill acquisition, able to learn complex skills and adapt to various scenarios, where traditional learning paradigms fail. Egocentric videos, captured from a first-person perspective, provide a rich and unique source of contextual information that can enhance the learning process. This research seeks to leverage this unique sensing approach to develop a framework for transferring acquired human skills to robots, enabling them to perform complex tasks in diverse environments. Major Objectives: Methodology: Significance: Keywords: Egocentric Videos, Skill Transfer, Robotics, Transfer Learning, Human-Robot Interaction. |
Required skills |
Outstanding passion and motivation for research. |