Artificial Intelligence and Digital Music
A New Advanced Training Program for the MUSIC4D Project

Preliminary training for the first two lessons has been completed. The final agenda for the remaining teaching units will be published in March.
Introduction
Starting on February 19, the MUSIC4D project launches an advanced training program dedicated to exploring the complex and compelling intersections between Artificial Intelligence, robotics, and musical creation. The course delivery, scheduled over the coming weeks in both synchronous and asynchronous formats, is designed to disseminate cutting-edge knowledge in generative AI and robotics applied to the field of music.
The training modules represent a core objective of the MUSIC4D project and a strategic commitment to qualifying its activities in alignment with the project’s stated goals.
Structured into five units for a total of 39 hours, the program has been designed to provide students, faculty members, and technical-administrative staff with the theoretical and practical competencies required to integrate emerging technologies into their creative and professional workflows.
The course is led by a team of experts in electronic music, robotics, and artificial intelligence: Professors Francesco Pupo, Riccardo Sarti, Sandro Mungianu, Caterina Perri, and Rashmi Chawla. Their guidance ensures an approach that combines academic rigor with hands-on application.
Structure and Objectives
The course is organized into five instructional units, each focusing on a specific area of expertise:
- Unit 1 – Foundations of AI, Intelligent Agents, and Robotics
The first week establishes the conceptual framework, examining intelligent agent architectures, reinforcement learning paradigms, and human–robot interaction (HRI) dynamics.
- Unit 2 – Generative Models and Prompt Engineering
The second unit focuses on the practice of prompt engineering, enabling participants to effectively guide content generation (text, images, and code) through Large Language Models (LLMs) and diffusion-based models. - Unit 3 – Audio Signal Fundamentals and Analysis
This asynchronous module delves into sound processing, addressing sampling theory, spectral analysis (FFT), filtering techniques, audio effects, and principal sound synthesis methodologies.
- Unit 4 – Digital Music and Artificial Intelligence
The fourth unit explores practical applications, ranging from AI-assisted composition and virtual orchestration to audio–video synchronization and the use of innovative tools such as “TuttiBot” for automated assessment. - Unit 5 – Final Integration (Project Work)
The program culminates in an intensive laboratory session in which participants design and develop a functional prototype of an “intelligent musical system,” supported by direct mentoring.
Methodology and Delivery Platform
The instructional model combines synchronous lectures with asynchronous MOOC modules, enabling flexible and adaptive learning pathways. The management and delivery of the entire program are hosted on the University of Calabria’s MOOC platform — a strategic choice that reinforces the university’s role as a key technological partner within the project.
UNICAL’s MOOC (Massive Open Online Courses) platform is a robust and scalable digital infrastructure specifically designed for large-scale course delivery. It provides all functionalities required to support a hybrid instructional model, including on-demand video lectures, discussion forums, downloadable teaching materials, and integrated assessment tools.
This solution ensures maximum flexibility and accessibility for all participants.
The course is structured as a comprehensive program designed to provide advanced technical competencies, foster critical reflection on the role of AI in creative processes, and equip music professionals with a new and powerful expressive framework.
Course Program: Artificial Intelligence, Robotics and Music
| # | Lesson Title | Video (h) | Study (h) | Total | Lecturer | Date |
|---|---|---|---|---|---|---|
| 1 | Introduction to Generative AI | 1 | 2 | 3 | Francesco Pupo | 19-feb |
| 2 | Intelligent Agent Architectures and Multi-Agent Interaction | 1 | 2 | 3 | Francesco Pupo | 20-feb |
| 3 | Generative Architectures and Transformers | 1 | 2 | 3 | Francesco Pupo | 24-mar |
| 4 | Reinforcement Learning | 1 | 2 | 3 | Francesco Pupo | 27-mar |
| 5 | Introduction to Signals and the Electroacoustic Chain | 1 | 2 | 3 | Riccardo Sarti | 30-mar |
| 6 | Sampling and Quantization | 1 | 2 | 3 | Riccardo Sarti | 31-mar |
| 7 | Frequency Analysis and Domain Transforms | 1 | 2 | 3 | Bimbi | 01-apr |
| 8 | STFT, Windowing, and Audio Coding | 1 | 2 | 3 | Bimbi | 13-apr |
| 9 | Digital Filters | 1 | 2 | 3 | Bimbi | 14-apr |
| 10 | First and Second-Order Filters | 1 | 2 | 3 | Sarti | 15-apr |
| 11 | Advanced Filters | 1 | 2 | 3 | Sarti | 16-apr |
| 12 | Audio Effects and Dynamics Processors | 1 | 2 | 3 | Lacamera | 17-apr |
| 13 | Spectral and Modulation Synthesis | 1 | 2 | 3 | Bimbi | 20-apr |
| 14 | Physical Modeling Synthesis and the Karplus–Strong Algorithm | 1 | 2 | 3 | Bimbi | 21-apr |
| 15 | Prompt Engineering: Foundations and Advanced Techniques | 1 | 2 | 3 | Perri – Pupo | 28-apr |
| 16 | Computer-Aided Composition and Musical AI | 1 | 2 | 3 | Mungianu | 29-apr |
| 17 | AI as a Support for the Compositional Process | 1 | 2 | 3 | Mungianu | 30-apr |
| 18 | Virtual Orchestration | 1 | 2 | 3 | Mungianu | 04-mag |
| 19 | Frame-by-Frame Method for Music and Images | 1 | 2 | 3 | Mungianu | 05-mag |
| 20 | Audio-Video Workflow in Professional DAWs | 1 | 2 | 3 | Mungianu | 06-mag |
| 21 | “TuttiBot” – the grading tool | 1 | 1 | 2 | Rashmi Chawla | 07-mag |
| 22 | Human-Robot Interaction (HRI) | 1 | 1 | 2 | Rashmi Chawla | 08-mag |
| 23 | System Concept and Design | 2 | 0 | 2 | Pupo – Sarti – Mungianu | 18-mag |
| 24 | AI Core Development and Audio Integration | 3 | 0 | 3 | Pupo – Sarti – Mungianu | 19-mag |
| 25 | Refinement and Final Rendering | 1 | 0 | 1 | Pupo – Sarti – Mungianu | 20-mag |
| 26 | Presentation and Review | 1 | 0 | 1 | Pupo – Sarti – Mungianu | 21-mag |