{"id":1382,"date":"2024-01-01T20:34:56","date_gmt":"2024-01-01T19:34:56","guid":{"rendered":"http:\/\/pages.di.unipi.it\/bacciu\/?page_id=1382"},"modified":"2024-01-02T11:19:07","modified_gmt":"2024-01-02T10:19:07","slug":"conferences-workshops","status":"publish","type":"page","link":"https:\/\/pages.di.unipi.it\/bacciu\/publications\/conferences-workshops\/","title":{"rendered":"Conferences &amp; Workshops"},"content":{"rendered":"\n<p><code><div class=\"teachpress_pub_list\"><form name=\"tppublistform\" method=\"get\"><a name=\"tppubs\" id=\"tppubs\"><\/a><\/form><div class=\"tablenav\"><div class=\"tablenav-pages\"><span class=\"displaying-num\">149 entries<\/span> <a class=\"page-numbers button disabled\">&laquo;<\/a> <a class=\"page-numbers button disabled\">&lsaquo;<\/a> 1 of 3 <a href=\"https:\/\/pages.di.unipi.it\/bacciu\/publications\/conferences-workshops\/?limit=2&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"next page\" class=\"page-numbers button\">&rsaquo;<\/a> <a href=\"https:\/\/pages.di.unipi.it\/bacciu\/publications\/conferences-workshops\/?limit=3&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"last page\" class=\"page-numbers button\">&raquo;<\/a> <\/div><\/div><div class=\"teachpress_publication_list\"><h3 class=\"tp_h3\" id=\"tp_h3_2024\">2024<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">1.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Gravina, Alessio;  Zambon, Daniele;  Bacciu, Davide;  Alippi, Cesare<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('261','tp_links')\" style=\"cursor:pointer;\">Temporal Graph ODEs for Irregularly-Sampled Time Series<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI 2024), <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_261\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('261','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_261\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('261','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_261\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('261','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_261\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{nokey,<br \/>\r\ntitle = {Temporal Graph ODEs for Irregularly-Sampled Time Series},<br \/>\r\nauthor = {Alessio Gravina and Daniele Zambon and Davide Bacciu and Cesare Alippi},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2404.19508, Arxiv},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-08-09},<br \/>\r\nurldate = {2024-08-09},<br \/>\r\nbooktitle = {Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI 2024)},<br \/>\r\nabstract = {Modern graph representation learning works mostly under the assumption of dealing with regularly sampled temporal graph snapshots, which is far from realistic, e.g., social networks and physical systems are characterized by continuous dynamics and sporadic observations. To address this limitation, we introduce the Temporal Graph Ordinary Differential Equation (TG-ODE) framework, which learns both the temporal and spatial dynamics from graph streams where the intervals between observations are not regularly spaced. We empirically validate the proposed approach on several graph benchmarks, showing that TG-ODE can achieve state-of-the-art performance in irregular graph stream tasks.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('261','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_261\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Modern graph representation learning works mostly under the assumption of dealing with regularly sampled temporal graph snapshots, which is far from realistic, e.g., social networks and physical systems are characterized by continuous dynamics and sporadic observations. To address this limitation, we introduce the Temporal Graph Ordinary Differential Equation (TG-ODE) framework, which learns both the temporal and spatial dynamics from graph streams where the intervals between observations are not regularly spaced. We empirically validate the proposed approach on several graph benchmarks, showing that TG-ODE can achieve state-of-the-art performance in irregular graph stream tasks.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('261','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_261\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2404.19508\" title=\"Arxiv\" target=\"_blank\">Arxiv<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('261','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Temporal Graph ODEs for Irregularly-Sampled Time Series\" src=\"https:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/ijcai.png\" width=\"80\" alt=\"Temporal Graph ODEs for Irregularly-Sampled Time Series\" \/><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_number\">2.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Massidda, Riccardo;  Magliacane, Sara;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('260','tp_links')\" style=\"cursor:pointer;\">Learning Causal Abstractions of Linear Structural Causal Models<\/a> <span class=\"tp_pub_type inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">The 40th Conference on Uncertainty in Artificial Intelligence, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_resource_link\"><a id=\"tp_links_sh_260\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('260','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_260\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('260','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_260\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{massidda2024learning,<br \/>\r\ntitle = {Learning Causal Abstractions of Linear Structural Causal Models},<br \/>\r\nauthor = {Riccardo Massidda and Sara Magliacane and Davide Bacciu},<br \/>\r\nurl = {https:\/\/openreview.net\/forum?id=XlFqI9TMhf},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-07-31},<br \/>\r\nurldate = {2024-07-31},<br \/>\r\nbooktitle = {The 40th Conference on Uncertainty in Artificial Intelligence},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('260','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_260\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/openreview.net\/forum?id=XlFqI9TMhf\" title=\"https:\/\/openreview.net\/forum?id=XlFqI9TMhf\" target=\"_blank\">https:\/\/openreview.net\/forum?id=XlFqI9TMhf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('260','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Learning Causal Abstractions of Linear Structural Causal Models\" src=\"https:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/08\/uai.png\" width=\"80\" alt=\"Learning Causal Abstractions of Linear Structural Causal Models\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">3.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Gravina, Alessio;  Lovisotto, Giulia;  Gallicchio, Claudio;  Bacciu, Davide;  Grohnfeldt, Claas<\/p><p class=\"tp_pub_title\">Long Range Propagation on Continuous-Time Dynamic Graphs <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the International Conference on Machine Learning (ICML 2024), <\/span><span class=\"tp_pub_additional_publisher\">PMLR, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_258\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('258','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_258\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('258','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_258\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{nokey,<br \/>\r\ntitle = {Long Range Propagation on Continuous-Time Dynamic Graphs},<br \/>\r\nauthor = {Alessio Gravina and Giulia Lovisotto and Claudio Gallicchio and Davide Bacciu and Claas Grohnfeldt},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-07-24},<br \/>\r\nurldate = {2024-07-24},<br \/>\r\nbooktitle = {Proceedings of the International Conference on Machine Learning (ICML 2024)},<br \/>\r\npublisher = {PMLR},<br \/>\r\nabstract = {Learning Continuous-Time Dynamic Graphs (C-TDGs) requires accurately modeling spatio-temporal information on streams of irregularly sampled events. While many methods have been proposed recently, we find that most message passing-, recurrent- or self-attention-based methods perform poorly on long-range tasks. These tasks require correlating information that occurred \"far\" away from the current event, either spatially (higher-order node information) or along the time dimension (events occurred in the past). To address long-range dependencies, we introduce Continuous-Time Graph Anti-Symmetric Network (CTAN). Grounded within the ordinary differential equations framework, our method is designed for efficient propagation of information. In this paper, we show how CTAN's (i) long-range modeling capabilities are substantiated by theoretical findings and how (ii) its empirical performance on synthetic long-range benchmarks and real-world benchmarks is superior to other methods. Our results motivate CTAN's ability to propagate long-range information in C-TDGs as well as the inclusion of long-range tasks as part of temporal graph models evaluation.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('258','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_258\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Learning Continuous-Time Dynamic Graphs (C-TDGs) requires accurately modeling spatio-temporal information on streams of irregularly sampled events. While many methods have been proposed recently, we find that most message passing-, recurrent- or self-attention-based methods perform poorly on long-range tasks. These tasks require correlating information that occurred &quot;far&quot; away from the current event, either spatially (higher-order node information) or along the time dimension (events occurred in the past). To address long-range dependencies, we introduce Continuous-Time Graph Anti-Symmetric Network (CTAN). Grounded within the ordinary differential equations framework, our method is designed for efficient propagation of information. In this paper, we show how CTAN's (i) long-range modeling capabilities are substantiated by theoretical findings and how (ii) its empirical performance on synthetic long-range benchmarks and real-world benchmarks is superior to other methods. Our results motivate CTAN's ability to propagate long-range information in C-TDGs as well as the inclusion of long-range tasks as part of temporal graph models evaluation.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('258','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Long Range Propagation on Continuous-Time Dynamic Graphs\" src=\"https:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/icml.png\" width=\"80\" alt=\"Long Range Propagation on Continuous-Time Dynamic Graphs\" \/><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_number\">4.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Bacciu, Davide;  Landolfi, Francesco<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('269','tp_links')\" style=\"cursor:pointer;\">Generalizing Convolution to Point Clouds<\/a> <span class=\"tp_pub_type inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">ICML 2024 Workshop on Differentiable Almost Everything: Differentiable Relaxations, Algorithms, Operators, and Simulators, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_269\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('269','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_269\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('269','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_269\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('269','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_269\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{bacciu2024generalizing,<br \/>\r\ntitle = {Generalizing Convolution to Point Clouds},<br \/>\r\nauthor = {Davide Bacciu and Francesco Landolfi},<br \/>\r\nurl = {https:\/\/openreview.net\/forum?id=TXwDtUmiaj},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-07-23},<br \/>\r\nurldate = {2024-01-01},<br \/>\r\nbooktitle = {ICML 2024 Workshop on Differentiable Almost Everything: Differentiable Relaxations, Algorithms, Operators, and Simulators},<br \/>\r\nabstract = {Convolution, a fundamental operation in deep learning for structured grid data like images, cannot be directly applied to point clouds due to their irregular and unordered nature. Many approaches in literature that perform convolution on point clouds achieve this by designing a convolutional operator from scratch, often with little resemblance to the one used on images. We present two point cloud convolutions that naturally follow from the convolution in its standard definition popular with images. We do so by relaxing the indexing of the kernel weights with a \"soft\" dictionary that resembles the attention mechanism of the transformers. Finally, experimental results demonstrate the effectiveness of the proposed relaxations on two benchmark point cloud classification tasks.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('269','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_269\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Convolution, a fundamental operation in deep learning for structured grid data like images, cannot be directly applied to point clouds due to their irregular and unordered nature. Many approaches in literature that perform convolution on point clouds achieve this by designing a convolutional operator from scratch, often with little resemblance to the one used on images. We present two point cloud convolutions that naturally follow from the convolution in its standard definition popular with images. We do so by relaxing the indexing of the kernel weights with a &quot;soft&quot; dictionary that resembles the attention mechanism of the transformers. Finally, experimental results demonstrate the effectiveness of the proposed relaxations on two benchmark point cloud classification tasks.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('269','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_269\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/openreview.net\/forum?id=TXwDtUmiaj\" title=\"https:\/\/openreview.net\/forum?id=TXwDtUmiaj\" target=\"_blank\">https:\/\/openreview.net\/forum?id=TXwDtUmiaj<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('269','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Generalizing Convolution to Point Clouds\" src=\"https:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/icml.png\" width=\"80\" alt=\"Generalizing Convolution to Point Clouds\" \/><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_number\">5.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Trenta, Alessandro;  Bacciu, Davide;  Cossu, Andrea;  Ferrero, Pietro<\/p><p class=\"tp_pub_title\">MultiSTOP: Solving Functional Equations with Reinforcement Learning <span class=\"tp_pub_type inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">ICLR 2024 Workshop on AI4DifferentialEquations In Science, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_262\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('262','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_262\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('262','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_262\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{trenta2024multistop,<br \/>\r\ntitle = {MultiSTOP: Solving Functional Equations with Reinforcement Learning},<br \/>\r\nauthor = {Alessandro Trenta and Davide Bacciu and Andrea Cossu and Pietro Ferrero},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-05-11},<br \/>\r\nurldate = {2024-05-11},<br \/>\r\nbooktitle = {ICLR 2024 Workshop on AI4DifferentialEquations In Science},<br \/>\r\nabstract = {We develop MultiSTOP, a Reinforcement Learning framework for solving functional equations in physics. This new methodology produces actual numerical solutions instead of bounds on them. We extend the original BootSTOP algorithm by adding multiple constraints derived from domain-specific knowledge, even in integral form, to improve the accuracy of the solution. We investigate a particular equation in a one-dimensional Conformal Field Theory.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('262','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_262\" style=\"display:none;\"><div class=\"tp_abstract_entry\">We develop MultiSTOP, a Reinforcement Learning framework for solving functional equations in physics. This new methodology produces actual numerical solutions instead of bounds on them. We extend the original BootSTOP algorithm by adding multiple constraints derived from domain-specific knowledge, even in integral form, to improve the accuracy of the solution. We investigate a particular equation in a one-dimensional Conformal Field Theory.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('262','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"MultiSTOP: Solving Functional Equations with Reinforcement Learning\" src=\"https:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/iclr.png\" width=\"80\" alt=\"MultiSTOP: Solving Functional Equations with Reinforcement Learning\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">6.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Massidda, Martina Cinquini Francesco Landolfi Riccardo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('254','tp_links')\" style=\"cursor:pointer;\">Constraint-Free Structure Learning with Smooth Acyclic Orientations<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">The Twelfth International Conference on Learning Representations, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_254\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('254','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_254\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('254','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_254\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('254','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_254\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{cosmo2024,<br \/>\r\ntitle = {Constraint-Free Structure Learning with Smooth Acyclic Orientations},<br \/>\r\nauthor = {Martina Cinquini Francesco Landolfi Riccardo Massidda},<br \/>\r\nurl = {https:\/\/openreview.net\/forum?id=KWO8LSUC5W},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-05-06},<br \/>\r\nurldate = {2024-01-01},<br \/>\r\nbooktitle = {The Twelfth International Conference on Learning Representations},<br \/>\r\nabstract = {The structure learning problem consists of fitting data generated by a Directed Acyclic Graph (DAG) to correctly reconstruct its arcs. In this context, differentiable approaches constrain or regularize an optimization problem with a continuous relaxation of the acyclicity property. The computational cost of evaluating graph acyclicity is cubic on the number of nodes and significantly affects scalability. In this paper, we introduce COSMO, a constraint-free continuous optimization scheme for acyclic structure learning. At the core of our method lies a novel differentiable approximation of an orientation matrix parameterized by a single priority vector. Differently from previous works, our parameterization fits a smooth orientation matrix and the resulting acyclic adjacency matrix without evaluating acyclicity at any step. Despite this absence, we prove that COSMO always converges to an acyclic solution. In addition to being asymptotically faster, our empirical analysis highlights how COSMO performance on graph reconstruction compares favorably with competing structure learning methods.<br \/>\r\n},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('254','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_254\" style=\"display:none;\"><div class=\"tp_abstract_entry\">The structure learning problem consists of fitting data generated by a Directed Acyclic Graph (DAG) to correctly reconstruct its arcs. In this context, differentiable approaches constrain or regularize an optimization problem with a continuous relaxation of the acyclicity property. The computational cost of evaluating graph acyclicity is cubic on the number of nodes and significantly affects scalability. In this paper, we introduce COSMO, a constraint-free continuous optimization scheme for acyclic structure learning. At the core of our method lies a novel differentiable approximation of an orientation matrix parameterized by a single priority vector. Differently from previous works, our parameterization fits a smooth orientation matrix and the resulting acyclic adjacency matrix without evaluating acyclicity at any step. Despite this absence, we prove that COSMO always converges to an acyclic solution. In addition to being asymptotically faster, our empirical analysis highlights how COSMO performance on graph reconstruction compares favorably with competing structure learning methods.<br \/>\r\n<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('254','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_254\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/openreview.net\/forum?id=KWO8LSUC5W\" title=\"https:\/\/openreview.net\/forum?id=KWO8LSUC5W\" target=\"_blank\">https:\/\/openreview.net\/forum?id=KWO8LSUC5W<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('254','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Constraint-Free Structure Learning with Smooth Acyclic Orientations\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/iclr.png\" width=\"80\" alt=\"Constraint-Free Structure Learning with Smooth Acyclic Orientations\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">7.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Pasquali, Alex;  Lomonaco, Vincenzo;  Bacciu, Davide;  Paganelli, Federica<\/p><p class=\"tp_pub_title\">Deep Reinforcement Learning for Network Slice Placement and the DeepNetSlice Toolkit <span class=\"tp_pub_type conference\">Conference<\/span> <span class=\"tp_pub_label_status forthcoming\">Forthcoming<\/span><\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the IEEE International Conference on Machine Learning for Communication and Networking 2024 (IEEE ICMLCN 2024), <\/span><span class=\"tp_pub_additional_publisher\">IEEE, <\/span>Forthcoming.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_251\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('251','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_251\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{nokey,<br \/>\r\ntitle = {Deep Reinforcement Learning for Network Slice Placement and the DeepNetSlice Toolkit},<br \/>\r\nauthor = {Alex Pasquali and Vincenzo Lomonaco and Davide Bacciu and Federica Paganelli},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-05-05},<br \/>\r\nurldate = {2024-05-05},<br \/>\r\nbooktitle = {Proceedings of the IEEE International Conference on Machine Learning for Communication and Networking 2024 (IEEE ICMLCN 2024)},<br \/>\r\npublisher = {IEEE},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {forthcoming},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('251','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Deep Reinforcement Learning for Network Slice Placement and the DeepNetSlice Toolkit\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/ICMLCN-Logo-164x70-1.png\" width=\"80\" alt=\"Deep Reinforcement Learning for Network Slice Placement and the DeepNetSlice Toolkit\" \/><\/div><\/div><div class=\"tp_publication tp_publication_workshop\"><div class=\"tp_pub_number\">8.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Ninniri, Matteo;  Podda, Marco;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('250','tp_links')\" style=\"cursor:pointer;\">Classifier-free graph diffusion for molecular property targeting<\/a> <span class=\"tp_pub_type workshop\">Workshop<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">4th workshop on Graphs and more Complex structures for Learning and Reasoning (GCLR) at AAAI 2024, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_250\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('250','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_250\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('250','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_250\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('250','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_250\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@workshop{Ninniri2024,<br \/>\r\ntitle = {Classifier-free graph diffusion for molecular property targeting},<br \/>\r\nauthor = {Matteo Ninniri and Marco Podda and Davide Bacciu},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2312.17397, Arxiv},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-02-27},<br \/>\r\nbooktitle = {4th workshop on Graphs and more Complex structures for Learning and Reasoning (GCLR) at AAAI 2024},<br \/>\r\nabstract = {This work focuses on the task of property targeting: that is, generating molecules conditioned on target chemical properties to expedite candidate screening for novel drug and materials development. DiGress is a recent diffusion model for molecular graphs whose distinctive feature is allowing property targeting through classifier-based (CB) guidance. While CB guidance may work to generate molecular-like graphs, we hint at the fact that its assumptions apply poorly to the chemical domain. Based on this insight we propose a classifier-free DiGress (FreeGress), which works by directly injecting the conditioning information into the training process. CF guidance is convenient given its less stringent assumptions and since it does not require to train an auxiliary property regressor, thus halving the number of trainable parameters in the model. We empirically show that our model yields up to 79% improvement in Mean Absolute Error with respect to DiGress on property targeting tasks on QM9 and ZINC-250k benchmarks. As an additional contribution, we propose a simple yet powerful approach to improve chemical validity of generated samples, based on the observation that certain chemical properties such as molecular weight correlate with the number of atoms in molecules. },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {workshop}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('250','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_250\" style=\"display:none;\"><div class=\"tp_abstract_entry\">This work focuses on the task of property targeting: that is, generating molecules conditioned on target chemical properties to expedite candidate screening for novel drug and materials development. DiGress is a recent diffusion model for molecular graphs whose distinctive feature is allowing property targeting through classifier-based (CB) guidance. While CB guidance may work to generate molecular-like graphs, we hint at the fact that its assumptions apply poorly to the chemical domain. Based on this insight we propose a classifier-free DiGress (FreeGress), which works by directly injecting the conditioning information into the training process. CF guidance is convenient given its less stringent assumptions and since it does not require to train an auxiliary property regressor, thus halving the number of trainable parameters in the model. We empirically show that our model yields up to 79% improvement in Mean Absolute Error with respect to DiGress on property targeting tasks on QM9 and ZINC-250k benchmarks. As an additional contribution, we propose a simple yet powerful approach to improve chemical validity of generated samples, based on the observation that certain chemical properties such as molecular weight correlate with the number of atoms in molecules. <\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('250','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_250\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2312.17397\" title=\"Arxiv\" target=\"_blank\">Arxiv<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('250','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Classifier-free graph diffusion for molecular property targeting\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/aaai.jpeg\" width=\"80\" alt=\"Classifier-free graph diffusion for molecular property targeting\" \/><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_number\">9.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Simone, Lorenzo;  Bacciu, Davide;  Gervasi, Vincenzo<\/p><p class=\"tp_pub_title\">Quasi-Orthogonal ECG-Frank XYZ Transformation with\u00a0Energy-Based Models and\u00a0Clinical Text <span class=\"tp_pub_type inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span> Finkelstein, Joseph;  Moskovitch, Robert;  Parimbelli, Enea (Ed.): <span class=\"tp_pub_additional_booktitle\">Artificial Intelligence in Medicine, <\/span><span class=\"tp_pub_additional_pages\">pp. 249\u2013253, <\/span><span class=\"tp_pub_additional_publisher\">Springer Nature Switzerland, <\/span><span class=\"tp_pub_additional_address\">Cham, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>, <span class=\"tp_pub_additional_isbn\">ISBN: 978-3-031-66535-6<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_256\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('256','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_256\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('256','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_256\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{10.1007\/978-3-031-66535-6_27,<br \/>\r\ntitle = {Quasi-Orthogonal ECG-Frank XYZ Transformation with\u00a0Energy-Based Models and\u00a0Clinical Text},<br \/>\r\nauthor = {Lorenzo Simone and Davide Bacciu and Vincenzo Gervasi},<br \/>\r\neditor = {Joseph Finkelstein and Robert Moskovitch and Enea Parimbelli},<br \/>\r\nisbn = {978-3-031-66535-6},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\nbooktitle = {Artificial Intelligence in Medicine},<br \/>\r\npages = {249\u2013253},<br \/>\r\npublisher = {Springer Nature Switzerland},<br \/>\r\naddress = {Cham},<br \/>\r\nabstract = {The transformation of 12-Lead electrocardiograms to 3D vectorcardiograms, along with its reverse process, offer numerous advantages for computer visualization, signal transmission and analysis. Recent literature has shown increasing interest in this structured representation, due to its effectiveness in various cardiac evaluations and machine learning-based arrhythmia prediction. Current transformation techniques utilize fixed matrices, often retrieved through regression methods which fail to correlate with patient's physical characteristics or ongoing diseases. In this paper, we propose the first quasi-orthogonal transformation handling multi-modal input (12-lead ECG and clinical annotations) through a conditional energy-based model. Within our novel probabilistic formulation, the model proposes multiple transformation coefficients without relying on a single fixed approximation to better highlight relationships between latent factors and structured output. The evaluation of our approach, conducted with a nested cross validation on PTB Diagnostic dataset, showcased improved reconstruction precision across various cardiac conditions compared to state-of-the-art techniques (Kors, Dower, and QSLV).},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('256','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_256\" style=\"display:none;\"><div class=\"tp_abstract_entry\">The transformation of 12-Lead electrocardiograms to 3D vectorcardiograms, along with its reverse process, offer numerous advantages for computer visualization, signal transmission and analysis. Recent literature has shown increasing interest in this structured representation, due to its effectiveness in various cardiac evaluations and machine learning-based arrhythmia prediction. Current transformation techniques utilize fixed matrices, often retrieved through regression methods which fail to correlate with patient's physical characteristics or ongoing diseases. In this paper, we propose the first quasi-orthogonal transformation handling multi-modal input (12-lead ECG and clinical annotations) through a conditional energy-based model. Within our novel probabilistic formulation, the model proposes multiple transformation coefficients without relying on a single fixed approximation to better highlight relationships between latent factors and structured output. The evaluation of our approach, conducted with a nested cross validation on PTB Diagnostic dataset, showcased improved reconstruction precision across various cardiac conditions compared to state-of-the-art techniques (Kors, Dower, and QSLV).<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('256','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_number\">10.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Ceni, Andrea;  Cossu, Andrea;  St\u00f6lzle, Maximilian W;  Liu, Jingyue;  Santina, Cosimo Della;  Bacciu, Davide;  Gallicchio, Claudio<\/p><p class=\"tp_pub_title\">Random Oscillators Network for Time Series Processing <span class=\"tp_pub_type inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">International Conference on Artificial Intelligence and Statistics, <\/span><span class=\"tp_pub_additional_pages\">pp. 4807\u20134815, <\/span><span class=\"tp_pub_additional_organization\">PMLR <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_263\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('263','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_263\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('263','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_263\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{ceni2024random,<br \/>\r\ntitle = {Random Oscillators Network for Time Series Processing},<br \/>\r\nauthor = {Andrea Ceni and Andrea Cossu and Maximilian W St\u00f6lzle and Jingyue Liu and Cosimo Della Santina and Davide Bacciu and Claudio Gallicchio},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\nurldate = {2024-01-01},<br \/>\r\nbooktitle = {International Conference on Artificial Intelligence and Statistics},<br \/>\r\npages = {4807\u20134815},<br \/>\r\norganization = {PMLR},<br \/>\r\nabstract = {We introduce the Random Oscillators Network (RON), a physically-inspired recurrent model derived from a network of heterogeneous oscillators. Unlike traditional recurrent neural networks, RON keeps the connections between oscillators untrained by leveraging on smart random initialisations, leading to exceptional computational efficiency. A rigorous theoretical analysis finds the necessary and sufficient conditions for the stability of RON, highlighting the natural tendency of RON to lie at the edge of stability, a regime of configurations offering particularly powerful and expressive models. Through an extensive empirical evaluation on several benchmarks, we show four main advantages of RON. 1) RON shows excellent long-term memory and sequence classification ability, outperforming other randomised approaches. 2) RON outperforms fully-trained recurrent models and state-of-the-art randomised models in chaotic time series forecasting. 3) RON provides expressive internal representations even in a small parametrisation regime making it amenable to be deployed on low-powered devices and at the edge. 4) RON is up to two orders of magnitude faster than fully-trained models. },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('263','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_263\" style=\"display:none;\"><div class=\"tp_abstract_entry\">We introduce the Random Oscillators Network (RON), a physically-inspired recurrent model derived from a network of heterogeneous oscillators. Unlike traditional recurrent neural networks, RON keeps the connections between oscillators untrained by leveraging on smart random initialisations, leading to exceptional computational efficiency. A rigorous theoretical analysis finds the necessary and sufficient conditions for the stability of RON, highlighting the natural tendency of RON to lie at the edge of stability, a regime of configurations offering particularly powerful and expressive models. Through an extensive empirical evaluation on several benchmarks, we show four main advantages of RON. 1) RON shows excellent long-term memory and sequence classification ability, outperforming other randomised approaches. 2) RON outperforms fully-trained recurrent models and state-of-the-art randomised models in chaotic time series forecasting. 3) RON provides expressive internal representations even in a small parametrisation regime making it amenable to be deployed on low-powered devices and at the edge. 4) RON is up to two orders of magnitude faster than fully-trained models. <\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('263','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Random Oscillators Network for Time Series Processing\" src=\"https:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/aistats.jpg\" width=\"80\" alt=\"Random Oscillators Network for Time Series Processing\" \/><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_number\">11.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Li, Lanpei;  Piccoli, Elia;  Cossu, Andrea;  Bacciu, Davide;  Lomonaco, Vincenzo<\/p><p class=\"tp_pub_title\">Calibration of Continual Learning Models <span class=\"tp_pub_type inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, <\/span><span class=\"tp_pub_additional_pages\">pp. 4160\u20134169, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_265\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('265','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_265\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('265','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_265\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{li2024calibration,<br \/>\r\ntitle = {Calibration of Continual Learning Models},<br \/>\r\nauthor = {Lanpei Li and Elia Piccoli and Andrea Cossu and Davide Bacciu and Vincenzo Lomonaco},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\nurldate = {2024-01-01},<br \/>\r\nbooktitle = {Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition},<br \/>\r\npages = {4160\u20134169},<br \/>\r\nabstract = {Continual Learning (CL) focuses on maximizing the predictive performance of a model across a non-stationary stream of data. Unfortunately CL models tend to forget previous knowledge thus often underperforming when compared with an offline model trained jointly on the entire data stream. Given that any CL model will eventually make mistakes it is of crucial importance to build calibrated CL models: models that can reliably tell their confidence when making a prediction. Model calibration is an active research topic in machine learning yet to be properly investigated in CL. We provide the first empirical study of the behavior of calibration approaches in CL showing that CL strategies do not inherently learn calibrated models. To mitigate this issue we design a continual calibration approach that improves the performance of post-processing calibration methods over a wide range of different benchmarks and CL strategies. CL does not necessarily need perfect predictive models but rather it can benefit from reliable predictive models. We believe our study on continual calibration represents a first step towards this direction.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('265','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_265\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Continual Learning (CL) focuses on maximizing the predictive performance of a model across a non-stationary stream of data. Unfortunately CL models tend to forget previous knowledge thus often underperforming when compared with an offline model trained jointly on the entire data stream. Given that any CL model will eventually make mistakes it is of crucial importance to build calibrated CL models: models that can reliably tell their confidence when making a prediction. Model calibration is an active research topic in machine learning yet to be properly investigated in CL. We provide the first empirical study of the behavior of calibration approaches in CL showing that CL strategies do not inherently learn calibrated models. To mitigate this issue we design a continual calibration approach that improves the performance of post-processing calibration methods over a wide range of different benchmarks and CL strategies. CL does not necessarily need perfect predictive models but rather it can benefit from reliable predictive models. We believe our study on continual calibration represents a first step towards this direction.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('265','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Calibration of Continual Learning Models\" src=\"https:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/cvpr.jpg\" width=\"80\" alt=\"Calibration of Continual Learning Models\" \/><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_number\">12.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Resta, Michele;  Bacciu, Davide<\/p><p class=\"tp_pub_title\">Self-generated Replay Memories for Continual Neural Machine Translation <span class=\"tp_pub_type inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), <\/span><span class=\"tp_pub_additional_pages\">pp. 175\u2013191, <\/span><span class=\"tp_pub_additional_year\">2024<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_268\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('268','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_268\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('268','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_268\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{resta2024self,<br \/>\r\ntitle = {Self-generated Replay Memories for Continual Neural Machine Translation},<br \/>\r\nauthor = {Michele Resta and Davide Bacciu},<br \/>\r\nyear  = {2024},<br \/>\r\ndate = {2024-01-01},<br \/>\r\nurldate = {2024-01-01},<br \/>\r\nbooktitle = {Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)},<br \/>\r\npages = {175\u2013191},<br \/>\r\nabstract = {Modern Neural Machine Translation systems exhibit strong performance in several different languages and are constantly improving. Their ability to learn continuously is, however, still severely limited by the catastrophic forgetting issue. In this work, we leverage a key property of encoder-decoder Transformers, i.e. their generative ability, to propose a novel approach to continually learning Neural Machine Translation systems. We show how this can effectively learn on a stream of experiences comprising different languages, by leveraging a replay memory populated by using the model itself as a generator of parallel sentences. We empirically demonstrate that our approach can counteract catastrophic forgetting without requiring explicit memorization of training data. Code will be publicly available upon publication. Code: https:\/\/github.com\/m-resta\/sg-rep},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('268','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_268\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Modern Neural Machine Translation systems exhibit strong performance in several different languages and are constantly improving. Their ability to learn continuously is, however, still severely limited by the catastrophic forgetting issue. In this work, we leverage a key property of encoder-decoder Transformers, i.e. their generative ability, to propose a novel approach to continually learning Neural Machine Translation systems. We show how this can effectively learn on a stream of experiences comprising different languages, by leveraging a replay memory populated by using the model itself as a generator of parallel sentences. We empirically demonstrate that our approach can counteract catastrophic forgetting without requiring explicit memorization of training data. Code will be publicly available upon publication. Code: https:\/\/github.com\/m-resta\/sg-rep<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('268','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Self-generated Replay Memories for Continual Neural Machine Translation\" src=\"https:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/08\/naacl-300x300.jpg\" width=\"80\" alt=\"Self-generated Replay Memories for Continual Neural Machine Translation\" \/><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2023\">2023<\/h3><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_number\">13.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Georgiev, Dobrik Georgiev;  Numeroso, Danilo;  Bacciu, Davide;  Li\u00f2, Pietro<\/p><p class=\"tp_pub_title\">Neural algorithmic reasoning for combinatorial optimisation <span class=\"tp_pub_type inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Learning on Graphs Conference, <\/span><span class=\"tp_pub_additional_pages\">pp. 28\u20131, <\/span><span class=\"tp_pub_additional_organization\">PMLR <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_264\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('264','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_264\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('264','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_264\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{georgiev2024neural,<br \/>\r\ntitle = {Neural algorithmic reasoning for combinatorial optimisation},<br \/>\r\nauthor = {Dobrik Georgiev Georgiev and Danilo Numeroso and Davide Bacciu and Pietro Li\u00f2},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-12-15},<br \/>\r\nurldate = {2023-12-15},<br \/>\r\nbooktitle = {Learning on Graphs Conference},<br \/>\r\npages = {28\u20131},<br \/>\r\norganization = {PMLR},<br \/>\r\nabstract = {Solving NP-hard\/complete combinatorial problems with neural networks is a challenging research area that aims to surpass classical approximate algorithms. The long-term objective is to outperform hand-designed heuristics for NP-hard\/complete problems by learning to generate superior solutions solely from training data. Current neural-based methods for solving CO problems often overlook the inherent\" algorithmic\" nature of the problems. In contrast, heuristics designed for CO problems, eg TSP, frequently leverage well-established algorithms, such as those for finding the minimum spanning tree. In this paper, we propose leveraging recent advancements in neural algorithmic reasoning to improve the learning of CO problems. Specifically, we suggest pre-training our neural model on relevant algorithms before training it on CO instances. Our results demonstrate that, using this learning setup, we achieve superior performance compared to non-algorithmically informed deep learning models.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('264','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_264\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Solving NP-hard\/complete combinatorial problems with neural networks is a challenging research area that aims to surpass classical approximate algorithms. The long-term objective is to outperform hand-designed heuristics for NP-hard\/complete problems by learning to generate superior solutions solely from training data. Current neural-based methods for solving CO problems often overlook the inherent&quot; algorithmic&quot; nature of the problems. In contrast, heuristics designed for CO problems, eg TSP, frequently leverage well-established algorithms, such as those for finding the minimum spanning tree. In this paper, we propose leveraging recent advancements in neural algorithmic reasoning to improve the learning of CO problems. Specifically, we suggest pre-training our neural model on relevant algorithms before training it on CO instances. Our results demonstrate that, using this learning setup, we achieve superior performance compared to non-algorithmically informed deep learning models.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('264','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Neural algorithmic reasoning for combinatorial optimisation\" src=\"https:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/log.png\" width=\"80\" alt=\"Neural algorithmic reasoning for combinatorial optimisation\" \/><\/div><\/div><div class=\"tp_publication tp_publication_workshop\"><div class=\"tp_pub_number\">14.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Gravina, Alessio;  Lovisotto, Giulio;  Gallicchio, Claudio;  Bacciu, Davide;  Grohnfeldt, Claas<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('246','tp_links')\" style=\"cursor:pointer;\">Effective Non-Dissipative Propagation for Continuous-Time Dynamic Graphs<\/a> <span class=\"tp_pub_type workshop\">Workshop<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Temporal Graph Learning Workshop, NeurIPS 2023, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_246\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('246','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_246\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('246','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_246\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('246','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_246\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@workshop{Gravina2023b,<br \/>\r\ntitle = {Effective Non-Dissipative Propagation for Continuous-Time Dynamic Graphs},<br \/>\r\nauthor = {Alessio Gravina and Giulio Lovisotto and Claudio Gallicchio and Davide Bacciu and Claas Grohnfeldt},<br \/>\r\nurl = {https:\/\/openreview.net\/forum?id=zAHFC2LNEe, PDF},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-12-11},<br \/>\r\nurldate = {2023-12-11},<br \/>\r\nbooktitle = {Temporal Graph Learning Workshop, NeurIPS 2023},<br \/>\r\nabstract = {Recent research on Deep Graph Networks (DGNs) has broadened the domain of learning on graphs to real-world systems of interconnected entities that evolve over time. This paper addresses prediction problems on graphs defined by a stream of events, possibly irregularly sampled over time, generally referred to as Continuous-Time Dynamic Graphs (C-TDGs). While many predictive problems on graphs may require capturing interactions between nodes at different distances, existing DGNs for C-TDGs are not designed to propagate and preserve long-range information - resulting in suboptimal performance. In this work, we present Continuous-Time Graph Anti-Symmetric Network (CTAN), a DGN for C-TDGs designed within the ordinary differential equations framework that enables efficient propagation of long-range dependencies. We show that our method robustly performs stable and non-dissipative information propagation over dynamically evolving graphs, where the number of ODE discretization steps allows scaling the propagation range. We empirically validate the proposed approach on several real and synthetic graph benchmarks, showing that CTAN leads to improved performance while enabling the propagation of long-range information},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {workshop}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('246','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_246\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Recent research on Deep Graph Networks (DGNs) has broadened the domain of learning on graphs to real-world systems of interconnected entities that evolve over time. This paper addresses prediction problems on graphs defined by a stream of events, possibly irregularly sampled over time, generally referred to as Continuous-Time Dynamic Graphs (C-TDGs). While many predictive problems on graphs may require capturing interactions between nodes at different distances, existing DGNs for C-TDGs are not designed to propagate and preserve long-range information - resulting in suboptimal performance. In this work, we present Continuous-Time Graph Anti-Symmetric Network (CTAN), a DGN for C-TDGs designed within the ordinary differential equations framework that enables efficient propagation of long-range dependencies. We show that our method robustly performs stable and non-dissipative information propagation over dynamically evolving graphs, where the number of ODE discretization steps allows scaling the propagation range. We empirically validate the proposed approach on several real and synthetic graph benchmarks, showing that CTAN leads to improved performance while enabling the propagation of long-range information<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('246','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_246\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/openreview.net\/forum?id=zAHFC2LNEe\" title=\"PDF\" target=\"_blank\">PDF<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('246','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Effective Non-Dissipative Propagation for Continuous-Time Dynamic Graphs\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2020\/11\/neurips.png\" width=\"80\" alt=\"Effective Non-Dissipative Propagation for Continuous-Time Dynamic Graphs\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">15.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Errica, Federico;  Gravina, Alessio;  Bacciu, Davide;  Micheli, Alessio<\/p><p class=\"tp_pub_title\">Hidden Markov Models for Temporal Graph Representation Learning <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning , <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_234\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('234','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_234\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Errica2023,<br \/>\r\ntitle = {Hidden Markov Models for Temporal Graph Representation Learning},<br \/>\r\nauthor = {Federico Errica and Alessio Gravina and Davide Bacciu and Alessio Micheli},<br \/>\r\neditor = {Michel Verleysen},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-10-04},<br \/>\r\nurldate = {2023-10-04},<br \/>\r\nbooktitle = {Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('234','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Hidden Markov Models for Temporal Graph Representation Learning\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/esann.png\" width=\"80\" alt=\"Hidden Markov Models for Temporal Graph Representation Learning\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">16.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Landolfi, Francesco;  Bacciu, Davide;  Numeroso, Danilo<\/p><p class=\"tp_pub_title\"> A Tropical View of Graph Neural Networks  <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning , <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_235\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('235','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_235\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Landolfi2023,<br \/>\r\ntitle = { A Tropical View of Graph Neural Networks },<br \/>\r\nauthor = {Francesco Landolfi and Davide Bacciu and Danilo Numeroso<br \/>\r\n<br \/>\r\n},<br \/>\r\neditor = {Michel Verleysen},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-10-04},<br \/>\r\nurldate = {2023-10-04},<br \/>\r\nbooktitle = {Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('235','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\" A Tropical View of Graph Neural Networks \" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/esann.png\" width=\"80\" alt=\" A Tropical View of Graph Neural Networks \" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">17.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Ceni, Andrea;  Bacciu, Davide;  Caro, Valerio De;  Gallicchio, Claudio;  Oneto, Luca<\/p><p class=\"tp_pub_title\"> Improving Fairness via Intrinsic Plasticity in Echo State Networks  <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning , <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_236\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('236','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_236\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Ceni2023,<br \/>\r\ntitle = { Improving Fairness via Intrinsic Plasticity in Echo State Networks },<br \/>\r\nauthor = {Andrea Ceni and Davide Bacciu and Valerio De Caro and Claudio Gallicchio and Luca Oneto<br \/>\r\n},<br \/>\r\neditor = {Michel Verleysen},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-10-04},<br \/>\r\nurldate = {2023-10-04},<br \/>\r\nbooktitle = {Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('236','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\" Improving Fairness via Intrinsic Plasticity in Echo State Networks \" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/esann.png\" width=\"80\" alt=\" Improving Fairness via Intrinsic Plasticity in Echo State Networks \" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">18.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Cossu, Andrea;  Spinnato, Francesco;  Guidotti, Riccardo;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"> A Protocol for Continual Explanation of SHAP  <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning , <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_237\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('237','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_237\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Cossu2023,<br \/>\r\ntitle = { A Protocol for Continual Explanation of SHAP },<br \/>\r\nauthor = {Andrea Cossu and Francesco Spinnato and Riccardo Guidotti and Davide Bacciu},<br \/>\r\neditor = {Michel Verleysen},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-10-04},<br \/>\r\nurldate = {2023-10-04},<br \/>\r\nbooktitle = {Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('237','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\" A Protocol for Continual Explanation of SHAP \" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/esann.png\" width=\"80\" alt=\" A Protocol for Continual Explanation of SHAP \" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">19.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Caro, Valerio De;  Mauro, Antonio Di;  Bacciu, Davide;  Gallicchio, Claudio<\/p><p class=\"tp_pub_title\"> Communication-Efficient Ridge Regression in Federated Echo State Networks  <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning , <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_238\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('238','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_238\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Caro2023,<br \/>\r\ntitle = { Communication-Efficient Ridge Regression in Federated Echo State Networks },<br \/>\r\nauthor = {Valerio De Caro and Antonio Di Mauro and Davide Bacciu and Claudio Gallicchio<br \/>\r\n<br \/>\r\n},<br \/>\r\neditor = {Michel Verleysen},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-10-04},<br \/>\r\nurldate = {2023-10-04},<br \/>\r\nbooktitle = {Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('238','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\" Communication-Efficient Ridge Regression in Federated Echo State Networks \" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/esann.png\" width=\"80\" alt=\" Communication-Efficient Ridge Regression in Federated Echo State Networks \" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">20.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Bacciu, Davide;  Errica, Federico;  Micheli, Alessio;  Navarin, Nicol\u00f2;  Pasa, Luca;  Podda, Marco;  Zambon, Daniele<\/p><p class=\"tp_pub_title\">Graph Representation Learning  <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning , <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_239\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('239','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_239\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Bacciu2023c,<br \/>\r\ntitle = {Graph Representation Learning },<br \/>\r\nauthor = {Davide Bacciu and Federico Errica and Alessio Micheli and Nicol\u00f2 Navarin and Luca Pasa and Marco Podda and Daniele Zambon<br \/>\r\n<br \/>\r\n},<br \/>\r\neditor = {Michel Verleysen},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-10-04},<br \/>\r\nurldate = {2023-10-04},<br \/>\r\nbooktitle = {Proceedings of the 31th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('239','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Graph Representation Learning \" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/esann.png\" width=\"80\" alt=\"Graph Representation Learning \" \/><\/div><\/div><div class=\"tp_publication tp_publication_workshop\"><div class=\"tp_pub_number\">21.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Ceni, Andrea;  Cossu, Andrea;  Liu, Jingyue;  St\u00f6lzle, Maximilian;  Santina, Cosimo Della;  Gallicchio, Claudio;  Bacciu, Davide<\/p><p class=\"tp_pub_title\">Randomly Coupled Oscillators <span class=\"tp_pub_type workshop\">Workshop<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the ECML\/PKDD Workshop on Deep Learning meets Neuromorphic Hardware, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_252\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('252','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_252\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@workshop{Ceni2023c,<br \/>\r\ntitle = {Randomly Coupled Oscillators},<br \/>\r\nauthor = {Andrea Ceni and Andrea Cossu and Jingyue Liu and Maximilian St\u00f6lzle and Cosimo Della Santina and Claudio Gallicchio and Davide Bacciu},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-09-18},<br \/>\r\nbooktitle = {Proceedings of the ECML\/PKDD Workshop on Deep Learning meets Neuromorphic Hardware},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {workshop}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('252','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Randomly Coupled Oscillators\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/ecml2023.png\" width=\"80\" alt=\"Randomly Coupled Oscillators\" \/><\/div><\/div><div class=\"tp_publication tp_publication_workshop\"><div class=\"tp_pub_number\">22.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Gravina, Alessio;  Gallicchio, Claudio;  Bacciu, Davide<\/p><p class=\"tp_pub_title\">Non-Dissipative Propagation by Randomized Anti-Symmetric Deep Graph Networks <span class=\"tp_pub_type workshop\">Workshop<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the ECML\/PKDD Workshop on Deep Learning meets Neuromorphic Hardware, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_253\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('253','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_253\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@workshop{Gravina2023c,<br \/>\r\ntitle = {Non-Dissipative Propagation by Randomized Anti-Symmetric Deep Graph Networks},<br \/>\r\nauthor = {Alessio Gravina and Claudio Gallicchio and Davide Bacciu},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-09-18},<br \/>\r\nurldate = {2023-09-18},<br \/>\r\nbooktitle = {Proceedings of the ECML\/PKDD Workshop on Deep Learning meets Neuromorphic Hardware},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {workshop}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('253','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Non-Dissipative Propagation by Randomized Anti-Symmetric Deep Graph Networks\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/ecml2023.png\" width=\"80\" alt=\"Non-Dissipative Propagation by Randomized Anti-Symmetric Deep Graph Networks\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">23.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Cosenza, Emanuele;  Valenti, Andrea;  Bacciu, Davide<\/p><p class=\"tp_pub_title\">Graph-based Polyphonic Multitrack Music Generation <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 32nd INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI 2023), <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_228\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('228','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_228\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Cosenza2023,<br \/>\r\ntitle = {Graph-based Polyphonic Multitrack Music Generation},<br \/>\r\nauthor = {Emanuele Cosenza and Andrea Valenti and Davide Bacciu },<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-08-19},<br \/>\r\nurldate = {2023-08-19},<br \/>\r\nbooktitle = {Proceedings of the 32nd INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI 2023)},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('228','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Graph-based Polyphonic Multitrack Music Generation\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/ijcai.png\" width=\"80\" alt=\"Graph-based Polyphonic Multitrack Music Generation\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">24.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Hemati, Hamed;  Lomonaco, Vincenzo;  Bacciu, Davide;  Borth, Damian<\/p><p class=\"tp_pub_title\">Partial Hypernetworks for Continual Learning <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the International Conference on Lifelong Learning Agents (CoLLAs 2023), <\/span><span class=\"tp_pub_additional_publisher\">Proceedings of Machine Learning Research, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_232\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('232','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_232\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Hemati2023,<br \/>\r\ntitle = {Partial Hypernetworks for Continual Learning},<br \/>\r\nauthor = {Hamed Hemati and Vincenzo Lomonaco and Davide Bacciu and Damian Borth},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-08-01},<br \/>\r\nurldate = {2023-08-01},<br \/>\r\nbooktitle = {Proceedings of the International Conference on Lifelong Learning Agents (CoLLAs 2023)},<br \/>\r\npublisher = {Proceedings of Machine Learning Research},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('232','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Partial Hypernetworks for Continual Learning\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/collas.png\" width=\"80\" alt=\"Partial Hypernetworks for Continual Learning\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">25.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Hemati, Hamed;  Cossu, Andrea;  Carta, Antonio;  Hurtado, Julio;  Pellegrini, Lorenzo;  Bacciu, Davide;  Lomonaco, Vincenzo;  Borth, Damian<\/p><p class=\"tp_pub_title\">Class-Incremental Learning with Repetition  <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the International Conference on Lifelong Learning Agents (CoLLAs 2023), <\/span><span class=\"tp_pub_additional_publisher\">Proceedings of Machine Learning Research, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_233\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('233','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_233\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Hemati2023b,<br \/>\r\ntitle = {Class-Incremental Learning with Repetition },<br \/>\r\nauthor = {Hamed Hemati and Andrea Cossu and Antonio Carta and Julio Hurtado and Lorenzo Pellegrini and Davide Bacciu and Vincenzo Lomonaco and Damian Borth},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-08-01},<br \/>\r\nurldate = {2023-08-01},<br \/>\r\nbooktitle = {Proceedings of the International Conference on Lifelong Learning Agents (CoLLAs 2023)},<br \/>\r\npublisher = {Proceedings of Machine Learning Research},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('233','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Class-Incremental Learning with Repetition \" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/collas.png\" width=\"80\" alt=\"Class-Incremental Learning with Repetition \" \/><\/div><\/div><div class=\"tp_publication tp_publication_workshop\"><div class=\"tp_pub_number\">26.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Caro, Valerio De;  Bacciu, Davide;  Gallicchio, Claudio<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('242','tp_links')\" style=\"cursor:pointer;\">Decentralized Plasticity in Reservoir Dynamical Networks for Pervasive Environments<\/a> <span class=\"tp_pub_type workshop\">Workshop<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 2023 ICML Workshop on Localized Learning: Decentralized Model Updates via Non-Global Objectives \r\n, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_resource_link\"><a id=\"tp_links_sh_242\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('242','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_242\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('242','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_242\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@workshop{nokey,<br \/>\r\ntitle = {Decentralized Plasticity in Reservoir Dynamical Networks for Pervasive Environments},<br \/>\r\nauthor = {Valerio De Caro and Davide Bacciu and Claudio Gallicchio<br \/>\r\n},<br \/>\r\nurl = {https:\/\/openreview.net\/forum?id=5hScPOeDaR, PDF},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-07-29},<br \/>\r\nurldate = {2023-07-29},<br \/>\r\nbooktitle = {Proceedings of the 2023 ICML Workshop on Localized Learning: Decentralized Model Updates via Non-Global Objectives <br \/>\r\n},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {workshop}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('242','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_242\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/openreview.net\/forum?id=5hScPOeDaR\" title=\"PDF\" target=\"_blank\">PDF<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('242','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Decentralized Plasticity in Reservoir Dynamical Networks for Pervasive Environments\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/icml.png\" width=\"80\" alt=\"Decentralized Plasticity in Reservoir Dynamical Networks for Pervasive Environments\" \/><\/div><\/div><div class=\"tp_publication tp_publication_workshop\"><div class=\"tp_pub_number\">27.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Ceni, Andrea;  Cossu, Andrea;  Liu, Jingyue;  St\u00f6lzle, Maximilian;  Santina, Cosimo Della;  Gallicchio, Claudio;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('240','tp_links')\" style=\"cursor:pointer;\">Randomly Coupled Oscillators for Time Series Processing<\/a> <span class=\"tp_pub_type workshop\">Workshop<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 2023 ICML Workshop on New Frontiers in Learning, Control, and Dynamical Systems , <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_resource_link\"><a id=\"tp_links_sh_240\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('240','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_240\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('240','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_240\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@workshop{Ceni2023b,<br \/>\r\ntitle = {Randomly Coupled Oscillators for Time Series Processing},<br \/>\r\nauthor = {Andrea Ceni and Andrea Cossu and Jingyue Liu and Maximilian St\u00f6lzle and Cosimo Della Santina and Claudio Gallicchio and Davide Bacciu},<br \/>\r\nurl = {https:\/\/openreview.net\/forum?id=fmn7PMykEb, PDF},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-07-28},<br \/>\r\nurldate = {2023-07-28},<br \/>\r\nbooktitle = {Proceedings of the 2023 ICML Workshop on New Frontiers in Learning, Control, and Dynamical Systems },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {workshop}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('240','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_240\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/openreview.net\/forum?id=fmn7PMykEb\" title=\"PDF\" target=\"_blank\">PDF<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('240','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Randomly Coupled Oscillators for Time Series Processing\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/icml.png\" width=\"80\" alt=\"Randomly Coupled Oscillators for Time Series Processing\" \/><\/div><\/div><div class=\"tp_publication tp_publication_workshop\"><div class=\"tp_pub_number\">28.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Massidda, Riccardo;  Landolfi, Francesco;  Cinquini, Martina;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('241','tp_links')\" style=\"cursor:pointer;\">Differentiable Causal Discovery with Smooth Acyclic Orientations<\/a> <span class=\"tp_pub_type workshop\">Workshop<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 2023 ICML Workshop on Differentiable Almost Everything: Differentiable Relaxations, Algorithms, Operators, and Simulators , <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_resource_link\"><a id=\"tp_links_sh_241\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('241','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_241\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('241','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_241\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@workshop{Massidda2023b,<br \/>\r\ntitle = {Differentiable Causal Discovery with Smooth Acyclic Orientations},<br \/>\r\nauthor = {Riccardo Massidda and Francesco Landolfi and Martina Cinquini and Davide Bacciu},<br \/>\r\nurl = {https:\/\/openreview.net\/forum?id=IVwWgscehR, PDF},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-07-28},<br \/>\r\nurldate = {2023-07-28},<br \/>\r\nbooktitle = {Proceedings of the 2023 ICML Workshop on Differentiable Almost Everything: Differentiable Relaxations, Algorithms, Operators, and Simulators },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {workshop}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('241','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_241\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/openreview.net\/forum?id=IVwWgscehR\" title=\"PDF\" target=\"_blank\">PDF<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('241','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Differentiable Causal Discovery with Smooth Acyclic Orientations\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/icml.png\" width=\"80\" alt=\"Differentiable Causal Discovery with Smooth Acyclic Orientations\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">29.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Simone, Lorenzo;  Bacciu, Davide<\/p><p class=\"tp_pub_title\">ECGAN: generative adversarial network for electrocardiography <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of Artificial Intelligence In Medicine 2023 (AIME 2023), <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_227\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('227','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_227\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{nokey,<br \/>\r\ntitle = {ECGAN: generative adversarial network for electrocardiography},<br \/>\r\nauthor = {Lorenzo Simone and Davide Bacciu },<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-06-12},<br \/>\r\nurldate = {2023-06-12},<br \/>\r\nbooktitle = {Proceedings of Artificial Intelligence In Medicine 2023 (AIME 2023)},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('227','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"ECGAN: generative adversarial network for electrocardiography\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/aime23.jpg\" width=\"80\" alt=\"ECGAN: generative adversarial network for electrocardiography\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">30.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Lomonaco, Vincenzo;  Caro, Valerio De;  Gallicchio, Claudio;  Carta, Antonio;  Sardianos, Christos;  Varlamis, Iraklis;  Tserpes, Konstantinos;  Coppola, Massimo;  Marpena, Mina;  Politi, Sevasti;  Schoitsch, Erwin;  Bacciu, Davide<\/p><p class=\"tp_pub_title\">AI-Toolkit: a Microservices Architecture for Low-Code Decentralized Machine Intelligence <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of 2023 IEEE International Conference on Acoustics, Speech and Signal Processing, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_230\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('230','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_230\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('230','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_230\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Lomonaco2023,<br \/>\r\ntitle = {AI-Toolkit: a Microservices Architecture for Low-Code Decentralized Machine Intelligence},<br \/>\r\nauthor = {Vincenzo Lomonaco and Valerio De Caro and Claudio Gallicchio and Antonio Carta and Christos Sardianos and Iraklis Varlamis and Konstantinos Tserpes and Massimo Coppola and Mina Marpena and Sevasti Politi and Erwin Schoitsch and Davide Bacciu},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-06-04},<br \/>\r\nurldate = {2023-06-04},<br \/>\r\nbooktitle = {Proceedings of 2023 IEEE International Conference on Acoustics, Speech and Signal Processing},<br \/>\r\nabstract = {Artificial Intelligence and Machine Learning toolkits such as Scikit-learn, PyTorch and Tensorflow provide today a solid starting point for the rapid prototyping of R&D solutions. However, they can be hardly ported to heterogeneous decentralised hardware and real-world production environments. A common practice involves outsourcing deployment solutions to scalable cloud infrastructures such as Amazon SageMaker or Microsoft Azure. In this paper, we proposed an open-source microservices-based architecture for decentralised machine intelligence which aims at bringing R&D and deployment functionalities closer following a low-code approach. Such an approach would guarantee flexible integration of cutting-edge functionalities while preserving complete control over the deployed solutions at negligible costs and maintenance efforts.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('230','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_230\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Artificial Intelligence and Machine Learning toolkits such as Scikit-learn, PyTorch and Tensorflow provide today a solid starting point for the rapid prototyping of R&amp;D solutions. However, they can be hardly ported to heterogeneous decentralised hardware and real-world production environments. A common practice involves outsourcing deployment solutions to scalable cloud infrastructures such as Amazon SageMaker or Microsoft Azure. In this paper, we proposed an open-source microservices-based architecture for decentralised machine intelligence which aims at bringing R&amp;D and deployment functionalities closer following a low-code approach. Such an approach would guarantee flexible integration of cutting-edge functionalities while preserving complete control over the deployed solutions at negligible costs and maintenance efforts.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('230','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"AI-Toolkit: a Microservices Architecture for Low-Code Decentralized Machine Intelligence\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/icassp23.png\" width=\"80\" alt=\"AI-Toolkit: a Microservices Architecture for Low-Code Decentralized Machine Intelligence\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">31.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Caro, Valerio De;  Danzinger, Herbert;  Gallicchio, Claudio;  K\u00f6ncz\u00f6l, Clemens;  Lomonaco, Vincenzo;  Marmpena, Mina;  Marpena, Mina;  Politi, Sevasti;  Veledar, Omar;  Bacciu, Davide<\/p><p class=\"tp_pub_title\">Prediction of Driver's Stress Affection in Simulated Autonomous Driving Scenarios <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of 2023 IEEE International Conference on Acoustics, Speech and Signal Processing, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_231\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('231','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_231\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('231','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_231\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{DeCaro2023,<br \/>\r\ntitle = {Prediction of Driver's Stress Affection in Simulated Autonomous Driving Scenarios},<br \/>\r\nauthor = {Valerio De Caro and Herbert Danzinger and Claudio Gallicchio and Clemens K\u00f6ncz\u00f6l and Vincenzo Lomonaco and Mina Marmpena and Mina Marpena and Sevasti Politi and Omar Veledar and Davide Bacciu},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-06-04},<br \/>\r\nurldate = {2023-06-04},<br \/>\r\nbooktitle = {Proceedings of 2023 IEEE International Conference on Acoustics, Speech and Signal Processing},<br \/>\r\nabstract = {We investigate the task of predicting stress affection from physiological data of users experiencing simulations of autonomous driving. We approach this task on two levels of granularity, depending on whether the prediction is performed at end of the simulation, or along the simulation. In the former, denoted as coarse-grained prediction, we employed Decision Trees. In the latter, denoted as fine-grained prediction, we employed Echo State Networks, a Recurrent Neural Network<br \/>\r\nthat allows efficient learning from temporal data and hence is<br \/>\r\nsuitable for pervasive environments. We conduct experiments on a private dataset of physiological data from people participating in multiple driving scenarios simulating different stressful events. The results show that the proposed model is capable of detecting conditions of event-related cognitive stress proving, the existence of a correlation between stressful events and the physiological data.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('231','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_231\" style=\"display:none;\"><div class=\"tp_abstract_entry\">We investigate the task of predicting stress affection from physiological data of users experiencing simulations of autonomous driving. We approach this task on two levels of granularity, depending on whether the prediction is performed at end of the simulation, or along the simulation. In the former, denoted as coarse-grained prediction, we employed Decision Trees. In the latter, denoted as fine-grained prediction, we employed Echo State Networks, a Recurrent Neural Network<br \/>\r\nthat allows efficient learning from temporal data and hence is<br \/>\r\nsuitable for pervasive environments. We conduct experiments on a private dataset of physiological data from people participating in multiple driving scenarios simulating different stressful events. The results show that the proposed model is capable of detecting conditions of event-related cognitive stress proving, the existence of a correlation between stressful events and the physiological data.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('231','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Prediction of Driver's Stress Affection in Simulated Autonomous Driving Scenarios\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/icassp23.png\" width=\"80\" alt=\"Prediction of Driver's Stress Affection in Simulated Autonomous Driving Scenarios\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">32.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Gravina, Alessio;  Bacciu, Davide;  Gallicchio, Claudio<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('225','tp_links')\" style=\"cursor:pointer;\">Anti-Symmetric DGN: a stable architecture for Deep Graph Networks<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the Eleventh International Conference on Learning Representations (ICLR 2023)  , <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_225\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('225','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_225\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('225','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_225\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('225','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_225\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Gravina2023,<br \/>\r\ntitle = {Anti-Symmetric DGN: a stable architecture for Deep Graph Networks},<br \/>\r\nauthor = {Alessio Gravina and Davide Bacciu and Claudio Gallicchio},<br \/>\r\nurl = {https:\/\/openreview.net\/pdf?id=J3Y7cgZOOS},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-05-01},<br \/>\r\nurldate = {2023-05-01},<br \/>\r\nbooktitle = {Proceedings of the Eleventh International Conference on Learning Representations (ICLR 2023)  },<br \/>\r\nabstract = {Deep Graph Networks (DGNs) currently dominate the research landscape of learning from graphs, due to their efficiency and ability to implement an adaptive message-passing scheme between the nodes. However, DGNs are typically limited in their ability to propagate and preserve long-term dependencies between nodes, i.e., they suffer from the over-squashing phenomena. As a result, we can expect them to under-perform, since different problems require to capture interactions at different (and possibly large) radii in order to be effectively solved. In this work, we present Anti-Symmetric Deep Graph Networks (A-DGNs), a framework for stable and non-dissipative DGN design, conceived through the lens of ordinary differential equations. We give theoretical proof that our method is stable and non-dissipative, leading to two key results: long-range information between nodes is preserved, and no gradient vanishing or explosion occurs in training. We empirically validate the proposed approach on several graph benchmarks, showing that A-DGN yields to improved performance and enables to learn effectively even when dozens of layers are used.ers are used.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('225','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_225\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Deep Graph Networks (DGNs) currently dominate the research landscape of learning from graphs, due to their efficiency and ability to implement an adaptive message-passing scheme between the nodes. However, DGNs are typically limited in their ability to propagate and preserve long-term dependencies between nodes, i.e., they suffer from the over-squashing phenomena. As a result, we can expect them to under-perform, since different problems require to capture interactions at different (and possibly large) radii in order to be effectively solved. In this work, we present Anti-Symmetric Deep Graph Networks (A-DGNs), a framework for stable and non-dissipative DGN design, conceived through the lens of ordinary differential equations. We give theoretical proof that our method is stable and non-dissipative, leading to two key results: long-range information between nodes is preserved, and no gradient vanishing or explosion occurs in training. We empirically validate the proposed approach on several graph benchmarks, showing that A-DGN yields to improved performance and enables to learn effectively even when dozens of layers are used.ers are used.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('225','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_225\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/openreview.net\/pdf?id=J3Y7cgZOOS\" title=\"https:\/\/openreview.net\/pdf?id=J3Y7cgZOOS\" target=\"_blank\">https:\/\/openreview.net\/pdf?id=J3Y7cgZOOS<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('225','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Anti-Symmetric DGN: a stable architecture for Deep Graph Networks\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/iclr.png\" width=\"80\" alt=\"Anti-Symmetric DGN: a stable architecture for Deep Graph Networks\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">33.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Numeroso, Danilo;  Bacciu, Davide;  Veli\u010dkovi\u0107, Petar<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('226','tp_links')\" style=\"cursor:pointer;\">Dual Algorithmic Reasoning<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the Eleventh International Conference on Learning Representations (ICLR 2023), <\/span><span class=\"tp_pub_additional_year\">2023<\/span><span class=\"tp_pub_additional_note\">, (Notable Spotlight paper)<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_226\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('226','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_226\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('226','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_226\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('226','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_226\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Numeroso2023,<br \/>\r\ntitle = {Dual Algorithmic Reasoning},<br \/>\r\nauthor = {Danilo Numeroso and Davide Bacciu and Petar Veli\u010dkovi\u0107},<br \/>\r\nurl = {https:\/\/openreview.net\/pdf?id=hhvkdRdWt1F},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-05-01},<br \/>\r\nurldate = {2023-05-01},<br \/>\r\nbooktitle = {Proceedings of the Eleventh International Conference on Learning Representations (ICLR 2023)},<br \/>\r\nabstract = {Neural Algorithmic Reasoning is an emerging area of machine learning which seeks to infuse algorithmic computation in neural networks, typically by training neural models to approximate steps of classical algorithms. In this context, much of the current work has focused on learning reachability and shortest path graph algorithms, showing that joint learning on similar algorithms is beneficial for generalisation. However, when targeting more complex problems, such \"similar\" algorithms become more difficult to find. Here, we propose to learn algorithms by exploiting duality of the underlying algorithmic problem. Many algorithms solve optimisation problems. We demonstrate that simultaneously learning the dual definition of these optimisation problems in algorithmic learning allows for better learning and qualitatively better solutions. Specifically, we exploit the max-flow min-cut theorem to simultaneously learn these two algorithms over synthetically generated graphs, demonstrating the effectiveness of the proposed approach. We then validate the real-world utility of our dual algorithmic reasoner by deploying it on a challenging brain vessel classification task, which likely depends on the vessels\u2019 flow properties. We demonstrate a clear performance gain when using our model within such a context, and empirically show that learning the max-flow and min-cut algorithms together is critical for achieving such a result.},<br \/>\r\nnote = {Notable Spotlight paper},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('226','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_226\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Neural Algorithmic Reasoning is an emerging area of machine learning which seeks to infuse algorithmic computation in neural networks, typically by training neural models to approximate steps of classical algorithms. In this context, much of the current work has focused on learning reachability and shortest path graph algorithms, showing that joint learning on similar algorithms is beneficial for generalisation. However, when targeting more complex problems, such &quot;similar&quot; algorithms become more difficult to find. Here, we propose to learn algorithms by exploiting duality of the underlying algorithmic problem. Many algorithms solve optimisation problems. We demonstrate that simultaneously learning the dual definition of these optimisation problems in algorithmic learning allows for better learning and qualitatively better solutions. Specifically, we exploit the max-flow min-cut theorem to simultaneously learn these two algorithms over synthetically generated graphs, demonstrating the effectiveness of the proposed approach. We then validate the real-world utility of our dual algorithmic reasoner by deploying it on a challenging brain vessel classification task, which likely depends on the vessels\u2019 flow properties. We demonstrate a clear performance gain when using our model within such a context, and empirically show that learning the max-flow and min-cut algorithms together is critical for achieving such a result.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('226','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_226\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/openreview.net\/pdf?id=hhvkdRdWt1F\" title=\"https:\/\/openreview.net\/pdf?id=hhvkdRdWt1F\" target=\"_blank\">https:\/\/openreview.net\/pdf?id=hhvkdRdWt1F<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('226','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Dual Algorithmic Reasoning\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/iclr.png\" width=\"80\" alt=\"Dual Algorithmic Reasoning\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">34.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Massidda, Riccardo;  Geiger, Atticus;  Icard, Thomas;  Bacciu, Davide<\/p><p class=\"tp_pub_title\">Causal Abstraction with Soft Interventions <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 2nd Conference on Causal Learning and Reasoning (CLeaR 2023), <\/span><span class=\"tp_pub_additional_publisher\">PMLR, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_223\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('223','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_223\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Massidda2023,<br \/>\r\ntitle = {Causal Abstraction with Soft Interventions},<br \/>\r\nauthor = {Riccardo Massidda and Atticus Geiger and Thomas Icard and Davide Bacciu},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-04-17},<br \/>\r\nurldate = {2023-04-17},<br \/>\r\nbooktitle = {Proceedings of the 2nd Conference on Causal Learning and Reasoning (CLeaR 2023)},<br \/>\r\npublisher = {PMLR},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('223','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Causal Abstraction with Soft Interventions\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/clear.png\" width=\"80\" alt=\"Causal Abstraction with Soft Interventions\" \/><\/div><\/div><div class=\"tp_publication tp_publication_workshop\"><div class=\"tp_pub_number\">35.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Gravina, Alessio;  Bacciu, Davide;  Gallicchio, Claudio<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('229','tp_links')\" style=\"cursor:pointer;\">Non-Dissipative Propagation by Anti-Symmetric Deep Graph Networks<\/a> <span class=\"tp_pub_type workshop\">Workshop<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedigns of the  Ninth International Workshop on Deep Learning on Graphs: Method and Applications (DLG-AAAI\u201923), <\/span><span class=\"tp_pub_additional_year\">2023<\/span><span class=\"tp_pub_additional_note\">, (Winner of the Best Student Paper Award at DLG-AAAI23)<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_229\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('229','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_229\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('229','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_229\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('229','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_229\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@workshop{nokey,<br \/>\r\ntitle = {Non-Dissipative Propagation by Anti-Symmetric Deep Graph Networks},<br \/>\r\nauthor = {Alessio Gravina and Davide Bacciu and Claudio Gallicchio},<br \/>\r\nurl = {https:\/\/drive.google.com\/file\/d\/1uPHhjwSa3g_hRvHwx6UnbMLgGN_cAqMu\/view. PDF},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-02-13},<br \/>\r\nurldate = {2023-02-13},<br \/>\r\nbooktitle = {Proceedigns of the  Ninth International Workshop on Deep Learning on Graphs: Method and Applications (DLG-AAAI\u201923)},<br \/>\r\nabstract = {Deep Graph Networks (DGNs) currently dominate the research landscape of learning from graphs, due to the efficiency of their adaptive message-passing scheme between nodes. However, DGNs are typically limited in their ability to propagate and preserve long-term dependencies between nodes, i.e., they suffer from the over-squashing phenomena. This reduces their effectiveness, since predictive problems may require to capture interactions at different, and possibly large, radii in order to be effectively solved. In this work, we present Anti-Symmetric DGN (A-DGN), a framework forstable and non-dissipative DGN design, conceived through the lens of ordinary differential equations. We give theoretical proof that our method is stable and non-dissipative, leading to two key results: long-range information between nodes is preserved, and no gradient vanishing or explosion occurs in training. We empirically validate the proposed approach on several graph benchmarks, showing that A-DGN yields to improved performance and enables to learn effectively even when dozens of layers are used.},<br \/>\r\nnote = {Winner of the Best Student Paper Award at DLG-AAAI23},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {workshop}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('229','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_229\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Deep Graph Networks (DGNs) currently dominate the research landscape of learning from graphs, due to the efficiency of their adaptive message-passing scheme between nodes. However, DGNs are typically limited in their ability to propagate and preserve long-term dependencies between nodes, i.e., they suffer from the over-squashing phenomena. This reduces their effectiveness, since predictive problems may require to capture interactions at different, and possibly large, radii in order to be effectively solved. In this work, we present Anti-Symmetric DGN (A-DGN), a framework forstable and non-dissipative DGN design, conceived through the lens of ordinary differential equations. We give theoretical proof that our method is stable and non-dissipative, leading to two key results: long-range information between nodes is preserved, and no gradient vanishing or explosion occurs in training. We empirically validate the proposed approach on several graph benchmarks, showing that A-DGN yields to improved performance and enables to learn effectively even when dozens of layers are used.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('229','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_229\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/drive.google.com\/file\/d\/1uPHhjwSa3g_hRvHwx6UnbMLgGN_cAqMu\/view. PDF\" title=\"https:\/\/drive.google.com\/file\/d\/1uPHhjwSa3g_hRvHwx6UnbMLgGN_cAqMu\/view. PDF\" target=\"_blank\">https:\/\/drive.google.com\/file\/d\/1uPHhjwSa3g_hRvHwx6UnbMLgGN_cAqMu\/view. PDF<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('229','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Non-Dissipative Propagation by Anti-Symmetric Deep Graph Networks\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/aaai.jpeg\" width=\"80\" alt=\"Non-Dissipative Propagation by Anti-Symmetric Deep Graph Networks\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">36.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Bacciu, Davide;  Conte, Alessio;  Landolfi, Francesco<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('219','tp_links')\" style=\"cursor:pointer;\">Generalizing Downsampling from Regular Data to Graphs<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, <\/span><span class=\"tp_pub_additional_year\">2023<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_219\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('219','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_219\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('219','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_219\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('219','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_219\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Bacciu2023,<br \/>\r\ntitle = {Generalizing Downsampling from Regular Data to Graphs},<br \/>\r\nauthor = {Davide Bacciu and Alessio Conte and Francesco Landolfi},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2208.03523, Arxiv},<br \/>\r\nyear  = {2023},<br \/>\r\ndate = {2023-02-07},<br \/>\r\nurldate = {2023-02-07},<br \/>\r\nbooktitle = {Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence},<br \/>\r\nabstract = {Downsampling produces coarsened, multi-resolution representations of data and it is used, for example, to produce lossy compression and visualization of large images, reduce computational costs, and boost deep neural representation learning. Unfortunately, due to their lack of a regular structure, there is still no consensus on how downsampling should apply to graphs and linked data. Indeed reductions in graph data are still needed for the goals described above, but reduction mechanisms do not have the same focus on preserving topological structures and properties, while allowing for resolution-tuning, as is the case in regular data downsampling. In this paper, we take a step in this direction, introducing a unifying interpretation of downsampling in regular and graph data. In particular, we define a graph coarsening mechanism which is a graph-structured counterpart of controllable equispaced coarsening mechanisms in regular data. We prove theoretical guarantees for distortion bounds on path lengths, as well as the ability to preserve key topological properties in the coarsened graphs. We leverage these concepts to define a graph pooling mechanism that we empirically assess in graph classification tasks, providing a greedy algorithm that allows efficient parallel implementation on GPUs, and showing that it compares favorably against pooling methods in literature. },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('219','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_219\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Downsampling produces coarsened, multi-resolution representations of data and it is used, for example, to produce lossy compression and visualization of large images, reduce computational costs, and boost deep neural representation learning. Unfortunately, due to their lack of a regular structure, there is still no consensus on how downsampling should apply to graphs and linked data. Indeed reductions in graph data are still needed for the goals described above, but reduction mechanisms do not have the same focus on preserving topological structures and properties, while allowing for resolution-tuning, as is the case in regular data downsampling. In this paper, we take a step in this direction, introducing a unifying interpretation of downsampling in regular and graph data. In particular, we define a graph coarsening mechanism which is a graph-structured counterpart of controllable equispaced coarsening mechanisms in regular data. We prove theoretical guarantees for distortion bounds on path lengths, as well as the ability to preserve key topological properties in the coarsened graphs. We leverage these concepts to define a graph pooling mechanism that we empirically assess in graph classification tasks, providing a greedy algorithm that allows efficient parallel implementation on GPUs, and showing that it compares favorably against pooling methods in literature. <\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('219','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_219\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2208.03523\" title=\"Arxiv\" target=\"_blank\">Arxiv<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('219','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Generalizing Downsampling from Regular Data to Graphs\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/aaai.jpeg\" width=\"80\" alt=\"Generalizing Downsampling from Regular Data to Graphs\" \/><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2022\">2022<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">37.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Caro, Valerio De;  Gallicchio, Claudio;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('210','tp_links')\" style=\"cursor:pointer;\">Federated Adaptation of Reservoirs via Intrinsic Plasticity<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 30th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning  (ESANN 2022), <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_210\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('210','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_210\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('210','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_210\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('210','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_210\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Caro2022,<br \/>\r\ntitle = {Federated Adaptation of Reservoirs via Intrinsic Plasticity},<br \/>\r\nauthor = {Valerio {De Caro} and Claudio Gallicchio and Davide Bacciu},<br \/>\r\neditor = {Michel Verleysen},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2206.11087, Arxiv},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-10-05},<br \/>\r\nurldate = {2022-10-05},<br \/>\r\nbooktitle = {Proceedings of the 30th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning  (ESANN 2022)},<br \/>\r\nabstract = {We propose a novel algorithm for performing federated learning with Echo State Networks (ESNs) in a client-server scenario. In particular, our proposal focuses on the adaptation of reservoirs by combining Intrinsic Plasticity with Federated Averaging. The former is a gradient-based method for adapting the reservoir's non-linearity in a local and unsupervised manner, while the latter provides the framework for learning in the federated scenario. We evaluate our approach on real-world datasets from human monitoring, in comparison with the previous approach for federated ESNs existing in literature. Results show that adapting the reservoir with our algorithm provides a significant improvement on the performance of the global model. },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('210','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_210\" style=\"display:none;\"><div class=\"tp_abstract_entry\">We propose a novel algorithm for performing federated learning with Echo State Networks (ESNs) in a client-server scenario. In particular, our proposal focuses on the adaptation of reservoirs by combining Intrinsic Plasticity with Federated Averaging. The former is a gradient-based method for adapting the reservoir's non-linearity in a local and unsupervised manner, while the latter provides the framework for learning in the federated scenario. We evaluate our approach on real-world datasets from human monitoring, in comparison with the previous approach for federated ESNs existing in literature. Results show that adapting the reservoir with our algorithm provides a significant improvement on the performance of the global model. <\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('210','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_210\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2206.11087\" title=\"Arxiv\" target=\"_blank\">Arxiv<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('210','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Federated Adaptation of Reservoirs via Intrinsic Plasticity\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/esann.png\" width=\"80\" alt=\"Federated Adaptation of Reservoirs via Intrinsic Plasticity\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">38.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Bacciu, Davide;  Errica, Federico;  Navarin, Nicol\u00f2;  Pasa, Luca;  Zambon, Daniele<\/p><p class=\"tp_pub_title\">Deep Learning for Graphs <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 30th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning  (ESANN 2022), <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_214\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('214','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_214\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{nokey,<br \/>\r\ntitle = {Deep Learning for Graphs},<br \/>\r\nauthor = {Davide Bacciu and Federico Errica and Nicol\u00f2 Navarin and Luca Pasa and Daniele Zambon},<br \/>\r\neditor = {Michel Verleysen},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-10-05},<br \/>\r\nurldate = {2022-10-05},<br \/>\r\nbooktitle = {Proceedings of the 30th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning  (ESANN 2022)},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('214','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Deep Learning for Graphs\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/esann.png\" width=\"80\" alt=\"Deep Learning for Graphs\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">39.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Valenti, Andrea;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('220','tp_links')\" style=\"cursor:pointer;\">Modular Representations for Weak Disentanglement<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 30th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN 2022), <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_220\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('220','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_220\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('220','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_220\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('220','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_220\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Valenti2022c,<br \/>\r\ntitle = {Modular Representations for Weak Disentanglement},<br \/>\r\nauthor = {Andrea Valenti and Davide Bacciu},<br \/>\r\neditor = {Michel Verleysen},<br \/>\r\nurl = {https:\/\/arxiv.org\/pdf\/2209.05336.pdf},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-10-05},<br \/>\r\nurldate = {2022-10-05},<br \/>\r\nbooktitle = {Proceedings of the 30th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN 2022)},<br \/>\r\nabstract = {The recently introduced weakly disentangled representations proposed to relax some constraints of the previous definitions of disentanglement, in exchange for more flexibility. However, at the moment, weak disentanglement can only be achieved by increasing the amount of supervision as the number of factors of variations of the data increase. In this paper, we introduce modular representations for weak disentanglement, a novel method that allows to keep the amount of supervised information constant with respect the number of generative factors. The experiments shows that models using modular representations can increase their performance with respect to previous work without the need of additional supervision.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('220','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_220\" style=\"display:none;\"><div class=\"tp_abstract_entry\">The recently introduced weakly disentangled representations proposed to relax some constraints of the previous definitions of disentanglement, in exchange for more flexibility. However, at the moment, weak disentanglement can only be achieved by increasing the amount of supervision as the number of factors of variations of the data increase. In this paper, we introduce modular representations for weak disentanglement, a novel method that allows to keep the amount of supervised information constant with respect the number of generative factors. The experiments shows that models using modular representations can increase their performance with respect to previous work without the need of additional supervision.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('220','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_220\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/pdf\/2209.05336.pdf\" title=\"https:\/\/arxiv.org\/pdf\/2209.05336.pdf\" target=\"_blank\">https:\/\/arxiv.org\/pdf\/2209.05336.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('220','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Modular Representations for Weak Disentanglement\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/esann.png\" width=\"80\" alt=\"Modular Representations for Weak Disentanglement\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">40.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Matteoni, Federico;  Cossu, Andrea;  Gallicchio, Claudio;  Lomonaco, Vincenzo;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('221','tp_links')\" style=\"cursor:pointer;\">Continual Learning for Human State Monitoring<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 30th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN 2022), <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_221\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('221','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_221\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('221','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_221\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('221','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_221\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Matteoni2022,<br \/>\r\ntitle = {Continual Learning for Human State Monitoring},<br \/>\r\nauthor = {Federico Matteoni and Andrea Cossu and Claudio Gallicchio and Vincenzo Lomonaco and Davide Bacciu},<br \/>\r\neditor = {Michel Verleysen},<br \/>\r\nurl = {https:\/\/arxiv.org\/pdf\/2207.00010, Arxiv},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-10-05},<br \/>\r\nurldate = {2022-10-05},<br \/>\r\nbooktitle = {Proceedings of the 30th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN 2022)},<br \/>\r\nabstract = {Continual Learning (CL) on time series data represents a promising but under-studied avenue for real-world applications. We propose two new CL benchmarks for Human State Monitoring. We carefully designed the benchmarks to mirror real-world environments in which new subjects are continuously added. We conducted an empirical evaluation to assess the ability of popular CL strategies to mitigate forgetting in our benchmarks. Our results show that, possibly due to the domain-incremental properties of our benchmarks, forgetting can be easily tackled even with a simple finetuning and that existing strategies struggle in accumulating knowledge over a fixed, held-out, test subject.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('221','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_221\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Continual Learning (CL) on time series data represents a promising but under-studied avenue for real-world applications. We propose two new CL benchmarks for Human State Monitoring. We carefully designed the benchmarks to mirror real-world environments in which new subjects are continuously added. We conducted an empirical evaluation to assess the ability of popular CL strategies to mitigate forgetting in our benchmarks. Our results show that, possibly due to the domain-incremental properties of our benchmarks, forgetting can be easily tackled even with a simple finetuning and that existing strategies struggle in accumulating knowledge over a fixed, held-out, test subject.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('221','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_221\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/pdf\/2207.00010\" title=\"Arxiv\" target=\"_blank\">Arxiv<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('221','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Continual Learning for Human State Monitoring\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/esann.png\" width=\"80\" alt=\"Continual Learning for Human State Monitoring\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">41.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Massidda, Riccardo;  Bacciu, Davide<\/p><p class=\"tp_pub_title\">Knowledge-Driven Interpretation of Convolutional Neural Networks <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 2022 European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD 2022), <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_217\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('217','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_217\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('217','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_217\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Massidda2022,<br \/>\r\ntitle = {Knowledge-Driven Interpretation of Convolutional Neural Networks},<br \/>\r\nauthor = {Riccardo Massidda and Davide Bacciu},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-09-20},<br \/>\r\nurldate = {2022-09-20},<br \/>\r\nbooktitle = {Proceedings of the 2022 European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD 2022)},<br \/>\r\nabstract = {Since the widespread adoption of deep learning solutions in critical environments, the interpretation of artificial neural networks has become a significant issue. To this end, numerous approaches currently try to align human-level concepts with the activation patterns of artificial neurons. Nonetheless, they often understate two related aspects: the distributed nature of neural representations and the semantic relations between concepts. We explicitly tackled this interrelatedness by defining a novel semantic alignment framework to align distributed activation patterns and structured knowledge. In particular, we detailed a solution to assign to both neurons and their linear combinations one or more concepts from the WordNet semantic network. Acknowledging semantic links also enabled the clustering of neurons into semantically rich and meaningful neural circuits. Our empirical analysis of popular convolutional networks for image classification found evidence of the emergence of such neural circuits. Finally, we discovered neurons in neural circuits to be pivotal for the network to perform effectively on semantically related tasks. We also contribute by releasing the code that implements our alignment framework.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('217','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_217\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Since the widespread adoption of deep learning solutions in critical environments, the interpretation of artificial neural networks has become a significant issue. To this end, numerous approaches currently try to align human-level concepts with the activation patterns of artificial neurons. Nonetheless, they often understate two related aspects: the distributed nature of neural representations and the semantic relations between concepts. We explicitly tackled this interrelatedness by defining a novel semantic alignment framework to align distributed activation patterns and structured knowledge. In particular, we detailed a solution to assign to both neurons and their linear combinations one or more concepts from the WordNet semantic network. Acknowledging semantic links also enabled the clustering of neurons into semantically rich and meaningful neural circuits. Our empirical analysis of popular convolutional networks for image classification found evidence of the emergence of such neural circuits. Finally, we discovered neurons in neural circuits to be pivotal for the network to perform effectively on semantically related tasks. We also contribute by releasing the code that implements our alignment framework.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('217','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Knowledge-Driven Interpretation of Convolutional Neural Networks\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/ecml22.png\" width=\"80\" alt=\"Knowledge-Driven Interpretation of Convolutional Neural Networks\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">42.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Lagani, Gabriele;  Bacciu, Davide;  Gallicchio, Claudio;  Falchi, Fabrizio;  Gennaro, Claudio;  Amato, Giuseppe<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('211','tp_links')\" style=\"cursor:pointer;\">Deep Features for CBIR with Scarce Data using Hebbian Learning<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proc. of the 19th International Conference on Content-based Multimedia Indexing (CBMI2022), <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_211\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('211','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_211\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('211','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_211\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('211','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_211\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{nokey,<br \/>\r\ntitle = {Deep Features for CBIR with Scarce Data using Hebbian Learning},<br \/>\r\nauthor = {Gabriele Lagani and Davide Bacciu and Claudio Gallicchio and Fabrizio Falchi and Claudio Gennaro and Giuseppe Amato},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2205.08935, Arxiv},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-09-14},<br \/>\r\nurldate = {2022-09-14},<br \/>\r\nbooktitle = {Proc. of the 19th International Conference on Content-based Multimedia Indexing (CBMI2022)},<br \/>\r\nabstract = { Features extracted from Deep Neural Networks (DNNs) have proven to be very effective in the context of Content Based Image Retrieval (CBIR). In recent work, biologically inspired textit{Hebbian} learning algorithms have shown promises for DNN training. In this contribution, we study the performance of such algorithms in the development of feature extractors for CBIR tasks. Specifically, we consider a semi-supervised learning strategy in two steps: first, an unsupervised pre-training stage is performed using Hebbian learning on the image dataset; second, the network is fine-tuned using supervised Stochastic Gradient Descent (SGD) training. For the unsupervised pre-training stage, we explore the nonlinear Hebbian Principal Component Analysis (HPCA) learning rule. For the supervised fine-tuning stage, we assume sample efficiency scenarios, in which the amount of labeled samples is just a small fraction of the whole dataset. Our experimental analysis, conducted on the CIFAR10 and CIFAR100 datasets shows that, when few labeled samples are available, our Hebbian approach provides relevant improvements compared to various alternative methods. },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('211','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_211\" style=\"display:none;\"><div class=\"tp_abstract_entry\"> Features extracted from Deep Neural Networks (DNNs) have proven to be very effective in the context of Content Based Image Retrieval (CBIR). In recent work, biologically inspired textit{Hebbian} learning algorithms have shown promises for DNN training. In this contribution, we study the performance of such algorithms in the development of feature extractors for CBIR tasks. Specifically, we consider a semi-supervised learning strategy in two steps: first, an unsupervised pre-training stage is performed using Hebbian learning on the image dataset; second, the network is fine-tuned using supervised Stochastic Gradient Descent (SGD) training. For the unsupervised pre-training stage, we explore the nonlinear Hebbian Principal Component Analysis (HPCA) learning rule. For the supervised fine-tuning stage, we assume sample efficiency scenarios, in which the amount of labeled samples is just a small fraction of the whole dataset. Our experimental analysis, conducted on the CIFAR10 and CIFAR100 datasets shows that, when few labeled samples are available, our Hebbian approach provides relevant improvements compared to various alternative methods. <\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('211','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_211\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2205.08935\" title=\"Arxiv\" target=\"_blank\">Arxiv<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('211','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Deep Features for CBIR with Scarce Data using Hebbian Learning\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/cbmi22.png\" width=\"80\" alt=\"Deep Features for CBIR with Scarce Data using Hebbian Learning\" \/><\/div><\/div><div class=\"tp_publication tp_publication_workshop\"><div class=\"tp_pub_number\">43.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Corti, Francesco;  Entezari, Rahim;  Hooker, Sara;  Bacciu, Davide;  Saukh, Olga<\/p><p class=\"tp_pub_title\">Studying the impact of magnitude pruning on contrastive learning methods <span class=\"tp_pub_type workshop\">Workshop<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">ICML 2022 workshop on Hardware Aware Efficient Training (HAET 2022), <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_216\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('216','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_216\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('216','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_216\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@workshop{nokey,<br \/>\r\ntitle = {Studying the impact of magnitude pruning on contrastive learning methods},<br \/>\r\nauthor = {Francesco Corti and Rahim Entezari and Sara Hooker and Davide Bacciu and Olga Saukh},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-07-23},<br \/>\r\nurldate = {2022-07-23},<br \/>\r\nbooktitle = {ICML 2022 workshop on Hardware Aware Efficient Training (HAET 2022)},<br \/>\r\nabstract = {We study the impact of different pruning techniques on the representation learned by deep neural networks trained with contrastive loss functions. Our work finds that at high sparsity levels, contrastive learning results in a higher number of misclassified examples relative to models trained with traditional cross-entropy loss. To understand this pronounced difference, we use metrics such as the number of PIEs, qscore and pdepth to measure the impact of pruning on the learned representation quality. Our analysis suggests the schedule of the pruning method implementation matters. We find that the negative impact of sparsity on the quality of the learned representation is the highest when pruning is introduced early-on in training phase.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {workshop}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('216','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_216\" style=\"display:none;\"><div class=\"tp_abstract_entry\">We study the impact of different pruning techniques on the representation learned by deep neural networks trained with contrastive loss functions. Our work finds that at high sparsity levels, contrastive learning results in a higher number of misclassified examples relative to models trained with traditional cross-entropy loss. To understand this pronounced difference, we use metrics such as the number of PIEs, qscore and pdepth to measure the impact of pruning on the learned representation quality. Our analysis suggests the schedule of the pruning method implementation matters. We find that the negative impact of sparsity on the quality of the learned representation is the highest when pruning is introduced early-on in training phase.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('216','tp_abstract')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Studying the impact of magnitude pruning on contrastive learning methods\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/icml.png\" width=\"80\" alt=\"Studying the impact of magnitude pruning on contrastive learning methods\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">44.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Sangermano, Matteo;  Carta, Antonio;  Cossu, Andrea;  Lomonaco, Vincenzo;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('204','tp_links')\" style=\"cursor:pointer;\">Sample Condensation in Online Continual Learning<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 2022 IEEE World Congress on Computational Intelligence, <\/span><span class=\"tp_pub_additional_publisher\">IEEE, <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_204\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('204','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_204\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('204','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_204\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('204','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_204\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Sangermano2022,<br \/>\r\ntitle = {Sample Condensation in Online Continual Learning},<br \/>\r\nauthor = {Matteo Sangermano and Antonio Carta and Andrea Cossu and Vincenzo Lomonaco and Davide Bacciu<br \/>\r\n},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2206.11849, Arxiv},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-07-18},<br \/>\r\nurldate = {2022-07-18},<br \/>\r\nbooktitle = {Proceedings of the 2022 IEEE World Congress on Computational Intelligence},<br \/>\r\npublisher = {IEEE},<br \/>\r\nabstract = {Online Continual learning is a challenging learning scenario where the model observes a non-stationary stream of data and learns online. The main challenge is to incrementally learn while avoiding catastrophic forgetting, namely the problem of forgetting previously acquired knowledge while learning from new data. A popular solution in these scenario is to use a small memory to retain old data and rehearse them over time. Unfortunately, due to the limited memory size, the quality of the memory will deteriorate over time. In this paper we propose OLCGM, a novel replay-based continual learning strategy that uses knowledge condensation techniques to continuously compress the memory and achieve a better use of its limited size. The sample condensation step compresses old samples, instead of removing them like other replay strategies. As a result, the experiments show that, whenever the memory budget is limited compared to the complexity of the data, OLCGM improves the final accuracy compared to state-of-the-art replay strategies.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('204','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_204\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Online Continual learning is a challenging learning scenario where the model observes a non-stationary stream of data and learns online. The main challenge is to incrementally learn while avoiding catastrophic forgetting, namely the problem of forgetting previously acquired knowledge while learning from new data. A popular solution in these scenario is to use a small memory to retain old data and rehearse them over time. Unfortunately, due to the limited memory size, the quality of the memory will deteriorate over time. In this paper we propose OLCGM, a novel replay-based continual learning strategy that uses knowledge condensation techniques to continuously compress the memory and achieve a better use of its limited size. The sample condensation step compresses old samples, instead of removing them like other replay strategies. As a result, the experiments show that, whenever the memory budget is limited compared to the complexity of the data, OLCGM improves the final accuracy compared to state-of-the-art replay strategies.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('204','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_204\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2206.11849\" title=\"Arxiv\" target=\"_blank\">Arxiv<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('204','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Sample Condensation in Online Continual Learning\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2022\/01\/WCCI2022-padua-logo.png\" width=\"80\" alt=\"Sample Condensation in Online Continual Learning\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">45.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Valenti, Andrea;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('205','tp_links')\" style=\"cursor:pointer;\"> Leveraging Relational Information for Learning Weakly Disentangled Representations <\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 2022 IEEE World Congress on Computational Intelligence, <\/span><span class=\"tp_pub_additional_publisher\">IEEE, <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_205\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('205','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_205\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('205','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_205\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('205','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_205\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Valenti2022,<br \/>\r\ntitle = { Leveraging Relational Information for Learning Weakly Disentangled Representations },<br \/>\r\nauthor = {Andrea Valenti and Davide Bacciu<br \/>\r\n},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2205.10056, Arxiv},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-07-18},<br \/>\r\nurldate = {2022-07-18},<br \/>\r\nbooktitle = {Proceedings of the 2022 IEEE World Congress on Computational Intelligence},<br \/>\r\npublisher = {IEEE},<br \/>\r\nabstract = {Disentanglement is a difficult property to enforce in neural representations. This might be due, in part, to a formalization of the disentanglement problem that focuses too heavily on separating relevant factors of variation of the data in single isolated dimensions of the neural representation. We argue that such a definition might be too restrictive and not necessarily beneficial in terms of downstream tasks. In this work, we present an alternative view over learning (weakly) disentangled representations, which leverages concepts from relational learning. We identify the regions of the latent space that correspond to specific instances of generative factors, and we learn the relationships among these regions in order to perform controlled changes to the latent codes. We also introduce a compound generative model that implements such a weak disentanglement approach. Our experiments shows that the learned representations can separate the relevant factors of variation in the data, while preserving the information needed for effectively generating high quality data samples.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('205','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_205\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Disentanglement is a difficult property to enforce in neural representations. This might be due, in part, to a formalization of the disentanglement problem that focuses too heavily on separating relevant factors of variation of the data in single isolated dimensions of the neural representation. We argue that such a definition might be too restrictive and not necessarily beneficial in terms of downstream tasks. In this work, we present an alternative view over learning (weakly) disentangled representations, which leverages concepts from relational learning. We identify the regions of the latent space that correspond to specific instances of generative factors, and we learn the relationships among these regions in order to perform controlled changes to the latent codes. We also introduce a compound generative model that implements such a weak disentanglement approach. Our experiments shows that the learned representations can separate the relevant factors of variation in the data, while preserving the information needed for effectively generating high quality data samples.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('205','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_205\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2205.10056\" title=\"Arxiv\" target=\"_blank\">Arxiv<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('205','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\" Leveraging Relational Information for Learning Weakly Disentangled Representations \" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2022\/01\/WCCI2022-padua-logo.png\" width=\"80\" alt=\" Leveraging Relational Information for Learning Weakly Disentangled Representations \" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">46.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Castellana, Daniele;  Errica, Federico;  Bacciu, Davide;  Micheli, Alessio<\/p><p class=\"tp_pub_title\">The Infinite Contextual Graph Markov Model <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 39th International Conference on Machine Learning (ICML 2022), <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_207\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('207','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_207\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{nokey,<br \/>\r\ntitle = {The Infinite Contextual Graph Markov Model},<br \/>\r\nauthor = {Daniele Castellana and Federico Errica and Davide Bacciu and Alessio Micheli<br \/>\r\n},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-07-18},<br \/>\r\nurldate = {2022-07-18},<br \/>\r\nbooktitle = {Proceedings of the 39th International Conference on Machine Learning (ICML 2022)},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('207','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"The Infinite Contextual Graph Markov Model\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/icml.png\" width=\"80\" alt=\"The Infinite Contextual Graph Markov Model\" \/><\/div><\/div><div class=\"tp_publication tp_publication_workshop\"><div class=\"tp_pub_number\">47.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Semola, Rudy;  Lomonaco, Vincenzo;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('222','tp_links')\" style=\"cursor:pointer;\">Continual-Learning-as-a-Service (CLaaS): On-Demand Efficient Adaptation of Predictive Models<\/a> <span class=\"tp_pub_type workshop\">Workshop<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proc. of the 1st International Workshop on Pervasive Artificial Intelligence,  2022 IEEE World Congress on Computational Intelligence, <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_222\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('222','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_222\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('222','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_222\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('222','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_222\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@workshop{Semola2022,<br \/>\r\ntitle = {Continual-Learning-as-a-Service (CLaaS): On-Demand Efficient Adaptation of Predictive Models},<br \/>\r\nauthor = {Rudy Semola and Vincenzo Lomonaco and Davide Bacciu},<br \/>\r\nurl = {https:\/\/arxiv.org\/pdf\/2206.06957.pdf},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-07-18},<br \/>\r\nurldate = {2022-07-18},<br \/>\r\nbooktitle = {Proc. of the 1st International Workshop on Pervasive Artificial Intelligence,  2022 IEEE World Congress on Computational Intelligence},<br \/>\r\nabstract = {Predictive machine learning models nowadays are often updated in a stateless and expensive way. The two main future trends for companies that want to build machine learning-based applications and systems are real-time inference and continual updating. Unfortunately, both trends require a mature infrastructure that is hard and costly to realize on-premise. This paper defines a novel software service and model delivery infrastructure termed Continual Learning-as-a-Service (CLaaS) to address these issues. Specifically, it embraces continual machine learning and continuous integration techniques. It provides support for model updating and validation tools for data scientists without an on-premise solution and in an efficient, stateful and easy-to-use manner. Finally, this CL model service is easy to encapsulate in any machine learning infrastructure or cloud system. This paper presents the design and implementation of a CLaaS instantiation, called LiquidBrain, evaluated in two real-world scenarios. The former is a robotic object recognition setting using the CORe50 dataset while the latter is a named category and attribute prediction using the DeepFashion-C dataset in the fashion domain. Our preliminary results suggest the usability and efficiency of the Continual Learning model services and the effectiveness of the solution in addressing real-world use-cases regardless of where the computation happens in the continuum Edge-Cloud.},<br \/>\r\nhowpublished = {CEUR-WS Proceedings},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {workshop}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('222','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_222\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Predictive machine learning models nowadays are often updated in a stateless and expensive way. The two main future trends for companies that want to build machine learning-based applications and systems are real-time inference and continual updating. Unfortunately, both trends require a mature infrastructure that is hard and costly to realize on-premise. This paper defines a novel software service and model delivery infrastructure termed Continual Learning-as-a-Service (CLaaS) to address these issues. Specifically, it embraces continual machine learning and continuous integration techniques. It provides support for model updating and validation tools for data scientists without an on-premise solution and in an efficient, stateful and easy-to-use manner. Finally, this CL model service is easy to encapsulate in any machine learning infrastructure or cloud system. This paper presents the design and implementation of a CLaaS instantiation, called LiquidBrain, evaluated in two real-world scenarios. The former is a robotic object recognition setting using the CORe50 dataset while the latter is a named category and attribute prediction using the DeepFashion-C dataset in the fashion domain. Our preliminary results suggest the usability and efficiency of the Continual Learning model services and the effectiveness of the solution in addressing real-world use-cases regardless of where the computation happens in the continuum Edge-Cloud.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('222','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_222\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/pdf\/2206.06957.pdf\" title=\"https:\/\/arxiv.org\/pdf\/2206.06957.pdf\" target=\"_blank\">https:\/\/arxiv.org\/pdf\/2206.06957.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('222','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Continual-Learning-as-a-Service (CLaaS): On-Demand Efficient Adaptation of Predictive Models\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2022\/01\/WCCI2022-padua-logo.png\" width=\"80\" alt=\"Continual-Learning-as-a-Service (CLaaS): On-Demand Efficient Adaptation of Predictive Models\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">48.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Carta, Antonio;  Cossu, Andrea;  Lomonaco, Vincenzo;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('202','tp_links')\" style=\"cursor:pointer;\">Ex-Model: Continual Learning from a Stream of Trained Models<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the CVPR 2022 Workshop on Continual Learning , <\/span><span class=\"tp_pub_additional_organization\">IEEE <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_202\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('202','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_202\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('202','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_202\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('202','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_202\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{carta2021ex,<br \/>\r\ntitle = {Ex-Model: Continual Learning from a Stream of Trained Models},<br \/>\r\nauthor = {Antonio Carta and Andrea Cossu and Vincenzo Lomonaco and Davide Bacciu},<br \/>\r\nurl = {https:\/\/arxiv.org\/pdf\/2112.06511.pdf, Arxiv},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-06-20},<br \/>\r\nurldate = {2022-06-20},<br \/>\r\nbooktitle = {Proceedings of the CVPR 2022 Workshop on Continual Learning },<br \/>\r\njournal = {arXiv preprint arXiv:2112.06511},<br \/>\r\npages = {3790-3799},<br \/>\r\norganization = {IEEE},<br \/>\r\nabstract = {Learning continually from non-stationary data streams is a challenging research topic of growing popularity in the last few years. Being able to learn, adapt, and generalize continually in an efficient, effective, and scalable way is fundamental for a sustainable development of Artificial Intelligent systems. However, an agent-centric view of continual learning requires learning directly from raw data, which limits the interaction between independent agents, the efficiency, and the privacy of current approaches. Instead, we argue that continual learning systems should exploit the availability of compressed information in the form of trained models. In this paper, we introduce and formalize a new paradigm named \"Ex-Model Continual Learning\" (ExML), where an agent learns from a sequence of previously trained models instead of raw data. We further contribute with three ex-model continual learning algorithms and an empirical setting comprising three datasets (MNIST, CIFAR-10 and CORe50), and eight scenarios, where the proposed algorithms are extensively tested. Finally, we highlight the peculiarities of the ex-model paradigm and we point out interesting future research directions. },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('202','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_202\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Learning continually from non-stationary data streams is a challenging research topic of growing popularity in the last few years. Being able to learn, adapt, and generalize continually in an efficient, effective, and scalable way is fundamental for a sustainable development of Artificial Intelligent systems. However, an agent-centric view of continual learning requires learning directly from raw data, which limits the interaction between independent agents, the efficiency, and the privacy of current approaches. Instead, we argue that continual learning systems should exploit the availability of compressed information in the form of trained models. In this paper, we introduce and formalize a new paradigm named &quot;Ex-Model Continual Learning&quot; (ExML), where an agent learns from a sequence of previously trained models instead of raw data. We further contribute with three ex-model continual learning algorithms and an empirical setting comprising three datasets (MNIST, CIFAR-10 and CORe50), and eight scenarios, where the proposed algorithms are extensively tested. Finally, we highlight the peculiarities of the ex-model paradigm and we point out interesting future research directions. <\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('202','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_202\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/pdf\/2112.06511.pdf\" title=\"Arxiv\" target=\"_blank\">Arxiv<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('202','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Ex-Model: Continual Learning from a Stream of Trained Models\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/cvpr.jpg\" width=\"80\" alt=\"Ex-Model: Continual Learning from a Stream of Trained Models\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">49.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Serramazza, Davide Italo;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('203','tp_links')\" style=\"cursor:pointer;\">Learning image captioning as a structured transduction task<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 23rd International Conference on Engineering Applications of Neural Networks (EANN 2022), <\/span><span class=\"tp_pub_additional_volume\">vol. 1600, <\/span><span class=\"tp_pub_additional_series\">Communications in Computer and Information Science  <\/span><span class=\"tp_pub_additional_publisher\">Springer, <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_203\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('203','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_203\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('203','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_203\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('203','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_203\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Serramazza2022,<br \/>\r\ntitle = {Learning image captioning as a structured transduction task},<br \/>\r\nauthor = {Davide Italo Serramazza and Davide Bacciu},<br \/>\r\ndoi = {doi.org\/10.1007\/978-3-031-08223-8_20},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-06-20},<br \/>\r\nurldate = {2022-06-20},<br \/>\r\nbooktitle = {Proceedings of the 23rd International Conference on Engineering Applications of Neural Networks (EANN 2022)},<br \/>\r\nvolume = {1600},<br \/>\r\npages = {235\u2013246},<br \/>\r\npublisher = {Springer},<br \/>\r\nseries = {Communications in Computer and Information Science },<br \/>\r\nabstract = {Image captioning is a task typically approached by deep encoder-decoder architectures, where the encoder component works on a flat representation of the image while the decoder considers a sequential representation of natural language sentences. As such, these encoder-decoder architectures implement a simple and very specific form of structured transduction, that is a generalization of a predictive problem where the input data and output predictions might have substantially different structures and topologies. In this paper, we explore a generalization of such an approach by addressing the problem as a general structured transduction problem. In particular, we provide a framework that allows considering input and output information with a tree-structured representation. This allows taking into account the hierarchical nature underlying both images and sentences. To this end, we introduce an approach to generate tree-structured representations from images along with an autoencoder working with this kind of data. We empirically assess our approach on both synthetic and realistic tasks.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('203','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_203\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Image captioning is a task typically approached by deep encoder-decoder architectures, where the encoder component works on a flat representation of the image while the decoder considers a sequential representation of natural language sentences. As such, these encoder-decoder architectures implement a simple and very specific form of structured transduction, that is a generalization of a predictive problem where the input data and output predictions might have substantially different structures and topologies. In this paper, we explore a generalization of such an approach by addressing the problem as a general structured transduction problem. In particular, we provide a framework that allows considering input and output information with a tree-structured representation. This allows taking into account the hierarchical nature underlying both images and sentences. To this end, we introduce an approach to generate tree-structured representations from images along with an autoencoder working with this kind of data. We empirically assess our approach on both synthetic and realistic tasks.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('203','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_203\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/doi.org\/10.1007\/978-3-031-08223-8_20\" title=\"Follow DOI:doi.org\/10.1007\/978-3-031-08223-8_20\" target=\"_blank\">doi:doi.org\/10.1007\/978-3-031-08223-8_20<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('203','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Learning image captioning as a structured transduction task\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/eann.jpg\" width=\"80\" alt=\"Learning image captioning as a structured transduction task\" \/><\/div><\/div><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_number\">50.<\/div><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Lucchesi, Nicol\u00f2;  Carta, Antonio;  Lomonaco, Vincenzo;  Bacciu, Davide<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('198','tp_links')\" style=\"cursor:pointer;\">Avalanche RL: a Continual Reinforcement Learning Library<\/a> <span class=\"tp_pub_type conference\">Conference<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">Proceedings of the 21st International Conference on Image Analysis and Processing (ICIAP 2021), <\/span><span class=\"tp_pub_additional_year\">2022<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_198\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('198','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_198\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('198','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_198\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('198','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_198\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{Lucchesi2022,<br \/>\r\ntitle = {Avalanche RL: a Continual Reinforcement Learning Library},<br \/>\r\nauthor = {Nicol\u00f2 Lucchesi and Antonio Carta and Vincenzo Lomonaco and Davide Bacciu},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2202.13657, Arxiv},<br \/>\r\nyear  = {2022},<br \/>\r\ndate = {2022-05-23},<br \/>\r\nurldate = {2022-05-23},<br \/>\r\nbooktitle = {Proceedings of the 21st International Conference on Image Analysis and Processing (ICIAP 2021)},<br \/>\r\nabstract = {Continual Reinforcement Learning (CRL) is a challenging setting where an agent learns to interact with an environment that is constantly changing over time (the stream of experiences). In this paper, we describe Avalanche RL, a library for Continual Reinforcement Learning which allows to easily train agents on a continuous stream of tasks. Avalanche RL is based on PyTorch and supports any OpenAI Gym environment. Its design is based on Avalanche, one of the more popular continual learning libraries, which allow us to reuse a large number of continual learning strategies and improve the interaction between reinforcement learning and continual learning researchers. Additionally, we propose Continual Habitat-Lab, a novel benchmark and a high-level library which enables the usage of the photorealistic simulator Habitat-Sim for CRL research. Overall, Avalanche RL attempts to unify under a common framework continual reinforcement learning applications, which we hope will foster the growth of the field. },<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('198','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_198\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Continual Reinforcement Learning (CRL) is a challenging setting where an agent learns to interact with an environment that is constantly changing over time (the stream of experiences). In this paper, we describe Avalanche RL, a library for Continual Reinforcement Learning which allows to easily train agents on a continuous stream of tasks. Avalanche RL is based on PyTorch and supports any OpenAI Gym environment. Its design is based on Avalanche, one of the more popular continual learning libraries, which allow us to reuse a large number of continual learning strategies and improve the interaction between reinforcement learning and continual learning researchers. Additionally, we propose Continual Habitat-Lab, a novel benchmark and a high-level library which enables the usage of the photorealistic simulator Habitat-Sim for CRL research. Overall, Avalanche RL attempts to unify under a common framework continual reinforcement learning applications, which we hope will foster the growth of the field. <\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('198','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_198\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2202.13657\" title=\"Arxiv\" target=\"_blank\">Arxiv<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('198','tp_links')\">Close<\/a><\/p><\/div><\/div><div class=\"tp_pub_image_right\"><img decoding=\"async\" name=\"Avalanche RL: a Continual Reinforcement Learning Library\" src=\"http:\/\/pages.di.unipi.it\/bacciu\/wp-content\/uploads\/sites\/12\/2024\/01\/iciap.jpg\" width=\"80\" alt=\"Avalanche RL: a Continual Reinforcement Learning Library\" \/><\/div><\/div><\/div><div class=\"tablenav\"><div class=\"tablenav-pages\"><span class=\"displaying-num\">149 entries<\/span> <a class=\"page-numbers button disabled\">&laquo;<\/a> <a class=\"page-numbers button disabled\">&lsaquo;<\/a> 1 of 3 <a href=\"https:\/\/pages.di.unipi.it\/bacciu\/publications\/conferences-workshops\/?limit=2&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"next page\" class=\"page-numbers button\">&rsaquo;<\/a> <a href=\"https:\/\/pages.di.unipi.it\/bacciu\/publications\/conferences-workshops\/?limit=3&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"last page\" class=\"page-numbers button\">&raquo;<\/a> <\/div><\/div><\/div><\/code><\/p>\n","protected":false},"excerpt":{"rendered":"","protected":false},"author":19,"featured_media":0,"parent":13,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-1382","page","type-page","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/pages.di.unipi.it\/bacciu\/wp-json\/wp\/v2\/pages\/1382","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/pages.di.unipi.it\/bacciu\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/pages.di.unipi.it\/bacciu\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/pages.di.unipi.it\/bacciu\/wp-json\/wp\/v2\/users\/19"}],"replies":[{"embeddable":true,"href":"https:\/\/pages.di.unipi.it\/bacciu\/wp-json\/wp\/v2\/comments?post=1382"}],"version-history":[{"count":8,"href":"https:\/\/pages.di.unipi.it\/bacciu\/wp-json\/wp\/v2\/pages\/1382\/revisions"}],"predecessor-version":[{"id":1430,"href":"https:\/\/pages.di.unipi.it\/bacciu\/wp-json\/wp\/v2\/pages\/1382\/revisions\/1430"}],"up":[{"embeddable":true,"href":"https:\/\/pages.di.unipi.it\/bacciu\/wp-json\/wp\/v2\/pages\/13"}],"wp:attachment":[{"href":"https:\/\/pages.di.unipi.it\/bacciu\/wp-json\/wp\/v2\/media?parent=1382"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}