Here you can find a consolidated (a.k.a. slowly updated) list of my publications. A frequently updated (and possibly noisy) list of works is available on my Google Scholar profile.
Please find below a short list of highlight publications for my recent activity.

Castellana, Daniele; Bacciu, Davide
A Tensor Framework for Learning in Structured Domains Journal Article
In: Neurocomputing, vol. 470, pp. 405-426, 2022.
@article{Castellana2021,
title = {A Tensor Framework for Learning in Structured Domains},
author = {Daniele Castellana and Davide Bacciu},
editor = {Kerstin Bunte and Niccolo Navarin and Luca Oneto},
doi = {10.1016/j.neucom.2021.05.110},
year = {2022},
date = {2022-01-22},
urldate = {2022-01-22},
journal = {Neurocomputing},
volume = {470},
pages = {405-426},
abstract = {Learning machines for structured data (e.g., trees) are intrinsically based on their capacity to learn representations by aggregating information from the multi-way relationships emerging from the structure topology. While complex aggregation functions are desirable in this context to increase the expressiveness of the learned representations, the modelling of higher-order interactions among structure constituents is unfeasible, in practice, due to the exponential number of parameters required. Therefore, the common approach is to define models which rely only on first-order interactions among structure constituents.
In this work, we leverage tensors theory to define a framework for learning in structured domains. Such a framework is built on the observation that more expressive models require a tensor parameterisation. This observation is the stepping stone for the application of tensor decompositions in the context of recursive models. From this point of view, the advantage of using tensor decompositions is twofold since it allows limiting the number of model parameters while injecting inductive biases that do not ignore higher-order interactions.
We apply the proposed framework on probabilistic and neural models for structured data, defining different models which leverage tensor decompositions. The experimental validation clearly shows the advantage of these models compared to first-order and full-tensorial models.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Learning machines for structured data (e.g., trees) are intrinsically based on their capacity to learn representations by aggregating information from the multi-way relationships emerging from the structure topology. While complex aggregation functions are desirable in this context to increase the expressiveness of the learned representations, the modelling of higher-order interactions among structure constituents is unfeasible, in practice, due to the exponential number of parameters required. Therefore, the common approach is to define models which rely only on first-order interactions among structure constituents.
In this work, we leverage tensors theory to define a framework for learning in structured domains. Such a framework is built on the observation that more expressive models require a tensor parameterisation. This observation is the stepping stone for the application of tensor decompositions in the context of recursive models. From this point of view, the advantage of using tensor decompositions is twofold since it allows limiting the number of model parameters while injecting inductive biases that do not ignore higher-order interactions.
We apply the proposed framework on probabilistic and neural models for structured data, defining different models which leverage tensor decompositions. The experimental validation clearly shows the advantage of these models compared to first-order and full-tensorial models.
Castellana, Daniele; Bacciu, Davide
Learning from Non-Binary Constituency Trees via Tensor Decomposition Conference
PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON COMPUTATIONAL LINGUISTICS (COLING 2020), 2020.
@conference{CastellanaCOLING2020,
title = {Learning from Non-Binary Constituency Trees via Tensor Decomposition},
author = {Daniele Castellana and Davide Bacciu},
year = {2020},
date = {2020-12-08},
urldate = {2020-12-08},
booktitle = {PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON COMPUTATIONAL LINGUISTICS (COLING 2020)},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Castellana, Daniele; Bacciu, Davide
Generalising Recursive Neural Models by Tensor Decomposition Conference
Proceedings of the 2020 IEEE World Congress on Computational Intelligence, 2020.
@conference{Wcci20Tensor,
title = {Generalising Recursive Neural Models by Tensor Decomposition},
author = {Daniele Castellana and Davide Bacciu},
url = {https://arxiv.org/abs/2006.10021, Arxiv},
year = {2020},
date = {2020-07-19},
urldate = {2020-07-19},
booktitle = {Proceedings of the 2020 IEEE World Congress on Computational Intelligence},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Bacciu, Davide; Mandic, Danilo
Tensor Decompositions in Deep Learning Conference
Proceedings of the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN'20), 2020.
@conference{esann20Tutorial,
title = {Tensor Decompositions in Deep Learning},
author = {Davide Bacciu and Danilo Mandic},
editor = {Michel Verleysen},
url = {https://arxiv.org/abs/2002.11835},
year = {2020},
date = {2020-04-21},
booktitle = {Proceedings of the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN'20)},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Castellana, Daniele; Bacciu, Davide
Tensor Decompositions in Recursive Neural Networks for Tree-Structured Data Conference
Proceedings of the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN'20), 2020.
@conference{esann20Castellana,
title = { Tensor Decompositions in Recursive Neural Networks for Tree-Structured Data },
author = {Daniele Castellana and Davide Bacciu},
editor = {Michel Verleysen},
url = {https://arxiv.org/pdf/2006.10619.pdf, Arxiv},
year = {2020},
date = {2020-04-21},
booktitle = {Proceedings of the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN'20)},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}