Here you can find a consolidated (a.k.a. slowly updated) list of my publications. A frequently updated (and possibly noisy) list of works is available on my Google Scholar profile.
Please find below a short list of highlight publications for my recent activity.
Serramazza, Davide Italo; Bacciu, Davide Learning image captioning as a structured transduction task Conference Proceedings of the 23rd International Conference on Engineering Applications of Neural Networks (EANN 2022), vol. 1600, Communications in Computer and Information Science Springer, 2022. Davide, Bacciu; Antonio, Bruno Text Summarization as Tree Transduction by Top-Down TreeLSTM Conference Proceedings of the 2018 IEEE Symposium Series on Computational Intelligence (SSCI'18), IEEE, 2018. Davide, Bacciu; Alessio, Micheli; Alessandro, Sperduti Integrating bi-directional contexts in a generative kernel for trees Conference Neural Networks (IJCNN), 2014 International Joint Conference on, IEEE, 2014. Davide, Bacciu; Alessio, Micheli; Alessandro, Sperduti An input–output hidden Markov model for tree transductions Journal Article In: Neurocomputing, vol. 112, pp. 34–46, 2013, ISSN: 0925-2312. Davide, Bacciu; Alessio, Micheli; Alessandro, Sperduti Input-Output Hidden Markov Models for Trees Conference ESANN 2012 - The 20th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning - Proceedings, Ciaco scrl - i6doc.com, 2012.@conference{Serramazza2022,
title = {Learning image captioning as a structured transduction task},
author = {Davide Italo Serramazza and Davide Bacciu},
doi = {doi.org/10.1007/978-3-031-08223-8_20},
year = {2022},
date = {2022-06-20},
urldate = {2022-06-20},
booktitle = {Proceedings of the 23rd International Conference on Engineering Applications of Neural Networks (EANN 2022)},
volume = {1600},
pages = {235–246},
publisher = {Springer},
series = {Communications in Computer and Information Science },
abstract = {Image captioning is a task typically approached by deep encoder-decoder architectures, where the encoder component works on a flat representation of the image while the decoder considers a sequential representation of natural language sentences. As such, these encoder-decoder architectures implement a simple and very specific form of structured transduction, that is a generalization of a predictive problem where the input data and output predictions might have substantially different structures and topologies. In this paper, we explore a generalization of such an approach by addressing the problem as a general structured transduction problem. In particular, we provide a framework that allows considering input and output information with a tree-structured representation. This allows taking into account the hierarchical nature underlying both images and sentences. To this end, we introduce an approach to generate tree-structured representations from images along with an autoencoder working with this kind of data. We empirically assess our approach on both synthetic and realistic tasks.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
@conference{ssci2018,
title = {Text Summarization as Tree Transduction by Top-Down TreeLSTM},
author = {Bacciu Davide and Bruno Antonio},
url = {https://arxiv.org/abs/1809.09096},
doi = {10.1109/SSCI.2018.8628873},
year = {2018},
date = {2018-11-18},
urldate = {2018-11-18},
booktitle = {Proceedings of the 2018 IEEE Symposium Series on Computational Intelligence (SSCI'18)},
pages = {1411-1418},
publisher = {IEEE},
abstract = {Extractive compression is a challenging natural language processing problem. This work contributes by formulating neural extractive compression as a parse tree transduction problem, rather than a sequence transduction task. Motivated by this, we introduce a deep neural model for learning structure-to-substructure tree transductions by extending the standard Long Short-Term Memory, considering the parent-child relationships in the structural recursion. The proposed model can achieve state of the art performance on sentence compression benchmarks, both in terms of accuracy and compression rate. },
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
@conference{11568_586070,
title = {Integrating bi-directional contexts in a generative kernel for trees},
author = {Bacciu Davide and Micheli Alessio and Sperduti Alessandro},
doi = {10.1109/IJCNN.2014.6889768},
year = {2014},
date = {2014-01-01},
urldate = {2014-01-01},
booktitle = {Neural Networks (IJCNN), 2014 International Joint Conference on},
pages = {4145--4151},
publisher = {IEEE},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
@article{bacciuNeuroComp2013,
title = {An input–output hidden Markov model for tree transductions},
author = {Bacciu Davide and Micheli Alessio and Sperduti Alessandro },
url = {http://www.sciencedirect.com/science/article/pii/S0925231213001914},
doi = {10.1016/j.neucom.2012.12.044},
issn = {0925-2312},
year = {2013},
date = {2013-01-01},
journal = {Neurocomputing},
volume = {112},
pages = {34--46},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
@conference{11568_152836,
title = {Input-Output Hidden Markov Models for Trees},
author = {Bacciu Davide and Micheli Alessio and Sperduti Alessandro},
year = {2012},
date = {2012-01-01},
urldate = {2012-01-01},
booktitle = {ESANN 2012 - The 20th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning - Proceedings},
pages = {25--30},
publisher = {Ciaco scrl - i6doc.com},
abstract = {The paper introduces an input-driven generative model for tree-structured data that extends the bottom-up hidden tree Markov model with non-homogenous transition and emission probabilities. The advantage of introducing an input-driven dynamics in structured-data pro- cessing is experimentally investigated. The results of this preliminary analysis suggest that input-driven models can capture more discrimina- tive structural information than non-input-driven approaches.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}