Hi! I am
Daniel, a double PhD student at the 🇮🇹 University of Milano - Bicocca and the 🇳🇱 University of Groningen working on interpretability, fairness and security of generative (and non-generative) Large Language Models. My supervisors are Elisabetta Fersini and Malvina Nissim.
My research focuses on the use of interpretability as a tool to make generative models safer, more reliable and less toxic in order to extend and improve their real-world applications.
Let the Models Respond: Interpreting Language Model Detoxification Through the Lens of Prompt Dependence2023
MIND at SemEval-2023 Task 11: From Uncertain Predictions to Subjective DisagreementIn Proceedings of the The 17th International Workshop on Semantic Evaluation (SemEval-2023), Jul 2023