AI Researcher | PhD Student
Hi, I'm Hassan, a PhD student at the UKP Lab at Technische Universität Darmstadt (TUD) in Germany, working under the guidance of Prof. Iryna Gurevych. I hold a Master’s degree in Computer Science from Saarland University, where I graduated with distinction. During my studies, I gained valuable industry experience at renowned organizations such as Bosch Center for AI (BCAI) in Germany, Amazon EU in Luxembourg, and the Max Planck Institute for Informatics, where I contributed as a Research Assistant.
My research centers on Large Language Models (LLMs) and Generative AI. I focus on developing innovative LLM-based applications, ranging from research prototypes to practical demonstrations. During my two-year tenure as a full-time AI Researcher at the German Research Center for Artificial Intelligence (DFKI) in Berlin, I worked on creating a chatbot tailored for graduate students, aimed at improving their understanding of university courses. Additionally, I contributed to enhancing the user experience of AI-driven phone assistants by utilizing the capabilities of LLMs.
For my Master's thesis, supervised by Prof. Dietrich Klakow in collaboration with the NLP and Semantic Reasoning group at Bosch Center for AI, I explored Cross-Domain Neural Entity Linking. My work involved investigating a Transformer-based model to streamline domain adaptation by identifying optimal fine-tuning data across knowledge bases.
I am deeply passionate about researching and experimenting with AI models, balancing trade-offs, and applying these models to impactful real-world use cases. Outside of work, I enjoy photography, staying active through fitness, delving into philosophy, and immersing myself in new cultures through travel and group activities.
Hassan Soliman
Updates
2024
[09-2024] Presented my publication with the title "Scalable Mentoring Support with a Large Language Model Chatbot" at the ECTEL'24 conference, and won the 2nd Best Demo Paper.
[09-2024] Presented my publication with the title "Generative KI zur Lernenbegleitung in den Bildungswissenschaften: Implementierung eines LLM-basierten Chatbots im Lehramtsstudium" at the DELFI'24 conference.
[07-2024] Presented my publication with the title "Using Large Language Models for Adaptive Dialogue Management in Digital Telephone Assistants" at the UMAP'24 conference (HAAPIE).
[07-2024] Attended the UMAP'24 (User Modelling Adaptation and Personalization) conference in Cagliari, Italy, and attended several keynotes and workshops.
[05-2024] Our demo paper in DFKI with the title "Scalable Mentoring Support with a Large Language Model Chatbot" got accepted to be published at the ECTEL'24 conference in September.
[05-2024] My work with chatbot design in DFKI with the title "Generative KI zur Lernenbegleitung in den Bildungswissenschaften: Implementierung eines LLM-basierten Chatbots im Lehramtsstudium" got accepted to be published at the DELFI'24 conference in September.
[04-2024] My work with dialogue utterance adaptation based on user's context in DFKI with the title "Using Large Language Models for Adaptive Dialogue Management in Digital Telephone Assistants" got accepted to be published at the UMAP'24 conference (HAAPIE).
[03-2024] Presented recent advancements in Chatbot Design for university-level courses during my work in DFKI in collaboration with our partners from the University of Leipzig.
Interests
Artificial Intelligence (AI)
Natural Language Processing (LLM)
Information Retrieval (IR)
Question Answering (QA)
Retrieval Augmented Generation (RAG)
Education
Awards
Master of Science in Computer Science
Saarland Informatics Campus, Saarland University
2019 — 2022
Thesis Title: Cross-Domain Neural Entity Linking
Supervision: Prof. Dietrich Klakow
Saarbrücken, Germany
Bachelor of Science in Computer and Communication Engineering
Faculty of Engineering, Alexandria University
2013 — 2018
Thesis Title: Egyptian Car License Plate Information Detection
Supervision: Prof. Marwan Torki
Alexandria, Egypt
Second Best Demo Paper
ECTEL
2024
Krems, Austria
First Class Honor Degree
Alexandria University
2018
Alexandria, Egypt
Philosophy Doctor in Computer Science
Uniquoutious Knowledge Processing Lab,
Technische Universität Darmstadt
2025 — Present
Thesis Title: Safeguarding Multi-Modal LLMs against Misleading Evidence Attacks
Supervision: Prof. Iryna Gurevych
Darmstadt, Germany
Selected Publications
[09-2024] Hassan Soliman, Miloš Kravčík, Alexander Tobias Neumann, Yue Yin, Norbert Pengel and Maike Haag. 2024. Scalable Mentoring Support with a Large Language Model Chatbot. Technology Enhanced Learning for Inclusive and Equitable Quality Education (ECTEL), September 16–20, 2024, Krems, Austria, 6 pages.
ECTEL'24
ACL'24
RepL4NLP
[05-2022] Hassan Soliman, Heike Adel, Mohamed H. Gad-Elrab, Dragan Milchevski, and Jannik Strötgen. 2022. A Study on Entity Linking Across Domains: Which Data is Best for Fine-Tuning?. In Proceedings of the 7th Workshop on Representation Learning for NLP, ACL, 184–190, Dublin, Ireland.
Experience
AI Researcher
German Research Center for AI (DFKI)
Jan 2023 — Dec 2024
Led two projects in the Educational Technology lab, managing technical implementation and supervising two students.
Developed a chatbot for a graduate-level course that answered student queries with 87% accuracy. One of the two papers published on the project was nominated for the Best Demo Award at ECTEL 2024.
Applied advanced Retrieval-Augmented Generation (RAG) techniques, including Hybrid Ensemble Search and Reranking Mechanism, to enhance chatbot interactions and improve the retrieval of course materials.
Supported mentoring-style conversations by leveraging flexible agentic workflows with LangGraph, utilizing multiple small open-source models hosted on Azure, and using databases for user usage tracking and monitoring.
Implemented a sub-module for adaptive dialogue systems, customizing responses based on user emotional state and demographics, and benchmarking performance using OpenAI LLMs and open-source models.
Berlin, Germany
Applied Scientist Intern
Bosch Center for AI (BCAI)
May 2022 — Aug 2022
Contributed to the NLP & Semantic Reasoning group, applying findings from my master’s thesis on Neural Entity Linking to a high-impact industrial project using real data at Bosch.
Refactored, tested, and documented production-level code for machine learning models, ensuring scalability and efficiency for real-world deployment, leveraging the in-house GPU cluster for model fine-tuning.
Trained and evaluated machine learning models on a large-scale domain-specific dataset, achieving 77% end-to-end recall for top-3 entity predictions, outperforming existing models.
Renningen, Germany
Research Assistant
Max Planck Institute for Informatics (MPII)
Nov 2020 — May 2021
Developed a model prototype within the Database & Information Systems group to identify diverse peer groups for entities, contributing to advanced set expansion techniques.
Implemented a baseline model for entity set expansion, leveraging Wikipedia lists as a knowledge source to enhance the model’s accuracy and comprehensiveness in the expanded sets.
Optimized the algorithm’s performance by achieving a 3x faster runtime using efficient sparse matrix multiplication techniques, significantly improving computational efficiency.
Saarbrücken, Germany
Master's Thesis Student
Bosch Center for AI (BCAI)
Jun 2021 — Jan 2022
Joined the NLP & Semantic Reasoning group and worked on a unified system for linking named entities to general-domain (Wikipedia) and domain-specific knowledge bases (KBs), using context-aware embeddings (BERT) to learn a joint vector space. A pre-print of the thesis is available on arXiv. https://arxiv.org/abs/2210.15616.
Optimized a state-of-the-art model for cross-domain applications, supporting domain extension and identifying optimal data sources for fine-tuning, and improved GPU memory utilization for efficient embedding calculations.
Achieved a 9% increase in Average Precision for the top-1 entity and a 20% gain in Mean Average Precision (MAP) for top-10 entity linking across four domain-specific KBs, resulting in a workshop publication at ACL 2022.
Renningen, Germany
Software Development Engineer Intern
Amazon
Aug 2019 — Feb 2020
Maintained a web-based simulation tool for the Fulfillment Acceleration team using the AWS cloud platform, working as a full-stack software engineer.
Enhanced delivery speed simulations for prime customers, contributing to a successful report on fulfillment operations and improving delivery efficiency.
Collaborated as a system administrator in an Agile environment, managing server infrastructure and providing technical support for team tools.
Luxembourg, Luxembourg
Projects
SmolLM: Implementing, Fine-Tuning, and Aligning a LLM for Grammatical Error Correction
Implemented the SmolLM-135M (by HuggingFace) language model architecture, including components like Rotary Positional Embeddings, KV Cache, and Grouped-Query Attention, RMS Normalization, and SwiGLU Activation.
Fine-tuned the model on the Grammatical Error Correction (GEC) task using the Grammarly CoEdIT dataset.
Applied RLAIF through Direct Preference Optimization (DPO) to align model outputs with desired corrections.
Created a Colab notebook to guide users through implementation, fine-tuning, and evaluation processes.
Achieved significant improvements in grammatical error correction accuracy, scoring an expected BLUE score of ∼ 0.48.
Leveraged Python libraries such as PyTorch, Transformers, Datasets, and TRL to build and train the model effectively.
The code is open-source and published on Github.
LinguaLexMatch: Enhanced Document Language Detection
Developed and evaluated three language detection models, including an Embedding-Based approach, a TF-IDF-based Multinomial Naive Bayes model, and a fine-tuned Transformer-Based methodology.
Implemented an embedding-based approach using the intfloat/multilingual-e5-large-instruct model by generating a representative embedding for each language and classifying documents based on cosine similarity.
Benchmarked models on the papluca/language-identification dataset, achieving 99.81% accuracy with the embedding-based model.
Analyzed performance metrics such as Accuracy, F1 Scores, and Confusion Matrices across 20 different languages.
Developed a Colab notebook for replicable implementation and evaluation of different language detection models.
Utilized Python libraries including Datasets, Transformers, and Scikit-learn for model development and evaluation.
The code is open-source and published on Github.
Scalable Mentoring Support with a LLM Chatbot
Designed and implemented a chatbot based on an LLM-based Agent to provide scalable educational support and timely feedback to students of education sciences, demonstrating significant potential of generative AI in education.
Utilized Advanced techniques in Retrieval Augmented Generation (RAG) to enhance chatbot interactions, e.g., Hybrid Ensemble Search and Reranking Mechanism, enabling it to retrieve and analyze course materials effectively.
The code is subject to a Non-Disclosure Agreement.
Using LLMs for Adaptive Dialogue Management
Adapted user-directed utterances using LLMs based on the user's parameters like gender, age, and sentiment, aiming to optimize user satisfaction in conversational AI systems, focusing on healthcare patient-practice interactions.
Evaluated different LLMs and open-source tools for effectiveness in utterance adaptation, in terms of speed, cost-effectiveness, and quality of the generated text based on the adaptation relevancy and adaptation adequacy.
The code is subject to a Non-Disclosure Agreement.