Title here
Summary here
Schema-Based Query Optimisation for Graph Databases. Proc. ACM on Management of Data. Vol. 3(1), 2025 (SIGMOD'25). ![]() |
Abstract: Recursive graph queries are increasingly popular for extracting information from interconnected data found in various domains such as social networks, life sciences, and business analytics. Graph data often come with schema information that describe how nodes and edges are organized. We propose a type inference mechanism that enriches recursive graph queries with relevant structural information contained in a graph schema. We show that this schema information can be useful in order to improve the performance when evaluating recursive graph queries. Furthermore, we prove that the proposed method is sound and complete, ensuring that the semantics of the query is preserved during the schema-enrichment process. |
BibTeX:
@article{geneves-sigmod2025, author = {Chandan Sharma and Pierre Genevès and Nils Gesbert and Nabil Layaïda}, title = {Schema-Based Query Optimisation for Graph Databases}, journal = {Proc. ACM on Management of Data.}, year = {2025}, volume = {3}, number = {1}, pages = {72:1--72:29}, url = {https://inria.hal.science/hal-04485125} } |
A Comparative Analysis of Neuro-symbolic Methods for Link Prediction. To appear in Proceedings of Machine Learning Research (PMLR), 2025 (19th Conference on Neurosymbolic Learning and Reasoning - NeSy 2025). ![]() |
Abstract: Link prediction on knowledge graphs is relevant to various applications, such as recommendation systems, question answering, and entity search. This task has been approached from different perspectives: symbolic methods leverage rule-based reasoning but struggle with scalability and noise, while knowledge graph embeddings (KGE) represent entities and relations in a continuous space, enabling scalability but often neglecting logical constraints from ontologies. Recently, neurosymbolic approaches have emerged to bridge this gap by integrating embedding-based learning with symbolic reasoning. This paper provides a structured review of state-of-the-art neurosymbolic methods for link prediction. Beyond a qualitative analysis, a key contribution of this work is a comprehensive experimental benchmarking, where we systematically compare these methods on the same datasets using the same metrics. This unified experimental setup allows for a fair assessment of their strengths and limitations, bringing elements of answers to following key questions: How accurate are these methods? How scalable are they? How beneficial are they for different levels of provided knowledge and to which extent are they robust to incorrect knowledge? |
BibTeX:
@article{geneves-nesy2025, author = { Guillaume Olivier Delplanque and Luisa Sophie Werner and Nabil Layaïda and Pierre Genevès}, title = {A Comparative Analysis of Neuro-symbolic Methods for Link Prediction}, journal = {To appear in Proceedings of Machine Learning Research (PMLR)}, year = {2025}, url = {https://hal.science/hal-04991391} } |
Distributed Evaluation of Graph Queries using Recursive Relational Algebra. In 41st IEEE International Conference on Data Engineering, ICDE 2025, Hong Kong, China May 19-23, 2025, May IEEE, 2025 (ICDE'25). |
Abstract: We present a method and its implementation Dist-mu-RA for the optimized distributed evaluation of recursive relational algebraic terms. This method provides a systematic parallelisation technique by means of fixpoint splitting, plan generation and selection. The goal is to offer expressivity for high-level queries while providing efficiency and reducing communication costs. Experimental results on both real and synthetic graphs show the effectiveness of the proposed approach compared to existing systems. |
BibTeX:
@inproceedings{geneves-icde2025, author = {Sarah Chlyah and Pierre Genevès and Nabil Layaïda}, title = {Distributed Evaluation of Graph Queries using Recursive Relational Algebra}, booktitle = {41st IEEE International Conference on Data Engineering, ICDE 2025, Hong Kong, China May 19-23, 2025}, publisher = {IEEE}, year = {2025}, pages = {to appear} } |
Efficient iterative programs with distributed data collections. J. Log. Algebraic Methods Program. Vol. 144, 2025. |
BibTeX:
@article{DBLP:journals/jlap/ChlyahGGL25, author = {Sarah Chlyah and Nils Gesbert and Pierre Genevès and Nabil Laya\ida}, title = {Efficient iterative programs with distributed data collections}, journal = {J. Log. Algebraic Methods Program.}, year = {2025}, volume = {144}, pages = {101047}, url = {https://doi.org/10.1016/j.jlamp.2025.101047}, doi = {10.1016/J.JLAMP.2025.101047} } |
Efficient Enumeration of Recursive Plans in Transformation-based Query Optimizers. Proc. VLDB Endow. Vol. 17(11), 2024 (VLDB'24). |
Abstract: Query optimizers built on the transformation-based Volcano/Cascades framework are used in many database systems. Transformations proposed earlier on the logical query dag (LQDAG) data structure, which is key in such a framework, focus only on recursion-free queries. In this paper, we propose the recursive logical query dag (RLQDAG) which extends the LQDAG with the ability to capture and transform recursive queries, leveraging recent developments in recursive relational algebra. Specifically, this extension includes: (i) the ability of capturing and transforming sets of recursive relational terms thanks to (ii) annotated equivalence nodes used for guiding transformations that are more complex in the presence of recursion; and (iii) RLQDAG rewrite rules that transform sets of subterms in a grouped manner, instead of transforming individual terms in a sequential manner; and that (iv) incrementally update the necessary annotations. Core concepts of the RLQDAG are formalized using a syntax and formal semantics with a particular focus on subterm sharing and recursion. The result is a clean generalization of the LQDAG transformation-based approach, enabling more efficient explorations of plan spaces for recursive queries. An implementation of the proposed approach shows significant performance gains compared to the state-of-the-art. |
BibTeX:
@proceedings{geneves-vldb2024, author = {Amela Fejza and Pierre Genevès and Nabil Layaïda}, title = {Efficient Enumeration of Recursive Plans in Transformation-based Query Optimizers}, journal = {Proc. VLDB Endow.}, year = {2024}, volume = {17}, number = {11}, pages = {3095--3108}, url = {https://www.vldb.org/pvldb/vol17/p3095-geneves.pdf} } |
A Fast Plan Enumerator for Recursive Queries. In 40th IEEE International Conference on Data Engineering, ICDE 2024, Utrecht, The Netherlands, May 13-16, 2024 IEEE, 2024 (ICDE'24). |
Abstract: Plan enumeration is one of the most crucial components in relational query optimization. We demonstrate RLQDAG, a system implementation of a top-down plan enumerator for the purpose of transforming sets of recursive relational terms efficiently. We describe a complete system of query optimization with parsers and compilers adapted for recursive queries over knowledge and property graphs. We focus on the enumeration component of this sytem, the RLQDAG, and especially on its efficiency in generating plans out of reach of other approaches. We show graphical representations of explored plan spaces for queries on real datasets. We demonstrate the plan enumerator and its benefits in finding more efficient query plans. |
BibTeX:
@proceedings{geneves-icde2024, author = {Amela Fejza and Pierre Genevès and Nabil Layaïda}, title = {A Fast Plan Enumerator for Recursive Queries}, booktitle = {40th IEEE International Conference on Data Engineering, ICDE 2024, Utrecht, The Netherlands, May 13-16, 2024}, publisher = {IEEE}, year = {2024}, pages = {5449--5452}, url = {https://doi.org/10.1109/ICDE60146.2024.00425}, doi = {10.1109/ICDE60146.2024.00425} } |
Reproduce, Replicate, Reevaluate. The Long but Safe Way to Extend Machine Learning Methods. In Thirty-Eighth AAAI Conference on Artificial Intelligence, AAAI 2024 (AAAI'24). |
Abstract: Reproducibility is a desirable property of scientific research. On the one hand, it increases confidence in results. On the other hand, reproducible results can be extended on a solid basis. In rapidly developing fields such as machine learning, the latter is particularly important to ensure the reliability of research. In this paper, we present a systematic approach to reproducing (using the available implementation), replicating (using an alternative implementation) and reevaluating (using different datasets) state-of-the-art experiments. This approach enables the early detection and correction of deficiencies and thus the development of more robust and transparent machine learning methods. We detail the independent reproduction, replication, and reevaluation of the initially published experiments with a method that we want to extend. For each step, we identify issues and draw lessons learned. We further discuss solutions that have proven effective in overcoming the encountered problems. This work can serve as a guide for further reproducibility studies and generally improve reproducibility in machine learning. |
BibTeX:
@inproceedings{geneves-aaai2024, author = {Luisa Werner and Nabil Laya\ida and Pierre Genevès and Jérôme Euzenat and Damien Graux}, editor = {Michael J. Wooldridge and Jennifer G. Dy and Sriraam Natarajan}, title = {Reproduce, Replicate, Reevaluate. The Long but Safe Way to Extend Machine Learning Methods}, booktitle = {Thirty-Eighth AAAI Conference on Artificial Intelligence, AAAI 2024, Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence, IAAI 2024, Fourteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2014, February 20-27, 2024, Vancouver, Canada}, publisher = {AAAI Press}, year = {2024}, pages = {15850--15858}, url = {https://doi.org/10.1609/aaai.v38i14.29515}, doi = {10.1609/AAAI.V38I14.29515} } |