RelBench: A Benchmark for Deep Learning on Relational Databases
Abstract
We present RelBench, a public benchmark for solving predictive tasks over relational databases with graph neural networks. RelBench provides databases and tasks spanning diverse domains and scales, and is intended to be a foundational infrastructure for future research. We use RelBench to conduct the first comprehensive study of Relational Deep Learning (RDL) (Fey et al., 2024), which combines graph neural network predictive models with (deep) tabular models that extract initial entity-level representations from raw tables. End-to-end learned RDL models fully exploit the predictive signal encoded in primary-foreign key links, marking a significant shift away from the dominant paradigm of manual feature engineering combined with tabular models. To thoroughly evaluate RDL against this prior gold-standard, we conduct an in-depth user study where an experienced data scientist manually engineers features for each task. In this study, RDL learns better models whilst reducing human work needed by more than an order of magnitude. This demonstrates the power of deep learning for solving predictive tasks over relational databases, opening up many new research opportunities enabled by RelBench.
Community
π Website: https://relbench.stanford.edu/
π Paper: https://arxiv.org/abs/2407.20060
π»GitHub: https://github.com/snap-stanford/relbench
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- TabReD: A Benchmark of Tabular Machine Learning in-the-Wild (2024)
- AnnotatedTables: A Large Tabular Dataset with Language Model Annotations (2024)
- rLLM: Relational Table Learning with LLMs (2024)
- LinkGPT: Teaching Large Language Models To Predict Missing Links (2024)
- Text-space Graph Foundation Models: Comprehensive Benchmarks and New Insights (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Cool! Very interested in the field and where it is going, good to see new benchmarks popping up. I also like that you compare GNNs with human performance in the experiments. The fact that many datasets do not overlap with the 4dbinfer https://arxiv.org/abs/2404.18209 is also nice (more datasets, good)
Models citing this paper 0
No model linking this paper