Charilaos I. Kanatsoulis
I am a Research Scientist in the Department of Computer Science at Stanford, working with Prof. Jure Leskovec at the SNAP Group. I am also an instructor for CS224W (Machine Learning with Graphs) and CS246 (Mining Massive Datasets).
I build foundation models for structured data, including relational databases, graphs, single-cell transcriptomics, and protein sequences. I also develop principled pretraining and adaptation methods for large language models.
What's New
- [04/2026]At ICLR 2026 in Rio de Janeiro 🇧🇷. Four papers: Relational Graph Transformer, Relational Transformer, RelBench v2 (DATA-FM Workshop), and GREmLN (MLGenX Workshop).
- [03/2026]🎉 Co-organizing the Graph Foundation Models: A New Era for Graph Machine Learning workshop, accepted at ICML 2026 (Seoul, S. Korea).
- [01/2026]CSE Colloquium talk at UC Santa Cruz: "Towards Relational Foundation Models: Zero-Shot Forecasting over Relational Databases."
- [12/2025]🏆 Oral at EurIPS Workshop on AI for Tabular Data for Relational Transformer!
- [11/2025]Invited talk on foundation models for relational data at UC Berkeley (Networks for Science Workshop).
- [10/2025]New preprint: Relational Transformer. Foundation model for forecasting on relational databases with zero-shot generalization.
- [10/2025]Co-organizing the Stanford Graph Learning Workshop 2025.
- [09/2025]🎉 KGGen at NeurIPS 2025, now with 1.1k+ GitHub stars.
pip install kg-gen. - [08/2025]🏆 Best Paper Award at KDD Temporal Graph Learning Workshop for Relational Graph Transformer!
- [08/2025]Panelist at the Temporal Graph Learning Workshop at KDD 2025.
- [07/2025]New preprint: GREmLN. First transformer foundation model for single-cell genomics with gene-regulatory networks. Collaboration with CZI.
- [06/2025]Co-organizing New Perspectives in Advancing Graph Machine Learning at NeurIPS 2025.
- [06/2025]Survey paper: Relational Deep Learning accepted at KDD 2025.
- [05/2025]Relational Deep Learning Tutorial accepted at ACM KDD 2025.
- [04/2025]Two papers at ICML 2025: RelGNN and Zero-shot Generalization of GNNs.
- [02/2025]New preprint: KGGen. Knowledge-graph extraction from text with LLMs. We also release MINE, the first benchmark for KG extraction.
- [02/2025]New preprint: RelGNN. Composite message passing for relational deep learning, up to 25% improvement over baselines.
- [01/2025]PEARL: Learning Efficient Positional Encodings with GNNs accepted at ICLR 2025.
- [12/2024]Invited talk on Relational Deep Learning at the Caltech AI Bootcamp.
- [11/2024]Invited talk on next-generation positional encodings at the LoG meetup, Stanford.
- [11/2024]Transferable Graph Autoencoder for Network Alignment accepted at LoG 2024.
- [11/2024]Talk on next-generation graph transformers at the Stanford Graph Learning Workshop 2024.
- [10/2024]Co-teaching CS 224W at Stanford with Jure.
- [10/2024]New preprint: LoRTA. Low-rank tensor adaptation for LLMs and protein models (Microsoft collaboration).
- [10/2024]Co-organizing the Stanford Graph Learning Workshop 2024.
- [01/2024]🚀 Joined Stanford as Research Scientist at the SNAP Group.
Recent Research Projects
Problem. Foundation models for tabular and relational data lag behind LLMs and vision FMs.
Contribution. A transformer architecture for forecasting on relational databases, enabling large-scale pretraining and zero-shot generalization across schemas.
Problem. Graph transformers don't naturally handle multi-table relational data.
Contribution. The first graph transformer architecture with relational attention and PEARL positional encoding tailored to multi-table data.
Problem. Standard GNNs underperform on relational deep learning tasks.
Contribution. A novel GNN architecture for predictive queries on relational databases achieving SOTA performance with up to 25% improvement over baselines.
Contribution. The latest on Relational Deep Learning: challenges, foundations, and a vision toward foundation models for relational databases.
Problem. Existing pipelines produce sparse, noisy KGs from text.
Contribution. A Python package (pip install kg-gen) producing high-fidelity KGs via LLM extraction + entity clustering. Released MINE, the first benchmark for KG extraction.
Impact. 1.1k+ GitHub stars; widely used for biomedical and scientific reasoning.
COVID-19
Problem. KG embeddings for biomedical discovery require fusing structured and textual evidence.
Contribution. Coupled tensor-matrix factorization for biomedical knowledge graphs, applied to COVID-19 drug repurposing, an early AI-driven hypothesis-generation pipeline.
CZI
Problem. Single-cell transcriptomics foundation models treat genes as independent tokens, ignoring regulatory structure.
Contribution. A multimodal graph transformer that injects gene-regulatory-network structure as positional information.
Key insight. Biological priors during pretraining capture long-range gene-token dependencies for cancer, Alzheimer's, and other downstream tasks.
Problem. LoRA's matrix factorization leaves parameter-efficiency on the table for multi-axis adaptation.
Contribution. Generalizes LoRA to tensor decomposition, reducing trainable parameters by up to two orders of magnitude at matched performance.
Applications. DPO, instruction tuning, vision, and protein folding fine-tuning.
Problem. Eigenvector-based positional encodings (e.g., Laplacian PEs) give strong inductive bias to graph transformers but require expensive eigendecomposition and suffer from instability and limited expressivity.
Contribution. PEARL is a learnable PE framework that approximates equivariant functions of eigenvectors with linear complexity, matching or surpassing full eigenvector PEs at one-to-two orders of magnitude lower cost.
Key insight. Message-passing GNNs initialized with random / basis-vector node features compute nonlinear maps of eigenvectors, unlocking expressive positional encodings without explicit eigendecomposition. A foundational primitive for our GFM work.
Problem. Common belief: Weisfeiler-Lehman bounds GNN expressivity.
Contribution. GNNs can discriminate the majority of real graphs; WL is not the real limit. We design convolutional architectures that are provably more expressive than WL.
Substructures
Problem. Counting substructures (triangles, cycles, paths, cliques) is fundamental to chemistry, biology, and graph reasoning, yet standard message-passing GNNs are bounded by 1-WL and provably cannot count most of them.
Contribution. A theoretical and architectural framework characterizing exactly when GNNs can count substructures, together with provably expressive architectures that go beyond 1-WL.
Key insight. Substructure counting is a much finer-grained yardstick for GNN expressivity than the 1-WL test, and reveals the latent power of message-passing models.
Teaching
See the full CV for workshop organization, invited talks, service, and complete awards list.