Overview

Date: December 8, 2018
Location: Palais des Congrès de Montréal, Montréal, Canada

Relational reasoning, i.e., learning and inference with relational data, is key to understanding how objects interact with each other and give rise to complex phenomena in the everyday world. Well-known applications include knowledge base completion and social network analysis. Although many relational datasets are available, integrating them directly into modern machine learning algorithms and systems that rely on continuous, gradient-based optimization and make strong i.i.d. assumptions is challenging. Relational representation learning has the potential to overcome these obstacles: it enables the fusion of recent advancements like deep learning and relational reasoning to learn from high-dimensional data. Success of such methods can facilitate novel applications of relational reasoning in areas such as scene understanding, visual question-answering, understanding chemical and biological processes, program synthesis and analysis, decision-making in multi-agent systems and many others.

How should we rethink classical representation learning theory for relational representations? Classical approaches based on dimensionality reduction techniques such as isoMap and spectral decompositions still serve as strong baselines and are slowly paving the way for modern methods in relational representation learning based on random walks over graphs, message-passing in neural networks, group-invariant deep architectures etc. amongst many others. How can systems be designed and potentially deployed for large scale representation learning? What are promising avenues, beyond traditional applications like knowledge base and social network analysis, that can benefit from relational representation learning?

This workshop aims to bring together researchers from both academia and industry interested in addressing various aspects of representation learning for relational reasoning.

Invited Speakers

Joan Bruna, New York University
Pedro Domingos, University of Washington
Lise Getoor, University of California, Santa Cruz
Timothy Lillicrap, Google Deepmind
Marina Meila, University of Washington
Maximilian Nickel, Facebook Artificial Intelligence Research
Oriol Vinyals, Google Deepmind

Schedule

Coming Soon!

Call for Papers

Topics include, but are not limited to:

We welcome and encourage position papers on this subject. We are also particularly interested in papers that introduce datasets and competitions to further progress in the field.

Submission Guidelines

Workshop papers should be at most 4 pages of content, including text and figures, excluding references. Authors can include an appendix of supplementary material after the references. However, reviewers will not be required to consult any appendices to make their decisions. The main 4-page paper should adequately describe the work and its contributions.

Papers should be anonymized and adhere to the NIPS conference format: https://nips.cc/Conferences/2018/PaperInformation/StyleFiles

Submission Site: https://cmt3.research.microsoft.com/R2L2018

Peer Review and Acceptance Criteria

All submissions will go through a double-blind peer review process. Accepted papers will be chosen based on techincal merit, interest, and novelty. The workshop allows submissions of papers that are under review or have been recently published in a conference or a journal. Authors should state any overlapping published work at the time of submission.

All accepted papers will be included in one of the two poster presentation and lightning talk sessions on the day of the workshop. Some accepted papers will be invited to give contributed oral talks. Final versions of accepted papers will be posted on the workshop website. These are archival but do not constitute a proceedings and can be submitted elsewhere.

Important Dates

Organizers

Aditya Grover, Stanford University
Paroma Varma, Stanford University
Fred Sala, Stanford University
Steven Holtzen, University of California, Los Angeles
Jennifer Neville, Purdue University
Stefano Ermon, Stanford University
Christopher Ré, Stanford University