Skip to content

Commit d06f0d1

Browse files
author
Miltos Allamanis
committed
Add Zugner et al.
1 parent 89a76c1 commit d06f0d1

1 file changed

Lines changed: 12 additions & 0 deletions

File tree

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,12 @@
1+
---
2+
layout: publication
3+
title: "Language-Agnostic Representation Learning of Source Code from Structure and Context"
4+
authors: Daniel Zügner, Tobias Kirschstein, Michele Catasta, Jure Leskovec, Stephan Günnemann
5+
conference: ICLR
6+
year: 2021
7+
bibkey: zugner2021language
8+
additional_links:
9+
- {name: "ArXiV", url: "https://arxiv.org/abs/2103.11318"}
10+
tags: ["transformers", "representation"]
11+
---
12+
Source code (Context) and its parsed abstract syntax tree (AST; Structure) are two complementary representations of the same computer program. Traditionally, designers of machine learning models have relied predominantly either on Structure or Context. We propose a new model, which jointly learns on Context and Structure of source code. In contrast to previous approaches, our model uses only language-agnostic features, i.e., source code and features that can be computed directly from the AST. Besides obtaining state-of-the-art on monolingual code summarization on all five programming languages considered in this work, we propose the first multilingual code summarization model. We show that jointly training on non-parallel data from multiple programming languages improves results on all individual languages, where the strongest gains are on low-resource languages. Remarkably, multilingual training only from Context does not lead to the same improvements, highlighting the benefits of combining Structure and Context for representation learning on code.

0 commit comments

Comments
 (0)