Amazon Web Services
AI Shanghai Lablet

Amazon Web Services
AI Shanghai Lablet

Businessman reading the newspaper on table

News and Blogs

DGL 1.0 : Empowering Graph Machine Learning for Everyone

We are thrilled to announce the arrival of DGL 1.0, a cutting-edge machine learning framework for deep learning on graphs. A couple of highlights here:

  • 100+ examples of state-of-the-art GNN models, 15+ top-ranked baselines on Open Graph Benchmark (OGB), available for learning and integration
  • 150+ GNN utilities including GNN layers, datasets, graph data transform modules, graph samplers, etc. for building new model architectures or GNN-based solutions
  • Flexible and efficient message passing and sparse matrix abstraction for developing new GNN building blocks
  • Multi-GPU and distributed training capability to scale to graphs of billions of nodes and edges
Learn More >

Feb. 23, 2023

One Main Conference Paper And Two Findings Accepted At EMNLP 2022

We have three papers accepted at EMNLP 2022 (1 main conference and 2 findings).

In the paper “DORE: Document Ordered Relation Extraction based on Generative Framework”, we develop a seq2seq approach for document-level relation extraction. Its main idea is converting the free text generation into a symbolic sequence generation problem and introducing the negative sampling to balance the training objective. We propose an reinforcement learning based approach in “ReLET: A Reinforcement Learning Based Approach for Explainable QA with Entailment Trees” for QA reasoning, which models the reasoning steps in a multi-step reinforcement learning framework, existing works model reasoning steps separately. A new dialog benchmark has been introduced in the paper “Dilogue Meaning Representation for Task-Oriented Dialogue Systems”, which demonstrates the graph structure to describe dialog states, enhancing the capability of representing multiple intents and nested structure.

Oct. 10, 2022

Accelerating Partitioning of Billion-Scale Graphs With DGL V0.9.1

In the latest DGL v0.9.1, we released a new pipeline for preprocess, partition and
dispatch graph of billions of nodes or edges for distributed GNN training. At
its core is a new data format called Chunked Graph Data Format (CGDF) which
stores graph data by chunks. The new pipeline processes data chunks in parallel
which not only reduces the memory requirement of each machine but also
significantly accelerates the entire procedure. 

Learn More >

Sep. 22, 2022

DGL Empowers Accurate Prediction on Organic Reactivity with Graph Neural Network

In the latest issue of Nature Machine Intelligence, Yousun Jung and his research team at Korea Institute of Science and Technology (KAIST) have proposed a novel generalized-template-based graph neural network called LocalTransform for accurate organic reactivity prediction . The generalized reaction templates demonstrate the built-in interpretability in addition to state-of-the-art prediction accuracy. LocalTransform has been implemented with DGL, DGL-LifeSci and PyTorch. 

Learn More >

Sep. 19, 2022

V0.9 Release Highlights

We're excited to announce a new major release DGL v0.9. Here are the highlights:

  • New functionality to bring highly efficient graph analytics into GNN via NVIDIA RAPIDS cuGraph.
  • Built-in FP16 & mixed precision support.
  • New command lines for model inference and graph prediction task.
Learn More >

July. 26, 2022

DGL May 2022 Update Note (0.8.2)

The new 0.8.2 update brings new additions for the GNN community. Here are the highlights:

  • Synthetic Datasets for Developing GNN Explainability Approaches
  • SIGN Diffusion Transform
  • Label Propagation
  • Directional Graph Network Layer
  • Graph Isomorphism Network Layer with Edge Features
  • Feature Masking
  • Row-Normalizer of Features
Learn More >

May. 31, 2022

DGL April 2022 Update Note (0.8.1)

V0.8.1 is a minor release that includes module updates, optimizations, new features and bug fixes.

The highlights:

  • Grouped Reversible Residual Connection for GNNs
  • Laplacian Positional Encoding
  • Random Walk Positional Encoding
  • GraphSAINT Samplers
  • E(n) Equivariant Convolutional Layer
  • Principal Neighbourhood Aggregation Layer
Learn More >

Apr. 22, 2022

V0.8 Release Highlights

We are excited to announce the release of DGL v0.8, which brings many new features as well as improvement on system performance.

The highlights:

  • A major update of the mini-batch sampling pipeline, better customizability, more optimizations.
  • Significant acceleration and code simplification of popular heterogeneous graph NN modules; 11 new off-the-shelf NN modules for building models for link prediction, heterogeneous graph learning and GNN explanation.
  • GNNLens: a DGL empowered tool to visualize and understand graph data using GNN explanation models.
  • New functions to create, transform and augment graph datasets, making it easier to conduct research on graph contrastive learning or repurposing a graph.
  • DGL-Go: a new GNN model training command line tool that utilizes a simple interface so that users can quickly apply GNNs to their problems and orchestrate experiments with state-of-the-art GNN models.
Learn More >

Mar. 4, 2022

There is an accelerated pace of industry adoption of GNNs. At the just-concluded GTC November 2021, CEO Jensen Huang highlighted the importance of GNN to the industry and announced the comprehensive partnership between NVIDIA and the DGL community. “Graphs — a key data structure in modern data science — can now be projected into deep-neural network frameworks with Deep Graph Library, or DGL, a new Python package.”, said Jensen.

DGL team is very excited to see the thriving growth of the DGL community. We and welcome all partnerships about this open-source project. We are HIRING both full-time employees as well as interns.

Learn More >

Nov. 12, 2021

DGL Won “The 2021 OSCAR Peak Open Source Case Award“

On September 17, 2021, at " the 2021 Open Source Industry Conference" hosted by China Academy of Information and Communications Technology (CAICT), DGL [Deep Graph Library] from Amazon Web Services AI Shanghai Lab won " The 2021 OSCAR Peak Open Source Case Award"-Open Source Community and Open Source Project, highlighting technical strength and open source contributions of Amazon Web Services. 
At annual open source industry conferences, the CAICT grandly announced the "OSCAR Peak Open Source Case" award, which aims to establish open source models and better promote the implementation of open source technology in China. In 2021, the award was divided into five categories: "Open Source People", "Open Source Community and Open Source Project", "Open Source Enterprise-Open Source Operations and Ecological Construction", "Open Source Enterprise-Open Source Governance", and "Open Source Innovation (Secondary Development)". In the category of " Open Source Community and Open Source Project", Amazon Web Services DGL won " The 2021 OSCAR Peak Open Source Case Award"-Open Source Community and Open Source Project, which is a National certification and recognition, among 170+ outstanding projects from big companies after reviewed by Cloud computing standards and Open source promotion committee members.
DGL is an easy-to-use, high performance and scalable Python package for deep learning on graphs, and it was announced to be Open Source at the Neurips conference in Dec. 2018. The development team of DGL originally came from NYU and NYU Shanghai. Now, DGL is led by Amazon Web Services AI Shanghai Lab which is led by Professor Zheng Zhang. The project has received widespread attentions and praises since its launch.
Learn More >

Sep. 17, 2021

Open-source Toolkit for OpenHGNN based on DGL Release Highlight

Today, the GAMMA Lab of Beijing University of Posts and Telecommunications officially released OpenHGNN-0.1.0, an Open Source Toolkit for Heterogeneous Graph Neural Network. OpenHGNN is based on DGL [Deep Graph Library] and PyTorch, and includes SOTA models of heterogeneous graph.

Key Features:

  • Easy-to-Use: OpenHGNN provides easy-to-use interfaces to run experiments with models and datasets, using optuna as hyperparameter optimization framework.
  • Extensibility: User can define customized task/model/dataset to apply new models to new scenarios.
  • Efficiency: Efficient APIs using DGL as the backend.
Learn More >

Sep. 14, 2021

Best Paper Award at KDD 2021 Deep Graph Learning Workshop

In collaboration with Shanghai Jiaotong University faculty and students, our new paper "Bag of Tricks for Node Classification with Graph Neural Networks" introduces and motivates several effective tricks-of-the-trade for training graph neural networks (GNN) models on node classification.

These relate to label usage, loss function formulations, and model design tweaks that can significantly improve performance, often to an extent that outweighs the gains from more dramatic changes in the underlying GNN architecture. Among other things, many of the top-ranked models on the Open Graph Bench-mark (OGB) leaderboard benefit from these techniques. This paper has also recently been awarded by the Best Paper at the Deep Learning on Graphs Workshop at KDD 2021.

Learn More >

Aug. 17, 2021

V0.7 Release Highlight

v0.7 brings improvements on the low-level system infrastructure as well as on the high-level user-facing utilities. Many of them involve contributions from the user community. We are grateful to see such a growing trend and welcome more in the future.
Learn More >

Jul. 26, 2021

A new paper accepted by the International Conference on Computer Vision (ICCV) 2021

Our paper "Learning Hierarchical Graph Neural Networks for Image Clustering" has been accepted by the International Conference on Computer Vision (ICCV) 2021. In this paper, we propose a hierarchical graph neural network (GNN) model that learns how to cluster a set of images into an unknown number of identities using a training set of images annotated with labels belonging to a disjoint set of identities.
Our hierarchical GNN uses a novel approach to merge connected components predicted at each level of the hierarchy to form a new graph at the next level. And unlike fully unsupervised hierarchical clustering, the choice of grouping and complexity criteria stems naturally from the supervision in the training set. The resulting method, Hi-LANDER, achieves an average of 54% improvement in F-score and 8% increase in Normalized Mutual Information (NMI) relative to current GNN-based clustering algorithms while also achieving a three-fold decrease in computational cost.

Learn More >

Jul. 22, 2021

A new paper accepted by ICML 2021

Our paper "Graph Neural Networks Inspired by Classical Iterative Algorithms" has been accepted by International Conference on Machine Learning (ICML) 2021 as the long oral talks. In this paper, we propose a novel graph learning framework TWIRLS from the energy minimization perspective. This framework interprets the message passing process on the graph as the energy minimization process with gradient descent methods.

We also derived a novel attention mechanism on the graph through this perspective, by adding the robustness constraint to the energy on graph and optimizing it by iterative reweighted least squares method. Experiments show that TWIRLS can effectively accomplish long-range information propagation problem without oversmoothing, and also has good performance on heterophily graph data and graph adversarial attack.

Learn More >

May 16, 2021

V0.6 Release Highlight

The recent DGL 0.6 release is a major update on many aspects of the project including documentation, APIs, system speed, and scalability. This article highlights some of the new features and enhancements.

Learn More >

Feb. 25, 2021

WebNLG Challenge 2020

We participated in WebNLG Challenge 2020 and ranked #1 in automatic evaluation for both graph-to-text (35 submissions) and text-to-graph (4 submissions) tracks; and our unsupervised approach cycleGT ranked #10 and #3, respectively. Besides, our graph-to-text submission also ranked #1 in the human evaluation. Combined with our first place finish in the automatic evaluation, it indicates that we achieved the top ranking in every possible category.

We also had two submissions accepted to the WebNLG 2020 workshop: our unsupervised approach CycleGT and our systems' report which won the challenge mentioned above. Both unsupervised and supervised system are open-sourced, unsupervised, supervised .

Learn More >

Dec. 18, 2020

DGL Empowers Service for Predictions on Connected Datasets with Graph Neural Networks

Dec 8, 2020, re:invent20, Amazon announced Amazon Neptune ML, a new capability of Amazon Neptune that uses Graph Neural Networks (GNNs), a machine learning (ML) technique purpose-built for graphs, to make easy, fast, and accurate predictions using graph data. Using the Deep Graph Library (DGL), an open-source library to which Amazon contributes, Neptune ML allows you to improve the accuracy of most predictions for graphs by over 50% when compared to making predictions using non-graph methods.

Learn More >

Dec. 08, 2020

V0.5.3 Patch Update

This is a patch release mainly for supporting CUDA 11.0. Now DGL supports CUDA 11.0 and PyTorch 1.7 on Linux/Windows/Mac.

Learn More >

Dec. 04, 2020

V0.5 Release Highlight

The recent DGL 0.5 release is a major update on many aspects of the project including documentation, APIs, system speed and scalability. This article highlights some of the new features and enhancements.

Learn More >

Aug. 26, 2020

Fighting COVID-19 with Deep Graph

Use the new Drug Repurposing Knowledge Graph (DRKG) for repurposing drugs for fighting COVID-19. A step-by-step tutorial on how to use knowledge graph embeddings learned by DGL-KE to make prediction and rank drugs, and some news coverage. DRKG is in SAIL’2021 top30 list.

Learn More >

Jun. 09, 2020

Introducing Amazon SageMaker Support for Deep Graph Library (DGL): Build and Train Graph Neural Networks

Dec 3, 2019, re:Invent19. Amazon SageMaker support for the Deep Graph Library (DGL) is now available. With DGL, you can improve the prediction accuracy of recommendation, fraud detection, and drug discovery systems using Graph Neural Networks (GNNs).

Learn More >

Dec. 03, 2019