Amazon Web Services
AI Shanghai Lablet

Amazon Web Services
AI Shanghai Lablet


News and Blogs
DGL 1.0 : Empowering Graph Machine Learning for Everyone
We are thrilled to announce the arrival of DGL 1.0, a cutting-edge machine learning framework for deep learning on graphs. A couple of highlights here:
- 100+ examples of state-of-the-art GNN models, 15+ top-ranked baselines on Open Graph Benchmark (OGB), available for learning and integration
- 150+ GNN utilities including GNN layers, datasets, graph data transform modules, graph samplers, etc. for building new model architectures or GNN-based solutions
- Flexible and efficient message passing and sparse matrix abstraction for developing new GNN building blocks
- Multi-GPU and distributed training capability to scale to graphs of billions of nodes and edges
Feb. 23, 2023
One Main Conference Paper And Two Findings Accepted At EMNLP 2022
We have three papers accepted at EMNLP 2022 (1 main conference and 2 findings).
In the paper “DORE: Document Ordered Relation Extraction based on Generative Framework”, we develop a seq2seq approach for document-level relation extraction. Its main idea is converting the free text generation into a symbolic sequence generation problem and introducing the negative sampling to balance the training objective. We propose an reinforcement learning based approach in “ReLET: A Reinforcement Learning Based Approach for Explainable QA with Entailment Trees” for QA reasoning, which models the reasoning steps in a multi-step reinforcement learning framework, existing works model reasoning steps separately. A new dialog benchmark has been introduced in the paper “Dilogue Meaning Representation for Task-Oriented Dialogue Systems”, which demonstrates the graph structure to describe dialog states, enhancing the capability of representing multiple intents and nested structure.
Oct. 10, 2022
Accelerating Partitioning of Billion-Scale Graphs With DGL V0.9.1
In the latest DGL v0.9.1, we released a new pipeline for preprocess, partition and
dispatch graph of billions of nodes or edges for distributed GNN training. At
its core is a new data format called Chunked Graph Data Format (CGDF) which
stores graph data by chunks. The new pipeline processes data chunks in parallel
which not only reduces the memory requirement of each machine but also
significantly accelerates the entire procedure.
Sep. 22, 2022
DGL Empowers Accurate Prediction on Organic Reactivity with Graph Neural Network
In the latest issue of Nature Machine Intelligence, Yousun Jung and his research team at Korea Institute of Science and Technology (KAIST) have proposed a novel generalized-template-based graph neural network called LocalTransform for accurate organic reactivity prediction . The generalized reaction templates demonstrate the built-in interpretability in addition to state-of-the-art prediction accuracy. LocalTransform has been implemented with DGL, DGL-LifeSci and PyTorch.
Sep. 19, 2022
V0.9 Release Highlights
We're excited to announce a new major release DGL v0.9. Here are the highlights:
- New functionality to bring highly efficient graph analytics into GNN via NVIDIA RAPIDS cuGraph.
- Built-in FP16 & mixed precision support.
- New command lines for model inference and graph prediction task.
July. 26, 2022
DGL May 2022 Update Note (0.8.2)
The new 0.8.2 update brings new additions for the GNN community. Here are the highlights:
- Synthetic Datasets for Developing GNN Explainability Approaches
- SIGN Diffusion Transform
- Label Propagation
- Directional Graph Network Layer
- Graph Isomorphism Network Layer with Edge Features
- Feature Masking
- Row-Normalizer of Features
May. 31, 2022
DGL April 2022 Update Note (0.8.1)
V0.8.1 is a minor release that includes module updates, optimizations, new features and bug fixes.
The highlights:
- Grouped Reversible Residual Connection for GNNs
- Laplacian Positional Encoding
- Random Walk Positional Encoding
- GraphSAINT Samplers
- E(n) Equivariant Convolutional Layer
- Principal Neighbourhood Aggregation Layer
Apr. 22, 2022
V0.8 Release Highlights
We are excited to announce the release of DGL v0.8, which brings many new features as well as improvement on system performance.
The highlights:
- A major update of the mini-batch sampling pipeline, better customizability, more optimizations.
- Significant acceleration and code simplification of popular heterogeneous graph NN modules; 11 new off-the-shelf NN modules for building models for link prediction, heterogeneous graph learning and GNN explanation.
- GNNLens: a DGL empowered tool to visualize and understand graph data using GNN explanation models.
- New functions to create, transform and augment graph datasets, making it easier to conduct research on graph contrastive learning or repurposing a graph.
- DGL-Go: a new GNN model training command line tool that utilizes a simple interface so that users can quickly apply GNNs to their problems and orchestrate experiments with state-of-the-art GNN models.
Mar. 4, 2022
DGL Recommended by NVIDIA CEO Jensen Huang in GTC November 2021 Keynote
There is an accelerated pace of industry adoption of GNNs. At the just-concluded GTC November 2021, CEO Jensen Huang highlighted the importance of GNN to the industry and announced the comprehensive partnership between NVIDIA and the DGL community. “Graphs — a key data structure in modern data science — can now be projected into deep-neural network frameworks with Deep Graph Library, or DGL, a new Python package.”, said Jensen.
DGL team is very excited to see the thriving growth of the DGL community. We and welcome all partnerships about this open-source project. We are HIRING both full-time employees as well as interns.
Nov. 12, 2021
DGL Won “The 2021 OSCAR Peak Open Source Case Award“
Sep. 17, 2021
Open-source Toolkit for OpenHGNN based on DGL Release Highlight
Today, the GAMMA Lab of Beijing University of Posts and Telecommunications officially released OpenHGNN-0.1.0, an Open Source Toolkit for Heterogeneous Graph Neural Network. OpenHGNN is based on DGL [Deep Graph Library] and PyTorch, and includes SOTA models of heterogeneous graph.
Key Features:
- Easy-to-Use: OpenHGNN provides easy-to-use interfaces to run experiments with models and datasets, using optuna as hyperparameter optimization framework.
- Extensibility: User can define customized task/model/dataset to apply new models to new scenarios.
- Efficiency: Efficient APIs using DGL as the backend.
Sep. 14, 2021
Best Paper Award at KDD 2021 Deep Graph Learning Workshop
In collaboration with Shanghai Jiaotong University faculty and students, our new paper "Bag of Tricks for Node Classification with Graph Neural Networks" introduces and motivates several effective tricks-of-the-trade for training graph neural networks (GNN) models on node classification.
These relate to label usage, loss function formulations, and model design tweaks that can significantly improve performance, often to an extent that outweighs the gains from more dramatic changes in the underlying GNN architecture. Among other things, many of the top-ranked models on the Open Graph Bench-mark (OGB) leaderboard benefit from these techniques. This paper has also recently been awarded by the Best Paper at the Deep Learning on Graphs Workshop at KDD 2021.
Aug. 17, 2021
V0.7 Release Highlight
Jul. 26, 2021
A new paper accepted by the International Conference on Computer Vision (ICCV) 2021
Our paper "Learning Hierarchical Graph Neural Networks for Image Clustering" has been accepted by the International Conference on Computer Vision (ICCV) 2021. In this paper, we propose a hierarchical graph neural network (GNN) model that learns how to cluster a set of images into an unknown number of identities using a training set of images annotated with labels belonging to a disjoint set of identities.
Our hierarchical GNN uses a novel approach to merge connected components predicted at each level of the hierarchy to form a new graph at the next level. And unlike fully unsupervised hierarchical clustering, the choice of grouping and complexity criteria stems naturally from the supervision in the training set. The resulting method, Hi-LANDER, achieves an average of 54% improvement in F-score and 8% increase in Normalized Mutual Information (NMI) relative to current GNN-based clustering algorithms while also achieving a three-fold decrease in computational cost.
Jul. 22, 2021
A new paper accepted by ICML 2021
Our paper "Graph Neural Networks Inspired by Classical Iterative Algorithms" has been accepted by International Conference on Machine Learning (ICML) 2021 as the long oral talks. In this paper, we propose a novel graph learning framework TWIRLS from the energy minimization perspective. This framework interprets the message passing process on the graph as the energy minimization process with gradient descent methods.
We also derived a novel attention mechanism on the graph through this perspective, by adding the robustness constraint to the energy on graph and optimizing it by iterative reweighted least squares method. Experiments show that TWIRLS can effectively accomplish long-range information propagation problem without oversmoothing, and also has good performance on heterophily graph data and graph adversarial attack.
May 16, 2021
V0.6 Release Highlight
The recent DGL 0.6 release is a major update on many aspects of the project including documentation, APIs, system speed, and scalability. This article highlights some of the new features and enhancements.
Feb. 25, 2021
WebNLG Challenge 2020
We participated in WebNLG Challenge 2020 and ranked #1 in automatic evaluation for both graph-to-text (35 submissions) and text-to-graph (4 submissions) tracks; and our unsupervised approach cycleGT ranked #10 and #3, respectively. Besides, our graph-to-text submission also ranked #1 in the human evaluation. Combined with our first place finish in the automatic evaluation, it indicates that we achieved the top ranking in every possible category.
We also had two submissions accepted to the WebNLG 2020 workshop: our unsupervised approach CycleGT and our systems' report which won the challenge mentioned above. Both unsupervised and supervised system are open-sourced, unsupervised, supervised .
Dec. 18, 2020
DGL Empowers Service for Predictions on Connected Datasets with Graph Neural Networks
Dec 8, 2020, re:invent20, Amazon announced Amazon Neptune ML, a new capability of Amazon Neptune that uses Graph Neural Networks (GNNs), a machine learning (ML) technique purpose-built for graphs, to make easy, fast, and accurate predictions using graph data. Using the Deep Graph Library (DGL), an open-source library to which Amazon contributes, Neptune ML allows you to improve the accuracy of most predictions for graphs by over 50% when compared to making predictions using non-graph methods.
Dec. 08, 2020
V0.5.3 Patch Update
This is a patch release mainly for supporting CUDA 11.0. Now DGL supports CUDA 11.0 and PyTorch 1.7 on Linux/Windows/Mac.
Dec. 04, 2020
V0.5 Release Highlight
The recent DGL 0.5 release is a major update on many aspects of the project including documentation, APIs, system speed and scalability. This article highlights some of the new features and enhancements.
Aug. 26, 2020
Fighting COVID-19 with Deep Graph
Use the new Drug Repurposing Knowledge Graph (DRKG) for repurposing drugs for fighting COVID-19. A step-by-step tutorial on how to use knowledge graph embeddings learned by DGL-KE to make prediction and rank drugs, and some news coverage. DRKG is in SAIL’2021 top30 list.
Jun. 09, 2020
Introducing Amazon SageMaker Support for Deep Graph Library (DGL): Build and Train Graph Neural Networks
Dec 3, 2019, re:Invent19. Amazon SageMaker support for the Deep Graph Library (DGL) is now available. With DGL, you can improve the prediction accuracy of recommendation, fraud detection, and drug discovery systems using Graph Neural Networks (GNNs).