Amazon Web Services
AI Shanghai Lablet

Amazon Web Services
AI Shanghai Lablet

Businessman reading the newspaper on table

News and Blogs

DGL 2.1: GPU Acceleration For Your GNN Data Pipeline

We are happy to announce the release of DGL 2.1. In this release, we are making GNN data loading lightning fast. We introduce GPU acceleration for the whole GNN data loading pipeline in GraphBolt, including the graph sampling and feature fetching stages.

Learn more.

March. 6, 2024

DGL 2.0: Streamlining Your GNN Data Pipeline From Bottleneck To Boost

We’re thrilled to announce the release of DGL 2.0, a major milestone in our mission to empower developers with cutting-edge tools for Graph Neural Networks (GNNs). Traditionally, data loading has been a significant bottleneck in GNN training. Complex graph structures and the need for efficient sampling often lead to slow data loading times and resource constraints. This can drastically hinder the training speed and scalability of your GNN models. DGL 2.0 breaks free from these limitations with the introduction of dgl.graphbolt, a revolutionary data loading framework that supercharges your GNN training by streamlining the data pipeline.

Learn more.

Jan. 26, 2024

BSChecker: A New Benchmark and Tool for Detecting Hallucinations in Large Language Models (LLMs)

We announces the launch of BSChecker, a new tool designed to identify subtle hallucinations in the outputs of large language models (LLMs). BSChecker provides

  • a hallucination detection benchmark that covers three distinctive settings differentiated by the quantity and quality of context provided for LLM responses;
  • a hallucination detection framework that checks hallucinations at finer granularity of knowledge triplets;
  • leaderboards by both human annotations and auto-checkers to evaluate LLMs on hallucinations.

Learn More

Dec. 7, 2023

Four Papers Accepted at ICCV 2023

Unsupervised Open-Vocabulary Object Localization in Videos: With the inductive bias of spatiotemporal consistency, we localize objects from videos in an unsupervised manner, further with a finetuned CLIP model we can also label semantics of each object.

Object-centric Multiple Object Tracking: OC-MOT significantly narrows the gap between the existing object-centric model and the fully supervised state-of-the-art MOT model by leveraging the versatile object-centric representation and the designed self-supervised memory module.

Coarse-to-Fine Amodal Segmentation with Shape Prior: we propose a novel approach, called Coarse-to-Fine Segmentation (C2F-Seg), to progressively model the amodal segmentation: initially reduces the learning space from the pixel-level image space to the VQ latent space, then refine with more details in pixel space.

Rethinking Amodal Video Segmentation from Learning Supervised Signals with Object-centric Representation: we propose a multi-view fusion layer based temporal module to fulfill sufficient object representation completion. Besides views from different frames, we design a translation module to project image features into the Bird’s-Eye View (BEV), which introduces 3D information to improve feature quality.

Oct. 2, 2023

Four Papers Accepted at ACL 2023

In the paper “An AMR-based Link Prediction Approach for Document-level Event Argument Extraction”, we proposed a method to build the Abstract Meaning Representation graph for a document whose nodes are normalized nouns and verbs. The edges are the pre-defined semantic roles, such as the subject and object of an action. We apply GNN+Transformer over an AMR graph to extract event arguments.

The idea in “Exploiting Abstract Meaning Representation for Open-Domain Question Answering” is similar to the above work but used in a different task with a different model architecture. Inspired by studies in Graph Learning area, we proposed a method to compress the information of edges as tokens and append these tokens to the text sequence. A transformer takes the concatenation of text tokens and graph tokens as inputs and generates the answer for the given question.

In the paper “Dual Cache for Long Document Neural Coreference Resolution”, we proposed a hybrid cache following LRU (Least recently used) and LFU (Least Frequently used) to maintain the candidate text span that is a reference of other entities when processing a document from left to right. Since the model only considers a fixed number of candidates in the cache, we save the memory and computation overhead without losing too much performance.

Since pre-defined and human-labeled discourse relations can be ambiguous when describing the semantics between sentences, we proposed to adopt a distributed representation to express the relationship between two sentences. We also propose an unsupervised learning method to obtain such distributed representations in the paper “Distributed Marker Representation for Ambiguous Discourse Markers and Entangled Relations”.

Jul. 18, 2023

DGL 1.0 : Empowering Graph Machine Learning for Everyone

We are thrilled to announce the arrival of DGL 1.0, a cutting-edge machine learning framework for deep learning on graphs. A couple of highlights here:

  • 100+ examples of state-of-the-art GNN models, 15+ top-ranked baselines on Open Graph Benchmark (OGB), available for learning and integration
  • 150+ GNN utilities including GNN layers, datasets, graph data transform modules, graph samplers, etc. for building new model architectures or GNN-based solutions
  • Flexible and efficient message passing and sparse matrix abstraction for developing new GNN building blocks
  • Multi-GPU and distributed training capability to scale to graphs of billions of nodes and edges
Learn More >

Feb. 23, 2023

One Main Conference Paper And Two Findings Accepted At EMNLP 2022

We have three papers accepted at EMNLP 2022 (1 main conference and 2 findings).

In the paper “DORE: Document Ordered Relation Extraction based on Generative Framework”, we develop a seq2seq approach for document-level relation extraction. Its main idea is converting the free text generation into a symbolic sequence generation problem and introducing the negative sampling to balance the training objective. We propose an reinforcement learning based approach in “ReLET: A Reinforcement Learning Based Approach for Explainable QA with Entailment Trees” for QA reasoning, which models the reasoning steps in a multi-step reinforcement learning framework, existing works model reasoning steps separately. A new dialog benchmark has been introduced in the paper “Dilogue Meaning Representation for Task-Oriented Dialogue Systems”, which demonstrates the graph structure to describe dialog states, enhancing the capability of representing multiple intents and nested structure.

Oct. 10, 2022

Accelerating Partitioning of Billion-Scale Graphs With DGL V0.9.1

In the latest DGL v0.9.1, we released a new pipeline for preprocess, partition and
dispatch graph of billions of nodes or edges for distributed GNN training. At
its core is a new data format called Chunked Graph Data Format (CGDF) which
stores graph data by chunks. The new pipeline processes data chunks in parallel
which not only reduces the memory requirement of each machine but also
significantly accelerates the entire procedure. 

Learn More >

Sep. 22, 2022

DGL Empowers Accurate Prediction on Organic Reactivity with Graph Neural Network

In the latest issue of Nature Machine Intelligence, Yousun Jung and his research team at Korea Institute of Science and Technology (KAIST) have proposed a novel generalized-template-based graph neural network called LocalTransform for accurate organic reactivity prediction . The generalized reaction templates demonstrate the built-in interpretability in addition to state-of-the-art prediction accuracy. LocalTransform has been implemented with DGL, DGL-LifeSci and PyTorch. 

Learn More >

Sep. 19, 2022

V0.9 Release Highlights

We're excited to announce a new major release DGL v0.9. Here are the highlights:

  • New functionality to bring highly efficient graph analytics into GNN via NVIDIA RAPIDS cuGraph.
  • Built-in FP16 & mixed precision support.
  • New command lines for model inference and graph prediction task.
Learn More >

July. 26, 2022

DGL May 2022 Update Note (0.8.2)

The new 0.8.2 update brings new additions for the GNN community. Here are the highlights:

  • Synthetic Datasets for Developing GNN Explainability Approaches
  • SIGN Diffusion Transform
  • Label Propagation
  • Directional Graph Network Layer
  • Graph Isomorphism Network Layer with Edge Features
  • Feature Masking
  • Row-Normalizer of Features
Learn More >

May. 31, 2022

DGL April 2022 Update Note (0.8.1)

V0.8.1 is a minor release that includes module updates, optimizations, new features and bug fixes.

The highlights:

  • Grouped Reversible Residual Connection for GNNs
  • Laplacian Positional Encoding
  • Random Walk Positional Encoding
  • GraphSAINT Samplers
  • E(n) Equivariant Convolutional Layer
  • Principal Neighbourhood Aggregation Layer
Learn More >

Apr. 22, 2022

V0.8 Release Highlights

We are excited to announce the release of DGL v0.8, which brings many new features as well as improvement on system performance.

The highlights:

  • A major update of the mini-batch sampling pipeline, better customizability, more optimizations.
  • Significant acceleration and code simplification of popular heterogeneous graph NN modules; 11 new off-the-shelf NN modules for building models for link prediction, heterogeneous graph learning and GNN explanation.
  • GNNLens: a DGL empowered tool to visualize and understand graph data using GNN explanation models.
  • New functions to create, transform and augment graph datasets, making it easier to conduct research on graph contrastive learning or repurposing a graph.
  • DGL-Go: a new GNN model training command line tool that utilizes a simple interface so that users can quickly apply GNNs to their problems and orchestrate experiments with state-of-the-art GNN models.
Learn More >

Mar. 4, 2022

There is an accelerated pace of industry adoption of GNNs. At the just-concluded GTC November 2021, CEO Jensen Huang highlighted the importance of GNN to the industry and announced the comprehensive partnership between NVIDIA and the DGL community. “Graphs — a key data structure in modern data science — can now be projected into deep-neural network frameworks with Deep Graph Library, or DGL, a new Python package.”, said Jensen.

DGL team is very excited to see the thriving growth of the DGL community. We and welcome all partnerships about this open-source project. We are HIRING both full-time employees as well as interns.

Learn More >

Nov. 12, 2021

DGL Won “The 2021 OSCAR Peak Open Source Case Award“

On September 17, 2021, at " the 2021 Open Source Industry Conference" hosted by China Academy of Information and Communications Technology (CAICT), DGL [Deep Graph Library] from Amazon Web Services AI Shanghai Lab won " The 2021 OSCAR Peak Open Source Case Award"-Open Source Community and Open Source Project, highlighting technical strength and open source contributions of Amazon Web Services. 
 
At annual open source industry conferences, the CAICT grandly announced the "OSCAR Peak Open Source Case" award, which aims to establish open source models and better promote the implementation of open source technology in China. In 2021, the award was divided into five categories: "Open Source People", "Open Source Community and Open Source Project", "Open Source Enterprise-Open Source Operations and Ecological Construction", "Open Source Enterprise-Open Source Governance", and "Open Source Innovation (Secondary Development)". In the category of " Open Source Community and Open Source Project", Amazon Web Services DGL won " The 2021 OSCAR Peak Open Source Case Award"-Open Source Community and Open Source Project, which is a National certification and recognition, among 170+ outstanding projects from big companies after reviewed by Cloud computing standards and Open source promotion committee members.
 
DGL is an easy-to-use, high performance and scalable Python package for deep learning on graphs, and it was announced to be Open Source at the Neurips conference in Dec. 2018. The development team of DGL originally came from NYU and NYU Shanghai. Now, DGL is led by Amazon Web Services AI Shanghai Lab which is led by Professor Zheng Zhang. The project has received widespread attentions and praises since its launch.
Learn More >

Sep. 17, 2021

Open-source Toolkit for OpenHGNN based on DGL Release Highlight

Today, the GAMMA Lab of Beijing University of Posts and Telecommunications officially released OpenHGNN-0.1.0, an Open Source Toolkit for Heterogeneous Graph Neural Network. OpenHGNN is based on DGL [Deep Graph Library] and PyTorch, and includes SOTA models of heterogeneous graph.

Key Features:

  • Easy-to-Use: OpenHGNN provides easy-to-use interfaces to run experiments with models and datasets, using optuna as hyperparameter optimization framework.
  • Extensibility: User can define customized task/model/dataset to apply new models to new scenarios.
  • Efficiency: Efficient APIs using DGL as the backend.
Learn More >

Sep. 14, 2021

Best Paper Award at KDD 2021 Deep Graph Learning Workshop

In collaboration with Shanghai Jiaotong University faculty and students, our new paper "Bag of Tricks for Node Classification with Graph Neural Networks" introduces and motivates several effective tricks-of-the-trade for training graph neural networks (GNN) models on node classification.

These relate to label usage, loss function formulations, and model design tweaks that can significantly improve performance, often to an extent that outweighs the gains from more dramatic changes in the underlying GNN architecture. Among other things, many of the top-ranked models on the Open Graph Bench-mark (OGB) leaderboard benefit from these techniques. This paper has also recently been awarded by the Best Paper at the Deep Learning on Graphs Workshop at KDD 2021.

Learn More >

Aug. 17, 2021

V0.7 Release Highlight

v0.7 brings improvements on the low-level system infrastructure as well as on the high-level user-facing utilities. Many of them involve contributions from the user community. We are grateful to see such a growing trend and welcome more in the future.
Learn More >

Jul. 26, 2021

A new paper accepted by the International Conference on Computer Vision (ICCV) 2021

Our paper "Learning Hierarchical Graph Neural Networks for Image Clustering" has been accepted by the International Conference on Computer Vision (ICCV) 2021. In this paper, we propose a hierarchical graph neural network (GNN) model that learns how to cluster a set of images into an unknown number of identities using a training set of images annotated with labels belonging to a disjoint set of identities.
Our hierarchical GNN uses a novel approach to merge connected components predicted at each level of the hierarchy to form a new graph at the next level. And unlike fully unsupervised hierarchical clustering, the choice of grouping and complexity criteria stems naturally from the supervision in the training set. The resulting method, Hi-LANDER, achieves an average of 54% improvement in F-score and 8% increase in Normalized Mutual Information (NMI) relative to current GNN-based clustering algorithms while also achieving a three-fold decrease in computational cost.

Learn More >

Jul. 22, 2021

A new paper accepted by ICML 2021

Our paper "Graph Neural Networks Inspired by Classical Iterative Algorithms" has been accepted by International Conference on Machine Learning (ICML) 2021 as the long oral talks. In this paper, we propose a novel graph learning framework TWIRLS from the energy minimization perspective. This framework interprets the message passing process on the graph as the energy minimization process with gradient descent methods.

We also derived a novel attention mechanism on the graph through this perspective, by adding the robustness constraint to the energy on graph and optimizing it by iterative reweighted least squares method. Experiments show that TWIRLS can effectively accomplish long-range information propagation problem without oversmoothing, and also has good performance on heterophily graph data and graph adversarial attack.

Learn More >

May 16, 2021

V0.6 Release Highlight

The recent DGL 0.6 release is a major update on many aspects of the project including documentation, APIs, system speed, and scalability. This article highlights some of the new features and enhancements.

Learn More >

Feb. 25, 2021

WebNLG Challenge 2020

We participated in WebNLG Challenge 2020 and ranked #1 in automatic evaluation for both graph-to-text (35 submissions) and text-to-graph (4 submissions) tracks; and our unsupervised approach cycleGT ranked #10 and #3, respectively. Besides, our graph-to-text submission also ranked #1 in the human evaluation. Combined with our first place finish in the automatic evaluation, it indicates that we achieved the top ranking in every possible category.

We also had two submissions accepted to the WebNLG 2020 workshop: our unsupervised approach CycleGT and our systems' report which won the challenge mentioned above. Both unsupervised and supervised system are open-sourced, unsupervised, supervised .

Learn More >

Dec. 18, 2020

DGL Empowers Service for Predictions on Connected Datasets with Graph Neural Networks

Dec 8, 2020, re:invent20, Amazon announced Amazon Neptune ML, a new capability of Amazon Neptune that uses Graph Neural Networks (GNNs), a machine learning (ML) technique purpose-built for graphs, to make easy, fast, and accurate predictions using graph data. Using the Deep Graph Library (DGL), an open-source library to which Amazon contributes, Neptune ML allows you to improve the accuracy of most predictions for graphs by over 50% when compared to making predictions using non-graph methods.

Learn More >

Dec. 08, 2020

V0.5.3 Patch Update

This is a patch release mainly for supporting CUDA 11.0. Now DGL supports CUDA 11.0 and PyTorch 1.7 on Linux/Windows/Mac.

Learn More >

Dec. 04, 2020

V0.5 Release Highlight

The recent DGL 0.5 release is a major update on many aspects of the project including documentation, APIs, system speed and scalability. This article highlights some of the new features and enhancements.

Learn More >

Aug. 26, 2020

Fighting COVID-19 with Deep Graph

Use the new Drug Repurposing Knowledge Graph (DRKG) for repurposing drugs for fighting COVID-19. A step-by-step tutorial on how to use knowledge graph embeddings learned by DGL-KE to make prediction and rank drugs, and some news coverage. DRKG is in SAIL’2021 top30 list.

Learn More >

Jun. 09, 2020

Introducing Amazon SageMaker Support for Deep Graph Library (DGL): Build and Train Graph Neural Networks

Dec 3, 2019, re:Invent19. Amazon SageMaker support for the Deep Graph Library (DGL) is now available. With DGL, you can improve the prediction accuracy of recommendation, fraud detection, and drug discovery systems using Graph Neural Networks (GNNs).

Learn More >

Dec. 03, 2019

Start to Build for Free with Amazon Web Services

Start to Build for Free with Amazon Web Services

Close
Hot Contact Us

Hotline Contact Us

1010 0766
Beijing Region
Operated By Sinnet
1010 0966
Ningxia Region
Operated By NWCD