CRYPTOREPORTCLUB
  • Crypto news
  • AI
  • Technologies
Tuesday, June 24, 2025
No Result
View All Result
CRYPTOREPORTCLUB
  • Crypto news
  • AI
  • Technologies
No Result
View All Result
CRYPTOREPORTCLUB

New framework reduces memory usage and boosts energy efficiency for large-scale AI graph analysis

June 23, 2025
150
0

June 23, 2025

The GIST New framework reduces memory usage and boosts energy efficiency for large-scale AI graph analysis

Related Post

AI is consuming more power than the grid can handle. Nuclear might be the answer

AI is consuming more power than the grid can handle. Nuclear might be the answer

June 24, 2025
Semantic watermarks for AI image recognition can be easily manipulated

Semantic watermarks for AI image recognition can be easily manipulated

June 23, 2025
Gaby Clark

scientific editor

Robert Egan

associate editor

Editors' notes

This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

trusted source

proofread

Real-time, large-scale graph neural network inference through BingoCGN
BingoCGN employs cross-partition message quantization to summarize inter-partition message flow, which eliminates the need for irregular off-chip memory access and utilizes a fine-grained structured strong lottery theory-based training algorithm to improve computational efficiency. Credit: Institute of Science Tokyo, Japan

BingoCGN, a scalable and efficient graph neural network accelerator that enables inference of real-time, large-scale graphs through graph partitioning, has been developed by researchers at the Institute of Science Tokyo, Japan. This breakthrough framework utilizes an innovative cross-partition message quantization technique and a novel training algorithm to significantly reduce memory demands and increase computational and energy efficiency.

Graph neural networks (GNNs) are powerful artificial intelligence (AI) models designed for analyzing complex, unstructured graph data. In such data, entities are represented as nodes and relationships between them are the edges. GNNs have been successfully employed in many real-world applications, including social networks, drug discovery, autonomous driving, and recommendation systems. Despite their potential, achieving real-time, large-scale GNN inference, critical for tasks like autonomous driving, remains challenging.

Large graphs require extensive memory, often overflowing on-chip buffers, which are memory regions integrated into a chip. This forces the system to rely on slower off-chip memory. Since graph data is stored irregularly, this leads to irregular memory access patterns, degrading computational efficiency and increasing energy consumption.

One promising solution is graph partitioning, where large graphs are divided into smaller graphs, each assigned its own on-chip buffer. This results in more localized memory access patterns and smaller buffer size requirements as the number of partitions increases.

However, this is only partially effective. As the number of partitions grows, the links between the partitions and inter-partition edges grow substantially. This requires increased off-chip memory access, limiting scalability.

To address this issue, a research team led by Associate Professor Daichi Fujiki from Institute of Science Tokyo, Japan, developed a novel, scalable and efficient GNN accelerator called BingoCGN. "BingoCGN employs a new technique called cross-partition message quantization (CMQ) that summarizes inter-partition message flow, eliminating irregular off-chip memory access, and a new training algorithm that significantly boosts computational efficiency," explains Fujiki. Their findings will be presented at the Proceedings of the 52nd Annual International Symposium on Computer Architecture (ISCA '25) from June 21–25, 2025.

CMQ uses a technique called vector quantization, which clusters inter-partition nodes and represents them using points called centroids. Nodes are clustered based on their distance, with each node assigned to its nearest centroid. For a given partition, these centroids replace the inter-partition nodes, effectively compressing node data. The centroids are stored in tables called codebooks, which reside directly in the on-chip buffer.

CMQ, therefore, allows inter-partition communication without the need for irregular and costly off-chip memory access. Additionally, since this method requires frequent reading and writing of nodes and centroids to memory, this method employs a hierarchical tree-like structure for codebooks, with parent and child centroids, reducing computation demands while maintaining accuracy.

While CMQ solves the memory bottleneck, it shifts the burden to computation. To counter this, the researchers developed a novel training algorithm based on strong lottery ticket theory. In this method, the GNN is initialized with random weights, generated on-chip using random number generators.

Then, unnecessary weights are pruned using a mask, forming a smaller, less dense or sparse sub-network that has comparable accuracy to the full GNN but is significantly more efficient to compute. Further, this method incorporates fine-grained (FG) structured pruning, which uses multiple masks with different levels of sparsity, to construct an even smaller and more efficient sub-network.

"Through these techniques, BingoCGN achieves high-performance GNN inference even on finely partitioned graph data, which was previously considered difficult," remarks Fujiki. "Our hardware implementation, tested on seven real-world datasets, achieves up to 65-fold speedup and up-to 107-fold increase in energy-efficiency compared to state-of-the-art accelerator FlowGNN."

This breakthrough opens the door to real-time processing of large-scale graph data, paving the way for diverse real-world applications of GNNs.

More information: Jiale Yan et al, BingoGCN: Towards Scalable and Efficient GNN Acceleration with Fine-Grained Partitioning and SLT, Proceedings of the 52nd Annual International Symposium on Computer Architecture (2025). DOI: 10.1145/3695053.3731115

Provided by Institute of Science Tokyo Citation: New framework reduces memory usage and boosts energy efficiency for large-scale AI graph analysis (2025, June 23) retrieved 23 June 2025 from https://techxplore.com/news/2025-06-framework-memory-usage-boosts-energy.html This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Novel out-of-core mechanism introduced for large-scale graph neural network training shares

Feedback to editors

Share212Tweet133ShareShare27ShareSend

Related Posts

AI is consuming more power than the grid can handle. Nuclear might be the answer
AI

AI is consuming more power than the grid can handle. Nuclear might be the answer

June 24, 2025
0

June 23, 2025 The GIST AI is consuming more power than the grid can handle. Nuclear might be the answer Lisa Lock scientific editor Andrew Zinin lead editor Editors' notes This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while...

Read moreDetails
Semantic watermarks for AI image recognition can be easily manipulated

Semantic watermarks for AI image recognition can be easily manipulated

June 23, 2025
Here’s why the public needs to challenge the ‘good AI’ myth pushed by tech companies

Here’s why the public needs to challenge the ‘good AI’ myth pushed by tech companies

June 23, 2025
AI applications are producing cleaner cities, smarter homes and more efficient transit

AI applications are producing cleaner cities, smarter homes and more efficient transit

June 23, 2025
The law relies on being precise. AI is disrupting that

The law relies on being precise. AI is disrupting that

June 23, 2025
Bilinear sequence regression model shows why AI excels at learning from word sequences

Bilinear sequence regression model shows why AI excels at learning from word sequences

June 20, 2025
AI image models gain creative edge by amplifying low-frequency features

AI image models gain creative edge by amplifying low-frequency features

June 20, 2025

Recent News

Apple iOS 26: Everything you need to know about the next iPhone operating system

Apple iOS 26: Everything you need to know about the next iPhone operating system

June 24, 2025
The Xbox PC app will soon integrate your various game libraries into one place

The Xbox PC app will soon integrate your various game libraries into one place

June 24, 2025

BTC and SOL Lead the Top Layer-1 Projects By Social Activity, ETH Follows With 139.3K Engaged Posts

June 24, 2025
AI is consuming more power than the grid can handle. Nuclear might be the answer

AI is consuming more power than the grid can handle. Nuclear might be the answer

June 24, 2025

TOP News

  • The best Android phones for 2023

    The best Android phones for 2023

    573 shares
    Share 229 Tweet 143
  • Google’s new AI Core update for Pixel 8 Pro will boost its powers and performance

    559 shares
    Share 224 Tweet 140
  • Shiba Inu Price Prediction Today

    618 shares
    Share 247 Tweet 155
  • North Korean Hackers Pose as South Korean Government Officials to Steal Crypto

    596 shares
    Share 238 Tweet 149
  • My go-to robot vacuum and mop is still $455 off following Cyber Monday

    549 shares
    Share 220 Tweet 137
  • About Us
  • Contact Us
  • Privacy Policy
  • Terms of Use
Advertising: digestmediaholding@gmail.com

Disclaimer: Information found on cryptoreportclub.com is those of writers quoted. It does not represent the opinions of cryptoreportclub.com on whether to sell, buy or hold any investments. You are advised to conduct your own research before making any investment decisions. Use provided information at your own risk.
cryptoreportclub.com covers fintech, blockchain and Bitcoin bringing you the latest crypto news and analyses on the future of money.

© 2023-2025 Cryptoreportclub. All Rights Reserved

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In
No Result
View All Result
  • Crypto news
  • AI
  • Technologies

Disclaimer: Information found on cryptoreportclub.com is those of writers quoted. It does not represent the opinions of cryptoreportclub.com on whether to sell, buy or hold any investments. You are advised to conduct your own research before making any investment decisions. Use provided information at your own risk.
cryptoreportclub.com covers fintech, blockchain and Bitcoin bringing you the latest crypto news and analyses on the future of money.

© 2023-2025 Cryptoreportclub. All Rights Reserved