📃
GPU Network Whitepaper
  • ABSTRACT
  • INTRODUCTION
    • Overview
    • Problem Statement
    • Vision
  • MARKET POTENTIAL
    • Market Potential
  • POTENTIAL USE CASES
    • AI/ML Training and Inference
    • Rendering and Animation
    • Gaming and Virtual Reality
    • Generative AI tools
    • Scientific Computing
    • Mining
  • TECHNICAL ARCHITECTURE
    • Overview
    • Technical Diagram
    • GPU Provider Nodes
    • GPU Consumer Nodes
    • Task Distribution and Execution
  • FULLY HOMOMORPHIC ENCRYPTION
    • FHE (Fully Homomorphic Encryption)
  • $GPU TOKEN
    • Tokenomics
    • $GPU Utility
    • Airdrop
  • Network Resilience/Fault Tolerance
    • Network Resilience/Fault Tolerance
  • CONCLUSION
    • Conclusion
    • Disclaimer
Powered by GitBook
On this page
  1. TECHNICAL ARCHITECTURE

Overview

We propose an efficient approximate top-k gradient sparsification algorithm on GPUs to compress the communication data with very slight computation overheads.

We present a novel hierarchical communication algorithm to aggregate sparsified gradients to better utilize the bandwidth on GPU clusters that are with fast interconnects within nodes and slow interconnects between nodes.

We perform distributed training experiments with two types of models, CNNs and Transformer, to demonstrate the effectiveness of the proposed techniques on a Tencent Cloud cluster with 16 nodes connected with 25 Gbps Ethernet (each node has 8 Nvidia Tesla V100- 32GB GPUs with NVLink). Experimental results show that our system achieves 25% − 40% faster than existing state-of-the-art systems.

PreviousMiningNextTechnical Diagram

Last updated 1 year ago