Dragon-Alpha cu32: A Java-based Tensor Computing Framework With its High-Performance CUDA Library

05/15/2023
by   Zhiyi Zhang, et al.
0

Java is very powerful, but in Deep Learning field, its capabilities probably has not been sufficiently exploited. Compared to the Java-based deep-learning-frameworks, the Python-based (PyTorch, TensorFlow, etc) are undoubtedly the mainstream, due to their easy-to-use, flexibility and better ecosystem. Dragon-Alpha is a Java-based Tensor Computing Framework, with easy-to-use, high-scalability and high-performance, trying to break Java's dilemma in deep learning field and make it more effective. Dragon-Alpha supports different levels of APIs, and can be used as a deep-learning-framework through its user-friendly high-level APIs. Dragon-Alpha has potential to aggregate computing-power across heterogeneous platforms and devices, based on its multi-layer architecture and Java's big-data ecosystem. Dragon-Alpha has its asynchronized APIs to improve parallelism, and highly-optimized CUDA library cu32 which adopts unique convolutionoperators for small feature maps. The experiments show that, compared to PyTorch cuDNN, Dragon-Alpha cu32 costs less time and memory (75.38 66.4 on Cifar-10.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/13/2020

Towards High Performance Java-based Deep Learning Frameworks

The advent of modern cloud services along with the huge volume of data p...
research
04/11/2018

Flexible and Scalable Deep Learning with MMLSpark

In this work we detail a novel open source library, called MMLSpark, tha...
research
05/08/2020

High Performance Cluster Computing for MapReduce

MapReduce is a technique used to vastly improve distributed processing o...
research
03/08/2023

SMaLL: A Software Framework for portable Machine Learning Libraries

Interest in deploying Deep Neural Network (DNN) inference on edge device...
research
09/23/2015

IllinoisSL: A JAVA Library for Structured Prediction

IllinoisSL is a Java library for learning structured prediction models. ...
research
09/28/2022

Accelerating netty-based applications through transparent InfiniBand support

Many big-data frameworks are written in Java, e.g. Apache Spark, Flink a...
research
07/27/2021

HPTMT: Operator-Based Architecture for Scalable High-Performance Data-Intensive Frameworks

Data-intensive applications impact many domains, and their steadily incr...

Please sign up or login with your details

Forgot password? Click here to reset