NVIDIA Magnum IONVIDIA
|
||||||
Related Products
|
||||||
About
NVIDIA Magnum IO is the architecture for parallel, intelligent data center I/O. It maximizes storage, network, and multi-node, multi-GPU communications for the world’s most important applications, using large language models, recommender systems, imaging, simulation, and scientific research. Magnum IO utilizes storage I/O, network I/O, in-network compute, and I/O management to simplify and speed up data movement, access, and management for multi-GPU, multi-node systems. It supports NVIDIA CUDA-X libraries and makes the best use of a range of NVIDIA GPU and networking hardware topologies to achieve optimal throughput and low latency. In multi-GPU, multi-node systems, slow CPU, single-thread performance is in the critical path of data access from local or remote storage devices. With storage I/O acceleration, the GPU bypasses the CPU and system memory, and accesses remote storage via 8x 200 Gb/s NICs, achieving up to 1.6 TB/s of raw storage bandwidth.
|
About
The world is moving to distributed applications and architectures, and your database should too. Learn how you can deploy where you want, when you want, and how you want with a distributed SQL database. Migrate existing SQL applications to a distributed, multi-node architecture that can dynamically scale out and in. Our Transaction Engines (TEs) and Storage Managers (SMs) work together to ensure ACID compliance across multiple nodes. Deploy in a distributed architecture. When you deploy your database with multiple nodes, the loss of one or multiple nodes will not result in the loss of database access. Deploy TEs and SMs to meet your variable workload needs, or deploy in the different environments the teams in your organization uses: in private and public clouds, in hybrid environments, and across clouds.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
AI researchers, data scientists, and HPC developers needing a tool to eliminate I/O bottlenecks in multi-GPU, multi-node environments
|
Audience
All types of businesses and organizations looking for a distributed SQL database solution
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
No information available.
Free Version
Free Trial
|
Pricing
No information available.
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationNVIDIA
Founded: 1993
United States
www.nvidia.com/en-us/data-center/magnum-io/
|
Company InformationNuoDB
Founded: 2010
United States
www.nuodb.com
|
|||||
Alternatives |
Alternatives |
|||||
|
|
|
|||||
|
|
||||||
|
|
||||||
|
|
|
|||||
Categories |
Categories |
|||||
Database Features
Backup and Recovery
Creation / Development
Data Migration
Data Replication
Data Search
Data Security
Database Conversion
Mobile Access
Monitoring
NOSQL
Performance Analysis
Queries
Relational Interface
Virtualization
Relational Database Features
ACID Compliance
Data Failure Recovery
Multi-Platform
Referential Integrity
SQL DDL Support
SQL DML Support
System Catalog
Unicode Support
|
||||||
Integrations
Apache Spark
CUDA
DQ Studio
DbVisualizer
NVIDIA NetQ
NVIDIA virtual GPU
Nucleon Database Master
|
Integrations
Apache Spark
CUDA
DQ Studio
DbVisualizer
NVIDIA NetQ
NVIDIA virtual GPU
Nucleon Database Master
|
|||||
|
|
|