Disk Inspection and Monitoring
smartmontools contains utility programs (smartctl, smartd) to control/monitor storage systems using the Self-Monitoring, Analysis and Reporting Technology System (S.M.A.R.T.) built into most modern ATA and SCSI disks. It is derived from smartsuite.
Your linux deployment tool!
m23 is a free software distribution system (license: GPL), that installs (via network, starting with partitioning and formatting) and administrates (updates, adds / removes software, adds / removes scripts) clients with Debian, (X/K)Ubuntu, LinuxMint, openSUSE, Fedora and CentOS. It is used for deployment of Linux clients in schools, institutions and enterprises. The m23 server is controlled via a web interface. A new m23 client can be installed easily in only three steps. Group functions and mass installation tools make managing a vast number of clients comfortable. The integration of existing clients (with .deb-based distributions) into the system is possible, too. Client backup and server backup are included to avoid data loss. With the integrated virtualisation software, m23 can create and manage virtual m23 clients, that run on real m23 clients or the m23 server. Scripts and software packages (for installation on clients) can be created directly from the m23 web interface.
Ganglia is a scalable distributed monitoring system for high-performance computing systems such as clusters and Grids. It is based on a hierarchical design targeted at federations of clusters. Supports clusters up to 2000 nodes in size.
DRBL provides diskless or systemless environment. It uses distributed hardware resources and makes it possible for clients to fully access local hardware. It also includes Clonezilla, a partition and disk cloning utility similar to Ghost.
Cluster administration tool
ClusterSSH controls a number of xterm windows via a single graphical console window to allow commands to be interactively run on multiple servers over an ssh connection.
The aim of the project is to develop an open source iSCSI target with professional features, works well in enterprise environment under real workload, and is scalable and versatile enough to meet the challenge of future storage needs and developements.
Collectl is a light-weight performance monitoring tool capable of reporting interactively as well as logging to disk. It reports statistics on cpu, disk, infiniband, lustre, memory, network, nfs, process, quadrics, slabs and more in easy to read format.
PelicanHPC is an iso-hybrid (CD or USB) image that let's you set up a high performance computing cluster in a few minutes. A Pelican cluster allows you to do parallel computing using MPI. You can run Pelican on a single multiple core machine to use all cores to solve a problem, or you can network multiple computers together to make a cluster. The frontend node (either a real computer or a virtual machine) boots from the image. The compute nodes boot by PXE, using the frontend node as the server. All of the nodes of the cluster get their filesystems from the same image, so it is guaranteed that all nodes run the the same software. Packages can be added to all nodes using apt-get, thanks to aufs. The bootable image is created by running a single script, which takes advantage of the Debian Live infrastructure.
Postgres-XC is a write-scalable synchronous multi-master PostgreSQL cluster with the following features. 1) Both read and write scalability. 2) Configured with more than one server. 3) Complete global transaction and visibility management.
The open source grid computing solution
JPPF makes it easy to parallelize computationally intensive tasks and execute them on a Grid.
Open HPI is an open source implementation of the SA Forum's Hardware Platform Interface (HPI). HPI provides an abstracted interface to managing computer hardware, typically for chassis and rack based servers.
Generic SCSI target subsystem for Linux SCST allows creation of sophisticated storage devices. It includes generic SCSI target core with SCSI target drivers and backend handlers. There are target drivers for iSCSI, Fibre Channel, SRP, SAS, FCoE, etc.
Based on the source code of Sun Grid Engine, we are continuing the open source development effort here on sourceforge.
Legacy Release only. Get latest Edition here: http://www.openqrm.com.
openQRM is a web-based open source datacenter management and hybrid cloud computing platform that integrates flexibly with existing components in enterprise data centers. Legacy Release available on Sourceforge. Please visit http://www.openqrm-enterprise.com/index-2.html for latest 5.3 free Community Release. Includes important security updates, bugfixes and enhancements especially for KVM and Cloud plugin. openQRM abstracts Virtualization and supports the following virtualization technologies: KVM, XEN, Citrix XenServer, VMWare ESX, LXC, OpenVZ. The Hybrid Cloud Connector in openQRM supports a range of private or public cloud providers to extend your infrastructure on demand via Amazon AWS, Eucalyptus or OpenStack. openQRM automates provisioning, virtualization, storage and configuration management, and it takes care of high-availability. A self-service cloud portal with integrated billing system enables end-users to request new servers and application stacks on-deman
Reliable group communication based on IP multicast and configurable protocol stack
Tools for the Linux Kernel's network block device, allowing you to use remote block devices over a TCP/IP network. Note that recent downloads are also available from github: http://github.com/yoe/nbd
Manage node sets, node groups and execute commands on cluster nodes in parallel. Provides an event-based Python library to improve administration of large compute clusters or server farms. Command line tools: clush and nodeset included.
Open Source High Availability Middleware Generally Based on SA Forum Specifications
UNFS3 is a user-space NFSv3 server
UNFS3 is a user-space implementation of the NFSv3 server specification. It provides a unified mountd and nfsd program and so far runs under Linux, NetBSD, and Solaris.
Fault tolerant, POSIX-compliant, network distributed filesystem
MooseFS (MFS) is a fault tolerant, highly performing, scaling-out, network distributed file system. It spreads data over several physical servers which are visible to the user as one resource. For standard file operations MooseFS mounted with FUSE acts as other Unix-alike file systems: * A hierarchical structure (directory tree) * Stores POSIX file attributes (permissions, last access and modification times) Supports special files (block and character devices, pipes and sockets) * Symbolic links (file names pointing to target files, not necessarily on MooseFS) and hard links (different names of files which refer to the same data on MooseFS) * Access to the file system can be limited based on IP address and/or password MooseFS on GitHub: https://github.com/moosefs/moosefs Source code, packages repo: https://moosefs.com/download.html Documentation: https://moosefs.com/documentation.html Recent changes: https://moosefs.com/documentation/changes-in-moosefs-3-0.html
openMosix is a Linux kernel extension for single-system image clustering. Taking n PC boxes, openMosix gives users and applications the illusion of one single computer with n CPUs. openMosix is perfectly scalable and adaptive.
The aoetools are programs for users of the ATA over Ethernet (AoE) network storage protocol, a simple protocol for using storage over an ethernet LAN. The vblade program (storage target) exports a block device using AoE.
Distributed and Parallel Computing with/for Python.
dispy is a comprehensive, yet easy to use framework for creating and using compute clusters to execute computations in parallel across multiple processors in a single machine (SMP), among many machines in a cluster, grid or cloud. dispy is well suited for data parallel (SIMD) paradigm where a computation (Python function or standalone program) is evaluated with different (large) datasets independently.
SystemInstaller is a tool to install a system with any Linux distribution. It works hand-in-hand with SystemImager and SystemConfigurator to build a cluster.
Cluster wide IPC providing datagram, connection, and bus messaging
The Transparent Inter Process Communication service can best be described as a cluster wide equivalent to the Unix Domain Socket IPC, but with much more features. Go to http://www.tipc.io for more details.