Automaton AI
With Automaton AI’s ADVIT, create, manage and develop high-quality training data and DNN models all in one place. Optimize the data automatically and prepare it for each phase of the computer vision pipeline. Automate the data labeling processes and streamline data pipelines in-house. Manage the structured and unstructured video/image/text datasets in runtime and perform automatic functions that refine your data in preparation for each step of the deep learning pipeline. Upon accurate data labeling and QA, you can train your own model. DNN training needs hyperparameter tuning like batch size, learning, rate, etc. Optimize and transfer learning on trained models to increase accuracy. Post-training, take the model to production. ADVIT also does model versioning. Model development and accuracy parameters can be tracked in run-time. Increase the model accuracy with a pre-trained DNN model for auto-labeling.
Learn more
Fabric for Deep Learning (FfDL)
Deep learning frameworks such as TensorFlow, PyTorch, Caffe, Torch, Theano, and MXNet have contributed to the popularity of deep learning by reducing the effort and skills needed to design, train, and use deep learning models. Fabric for Deep Learning (FfDL, pronounced “fiddle”) provides a consistent way to run these deep-learning frameworks as a service on Kubernetes. The FfDL platform uses a microservices architecture to reduce coupling between components, keep each component simple and as stateless as possible, isolate component failures, and allow each component to be developed, tested, deployed, scaled, and upgraded independently. Leveraging the power of Kubernetes, FfDL provides a scalable, resilient, and fault-tolerant deep-learning framework. The platform uses a distribution and orchestration layer that facilitates learning from a large amount of data in a reasonable amount of time across compute nodes.
Learn more
Google Deep Learning Containers
Build your deep learning project quickly on Google Cloud:
Quickly prototype with a portable and consistent environment for developing, testing, and deploying your AI applications with Deep Learning Containers. These Docker images use popular frameworks and are performance optimized, compatibility tested, and ready to deploy. Deep Learning Containers provide a consistent environment across Google Cloud services, making it easy to scale in the cloud or shift from on-premises. You have the flexibility to deploy on Google Kubernetes Engine (GKE), AI Platform, Cloud Run, Compute Engine, Kubernetes, and Docker Swarm.
Learn more
DeepCube
DeepCube focuses on the research and development of deep learning technologies that result in improved real-world deployment of AI systems. The company’s numerous patented innovations include methods for faster and more accurate training of deep learning models and drastically improved inference performance. DeepCube’s proprietary framework can be deployed on top of any existing hardware in both datacenters and edge devices, resulting in over 10x speed improvement and memory reduction. DeepCube provides the only technology that allows efficient deployment of deep learning models on intelligent edge devices. After the deep learning training phase, the resulting model typically requires huge amounts of processing and consumes lots of memory. Due to the significant amount of memory and processing requirements, today’s deep learning deployments are limited mostly to the cloud.
Learn more