Alternatives to Daft

Compare Daft alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Daft in 2024. Compare features, ratings, user reviews, pricing, and more from Daft competitors and alternatives in order to make an informed decision for your business.

  • 1
    Vertex AI
    Build, deploy, and scale machine learning (ML) models faster, with fully managed ML tools for any use case. Through Vertex AI Workbench, Vertex AI is natively integrated with BigQuery, Dataproc, and Spark. You can use BigQuery ML to create and execute machine learning models in BigQuery using standard SQL queries on existing business intelligence tools and spreadsheets, or you can export datasets from BigQuery directly into Vertex AI Workbench and run your models from there. Use Vertex Data Labeling to generate highly accurate labels for your data collection.
    Compare vs. Daft View Software
    Visit Website
  • 2
    Domo

    Domo

    Domo

    Domo puts data to work for everyone so they can multiply their impact on the business. Our cloud-native data experience platform goes beyond traditional business intelligence and analytics, making data visible and actionable with user-friendly dashboards and apps. Underpinned by a secure data foundation that connects with existing cloud and legacy systems, Domo helps companies optimize critical business processes at scale and in record time to spark the bold curiosity that powers exponential business results.
    Leader badge
    Compare vs. Daft View Software
    Visit Website
  • 3
    Posit

    Posit

    Posit

    At Posit, our goal is to make data science more open, intuitive, accessible, and collaborative. We provide tools that make it easy for individuals, teams, and enterprises to leverage powerful analytics and gain the insights they need to make a lasting impact. From the beginning, we’ve invested in open-source software like the RStudio IDE, Shiny, and tidyverse. Because we believe in putting the power of data science tools in the hands of everyone. We develop R and Python-based tools to help you produce higher-quality analysis faster. Securely share data-science applications across your team and the enterprise. Our code is your code. Build on it. Share it. Improve people’s lives with it. Take the time and effort out of uploading, storing, accessing, and sharing your work. We love hearing about the amazing work being done with our tools around the world. And we really love sharing those stories.
  • 4
    Azure Data Science Virtual Machines
    DSVMs are Azure Virtual Machine images, pre-installed, configured and tested with several popular tools that are commonly used for data analytics, machine learning and AI training. Consistent setup across team, promote sharing and collaboration, Azure scale and management, Near-Zero Setup, full cloud-based desktop for data science. Quick, Low friction startup for one to many classroom scenarios and online courses. Ability to run analytics on all Azure hardware configurations with vertical and horizontal scaling. Pay only for what you use, when you use it. Readily available GPU clusters with Deep Learning tools already pre-configured. Examples, templates and sample notebooks built or tested by Microsoft are provided on the VMs to enable easy onboarding to the various tools and capabilities such as Neural Networks (PYTorch, Tensorflow, etc.), Data Wrangling, R, Python, Julia, and SQL Server.
    Starting Price: $0.005
  • 5
    Dask

    Dask

    Dask

    Dask is open source and freely available. It is developed in coordination with other community projects like NumPy, pandas, and scikit-learn. Dask uses existing Python APIs and data structures to make it easy to switch between NumPy, pandas, scikit-learn to their Dask-powered equivalents. Dask's schedulers scale to thousand-node clusters and its algorithms have been tested on some of the largest supercomputers in the world. But you don't need a massive cluster to get started. Dask ships with schedulers designed for use on personal machines. Many people use Dask today to scale computations on their laptop, using multiple cores for computation and their disk for excess storage. Dask exposes lower-level APIs letting you build custom systems for in-house applications. This helps open source leaders parallelize their own packages and helps business leaders scale custom business logic.
  • 6
    JetBrains DataSpell
    Switch between command and editor modes with a single keystroke. Navigate over cells with arrow keys. Use all of the standard Jupyter shortcuts. Enjoy fully interactive outputs – right under the cell. When editing code cells, enjoy smart code completion, on-the-fly error checking and quick-fixes, easy navigation, and much more. Work with local Jupyter notebooks or connect easily to remote Jupyter, JupyterHub, or JupyterLab servers right from the IDE. Run Python scripts or arbitrary expressions interactively in a Python Console. See the outputs and the state of variables in real-time. Split Python scripts into code cells with the #%% separator and run them individually as you would in a Jupyter notebook. Browse DataFrames and visualizations right in place via interactive controls. All popular Python scientific libraries are supported, including Plotly, Bokeh, Altair, ipywidgets, and others.
    Starting Price: $229
  • 7
    NVIDIA RAPIDS
    The RAPIDS suite of software libraries, built on CUDA-X AI, gives you the freedom to execute end-to-end data science and analytics pipelines entirely on GPUs. It relies on NVIDIA® CUDA® primitives for low-level compute optimization, but exposes that GPU parallelism and high-bandwidth memory speed through user-friendly Python interfaces. RAPIDS also focuses on common data preparation tasks for analytics and data science. This includes a familiar DataFrame API that integrates with a variety of machine learning algorithms for end-to-end pipeline accelerations without paying typical serialization costs. RAPIDS also includes support for multi-node, multi-GPU deployments, enabling vastly accelerated processing and training on much larger dataset sizes. Accelerate your Python data science toolchain with minimal code changes and no new tools to learn. Increase machine learning model accuracy by iterating on models faster and deploying them more frequently.
  • 8
    Microsoft R Open
    Microsoft continues its commitment and development in R, not only in the latest Machine Learning Server release, but also in the newest Microsoft R Client and Microsoft R Open releases. You can also find R and Python support in SQL Server Machine Learning Services on Windows and Linux, and R support in Azure SQL Database. R components are backwards compatible. You should be able to run existing R script on newer versions, with the exception of dependencies on packages or platforms that are no longer supported, or known issues that require a workaround or code change. Microsoft R Open is the enhanced distribution of R from Microsoft Corporation. The current release, Microsoft R Open 4.0.2, is based the statistical language R-4.0.2 and includes additional capabilities for improved performance, reproducibility and platform support. Compatibility with all packages, scripts and applications that work with R-4.0.2.
  • 9
    Oracle Machine Learning
    Machine learning uncovers hidden patterns and insights in enterprise data, generating new value for the business. Oracle Machine Learning accelerates the creation and deployment of machine learning models for data scientists using reduced data movement, AutoML technology, and simplified deployment. Increase data scientist and developer productivity and reduce their learning curve with familiar open source-based Apache Zeppelin notebook technology. Notebooks support SQL, PL/SQL, Python, and markdown interpreters for Oracle Autonomous Database so users can work with their language of choice when developing models. A no-code user interface supporting AutoML on Autonomous Database to improve both data scientist productivity and non-expert user access to powerful in-database algorithms for classification and regression. Data scientists gain integrated model deployment from the Oracle Machine Learning AutoML User Interface.
  • 10
    Saturn Cloud

    Saturn Cloud

    Saturn Cloud

    Saturn Cloud is an award-winning ML platform for any cloud with 100,000+ users, including NVIDIA, CFA Institute, Snowflake, Flatiron School, Nestle, and more. It is an all-in-one solution for data science & ML development, deployment, and data pipelines in the cloud. Users can spin up a notebook with 4TB of RAM, add a GPU, connect to a distributed cluster of workers, build large language models, and more in a completely hosted environment. Data professionals can use your preferred languages, IDEs, and machine-learning libraries in Saturn Cloud. We offer full Git integration, shared custom images, and secure credential storage, making scaling and building your team in the cloud easy. We support the entire machine learning lifecycle from experimentation to production with features like jobs and deployments. These features and built-in tools are easily shareable within teams, so time is saved and work is reproducible.
    Leader badge
    Starting Price: $0.005 per GB per hour
  • 11
    Plotly Dash
    Dash & Dash Enterprise let you build & deploy analytic web apps using Python, R, and Julia. No JavaScript or DevOps required. Through Dash, the world's largest companies elevate AI, ML, and Python analytics to business users at 5% the cost of a full-stack development approach. Deliver apps and dashboards that run advanced analytics: ML, NLP, forecasting, computer vision and more. Work in the languages you love: Python, R, and Julia. Reduce costs by migrating legacy, per-seat licensed software to Dash Enterprise's open-core, unlimited end-user pricing model. Move faster by deploying and updating Dash apps without an IT or DevOps team. Create pixel-perfect dashboards & web apps, without writing any CSS. Scale effortlessly with Kubernetes. Support mission-critical Python applications with high availability.
  • 12
    IBM Watson Studio
    Build, run and manage AI models, and optimize decisions at scale across any cloud. IBM Watson Studio empowers you to operationalize AI anywhere as part of IBM Cloud Pak® for Data, the IBM data and AI platform. Unite teams, simplify AI lifecycle management and accelerate time to value with an open, flexible multicloud architecture. Automate AI lifecycles with ModelOps pipelines. Speed data science development with AutoAI. Prepare and build models visually and programmatically. Deploy and run models through one-click integration. Promote AI governance with fair, explainable AI. Drive better business outcomes by optimizing decisions. Use open source frameworks like PyTorch, TensorFlow and scikit-learn. Bring together the development tools including popular IDEs, Jupyter notebooks, JupterLab and CLIs — or languages such as Python, R and Scala. IBM Watson Studio helps you build and scale AI with trust and transparency by automating AI lifecycle management.
  • 13
    Dataiku DSS
    Bring data analysts, engineers, and scientists together. Enable self-service analytics and operationalize machine learning. Get results today and build for tomorrow. Dataiku DSS is the collaborative data science software platform for teams of data scientists, data analysts, and engineers to explore, prototype, build, and deliver their own data products more efficiently. Use notebooks (Python, R, Spark, Scala, Hive, etc.) or a customizable drag-and-drop visual interface at any step of the predictive dataflow prototyping process – from wrangling to analysis to modeling. Profile the data visually at every step of the analysis. Interactively explore and chart your data using 25+ built-in charts. Prepare, enrich, blend, and clean data using 80+ built-in functions. Leverage Machine Learning technologies (Scikit-Learn, MLlib, TensorFlow, Keras, etc.) in a visual UI. Build & optimize models in Python or R and integrate any external ML library through code APIs.
  • 14
    Oracle Cloud Infrastructure Data Flow
    Oracle Cloud Infrastructure (OCI) Data Flow is a fully managed Apache Spark service to perform processing tasks on extremely large data sets without infrastructure to deploy or manage. This enables rapid application delivery because developers can focus on app development, not infrastructure management. OCI Data Flow handles infrastructure provisioning, network setup, and teardown when Spark jobs are complete. Storage and security are also managed, which means less work is required for creating and managing Spark applications for big data analysis. With OCI Data Flow, there are no clusters to install, patch, or upgrade, which saves time and operational costs for projects. OCI Data Flow runs each Spark job in private dedicated resources, eliminating the need for upfront capacity planning. With OCI Data Flow, IT only needs to pay for the infrastructure resources that Spark jobs use while they are running.
    Starting Price: $0.0085 per GB per hour
  • 15
    Zepl

    Zepl

    Zepl

    Sync, search and manage all the work across your data science team. Zepl’s powerful search lets you discover and reuse models and code. Use Zepl’s enterprise collaboration platform to query data from Snowflake, Athena or Redshift and build your models in Python. Use pivoting and dynamic forms for enhanced interactions with your data using heatmap, radar, and Sankey charts. Zepl creates a new container every time you run your notebook, providing you with the same image each time you run your models. Invite team members to join a shared space and work together in real time or simply leave their comments on a notebook. Use fine-grained access controls to share your work. Allow others have read, edit, and run access as well as enable collaboration and distribution. All notebooks are auto-saved and versioned. You can name, manage and roll back all versions through an easy-to-use interface, and export seamlessly into Github.
  • 16
    Brilent

    Brilent

    Brilent

    Brilent is a data science tech company developing a SaaS solution to help talent seekers quickly and effectively identify the right talent to hire. The exciting part about this intelligent technology is its simplicity. No tricks. It leverages the components recruiters find quite relevant. At the core of the Brilent engine are three simple elements: the job requirements; the candidate profile; and our unique repository of market data. Next comes the fun part. Our system gathers all the relevant data from the job requirements and candidate profiles. Using hundreds of variables from these familiar recruiting elements and the market data, we use our well-honed experience to apply artificial intelligence and machine learning algorithms to predict how likely a candidate is a good fit for a particular job. In other words, a whole lot of data crunching, and completed in seconds. Recruiters get a ranking of candidates according to the desired specifications.
  • 17
    Obviously AI

    Obviously AI

    Obviously AI

    The entire process of building machine learning algorithms and predicting outcomes, packed in one single click. Not all data is built to be ready for ML, use the Data Dialog to seamlessly shape your dataset without wrangling your files. Share your prediction reports with your team or make them public. Allow anyone to start making predictions on your model. Bring dynamic ML predictions into your own app using our low-code API. Predict willingness to pay, score leads and much more in real-time. Obviously AI puts the world’s most cutting-edge algorithms in your hands, without compromising on performance. Forecast revenue, optimize supply chain, personalize marketing. You can now know what happens next. Add a CSV file OR integrate with your favorite data sources in minutes. Pick your prediction column from a dropdown, we'll auto build the AI. Beautifully visualize predicted results, top drivers and simulate "what-if" scenarios.
    Starting Price: $75 per month
  • 18
    Comet

    Comet

    Comet

    Manage and optimize models across the entire ML lifecycle, from experiment tracking to monitoring models in production. Achieve your goals faster with the platform built to meet the intense demands of enterprise teams deploying ML at scale. Supports your deployment strategy whether it’s private cloud, on-premise servers, or hybrid. Add two lines of code to your notebook or script and start tracking your experiments. Works wherever you run your code, with any machine learning library, and for any machine learning task. Easily compare experiments—code, hyperparameters, metrics, predictions, dependencies, system metrics, and more—to understand differences in model performance. Monitor your models during every step from training to production. Get alerts when something is amiss, and debug your models to address the issue. Increase productivity, collaboration, and visibility across all teams and stakeholders.
    Starting Price: $179 per user per month
  • 19
    Quadratic

    Quadratic

    Quadratic

    Quadratic enables your team to work together on data analysis to deliver faster results. You already know how to use a spreadsheet, but you’ve never had this much power. Quadratic speaks Formulas and Python (SQL & JavaScript coming soon). Use the language you and your team already know. Single-line formulas are hard to read. In Quadratic you can expand your recipes to as many lines as you need. Quadratic has Python library support built-in. Bring the latest open-source tools directly to your spreadsheet. The last line of code is returned to the spreadsheet. Raw values, 1/2D arrays, and Pandas DataFrames are supported by default. Pull or fetch data from an external API, and it updates automatically in Quadratic's cells. Navigate with ease, zoom out for the big picture, and zoom in to focus on the details. Arrange and navigate your data how it makes sense in your head, not how a tool forces you to do it.
  • 20
    Streamlit

    Streamlit

    Streamlit

    Streamlit. The fastest way to build and share data apps. Turn data scripts into sharable web apps in minutes. All in Python. All for free. No front-end experience required. Streamlit combines three simple ideas. Embrace Python scripting. Build an app in a few lines of code with our magically simple API. Then see it automatically update as you save the source file. Weave in interaction. Adding a widget is the same as declaring a variable. No need to write a backend, define routes, handle HTTP requests, etc. Deploy instantly. Use Streamlit’s sharing platform to effortlessly share, manage, and collaborate on your apps. A minimal framework for powerful apps. Face-GAN explorer. App that uses Shaobo Guan’s TL-GAN project from Insight Data Science, TensorFlow, and NVIDIA's PG-GAN to generate faces that match selected attributes. Real time object detection. An image browser for the Udacity self-driving-car dataset with real-time object detection.
  • 21
    Cloudera Data Science Workbench
    Accelerate machine learning from research to production with a consistent experience built for your traditional platform. With Python, R, and Scala directly in the web browser, Cloudera Data Science Workbench (CDSW) delivers a self-service experience data scientists will love. Download and experiment with the latest libraries and frameworks in customizable project environments that work just like your laptop. Cloudera Data Science Workbench provides connectivity not only to CDH and HDP but also to the systems your data science teams rely on for analysis. Cloudera Data Science Workbench lets data scientists manage their own analytics pipelines, including built-in scheduling, monitoring, and email alerting. Quickly develop and prototype new machine learning projects and easily deploy them to production.
  • 22
    Metaflow

    Metaflow

    Metaflow

    Successful data science projects are delivered by data scientists who can build, improve, and operate end-to-end workflows independently, focusing more on data science, less on engineering. Use Metaflow with your favorite data science libraries, such as Tensorflow or SciKit Learn, and write your models in idiomatic Python code with not much new to learn. Metaflow also supports the R language. Metaflow helps you design your workflow, run it at scale, and deploy it to production. It versions and tracks all your experiments and data automatically. It allows you to inspect results easily in notebooks. Metaflow comes packaged with the tutorials, so getting started is easy. You can make copies of all the tutorials in your current directory using the metaflow command line interface.
  • 23
    IBM Analytics for Apache Spark
    IBM Analytics for Apache Spark is a flexible and integrated Spark service that empowers data science professionals to ask bigger, tougher questions, and deliver business value faster. It’s an easy-to-use, always-on managed service with no long-term commitment or risk, so you can begin exploring right away. Access the power of Apache Spark with no lock-in, backed by IBM’s open-source commitment and decades of enterprise experience. A managed Spark service with Notebooks as a connector means coding and analytics are easier and faster, so you can spend more of your time on delivery and innovation. A managed Apache Spark services gives you easy access to the power of built-in machine learning libraries without the headaches, time and risk associated with managing a Sparkcluster independently.
  • 24
    Kedro

    Kedro

    Kedro

    Kedro is the foundation for clean data science code. It borrows concepts from software engineering and applies them to machine-learning projects. A Kedro project provides scaffolding for complex data and machine-learning pipelines. You spend less time on tedious "plumbing" and focus instead on solving new problems. Kedro standardizes how data science code is created and ensures teams collaborate to solve problems easily. Make a seamless transition from development to production with exploratory code that you can transition to reproducible, maintainable, and modular experiments. A series of lightweight data connectors is used to save and load data across many different file formats and file systems.
    Starting Price: Free
  • 25
    SAS Viya
    SAS® Viya® data science offerings provide a comprehensive, scalable analytics environment that's quick and easy to deploy, enabling you to meet diverse business needs. Automatically generated insights enable you to identify the most common variables across all models, the most important variables selected across models and assessment results for all models. Natural language generation capabilities are used to create project summaries written in plain language, enabling you to easily interpret reports. Analytics team members can add project notes to the insights report to facilitate communication and collaboration among team members. SAS lets you embed open source code within an analysis and call open source algorithms seamlessly within its environment. This facilitates collaboration across your organization because users can program in their language of choice. You can also take advantage of SAS Deep Learning with Python (DLPy), our open-source package on GitHub.
  • 26
    Outerbounds

    Outerbounds

    Outerbounds

    Design and develop data-intensive projects with human-friendly, open-source Metaflow. Run, scale, and deploy them reliably on the fully managed Outerbounds platform. One platform for all your ML and data science projects. Access data securely from your existing data warehouses. Compute with a cluster optimized for scale and cost. 24/7 managed orchestration for production workflows. Use results to power any application. Give your data scientists superpowers, approved by your engineers. Outerbounds Platform allows data scientists to develop rapidly, experiment at scale, and deploy to production confidently. All within the outer bounds of policies and processes defined by your engineers, running on your cloud account, fully managed by us. Security is in our DNA, not at the perimeter. The platform adapts to your policies and compliance requirements through multiple layers of security. Centralized auth, a strict permission boundary, and granular task execution roles.
  • 27
    MATLAB

    MATLAB

    The MathWorks

    MATLAB® combines a desktop environment tuned for iterative analysis and design processes with a programming language that expresses matrix and array mathematics directly. It includes the Live Editor for creating scripts that combine code, output, and formatted text in an executable notebook. MATLAB toolboxes are professionally developed, rigorously tested, and fully documented. MATLAB apps let you see how different algorithms work with your data. Iterate until you’ve got the results you want, then automatically generate a MATLAB program to reproduce or automate your work. Scale your analyses to run on clusters, GPUs, and clouds with only minor code changes. There’s no need to rewrite your code or learn big data programming and out-of-memory techniques. Automatically convert MATLAB algorithms to C/C++, HDL, and CUDA code to run on your embedded processor or FPGA/ASIC. MATLAB works with Simulink to support Model-Based Design.
  • 28
    Darwin

    Darwin

    SparkCognition

    Darwin is an automated machine learning product that enables your data science and business analytics teams to move more quickly from data to meaningful results. Darwin helps organizations scale the adoption of data science across teams, and the implementation of machine learning applications across operations, becoming data-driven enterprises.
    Starting Price: $4000
  • 29
    Shapelets

    Shapelets

    Shapelets

    Powerful computing at your fingertips. Parallel computing, groundbreaking algorithms, so what are you waiting for? Designed to empower data scientists in business. Get the fastest computing in an all-inclusive time-series platform. Shapelets provides you with analytical features such as causality, discords and motif discovery, forecasting, clustering, etc. Run, extend and integrate your own algorithms into the Shapelets platform to make the most of Big Data analysis. Shapelets integrates seamlessly with any data collection and storage solution. It also integrates with MS Office and any other visualization tool to simplify and share insights without any technical acumen. Our UI works with the server to bring you interactive visualizations. You can make the most of your metadata and represent it in the many different visual graphs provided by our modern interface. Shapelets enables users from the oil, gas, and energy industry to perform real-time analysis of operational data.
  • 30
    cnvrg.io

    cnvrg.io

    cnvrg.io

    Scale your machine learning development from research to production with an end-to-end solution that gives your data science team all the tools they need in one place. As the leading data science platform for MLOps and model management, cnvrg.io is a pioneer in building cutting-edge machine learning development solutions so you can build high-impact machine learning models in half the time. Bridge science and engineering teams in a clear and collaborative machine learning management environment. Communicate and reproduce results with interactive workspaces, dashboards, dataset organization, experiment tracking and visualization, a model repository and more. Focus less on technical complexity and more on building high impact ML models. Cnvrg.io container-based infrastructure helps simplify engineering heavy tasks like tracking, monitoring, configuration, compute resource management, serving infrastructure, feature extraction, and model deployment.
  • 31
    Anaconda

    Anaconda

    Anaconda

    Empowering the enterprise to do real data science at speed and scale with a full-featured machine learning platform. Spend less time managing tools and infrastructure, so you can focus on building machine learning applications that move your business forward. Anaconda Enterprise takes the headache out of ML operations, puts open-source innovation at your fingertips, and provides the foundation for serious data science and machine learning production without locking you into specific models, templates, or workflows. Software developers and data scientists can work together with AE to build, test, debug, and deploy models using their preferred languages and tools. AE provides access to both notebooks and IDEs so developers and data scientists can work together more efficiently. They can also choose from example projects and preconfigured projects. AE projects are automatically containerized so they can be moved between environments with ease.
  • 32
    Algopine

    Algopine

    Algopine

    We develop, manage and run predictive software services based on data science and machine learning. Software service for large e-commerce businesses and retail chains which uses machine learning to forecast sales and optimize stock distribution among retail stores and warehouses. Personalized product recommender for e-commerce web sites which uses real-time Bayes nets to display relevant recommended products for e-shop visitors. Software service which automatically suggests product price movements to improve profit by using statistical price and demand elasticity models. API for computing optimal path routes for batch picking time optimization in a retailer's warehouse, which is built on shortest path graph algorithms.
  • 33
    Wolfram|One
    Wolfram|One is the world's first fully cloud-desktop hybrid, integrated computation platform, the ideal entry point to using the full capabilities of the Wolfram technology stack. From data analytics to modeling (with our curated data or yours), from publishing an API to live presenting your latest R&D, from instant scratchpad to rapidly programming your prototype, Wolfram|One is the culmination of 30 years' experience in one easy-to-use, get-started-now product from the world's leading computation company. From basic web forms to large-scale data analytics, Wolfram technology includes the functionality for any kind of computational task. Underlying everything we do is the Wolfram Language. Designed for the new generation of programmers, the Wolfram Language has a vast depth of built-in algorithms and knowledge, all automatically accessible through its elegant unified symbolic language. Scalable for programs from tiny to huge, with immediate deployment locally and in the cloud.
    Starting Price: $148 per month
  • 34
    TrueFoundry

    TrueFoundry

    TrueFoundry

    TrueFoundry is a Cloud-native Machine Learning Training and Deployment PaaS on top of Kubernetes that enables Machine learning teams to train and Deploy models at the speed of Big Tech with 100% reliability and scalability - allowing them to save cost and release Models to production faster. We abstract out the Kubernetes for Data Scientists and enable them to operate in a way they are comfortable. It also allows teams to deploy and fine-tune large language models seamlessly with full security and cost optimization. TrueFoundry is open-ended, API Driven and integrates with the internal systems, deploys on a company's internal infrastructure and ensures complete Data Privacy and DevSecOps practices.
    Starting Price: $5 per month
  • 35
    HPE Ezmeral

    HPE Ezmeral

    Hewlett Packard Enterprise

    Run, manage, control and secure the apps, data and IT that run your business, from edge to cloud. HPE Ezmeral advances digital transformation initiatives by shifting time and resources from IT operations to innovations. Modernize your apps. Simplify your Ops. And harness data to go from insights to impact. Accelerate time-to-value by deploying Kubernetes at scale with integrated persistent data storage for app modernization on bare metal or VMs, in your data center, on any cloud or at the edge. Harness data and get insights faster by operationalizing the end-to-end process to build data pipelines. Bring DevOps agility to the machine learning lifecycle, and deliver a unified data fabric. Boost efficiency and agility in IT Ops with automation and advanced artificial intelligence. And provide security and control to eliminate risk and reduce costs. HPE Ezmeral Container Platform provides an enterprise-grade platform to deploy Kubernetes at scale for a wide range of use cases.
  • 36
    NVIDIA Merlin
    NVIDIA Merlin empowers data scientists, machine learning engineers, and researchers to build high-performing recommenders at scale. Merlin includes libraries, methods, and tools that streamline the building of recommenders by addressing common preprocessing, feature engineering, training, inference, and deploying to production challenges. Merlin components and capabilities are optimized to support the retrieval, filtering, scoring, and ordering of hundreds of terabytes of data, all accessible through easy-to-use APIs. With Merlin, better predictions, increased click-through rates, and faster deployment to production are within reach. NVIDIA Merlin, as part of NVIDIA AI, advances our commitment to supporting innovative practitioners doing their best work. As an end-to-end solution, NVIDIA Merlin components are designed to be interoperable within existing recommender workflows that utilize data science, and machine learning (ML).
  • 37
    SAS Visual Data Science Decisioning
    Integrate analytics into real-time ​interactions and event-based capabilities​. SAS Visual Data Science Decisioning features robust data management, visualization, advanced analytics and model management. It supports decisions by creating, embedding and governing analytically driven decision flows at scale in real-time or batch. It also deploys analytics and decisions in the stream to help you discover insights. Solve complex analytical problems with a comprehensive visual interface that handles all tasks in the analytics life cycle. SAS Visual Data Mining and Machine Learning, which runs in SAS® Viya®, combines data wrangling, exploration, feature engineering, and modern statistical, data mining, and machine learning techniques in a single, scalable in-memory processing environment. Access data files, libraries and existing programs, or write new ones, with this developmental web application accessible through your browser.
  • 38
    Appsilon

    Appsilon

    Appsilon

    Appsilon provides innovative data analytics, machine learning, and managed services solutions for Fortune 500 companies, NGOs, and non-profit organizations. We deliver the world’s most advanced R Shiny applications, with a unique ability to rapidly develop and scale enterprise Shiny dashboards. Our proprietary machine learning frameworks allow us to deliver Computer Vision, NLP, and fraud detection prototypes in as little as one week. Above all, we are committed to making a positive impact on the world. Through our AI For Good Initiative, we routinely contribute our skills to projects that support the preservation of human life and the conservation of animal populations all over the globe. Recently, our team has worked to mitigate poaching in Africa with computer vision, provide satellite image analysis for assessing damage after natural disasters, and build tools to help with COVID-19 risk assessment. Appsilon is also a pioneer in open source.
  • 39
    Peak

    Peak

    Peak

    A new decision intelligence system, putting AI in the hands of commercial leaders to drive great decision making. CODI, our Connected Decision Intelligence system, has been engineered by Peak to become a layer of intelligence that sits between your other systems, unleashing the power of your data for the first time. CODI enables you to quickly deploy AI solutions, harnessing the true potential of your data thanks to its unique full-stack features. It allows data science and data engineering teams to take full control of every aspect of building and deploying AI solutions, at speed and at scale. With CODI, AI projects move beyond being just experiments and become fully-deployed solutions that deliver real world value and results. Built on enterprise-grade infrastructure, CODI is capable of handling data at scale, and seamlessly integrates into existing tech stacks. Surface more insight and combine data from across your organization.
  • 40
    Bitfount

    Bitfount

    Bitfount

    Bitfount is a platform for distributed data science. We power deep data collaborations without data sharing. Distributed data science sends algorithms to data, instead of the other way around. Set up a federated privacy-preserving analytics and machine learning network in minutes, and let your team focus on insights and innovation instead of bureaucracy. Your data team has the skills to solve your biggest challenges and innovate, but they are held back by barriers to data access. Is complex data pipeline infrastructure messing with your plans? Are compliance processes taking too long? Bitfount has a better way to unleash your data experts. Connect siloed and multi-cloud datasets while preserving privacy and respecting commercial sensitivity. No expensive, time-consuming data lift-and-shift. Usage-based access controls to ensure teams only perform the analysis you want, on the data you want. Transfer management of access controls to the teams who control the data.
  • 41
    Metacoder

    Metacoder

    Wazoo Mobile Technologies LLC

    Metacoder makes processing data faster and easier. Metacoder gives analysts needed flexibility and tools to facilitate data analysis. Data preparation steps such as cleaning are managed reducing the manual inspection time required before you are up and running. Compared to alternatives, is in good company. Metacoder beats similar companies on price and our management is proactively developing based on our customers' valuable feedback. Metacoder is used primarily to assist predictive analytics professionals in their job. We offer interfaces for database integrations, data cleaning, preprocessing, modeling, and display/interpretation of results. We help organizations distribute their work transparently by enabling model sharing, and we make management of the machine learning pipeline easy to make tweaks. Soon we will be including code free solutions for image, audio, video, and biomedical data.
    Starting Price: $89 per user/month
  • 42
    INQDATA

    INQDATA

    INQDATA

    Cloud-based Data Science platform delivering intelligently curated and cleansed data, ready to be consumed. Firms face significant challenges, resource constraints, and high costs when managing their data before they can start adding any value. The data is ingested, cleansed, stored, accessed, and only then analyzed. But the analysis is where the value is. Our solution allows clients to focus on core business activities, not on the expensive, resource heavy data lifecycle. We take care of that. Cloud-native platform for real-time streaming analytics that fully leverages the benefits of cloud architecture to enable INQDATA to deliver fast scalable historical and real-time data without the complexity of infrastructure.
  • 43
    StreamFlux

    StreamFlux

    Fractal

    Data is crucial when it comes to building, streamlining and growing your business. However, getting the full value out of data can be a challenge, many organizations are faced with poor access to data, incompatible tools, spiraling costs and slow results. Simply put, leaders who can turn raw data into real results will thrive in today’s landscape. The key to this is empowering everyone across your business to be able to analyze, build and collaborate on end-to-end AI and machine learning solutions in one place, fast. Streamflux is a one-stop shop to meet your data analytics and AI challenges. Our self-serve platform allows you the freedom to build end-to-end data solutions, uses models to answer complex questions and assesses user behaviors. Whether you’re predicting customer churn and future revenue, or generating recommendations, you can go from raw data to genuine business impact in days, not months.
  • 44
    PySpark

    PySpark

    PySpark

    PySpark is an interface for Apache Spark in Python. It not only allows you to write Spark applications using Python APIs, but also provides the PySpark shell for interactively analyzing your data in a distributed environment. PySpark supports most of Spark’s features such as Spark SQL, DataFrame, Streaming, MLlib (Machine Learning) and Spark Core. Spark SQL is a Spark module for structured data processing. It provides a programming abstraction called DataFrame and can also act as distributed SQL query engine. Running on top of Spark, the streaming feature in Apache Spark enables powerful interactive and analytical applications across both streaming and historical data, while inheriting Spark’s ease of use and fault tolerance characteristics.
  • 45
    Visplore

    Visplore

    Visplore

    Visplore is a plug-and-play software solution for rapid advanced analytics of process and asset data. Easy-to-use visualization and automated analytics provide process and maintenance engineers with answers for data-driven decision-making. Increase the speed and value of data analytics by 10x – 100x and master the digital transformation with your subject-matter experts. Highlights: - Work with millions of data records without delay (zooming etc.). - Select, cleanse, label and export data interactively - Connect with Python, R, Matlab, CSV, databases and OSISoft PI to get started in 1 minute.
  • 46
    Hex

    Hex

    Hex

    Hex brings together the best of notebooks, BI, and docs into a seamless, collaborative UI. Hex is a modern Data Workspace. It makes it easy to connect to data, analyze it in collaborative SQL and Python-powered notebooks, and share work as interactive data apps and stories. Your default landing page in Hex is the Projects page. You can quickly find projects you created, as well as those shared with you and your workspace. The outline provides an easy-to-browse overview of all the cells in a project's Logic View. Every cell in the outline lists the variables it defines, and cells that return a displayed output (chart cells, Input Parameters, markdown cells, etc.) display a preview of that output. You can click any cell in the outline to automatically jump to that position in the logic.
    Starting Price: $24 per user per month
  • 47
    esDynamic
    Maximize your security testing journey, from setting up your bench to analyzing your data processing results, esDynamic saves you valuable time and effort, empowering you to unleash the full potential of your attack workflow. Discover the flexible and comprehensive Python-based platform, perfectly suited for every phase of your security analysis. Customize your research space to meet your unique requirements by effortlessly adding new equipment, integrating tools, and modifying data. Additionally, esDynamic features an extensive collection of materials on complex topics that would typically require extensive research or a team of specialists, granting you instant access to expertise. Say goodbye to scattered data and fragmented knowledge. Welcome a cohesive workspace where your team can effortlessly share data and insights, fostering collaboration and accelerating discoveries. Centralize and solidify your efforts in JupyterLab notebooks to share with your team.
    Starting Price: Free
  • 48
    Deepnote

    Deepnote

    Deepnote

    Deepnote is building the best data science notebook for teams. In the notebook, users can connect their data, explore, and analyze it with real-time collaboration and version control. Users can easily share project links with team collaborators, or with end-users to present polished assets. All of this is done through a powerful, browser-based UI that runs in the cloud. We built Deepnote because data scientists don't work alone. Features: - Sharing notebooks and projects via URL - Inviting others to view, comment and collaborate, with version control - Publishing notebooks with visualizations for presentations - Sharing datasets between projects - Set team permissions to decide who can edit vs view code - Full linux terminal access - Code completion - Automatic python package management - Importing from github - PostgreSQL DB connection
    Starting Price: Free
  • 49
    Zerve AI

    Zerve AI

    Zerve AI

    Merging the best of a notebook and an IDE into one integrated coding environment, experts can explore their data and write stable code at the same time with fully automated cloud infrastructure. Zerve’s data science development environment gives data science and ML teams a unified space to explore, collaborate, build, and deploy data science & AI projects like never before. Zerve offers true language interoperability, meaning that as well as being able to use Python, R, SQL, or Markdown all in the same canvas, users can connect these code blocks to each other. No more long-running code blocks or containers, with Zerve enjoying unlimited parallelization at any stage of the development journey. Analysis artifacts are automatically serialized, versioned, stored, and preserved for later use, meaning easily changing a step in the data flow without needing to rerun any preceding steps. Fine-grained selection of compute resources and extra memory for complex data transformation.
  • 50
    Coder

    Coder

    Coder

    Coder delivers self-hosted cloud development environments consistently provisioned as code and pre-deployed for developer activity on day one. Preferred by enterprises, Coder is open source and runs air-gapped on-premise or in your cloud, giving developers access to powerful infrastructure without compromising governance. Coder moves local development and source code to your centralized infrastructure where developers can access their remote environments via their favorite desktop or web-based IDE. This is the optimal intersection of improved developer experience, productivity, and security. Coder’s ephemeral development environments are provisioned as code from pre-defined templates, which means developers can create a new workspace with a click. Now, they can avoid days or weeks struggling with local dependency versioning conflicts or security approvals. With Coder, developers can onboard or shift between projects in minutes.