Are you training large models?Explore Neptune Scale: The experiment tracker for foundation model trainingRead more

Compare

Neptune vs Amazon SageMaker

neptune-logo
vs
aws sagemaker
Show differences only

Commercial Requirements

Commercial requirements chevron
Standalone component or a part of a broader ML platform?

Standalone component

Part of the AWS SageMaker ecosystem

Is the product delivered as commercial software, open-source software, or a managed cloud service?

Managed cloud service

Managed cloud service

What is the pricing model?

Pricing depends on usage. You can get started for free with AWS free tier. Costs can be estimated using the AWS Pricing Calculator. Read more here

SLOs / SLAs: Does the vendor provide guarantees around service levels?

Yes

Support: Does the vendor provide 24×7 support?
SSO, ACL: does the vendor provide user access management?
Security policy and compliance

Yes

General Information

Deployment chevron
Cloud (SaaS)

Yes

Yes

Self-hosted (your infrastructure)

Yes

It cannot be deployed on-premises. You can, however, deploy it on your Virtual Private Cloud on AWS.

– On-prem bare metal

Yes

No

– Private Cloud: Amazon Web Services (AWS)

Yes

It cannot be deployed on-premises. You can, however, deploy it on your Virtual Private Cloud on AWS.

– Private Cloud: Google Cloud Platform (GCP)

Yes

No

– Private Cloud: Microsoft Azure

Yes

No

Setup chevron
What are the infrastructure requirements?

No special requirements other than having the neptune-client installed and access to the internet if using managed hosting. Check here for infrastructure requirements for on-prem deployment.

Just internet access is needed to access the AWS ecosystem

How much do you have to change in your training process?

Minimal. Just a few lines of code needed for tracking. Read more

Does it integrate with the training process via CLI/YAML/Client library?

Yes, through the neptune-client library

Does it come with a web UI or is it console-based?

Both web UI and CLI

Serverless UI

No

No

Flexibility, speed, and accessibility chevron
Customizable metadata structure

Yes

No

How can you access model metadata?
– gRPC API

No

Yes

– CLI / custom API

Yes

Yes

– REST API

No

Requires using additional AWS services like Lambda or API Gateway

– Python SDK

Yes

Yes

– R SDK

Yes

– Java SDK

No

Yes

– Julia SDK

No

Yes

Supported operations
– Search

Yes

Yes

– Update

Yes

No

– Delete

Yes

– Download

Yes

Yes

Distributed training support

Yes

Yes

Pipelining support

Yes

Yes

Logging modes
– Offline

Yes

No

– Disabled/off

Yes

No

– Asynchronous

Yes

Yes

– Synchronous

Yes

No

Live monitoring

Yes

Yes

Mobile support

No

No

Webhooks and notifications

No

No

Capabilities

Log and display of metadata chevron
Dataset
– location (path/s3)

Yes

Yes

– hash (md5)

Yes

No

– Preview table

Yes

No

– Preview image

Yes

Yes

– Preview text

Yes

No

– Preview rich media

Yes

– Multifile support

Yes

No

Code versions
– Git

Yes

– Source

Yes

No

– Notebooks

Yes

Yes

Parameters

Yes

Yes

Metrics and losses
– Single values

Yes

Yes

– Series values

Yes

Yes

– Series aggregates (min/max/avg/var/last)

Yes

Yes

Tags

Yes

Yes

Descriptions/comments

Yes

Rich format
– Images (support for labels and descriptions)

Yes

No

– Plots

Yes

Yes

– Interactive visualizations (widgets and plugins)

Yes

No

– Video

Yes

– Audio

Yes

– Neural Network Histograms

No

No

– Prediction visualization (tabular)

No

– Prediction visualization (image)

No

No

– Prediction visualization (image – interactive confusion matrix for image classification)

No

NA

– Prediction visualization (image – overlayed prediction masks for image segmentation)

No

NA

– Prediction visualization (image – overlayed prediction bounding boxes for object detection)

No

NA

Hardware consumption
– CPU

Yes

Yes

– GPU

Yes

Yes

– TPU

No

No

– Memory

Yes

Yes

System information
– Console logs (Stderr, Stdout)

Yes

Yes

– Error stack trace

Yes

Yes

– Execution command

Yes

No

– System details (host, user, hardware specs)

Yes

Yes

Environment config
– pip requirements.txt

Yes

No

– conda env.yml

Yes

No

– Docker Dockerfile

Yes

No

Files
– External file reference (s3 buckets)

Yes

Explanations (SHAP, DALEX)

Yes

Comparing experiments chevron
Table format diff

Yes

No

Overlayed learning curves

Yes

Yes

Parameters and metrics
– Groupby on experiment values (parameters)

Yes

Yes

– Parallel coordinates plots

Yes

No

– Parameter Importance plot

No

No

Rich format (side by side)
– Image

Yes

No

– Video

No

No

– Audio

No

No

– Plots

No

No

– Interactive visualization (HTML)

No

No

– Text

Yes

No

– Neural Network Histograms

No

No

– Prediction visualization (tabular)

Yes

No

– Prediction visualization (image, video, audio)

No

No

Code
– Git

No

No

– Source files

No

No

– Notebooks

Yes

No

Environment
– pip requirements.txt

No

No

– conda env.yml

No

No

– Docker Dockerfile

No

No

Hardware
– CPU

Yes

No

– GPU

Yes

No

– Memory

Yes

No

System information
– Console logs (Stderr, Stdout)

No

No

– Error stack trace

No

No

– Execution command

No

No

– System details (host, owner)

Yes

No

Data versions
– Location

Yes

No

– Hash

Yes

No

– Dataset diff

Yes

No

– External reference version diff (s3)

Yes

No

Files
– Models

No

No

– CSV

No

No

Custom compare dashboards
– Combining multiple metadata types (image, learning curve, hardware)

Yes

No

– Logging custom comparisons from notebooks/code

No

No

– Compare/diff of multiple (3+) experiments/runs

Yes

Organizing and searching experiments and metadata chevron
Experiment table customization
– Adding/removing columns

Yes

No

– Renaming columns in the UI

Yes

No

– Adding colors to columns

Yes

No

– Displaying aggregate (min/max/avg/var/last) for series like training metrics in a table

Yes

No

– Automagical column suggestion

Yes

No

Experiment filtering and searching
– Searching on multiple criteria

Yes

Yes

– Query language vs fixed selectors
– Saving filters and search history

Yes

No

Custom dashboards for a single experiment
– Can combine different metadata types in one view

Yes

Yes

– Saving experiment table views

Yes

No

– Logging project-level metadata

Yes

No

– Custom widgets and plugins

No

No

Tagging and searching on tags

Yes

Yes

Nested metadata structure support in the UI

Yes

No

Reproducibility and traceability chevron
One-command experiment re-run

No

No

Experiment lineage
– List of datasets used downstream

No

Yes

– List of other artifacts (models) used downstream

No

Yes

– Downstream artifact dependency graph

No

Yes

Reproducibility protocol

No

Is environment versioned and reproducible

Yes

Yes

Saving/fetching/caching datasets for experiments

No

No

Collaboration and knowledge sharing chevron
User groups and ACL

Yes

Sharing UI links with project members

Yes

Yes

Sharing UI links with external people

Yes

No

Commenting

No

Interactive project-level reports

No

No

Model lineage and evaluation history chevron
History of evaluation/testing runs

No

Yes

Support for continuous testing

No

Yes

Users who created a model or downstream experiments

No

Yes

Access control, model review, and promoting models chevron
Locking model version and downstream runs, experiments, and artifacts

No

No

Adding annotations/comments and approvals from the UI

Yes

Model compare (current vs challenger etc)
Compatibility audit (input/output schema)

No

No

Compliance audit (datasets used, creation process approvals, results/explanations approvals)

No

CI/CD/CT compatibility chevron
Webhooks

No

No

Model accessibility

No

No

Support for continuous testing

No

Yes

Integrations with CI/CD tools

No

Model packaging chevron
Native packaging system

No

Yes

Compatibility with packaging protocols (ONNX, etc)

No

Yes

One model one file or flexible structure

No

One model one file

Integrations with packaging frameworks

No

Yes

Integrations and Support

Languages chevron
Java

No

Yes

Julia

No

Yes

Python

Yes

Yes

R

No

Yes

REST API

No

No

Model training chevron
Catalyst

Yes

No

CatBoost

Yes

Yes

fastai

Yes

Yes

Gluon

No

Yes

HuggingFace

Yes

Yes

H2O

No

Yes

LightGBM

Yes

Yes

Paddle

No

Yes

PyTorch

Yes

Yes

PyTorch Ignite

Yes

No

PyTorch Lightning

Yes

Yes

Scikit Learn

Yes

Yes

Skorch

Yes

No

Spacy

No

No

Spark MLlib

No

Yes

Statsmodel

No

No

TesorFlow / Keras

Yes

Yes

XGBoost

Yes

Yes

Hyperparameter Optimization chevron
Hyperopt

No

Yes

Keras Tuner

No

Optuna

Yes

Yes

Ray Tune

No

No

Scikit-Optimize

No

Model visualization and debugging chevron
DALEX

Yes

No

Netron

No

No

SHAP

No

Yes

TensorBoard

Yes

Yes

IDEs and Notebooks chevron
JupyterLab and Jupyter Notebook

Yes

Yes

Google Colab

Yes

No

Deepnote

Yes

No

AWS SageMaker

Yes

N/A

Data versioning chevron
DVC

Yes

Yes

Orchestration and pipelining chevron
Airflow

Yes

Yes

Argo

No

No

Kedro

Yes

Yes

Kubeflow

No

Yes

ZenML

Yes

Yes

Experiment tracking tools chevron
MLflow

Yes

No

Sacred

Yes

No

TensorBoard

Yes

CI/CD chevron
GitHub Actions

Yes

Yes

Gitlab CI

No

Yes

CircleCI

No

Yes

Travis

No

No

Jenkins

No

No

Model serving chevron
Seldon

No

Yes

Databricks

No

No

Model versioning chevron
Seldon

No

No

Fiddler.ai

No

Yes

Arthur.ai

No

No

LLMs chevron
LangChain

No

Yes

This table has been updated on 6 November 2023. Some information may be outdated.
Report outdated information here.

What are the key advantages of Neptune then?

  • Standalone component, easy-to-integrate with multiple ML frameworks
  • Customizable metadata structure and custom compare dashboards
  • Dataset and model versioning
  • More developed collaboration and sharing features

Already using Amazon SageMaker?

You can use Neptune to improve the tracking component. SageMaker and Neptune can be easily integrated and provide even more value together.

Many teams have successfully integrated Neptune with their SageMaker pipelines and achieved better results this way.

It only takes 5 minutes to integrate Neptune with your code

Don’t overthink it