trainML Notebooks are full instances of JupyterLab running on up to 4 dedicated GPUs. Our pre-built conda environments are designed specifically for machine learning model training on GPUs, with the latest Tensorflow, PyTorch, MXNet, and others are pre-installed.
Prebuilt, Optimized Environments
trainML's Notebooks run in your choice of pre-built, conda-based Python environments, configured with all the popular deep learning frameworks, their dependencies, and a multitude of other packages to facilitate data analysis and model training. Everything you need for GPU acceleration is already installed to ensure version compatibility with the deep learning frameworks. Pre-installed frameworks include:
- Apache MXNet
GPU acceleration libraries included:
- NVIDIA Driver
Load Models and Data Automatically
Why waste time and effort loading your data and model code into the instance's local storage? trainML's unique job environment can automatically download a git repository, configure access keys, and attach datasets for you. Attached datasets are fully cached on local NVMe storage and do not incur additional storage cost no matter how many notebook instances they are attached to.
Dynamic Instance Type Changes
Unlike other cloud Notebook providers, your notebook instance isn't locked into a specific number of GPUs. With trainML, you can start testing your model on 1 GPU, and when you are ready to scale up for training, you can add more GPUs to the existing notebook job. Once training is done, you scale back down to do some analysis. Since all trainML notebook instances are fully persistent, the notebook can transition seamlessly through these resource changes, allowing you to minimize your expense without compromising time.
Forking and Converting for Rapid Parallel Experimentation
Running a new copy of a notebook on a new dedicated GPU isn't as simple as copying a file. Normally, you have to provision a whole new GPU-enabled computing environment to run the notebook. If you had staged data, manually modified code, or updated configuration data, all that work as to be replicated in the new environment. You also have to remember to reinstall all your packages and libraries manually.
With trainML notebooks, you can create a new copy of a notebook environment in only 3 clicks. Unlike other cloud notebooks, when you fork a trainML notebook, the entire working directory is copied. All datasets, checkpoints, and other data are copied into the new notebook automatically. Since trainML jobs automatically install your requirements files and other packages, these will already be setup before you open the notebook. If your code can run as a script instead of interactively, you can even convert it to an independent training job that can run autonomously, send its output to a location you specify, and automatically terminate when training finishes.