No need to fear from Cuda and Nvidia installation
Use Cases and Deployment Scope
We use AWS EC2 for training machine learning models. For spinning up these instances, we use prebuilt AMIs specific for Deep Learning. These AMIs help a lot in setting up the environment very easily without any worries of installing CUDA, Nvidia Drivers, and then specific libraries like PyTorch, and Tensorflow. Support for various Conda environments is also available. The amazing part of this is that it provides support for different types of machines like Ubuntu, Linux, Amazon OS, etc.
Pros
- Setting up environment
- Support for different types of machines
- Perfect for Machine Learning / Deep Learning use cases
- Nvidia / Cuda / Conda support easily
Cons
- Simpler documentation of different types of AMIs
- Clearly listing out different types of machines as I got confused and spinned up an AMI in Amazon Linux machine instead of Ubuntu
- Support for latest version of libraries, to avoid manually updating them after launch
Likelihood to Recommend
Suitable:
1. Best for quickly setting up an instance with pre-installed libraries.
2. Ideal for people in Deep Learning space who struggle with Cuda / Nvidia driver installations.
Not suitable:
1. People who want to install custom libraries or different version of those.
2. In these cases, updating the version of libraries many times leads to version mismatch which can cause many errors.