CONDA IS NOT NEEDED AS A PACKAGE MANAGER. All setup is done using the Python Software Foundation recommended tools: virtualenv and pip and mainstream production tools Docker. Please see PEP 453 "officially recommend the use of pip as the default installer for Python packages"
GitHub Codespaces are FREE for education and as are GPU Codespaces as of this writing in December 2022
- First thing to do on launch is to open a new shell and verify virtualenv is sourced.
Things included are:
-
Makefile
-
Pytest
-
pandas
-
Pylint
-
Dockerfile
-
GitHub copilot
-
jupyter
andipython
-
Most common Python libraries for ML/DL and Hugging Face
-
githubactions
- Zero-shot classification: ./hugging-face/zero_shot_classification.py classify
- Yake for candidate label creation: ./utils/kw_extract.py
docker run -it --rm -p 8888:8888 -p 3000:3000 -p 3001:3001 bentoml/quickstart:latest
The following examples test out the GPU (including Docker GPU)
- run pytorch training test:
python utils/quickstart_pytorch.py
- run pytorch CUDA test:
python utils/verify_cuda_pytorch.py
- run tensorflow training test:
python utils/quickstart_tf2.py
- run nvidia monitoring test:
nvidia-smi -l 1
it should show a GPU - run whisper transcribe test
./utils/transcribe-whisper.sh
and verify GPU is working withnvidia-smi -l 1
- run
lspci | grep -i nvidia
you should see something like:0001:00:00.0 3D controller: NVIDIA Corporation GV100GL [Tesla V100 PCIe 16GB] (rev a1)
Additionally, this workspace is setup to fine-tune Hugging Face
python hugging-face/hf_fine_tune_hello_world.py
Because of potential versioning conflicts between PyTorch and Tensorflow it is recommended to run Tensorflow via GPU Container and PyTorch via default environment.
See TensorFlow GPU documentation
-
Run
docker run --gpus all -it --rm tensorflow/tensorflow:latest-gpu \ python -c "import tensorflow as tf; print(tf.reduce_sum(tf.random.normal([1000, 1000])))"
-
Also interactively explore:
docker run --gpus all -it --rm tensorflow/tensorflow:latest-gpu
, then when inside run:apt-get update && apt-get install pciutils
thenlspci | grep -i nvidia
-
To mount the code into your container:
docker run --gpus all -it --rm -v $(pwd):/tmp tensorflow/tensorflow:latest-gpu /bin/bash
. Then doapt-get install -y git && cd /tmp
. Then all you need to do is runmake install
. Now you can verify you can train deep learning models by doingpython utils/quickstart_tf2.py
https://www.tensorflow.org/resources/recommendation-systems
# Deploy the retrieval model with TensorFlow Serving
docker run -t --rm -p 8501:8501 \
-v "RETRIEVAL/MODEL/PATH:/models/retrieval" \
-e MODEL_NAME=retrieval tensorflow/serving &
Used as the base and customized in the following Duke MLOps and Applied Data Engineering Coursera Labs:
- MLOPs-C2-Lab1-CICD
- MLOps-C2-Lab2-PokerSimulator
- MLOps-C2-Final-HuggingFace
- Coursera-MLOps-C2-lab3-probability-simulations
- Coursera-MLOps-C2-lab4-greedy-optimization
- Docker "one-liners" for Tensorflow recommenders
- Watch GitHub Universe Talk: Teaching MLOps at scale with Github
- Building Cloud Computing Solutions at Scale Specialization
- Python, Bash and SQL Essentials for Data Engineering Specialization
- Implementing MLOps in the Enterprise
- Practical MLOps: Operationalizing Machine Learning Models
- Coursera-Dockerfile