This toy-submission contains a dockerfile that exposes a HTTP server. Requests will be made against this server during the evaluation phase of the competition
Make sure you have recursively cloned the top this repository in order to get lit-gpt.
❗ Make sure the repo is cloned with git submodule support either:
git clone --recurse-submodules ...
or if you cloned the repo but are missing the lit-gpt
folder
git submodule update --init --recursive
- lit-gpt/
- unmodified submodule that contains a hackable
torch.nn.Module
GPT definition as well as optional fine-tuning and inference code.
- unmodified submodule that contains a hackable
- main.py
- The process/ and tokenize/ endpoints are defined here
- helper.py
- Applies logic on top of lit-gpt's generate in order to produce responses in accordance with the spec.
- api.py
- Defines the pydantic classes for the FASTapi server
- Dockerfile
- Definition of the image that will set-up the server used for submissions
Follow this guide to install nvidia-ctk.
nvidia-ctk runtime configure
systemctl restart docker
docker build -t toy_submission .
docker run --gpus all -p 8080:80 toy_submission
curl -X POST -H "Content-Type: application/json" -d '{"prompt": "The capital of france is "}' http://localhost:8080/process