First, create a .streamlit/secrets.toml
file such that:
# fill in <your value>
DSBA_LLAMA3_KEY="<your key>"
MODAL_BASE_URL="https://<your url>--vllm-openai-compatible-serve.modal.run"
$ python3.11 -m venv venv
$ source venv/bin/activate
$ python -m pip install -r requirements.txt
$ python -m streamlit run app.py
Make sure you have a Modal account.
First, sign in:
# sign in
$ python -m modal setup
Then set Modal secrets first as dsba-llama3-key
with the secret name DSBA_LLAMA3_KEY
and modal-base-url
as MODAL_BASE_URL
which is your LLM serving endpoint (not including v1/
).
You can run a temporary "dev" environment to test:
# to test
$ modal serve modal/serve_streamlit.py
Or deploy it as a new app to modal:
# when ready to deploy
$ modal deploy modal/serve_streamlit.py