What is the simplest way to run inference/ inference endopint with Fin-GPT? #133
Unanswered
datainvestor
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I want to try out the fingpt pre-trained model with some of my prompts. I want to do this as easy as possible and cost-efficient as well.
First I tried running this script in jupyter-notebook deployed on my VM, but its crashing saying that its running out of GPU memory (16GB)
So then I tried what I found on huggingface:
But its taking quite long to load and my VM died before finishing so not sure if this is correct approach.
I was also thinking about just creating inference endpoint api which I can access like for exampl this Mistral model:
https://docs.mistral.ai/cloud-deployment/skypilot
But I am not sure how this is achievable with pretrained models like Fin GPT where you need to load base model first.
Could someone provide some explanation or recommendations how to achieve this? Basically I just want to run either simple script with my prompt using fin-gpt model or even better run an API like gpt-4 OpenAPI or similar.
Beta Was this translation helpful? Give feedback.
All reactions