-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
merge cog-trt-llm into this repo #38
Conversation
yorickvP
commented
May 14, 2024
- subtree merge cog-trt-llm
- ignore cog-trt-llm subdir in runners
- use cog-trt-llm as root dir in builder
…ATH to fix libnvinfer.so.9 import
Update README.md
Co-authored-by: Nathan Raw <[email protected]>
Make instructions more noob friendly & fix papercuts
Read TRTLLM_DIR from env with fallback
# copy cog-trt-llm source into /src | ||
cognix.postCopyCommands = '' | ||
cp ${config.deps.cog-trt-llm}/{*.py,cog-trt-llm-config.yaml} $out/src/ | ||
''; | ||
cognix.rootPath = lib.mkForce "${./cog-trt-llm}"; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
doesn't this result in a pretty different image?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
/src
will be different, just containing the data from cog-trt-llm
instead of a weird merge. But that should be it.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I feel like I'd prefer to keep the weird merge as it is, I think
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@technillogue, could you tldr the costs/benefits as you see them? I don't think I understand your preference, but I'd like to :)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I was just thinking of keeping the output images the way things are now, but I don't actually care that much
This seems good to me. Presumably you've tested the builds? |
[submodule "tensorrtllm_backend"] | ||
path = tensorrtllm_backend | ||
url = https://github.com/triton-inference-server/tensorrtllm_backend.git |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
we should probably drop this
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
haven't dropped it because that would break the 'legacy' dockerfile cog-trt-llm build
*.bin | ||
*.safetensors | ||
*.cog | ||
*.hypothesis | ||
*.pytest_cache | ||
__pycache__/ | ||
*.pyc | ||
models/* |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this doesn't do anything