Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Technical requirements for LLM model #238

Open
sellmon opened this issue Jul 12, 2024 · 0 comments
Open

Technical requirements for LLM model #238

sellmon opened this issue Jul 12, 2024 · 0 comments

Comments

@sellmon
Copy link

sellmon commented Jul 12, 2024

I have a few questions about the technical requirements for Codellama model:

  1. What are the minimum and recommended amounts of RAM needed to run the model 70b effectively? How does the amount of RAM affect the performance of the model? For example, if the model takes up 70GB of memory, will having 96GB of RAM provide the same speed as 256GB? Or would more RAM potentially increase performance?

  2. Is there any specific GPU configuration that would be optimal for this model? If I connect 5 GPUs using a rig, but the rig is connected to my computer via 4x PCI x1 and 1x PCI x16 slots, will it improve performance significantly? I do not plan to train the model, I only need it for answers + also taking into account the context from additional documents

  3. Do I need to use PyTorch for this setup or can I use another framework like TensorFlow?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant