mirror of
https://github.com/jehna/humanify.git
synced 2026-04-27 09:35:58 +03:00
[GH-ISSUE #16] High memory usage when using local inference with llama2-7b model #15
Labels
No labels
bug
enhancement
pull-request
wontfix
No milestone
No project
No assignees
1 participant
Notifications
Due date
No due date set.
Dependencies
No dependencies set.
Reference
starred/humanify#15
Loading…
Add table
Add a link
Reference in a new issue
No description provided.
Delete branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @fear-rush on GitHub (Jun 20, 2024).
Original GitHub issue: https://github.com/jehna/humanify/issues/16
Is there any workarounds that i can do to reduce the memory usage? For the context i'm using Macbook M1 Pro with 16GB memory and the memory usage when running local inference is too high. It uses around 34GB of memory. I tried to Convert model parameters to half-precision (float16) to reduce memory usage but still doesnt work.
@jehna commented on GitHub (Aug 12, 2024):
@fear-rush The new v2 now has a much smaller default model that should run on 8gb RAM.