mirror of
https://github.com/jehna/humanify.git
synced 2026-04-27 09:35:58 +03:00
[GH-ISSUE #53] Better error handling/user guidance for missing local models #27
Labels
No labels
bug
enhancement
pull-request
wontfix
No milestone
No project
No assignees
1 participant
Notifications
Due date
No due date set.
Dependencies
No dependencies set.
Reference
starred/humanify#27
Loading…
Add table
Add a link
Reference in a new issue
No description provided.
Delete branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @0xdevalias on GitHub (Aug 23, 2024).
Original GitHub issue: https://github.com/jehna/humanify/issues/53
Currently when trying to run with a local model that isn't downloaded, the app crashes with an error such as the following:
It might be useful to give a more user friendly error that explains how to resolve the issue.
I ran into this while trying to test/replicate the following:
@0xdevalias commented on GitHub (Aug 23, 2024):
I can see that the instructions are in the README here:
Which suggests I need to run
humanify download 2bfirst.I wonder if it might make more sense to have the local model download as a sub-command of
humanify local, as that's where I was first looking for help for how to download the models, and it didn't even occur to me to check the root level command, since things local things seemed to be 'scoped' under thelocalcommand:There also seems to be very minimal information output during the download. It might be nice to know a bit more about which model is being downloaded, from where, where it's being saved, how large it is, etc:
I guess it does provide slightly more info when the download is completed:
I can see it's downloaded here:
And the code for that is here:
github.com/jehna/humanify@85d17e73d6/src/local-models.ts (L13-L25)I also notice that
MODEL_DIRECTORYis hardcoded currently. I wonder if that would be something useful to be able to specify/customize via a CLI arg/env variable/etc.It seems the
humanify localcommand usesgetModelPath:github.com/jehna/humanify@85d17e73d6/src/plugins/local-llm-rename/llama.ts (L19-L22)Which only seems to work for model aliases defined in
MODELS:github.com/jehna/humanify@85d17e73d6/src/local-models.ts (L69-L75)Even though the error text for
humanify downloadsounds as though it would be capable of downloading any named model:github.com/jehna/humanify@85d17e73d6/src/local-models.ts (L77-L85)And usually for LLM apps, the
--modelparam would let us specify arbitrary models from huggingface or similar.Edit: Created a new issue related to the download progress/etc: