-
Notifications
You must be signed in to change notification settings - Fork 243
Conversation
mentat/llm_api_handler.py
Outdated
@@ -170,6 +175,10 @@ class Model: | |||
"text-embedding-ada-002": Model( | |||
"text-embedding-ada-002", 8191, 0.0001, 0, embedding_model=True | |||
), | |||
# Fine-tuned on Jan-6 2024 with `sampler-one-hundred-v1.jsonl` data | |||
"ft:gpt-3.5-turbo-1106:abante::8dsQMc4F": Model( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Unfortunately only abante ai org ai keys will be able to use this model so we shouldn't add it to the list.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm not sure this makes much sense; like @jakethekoenig mentioned, the finetuned models can only be used by us. How about we just makes the include_system_prompt a config option?
I get what you mean about the fine-tuned 3.5, makes sense ya I'll remove that one for now. I do think it's good to build-in support for it though. I was planning to do a lil video about the end-to-end process of making our fine-tuned 3.5, with the thought that any of our users could do the same thing with their data. Maybe |
I like this idea. Back when I was doing this I thought we could eventually integrate it such that you could run a
I like this idea. Mentat should know the costs to use a fine tuned gpt-3.5. |
Added a wrapper class around Removed the new
|
mentat/config.py
Outdated
@@ -36,17 +36,19 @@ class Config: | |||
# Model specific settings | |||
model: str = attr.field( | |||
default="gpt-4-1106-preview", | |||
metadata={"auto_completions": list(known_models.keys())}, | |||
metadata={"auto_completions": list(known_models.asdict().keys())}, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we have to do asdict? Isn't it technically a dict since it extends Dict?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thb I spent an hour trying to figure out how to have an overridden keys
method return a type that my linter was happy with. Tried typing.KeyView
, tried returning self.model.keys() directly - it wouldn't take.
attrs also has an asdict method which we use here and there so it seemed ok.
EDIT: nevermind I figured it out.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Fix the fine-tuning output format, add a way for fine-tune models to exclude the system_prompt, and patch up some issues I found.
Pull Request Checklist