-
Notifications
You must be signed in to change notification settings - Fork 256
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DOCS-1186 AI Accelerator updates for 2.1.1 #6445
base: develop
Are you sure you want to change the base?
Conversation
Signed-off-by: Dj Walker-Morgan <dj.walker-morgan@enterprisedb.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
2.1.1 release is scheduled for Monday (Feb 3rd) right now.
Signed-off-by: Dj Walker-Morgan <dj.walker-morgan@enterprisedb.com>
FYI; this release includes the first pieces for an upcoming "document prep pipeline"; i.e. SQL primitives to chunk text and similar. We're dark-launching these for now so they should not be mentioned in the release notes. Interfaces are subject to change and not everything is finished yet. ... as discussed with @noahbaculi |
@@ -21,7 +21,7 @@ The starting point for this process is creating a model. When you create a model | |||
```sql | |||
select aidb.create_model( | |||
'my_local_ollama', | |||
'openai_embeddings', | |||
'embeddings', | |||
'{"model":"llama3.3", "url":"http://llama.local:11434/v1/embeddings", "dimensions":8192}'::JSONB, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I believe we need to add a warning somewhere after this example about the pgvector limitation. Since it doesn't support storing over 2000 vectors. Otherwise, this example can be misleading.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Trying to remember how I got to 8192... will bring the local box up and see about shrinking it
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The code may indicate the maximum is 8192, but there. is a non-obvious technical limitation of 2000.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good, only one minor nit on language.
|
||
## About Completions | ||
|
||
Completions is a text completion model that enables use of any OpenAI API compatible text generation model. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Since this is an interface for a common API format I would probably drop the first half of this sentence and say something like:
Enables use of any OpenAI API compatible text generation model.
What Changed?
Updates for credentials and replacement of credentials.