v0.6.0
版本发布时间: 2024-08-02 17:28:42
zylon-ai/private-gpt最新发布版本:v0.6.2(2024-08-09 00:16:58)
0.6.0 (2024-08-02)
What's new
Introducing Recipes!
Recipes are high-level APIs that represent AI-native use cases. Under the hood, recipes execute complex pipelines to get the work done.
With the introduction of the first recipe, summarize
, our aim is not only to include that useful use case in PrivateGPT but also getting the project ready to onboard community-built recipes!
Summarization Recipe
summarize
is the first recipe
included in PrivateGPT. The new API lets users summarize ingested documents, customize the resulting summary and use it as streaming. Read the full documentation here.
POST /v1/summarize
Improved cold-start
We've put a lot of effort to run PrivateGPT from a fresh clone as straightforward as possible, defaulting to Ollama, auto-pulling models, making the tokenizer optional...
More models and databases support
Support for Gemini (both LLM and Embeddings) and for Milvus and Clickhouse vector databases.
Breaking changes
-
The minimum required Python version is now 3.11.9, and Poetry must be >= 1.7.1. However, we recommend updating to Poetry 1.8.3. Instructions for updating:
-
Python 3.11.9:
-
Before proceeding, make sure
pyenv
is installed on your system. If it isn't, you can install it by following the instructions on the PrivateGPT documentation. -
Use
pyenv
to install the specific version of Python.pyenv install 3.11.9
-
Verify the installation by running
python --version
in your terminal.
-
-
Poetry 1.8.3:
-
Update Poetry if already installed:
poetry self update 1.8.3
-
Verify the installation by running
poetry --version
in your terminal.
-
-
-
Default LLM model to LLaMA 3.1 for both Ollama and Llamacpp local setups. If you want to keep on using v0.5.0 defaults, place this settings-legacy.yaml file next to your settings.yaml file and run privateGPT with
PGPT_PROFILES=legacy make run
. Learn more about profiles here. -
Default Embeddings to
nomic-embed-text
for both Ollama and Llamacpp local setups. This embeddings model may work with a different dimension than the one you were using before, making it incompatible with already ingested files. If you want to keep on using v0.5.0 defaults to not lose your ingested files, place this settings-legacy.yaml file next to your settings.yaml file and run privateGPT withPGPT_PROFILES=legacy make run
. Learn more about profiles here. As an alternative, if you prefer to start fresh, you can just wipe your existing vector database by removing thelocal_data
folder.
Full Changelog
Features
- bump dependencies (#1987) (b687dc8)
- docs: add privategpt-ts sdk (#1924) (d13029a)
- docs: Fix setup docu (#1926) (067a5f1)
- docs: update doc for ipex-llm (#1968) (19a7c06)
- docs: update documentation and fix preview-docs (#2000) (4523a30)
- llm: add progress bar when ollama is pulling models (#2031) (cf61bf7)
- llm: autopull ollama models (#2019) (20bad17)
- llm: Support for Google Gemini LLMs and Embeddings (#1965) (fc13368)
- make llama3.1 as default (#2022) (9027d69)
- prompt_style applied to all LLMs + extra LLM params. (#1835) (e21bf20)
-
recipe: add our first recipe
Summarize
(#2028) (8119842) - vectordb: Milvus vector db Integration (#1996) (43cc31f)
- vectorstore: Add clickhouse support as vectore store (#1883) (2612928)
Bug Fixes
- "no such group" error in Dockerfile, added docx2txt and cryptography deps (#1841) (947e737)
- config: make tokenizer optional and include a troubleshooting doc (#1998) (01b7ccd)
- docs: Fix concepts.mdx referencing to installation page (#1779) (dde0224)
- docs: Update installation.mdx (#1866) (c1802e7)
- ffmpy dependency (#2020) (dabf556)
- light mode (#2025) (1020cd5)
- LLM: mistral ignoring assistant messages (#1954) (c7212ac)
- llm: special tokens and leading space (#1831) (347be64)
- make embedding_api_base match api_base when on docker (#1859) (2a432bf)
- nomic embeddings (#2030) (5465958)
- prevent to ingest local files (by default) (#2010) (e54a8fe)
- Replacing unsafe
eval()
withjson.loads()
(#1890) (9d0d614) - settings: enable cors by default so it will work when using ts sdk (spa) (#1925) (966af47)
- ui: gradio bug fixes (#2021) (d4375d0)
- unify embedding models (#2027) (40638a1)