v2.3.0-rc1
版本发布时间: 2024-07-08 21:02:27
deepset-ai/haystack最新发布版本:v2.4.0(2024-08-15 17:39:00)
Release Notes
Highlights
Adding the DocxToDocument component to convert Docx files to Documents.
⬆️ Upgrade Notes
-
trafilatura must now be manually installed with pip install trafilatura to use the HTMLToDocument Component.
-
The deprecated converter_name parameter has been removed from PyPDFToDocument.
To specify a custom converter for PyPDFToDocument, use the converter initialization parameter and pass an instance of a class that implements the PyPDFConverter protocol.
The PyPDFConverter protocol defines the methods convert, to_dict and from_dict. A default implementation of PyPDFConverter is provided in the DefaultConverter class.
-
Deprecated HuggingFaceTEITextEmbedder and HuggingFaceTEIDocumentEmbedder have been removed. Use HuggingFaceAPITextEmbedder and HuggingFaceAPIDocumentEmbedder instead.
-
Deprecated HuggingFaceTGIGenerator and HuggingFaceTGIChatGenerator have been removed. Use HuggingFaceAPIGenerator and HuggingFaceAPIChatGenerator instead.
🚀 New Features
- Added custom filters support to ConditionalRouter. Users can now pass in one or more custom Jinja2 filter callables and be able to access those filters when defining condition expressions in routes.
- Added a new mode in JoinDocuments, Distribution-based rank fusion as [the article](https://medium.com/plain-simple-software/distribution-based-score-fusion-dbsf-a-new-approach-to-vector-search-ranking-f87c37488b18)
- Adding the DocxToDocument component inside the converters category. It uses the python-docx library to convert Docx files to haystack Documents.
- Added haystack-experimental to the project's dependencies to enable automatic use of cutting-edge features from Haystack. Users can now access components from haystack-experimental by simply importing them from haystack_experimental instead of haystack. For more information, visit https://github.com/deepset-ai/haystack-experimental.
- Add a PPTX to Document converter using the python-pptx library. Extracts all text from each slide. Each slide is separated with a page break "f" so a Document Splitter could split by slide.
- The DocumentSplitter now has support for the split_id and split_overlap to allow for more control over the splitting process.
- Introduces the TransformersTextRouter! This component uses a transformers text classification pipeline to route text inputs onto different output connections based on the labels of the chosen text classification model.
- Add memory sharing between different instances of InMemoryDocumentStore. Setting the same index argument as another instance will make sure that the memory is shared. e.g.
`python index = "my_personal_index" document_store_1 = InMemoryDocumentStore(index=index) document_store_2 = InMemoryDocumentStore(index=index) assert document_store_1.count_documents() == 0 assert document_store_2.count_documents() == 0 document_store_1.write_documents([Document(content="Hello world")]) assert document_store_1.count_documents() == 1 assert document_store_2.count_documents() == 1
` - Add a new missing_meta param to MetaFieldRanker, which determines what to do with documents that lack the ranked meta field. Supported values are "bottom" (which puts documents with missing meta at the bottom of the sorted list), "top" (which puts them at the top), and "drop" (which removes them from the results entirely).
⚡️ Enhancement Notes
- Added the apply_filter_policy function to standardize the application of filter policies across all document store-specific retrievers, allowing for consistent handling of initial and runtime filters based on the chosen policy (replace or merge).
- Added a new parameter to EvaluationRunResult.comparative_individual_scores_report() to specify columns to keep in the comparative DataFrame.
- Added the 'remove_component' method in 'PipelineBase' to delete components and its connections.
- Added serialization methods save_to_disk and write_to_disk to InMemoryDocumentStore.
- When using "openai" for the LLM-based evaluators the metadata from OpenAI will be in the output dictionary, under the key "meta".
- Remove trafilatura as direct dependency and make it a lazily imported one
- Renamed component from DocxToDocument to DOCXToDocument to follow the naming convention of other converter components.
- Made JSON schema validator compatible with all LLM by switching error template handling to a single user message. Also reduce cost by only including last error instead of full message history.
- Enhanced flexibility in HuggingFace API environment variable names across all related components to support both 'HF_API_TOKEN' and 'HF_TOKEN', improving compatibility with the widely used HF environmental variable naming conventions.
- Updated the ContextRelevance evaluator prompt, explicitly asking to score each statement.
- Improve LinkContentFetcher to support a broader range of content types, including glob patterns for text, application, audio, and video types. This update introduces a more flexible content handler resolution mechanism, allowing for direct matches and pattern matching, thereby greatly improving the handler's adaptability to various content types encountered on the web.
- Add max_retries to AzureOpenAIGenerator. AzureOpenAIGenerator can now be initialised by setting max_retries. If not set, it is inferred from the OPENAI_MAX_RETRIES environment variable or set to 5. The timeout for AzureOpenAIGenerator, if not set, it is inferred from the OPENAI_TIMEOUT environment variable or set to 30.
- Introduced a 'filter_policy' init parameter for both InMemoryBM25Retriever and InMemoryEmbeddingRetriever, allowing users to define how runtime filters should be applied with options to either 'replace' the initial filters or 'merge' them, providing greater flexibility in filtering query results.
- Pipeline serialization to YAML now supports tuples as field values.
- Add support for [structlog context variables](https://www.structlog.org/en/24.2.0/contextvars.html) to structured logging.
- AnswerBuilder can now accept ChatMessages as input in addition to strings. When using ChatMessages, metadata will be automatically added to the answer.
- Update the error message when the sentence-transformers library is not installed and the used component requires it.
⚠️ Deprecation Notes
- The output of the ContextRelevanceEvaluator will change in Haystack 2.4.0. Contexts will be scored as a whole instead of individual statements and only the relevant sentences will be returned. A score of 1 is now returned if a relevant sentence is found, and 0 otherwise.
🐛 Bug Fixes
- SASEvaluator now raises a ValueError if a None value is contained in the predicted_answers input.
- Auto enable tracing upon import if ddtrace or opentelemetry is installed.
- Meta handling of bytestreams in Azure OCR has been fixed.
- Use new filter syntax in the CacheChecker component instead of legacy one.
- Solve serialization bug on 'ChatPromptBuilder' by creating 'to_dict' and 'from_dict' methods on 'ChatMessage' and 'ChatPromptBuilder'.
- Fix some bugs running a Pipeline that has Components with conditional outputs. Some branches that were expected not to run would run anyway, even if they received no inputs. Some branches instead would cause the Pipeline to get stuck waiting to run that branch, even if they received no inputs. The behaviour would depend whether the Component not receiving the input has a optional input or not.
- Fixed the calculation for MRR and MAP scores.
- Fix the deserialization of pipelines containing evaluator components that were subclasses of LLMEvaluator.
- Fix recursive JSON type conversion in the schema validator to be less aggressive (no infinite recursion).
- Adds the missing 'organization' parameter to the serialization function.
- Correctly serialize tuples and types in the init parameters of the LLMEvaluator component and its subclasses.
- Pin numpy<2 to avoid breaking changes that cause several core integrations to fail. Pin tenacity too (8.4.0 is broken).