| |
| |
| |
|
|
| import re |
| from copy import deepcopy |
| from functools import partial, reduce |
| from itertools import chain |
| from typing import Generator, List, Literal, Optional, Set |
| from unicodedata import normalize |
|
|
| from haystack import Document, component, logging |
|
|
| logger = logging.getLogger(__name__) |
|
|
|
|
| @component |
| class DocumentCleaner: |
| """ |
| Cleans the text in the documents. |
| |
| It removes extra whitespaces, |
| empty lines, specified substrings, regexes, |
| page headers and footers (in this order). |
| |
| ### Usage example: |
| |
| ```python |
| from haystack import Document |
| from haystack.components.preprocessors import DocumentCleaner |
| |
| doc = Document(content="This is a document to clean\\n\\n\\nsubstring to remove") |
| |
| cleaner = DocumentCleaner(remove_substrings = ["substring to remove"]) |
| result = cleaner.run(documents=[doc]) |
| |
| assert result["documents"][0].content == "This is a document to clean " |
| ``` |
| """ |
|
|
| def __init__( |
| self, |
| remove_empty_lines: bool = True, |
| remove_extra_whitespaces: bool = True, |
| remove_repeated_substrings: bool = False, |
| keep_id: bool = False, |
| remove_substrings: Optional[List[str]] = None, |
| remove_regex: Optional[str] = None, |
| unicode_normalization: Optional[Literal["NFC", "NFKC", "NFD", "NFKD"]] = None, |
| ascii_only: bool = False, |
| ): |
| """ |
| Initialize DocumentCleaner. |
| |
| :param remove_empty_lines: If `True`, removes empty lines. |
| :param remove_extra_whitespaces: If `True`, removes extra whitespaces. |
| :param remove_repeated_substrings: If `True`, removes repeated substrings (headers and footers) from pages. |
| Pages must be separated by a form feed character "\\f", |
| which is supported by `TextFileToDocument` and `AzureOCRDocumentConverter`. |
| :param remove_substrings: List of substrings to remove from the text. |
| :param remove_regex: Regex to match and replace substrings by "". |
| :param keep_id: If `True`, keeps the IDs of the original documents. |
| :param unicode_normalization: Unicode normalization form to apply to the text. |
| Note: This will run before any other steps. |
| :param ascii_only: Whether to convert the text to ASCII only. |
| Will remove accents from characters and replace them with ASCII characters. |
| Other non-ASCII characters will be removed. |
| Note: This will run before any pattern matching or removal. |
| """ |
|
|
| self._validate_params(unicode_normalization=unicode_normalization) |
|
|
| self.remove_empty_lines = remove_empty_lines |
| self.remove_extra_whitespaces = remove_extra_whitespaces |
| self.remove_repeated_substrings = remove_repeated_substrings |
| self.remove_substrings = remove_substrings |
| self.remove_regex = remove_regex |
| self.keep_id = keep_id |
| self.unicode_normalization = unicode_normalization |
| self.ascii_only = ascii_only |
|
|
| def _validate_params(self, unicode_normalization: Optional[str]): |
| """ |
| Validate the parameters of the DocumentCleaner. |
| |
| :param unicode_normalization: Unicode normalization form to apply to the text. |
| :raises ValueError: if the parameters are not valid. |
| """ |
| if unicode_normalization and unicode_normalization not in ["NFC", "NFKC", "NFD", "NFKD"]: |
| raise ValueError("unicode_normalization must be one of 'NFC', 'NFKC', 'NFD', 'NFKD'.") |
|
|
| @component.output_types(documents=List[Document]) |
| def run(self, documents: List[Document]): |
| """ |
| Cleans up the documents. |
| |
| :param documents: List of Documents to clean. |
| |
| :returns: A dictionary with the following key: |
| - `documents`: List of cleaned Documents. |
| |
| :raises TypeError: if documents is not a list of Documents. |
| """ |
| if not isinstance(documents, list) or documents and not isinstance(documents[0], Document): |
| raise TypeError("DocumentCleaner expects a List of Documents as input.") |
|
|
| cleaned_docs = [] |
| for doc in documents: |
| if doc.content is None: |
| logger.warning( |
| "DocumentCleaner only cleans text documents but document.content for document ID" |
| " %{document_id} is None.", |
| document_id=doc.id, |
| ) |
| cleaned_docs.append(doc) |
| continue |
| text = doc.content |
|
|
| if self.unicode_normalization: |
| text = self._normalize_unicode(text, self.unicode_normalization) |
| if self.ascii_only: |
| text = self._ascii_only(text) |
| if self.remove_extra_whitespaces: |
| text = self._remove_extra_whitespaces(text) |
| if self.remove_empty_lines: |
| text = self._remove_empty_lines(text) |
| if self.remove_substrings: |
| text = self._remove_substrings(text, self.remove_substrings) |
| if self.remove_regex: |
| text = self._remove_regex(text, self.remove_regex) |
| if self.remove_repeated_substrings: |
| text = self._remove_repeated_substrings(text) |
|
|
| cleaned_docs.append(Document(content=text, meta=deepcopy(doc.meta), id=doc.id if self.keep_id else "")) |
|
|
| return {"documents": cleaned_docs} |
|
|
| def _normalize_unicode(self, text: str, form: Literal["NFC", "NFKC", "NFD", "NFKD"]) -> str: |
| """ |
| Normalize the unicode of the text. |
| |
| :param text: Text to normalize. |
| :param form: Unicode normalization form to apply to the text. |
| Options: "NFC", "NFKC", "NFD", "NFKD". |
| :returns: The normalized text. |
| """ |
| return normalize(form, text) |
|
|
| def _ascii_only(self, text: str) -> str: |
| """ |
| Convert the text to ASCII only. |
| |
| Will remove accents from characters and replace them with ASCII characters. |
| Other non-ASCII characters will be removed. |
| |
| :param text: Text to convert to ASCII only. |
| :returns: The text in ASCII only. |
| """ |
|
|
| |
| |
| return self._normalize_unicode(text, "NFKD").encode("ascii", "ignore").decode("utf-8") |
|
|
| def _remove_empty_lines(self, text: str) -> str: |
| """ |
| Remove empty lines and lines that contain nothing but whitespaces from text. |
| |
| :param text: Text to clean. |
| :returns: The text without empty lines. |
| """ |
| pages = text.split("\f") |
| cleaned_pages = ["\n".join(line for line in page.split("\n") if line.strip()) for page in pages] |
| return "\f".join(cleaned_pages) |
|
|
| def _remove_extra_whitespaces(self, text: str) -> str: |
| """ |
| Remove extra whitespaces from text. |
| |
| :param text: Text to clean. |
| :returns: The text without extra whitespaces. |
| """ |
| texts = text.split("\f") |
| cleaned_text = [re.sub(r"\s\s+", " ", text).strip() for text in texts] |
| return "\f".join(cleaned_text) |
|
|
| def _remove_regex(self, text: str, regex: str) -> str: |
| """ |
| Remove substrings that match the specified regex from the text. |
| |
| :param text: Text to clean. |
| :param regex: Regex to match and replace substrings by "". |
| :returns: The text without the substrings that match the regex. |
| """ |
| texts = text.split("\f") |
| cleaned_text = [re.sub(regex, "", text).strip() for text in texts] |
| return "\f".join(cleaned_text) |
|
|
| def _remove_substrings(self, text: str, substrings: List[str]) -> str: |
| """ |
| Remove all specified substrings from the text. |
| |
| :param text: Text to clean. |
| :param substrings: Substrings to remove. |
| :returns: The text without the specified substrings. |
| """ |
| for substring in substrings: |
| text = text.replace(substring, "") |
| return text |
|
|
| def _remove_repeated_substrings(self, text: str) -> str: |
| """ |
| Remove any substrings from the text that occur repeatedly on every page. For example headers or footers. |
| |
| Pages in the text need to be separated by form feed character "\f". |
| :param text: Text to clean. |
| :returns: The text without the repeated substrings. |
| """ |
| return self._find_and_remove_header_footer( |
| text, n_chars=300, n_first_pages_to_ignore=1, n_last_pages_to_ignore=1 |
| ) |
|
|
| def _find_and_remove_header_footer( |
| self, text: str, n_chars: int, n_first_pages_to_ignore: int, n_last_pages_to_ignore: int |
| ) -> str: |
| """ |
| Heuristic to find footers and headers across different pages by searching for the longest common string. |
| |
| Pages in the text need to be separated by form feed character "\f". |
| For headers, we only search in the first n_chars characters (for footer: last n_chars). |
| Note: This heuristic uses exact matches and therefore works well for footers like "Copyright 2019 by XXX", |
| but won't detect "Page 3 of 4" or similar. |
| |
| :param n_chars: The number of first/last characters where the header/footer shall be searched in. |
| :param n_first_pages_to_ignore: The number of first pages to ignore |
| (e.g. TOCs often don't contain footer/header). |
| :param n_last_pages_to_ignore: The number of last pages to ignore. |
| :returns: The text without the found headers and footers. |
| """ |
|
|
| pages = text.split("\f") |
|
|
| |
| start_of_pages = [p[:n_chars] for p in pages[n_first_pages_to_ignore:-n_last_pages_to_ignore]] |
| found_header = self._find_longest_common_ngram(start_of_pages) |
| if found_header: |
| pages = [page.replace(found_header, "") for page in pages] |
|
|
| |
| end_of_pages = [p[-n_chars:] for p in pages[n_first_pages_to_ignore:-n_last_pages_to_ignore]] |
| found_footer = self._find_longest_common_ngram(end_of_pages) |
| if found_footer: |
| pages = [page.replace(found_footer, "") for page in pages] |
|
|
| logger.debug( |
| "Removed header '{header}' and footer '{footer}' in document", header=found_header, footer=found_footer |
| ) |
| text = "\f".join(pages) |
| return text |
|
|
| def _ngram(self, seq: str, n: int) -> Generator[str, None, None]: |
| """ |
| Return all ngrams of length n from a text sequence. Each ngram consists of n words split by whitespace. |
| |
| :param seq: The sequence to generate ngrams from. |
| :param n: The length of the ngrams to generate. |
| :returns: A Generator generating all ngrams of length n from the given sequence. |
| """ |
|
|
| |
| |
| seq = seq.replace("\n", " \n") |
| seq = seq.replace("\t", " \t") |
|
|
| words = seq.split(" ") |
| ngrams = ( |
| " ".join(words[i : i + n]).replace(" \n", "\n").replace(" \t", "\t") for i in range(0, len(words) - n + 1) |
| ) |
|
|
| return ngrams |
|
|
| def _allngram(self, seq: str, min_ngram: int, max_ngram: int) -> Set[str]: |
| """ |
| Generates all possible ngrams from a given sequence of text. |
| |
| Considering all ngram lengths between the minimum and maximum length. |
| |
| :param seq: The sequence to generate ngrams from. |
| :param min_ngram: The minimum length of ngram to consider. |
| :param max_ngram: The maximum length of ngram to consider. |
| :returns: A set of all ngrams from the given sequence. |
| """ |
| lengths = range(min_ngram, max_ngram) if max_ngram else range(min_ngram, len(seq)) |
| ngrams = map(partial(self._ngram, seq), lengths) |
| res = set(chain.from_iterable(ngrams)) |
| return res |
|
|
| def _find_longest_common_ngram(self, sequences: List[str], min_ngram: int = 3, max_ngram: int = 30) -> str: |
| """ |
| Find the longest common ngram across a list of text sequences (e.g. start of pages). |
| |
| Considering all ngram lengths between the minimum and maximum length. Helpful for finding footers, headers etc. |
| Empty sequences are ignored. |
| |
| :param sequences: The list of strings that shall be searched for common n_grams. |
| :param max_ngram: The maximum length of ngram to consider. |
| :param min_ngram: The minimum length of ngram to consider. |
| :returns: The longest ngram that all sequences have in common. |
| """ |
| sequences = [s for s in sequences if s] |
| if not sequences: |
| return "" |
| seqs_ngrams = map(partial(self._allngram, min_ngram=min_ngram, max_ngram=max_ngram), sequences) |
| intersection = reduce(set.intersection, seqs_ngrams) |
|
|
| longest = max(intersection, key=len, default="") |
| return longest if longest.strip() else "" |
|
|