mirror of
https://github.com/hwchase17/langchain.git
synced 2025-06-25 08:03:39 +00:00
Description: Adds tags and dataview fields to ObsidianLoader doc metadata. - Issue: #9800, #4991 - Dependencies: none - Tag maintainer: My best guess is @hwchase17 looking through the git logs - Twitter handle: I don't use twitter, sorry!
This commit is contained in:
parent
ce47124e8f
commit
7d1b0fbe79
@ -1,15 +1,24 @@
|
||||
import logging
|
||||
import re
|
||||
from pathlib import Path
|
||||
from typing import List
|
||||
|
||||
import yaml
|
||||
|
||||
from langchain.docstore.document import Document
|
||||
from langchain.document_loaders.base import BaseLoader
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class ObsidianLoader(BaseLoader):
|
||||
"""Load `Obsidian` files from directory."""
|
||||
|
||||
FRONT_MATTER_REGEX = re.compile(r"^---\n(.*?)\n---\n", re.MULTILINE | re.DOTALL)
|
||||
FRONT_MATTER_REGEX = re.compile(r"^---\n(.*?)\n---\n", re.DOTALL)
|
||||
TAG_REGEX = re.compile(r"[^\S\/]#([a-zA-Z_]+[-_/\w]*)")
|
||||
DATAVIEW_LINE_REGEX = re.compile(r"^\s*(\w+)::\s*(.*)$", re.MULTILINE)
|
||||
DATAVIEW_INLINE_BRACKET_REGEX = re.compile(r"\[(\w+)::\s*(.*)\]", re.MULTILINE)
|
||||
DATAVIEW_INLINE_PAREN_REGEX = re.compile(r"\((\w+)::\s*(.*)\)", re.MULTILINE)
|
||||
|
||||
def __init__(
|
||||
self, path: str, encoding: str = "UTF-8", collect_metadata: bool = True
|
||||
@ -30,18 +39,64 @@ class ObsidianLoader(BaseLoader):
|
||||
"""Parse front matter metadata from the content and return it as a dict."""
|
||||
if not self.collect_metadata:
|
||||
return {}
|
||||
|
||||
match = self.FRONT_MATTER_REGEX.search(content)
|
||||
front_matter = {}
|
||||
if match:
|
||||
lines = match.group(1).split("\n")
|
||||
for line in lines:
|
||||
if ":" in line:
|
||||
key, value = line.split(":", 1)
|
||||
front_matter[key.strip()] = value.strip()
|
||||
else:
|
||||
# Skip lines without a colon
|
||||
continue
|
||||
return front_matter
|
||||
if not match:
|
||||
return {}
|
||||
|
||||
try:
|
||||
front_matter = yaml.safe_load(match.group(1))
|
||||
|
||||
# If tags are a string, split them into a list
|
||||
if "tags" in front_matter and isinstance(front_matter["tags"], str):
|
||||
front_matter["tags"] = front_matter["tags"].split(", ")
|
||||
|
||||
return front_matter
|
||||
except yaml.parser.ParserError:
|
||||
logger.warning("Encountered non-yaml frontmatter")
|
||||
return {}
|
||||
|
||||
def _to_langchain_compatible_metadata(self, metadata: dict) -> dict:
|
||||
"""Convert a dictionary to a compatible with langchain."""
|
||||
result = {}
|
||||
for key, value in metadata.items():
|
||||
if type(value) in {str, int, float}:
|
||||
result[key] = value
|
||||
else:
|
||||
result[key] = str(value)
|
||||
return result
|
||||
|
||||
def _parse_document_tags(self, content: str) -> set:
|
||||
"""Return a set of all tags in within the document."""
|
||||
if not self.collect_metadata:
|
||||
return set()
|
||||
|
||||
match = self.TAG_REGEX.findall(content)
|
||||
if not match:
|
||||
return set()
|
||||
|
||||
return {tag for tag in match}
|
||||
|
||||
def _parse_dataview_fields(self, content: str) -> dict:
|
||||
"""Parse obsidian dataview plugin fields from the content and return it
|
||||
as a dict."""
|
||||
if not self.collect_metadata:
|
||||
return {}
|
||||
|
||||
return {
|
||||
**{
|
||||
match[0]: match[1]
|
||||
for match in self.DATAVIEW_LINE_REGEX.findall(content)
|
||||
},
|
||||
**{
|
||||
match[0]: match[1]
|
||||
for match in self.DATAVIEW_INLINE_PAREN_REGEX.findall(content)
|
||||
},
|
||||
**{
|
||||
match[0]: match[1]
|
||||
for match in self.DATAVIEW_INLINE_BRACKET_REGEX.findall(content)
|
||||
},
|
||||
}
|
||||
|
||||
def _remove_front_matter(self, content: str) -> str:
|
||||
"""Remove front matter metadata from the given content."""
|
||||
@ -51,22 +106,29 @@ class ObsidianLoader(BaseLoader):
|
||||
|
||||
def load(self) -> List[Document]:
|
||||
"""Load documents."""
|
||||
ps = list(Path(self.file_path).glob("**/*.md"))
|
||||
paths = list(Path(self.file_path).glob("**/*.md"))
|
||||
docs = []
|
||||
for p in ps:
|
||||
with open(p, encoding=self.encoding) as f:
|
||||
for path in paths:
|
||||
with open(path, encoding=self.encoding) as f:
|
||||
text = f.read()
|
||||
|
||||
front_matter = self._parse_front_matter(text)
|
||||
tags = self._parse_document_tags(text)
|
||||
dataview_fields = self._parse_dataview_fields(text)
|
||||
text = self._remove_front_matter(text)
|
||||
metadata = {
|
||||
"source": str(p.name),
|
||||
"path": str(p),
|
||||
"created": p.stat().st_ctime,
|
||||
"last_modified": p.stat().st_mtime,
|
||||
"last_accessed": p.stat().st_atime,
|
||||
**front_matter,
|
||||
"source": str(path.name),
|
||||
"path": str(path),
|
||||
"created": path.stat().st_ctime,
|
||||
"last_modified": path.stat().st_mtime,
|
||||
"last_accessed": path.stat().st_atime,
|
||||
**self._to_langchain_compatible_metadata(front_matter),
|
||||
**dataview_fields,
|
||||
}
|
||||
|
||||
if tags or front_matter.get("tags"):
|
||||
metadata["tags"] = ",".join(tags | set(front_matter.get("tags", [])))
|
||||
|
||||
docs.append(Document(page_content=text, metadata=metadata))
|
||||
|
||||
return docs
|
||||
|
@ -0,0 +1,9 @@
|
||||
---
|
||||
anArray:
|
||||
one
|
||||
- two
|
||||
- three
|
||||
tags: 'onetag', 'twotag' ]
|
||||
---
|
||||
|
||||
A document with frontmatter that isn't valid.
|
@ -0,0 +1,5 @@
|
||||
---
|
||||
tags: journal/entry, obsidian
|
||||
---
|
||||
|
||||
No other content than the frontmatter.
|
@ -0,0 +1,14 @@
|
||||
### Description
|
||||
#recipes #dessert #cookies
|
||||
|
||||
A document with HR elements that might trip up a front matter parser:
|
||||
|
||||
---
|
||||
|
||||
### Ingredients
|
||||
|
||||
- 3/4 cup (170g) **unsalted butter**, slightly softened to room temperature.
|
||||
- 1 and 1/2 cups (180g) **confectioners’ sugar**
|
||||
|
||||
---
|
||||
|
@ -0,0 +1 @@
|
||||
A markdown document with no additional metadata.
|
@ -0,0 +1,35 @@
|
||||
---
|
||||
aFloat: 13.12345
|
||||
anInt: 15
|
||||
aBool: true
|
||||
aString: string value
|
||||
anArray:
|
||||
- one
|
||||
- two
|
||||
- three
|
||||
aDict:
|
||||
dictId1: '58417'
|
||||
dictId2: 1500
|
||||
tags: [ 'onetag', 'twotag' ]
|
||||
---
|
||||
|
||||
# Tags
|
||||
|
||||
()#notatag
|
||||
#12345
|
||||
#read
|
||||
something #tagWithCases
|
||||
- #tag-with-dash
|
||||
#tag_with_underscore #tag/with/nesting
|
||||
|
||||
# Dataview
|
||||
|
||||
Here is some data in a [dataview1:: a value] line.
|
||||
Here is even more data in a (dataview2:: another value) line.
|
||||
dataview3:: more data
|
||||
notdataview4: this is not a field
|
||||
notdataview5: this is not a field
|
||||
|
||||
# Text content
|
||||
|
||||
https://example.com/blog/#not-a-tag
|
@ -0,0 +1,128 @@
|
||||
from pathlib import Path
|
||||
|
||||
from langchain.document_loaders.obsidian import ObsidianLoader
|
||||
|
||||
OBSIDIAN_EXAMPLE_PATH = Path(__file__).parent / "sample_documents" / "obsidian"
|
||||
STANDARD_METADATA_FIELDS = {
|
||||
"created",
|
||||
"path",
|
||||
"source",
|
||||
"last_accessed",
|
||||
"last_modified",
|
||||
}
|
||||
|
||||
loader = ObsidianLoader(str(OBSIDIAN_EXAMPLE_PATH))
|
||||
docs = loader.load()
|
||||
|
||||
|
||||
def test_page_content_loaded() -> None:
|
||||
"""Verify that all docs have page_content"""
|
||||
assert len(docs) == 5
|
||||
assert all(doc.page_content for doc in docs)
|
||||
|
||||
|
||||
def test_disable_collect_metadata() -> None:
|
||||
"""If collect_metadata is False, no additional metadata should be collected."""
|
||||
loader_without_metadata = ObsidianLoader(
|
||||
str(OBSIDIAN_EXAMPLE_PATH), collect_metadata=False
|
||||
)
|
||||
docs_wo = loader_without_metadata.load()
|
||||
assert len(docs_wo) == 5
|
||||
assert all(doc.page_content for doc in docs_wo)
|
||||
assert all(set(doc.metadata) == STANDARD_METADATA_FIELDS for doc in docs_wo)
|
||||
|
||||
|
||||
def test_metadata_without_frontmatter() -> None:
|
||||
"""Verify docs without frontmatter, still have basic metadata."""
|
||||
doc = next(doc for doc in docs if doc.metadata["source"] == "no_metadata.md")
|
||||
assert set(doc.metadata) == STANDARD_METADATA_FIELDS
|
||||
|
||||
|
||||
def test_metadata_with_frontmatter() -> None:
|
||||
"""Verify a standard frontmatter field is loaded."""
|
||||
doc = next(doc for doc in docs if doc.metadata["source"] == "frontmatter.md")
|
||||
assert set(doc.metadata) == STANDARD_METADATA_FIELDS | {"tags"}
|
||||
assert set(doc.metadata["tags"].split(",")) == {"journal/entry", "obsidian"}
|
||||
|
||||
|
||||
def test_metadata_with_bad_frontmatter() -> None:
|
||||
"""Verify a doc with non-yaml frontmatter."""
|
||||
doc = next(doc for doc in docs if doc.metadata["source"] == "bad_frontmatter.md")
|
||||
assert set(doc.metadata) == STANDARD_METADATA_FIELDS
|
||||
|
||||
|
||||
def test_metadata_with_tags_and_frontmatter() -> None:
|
||||
"""Verify a doc with frontmatter and tags/dataview tags are all added to
|
||||
metadata."""
|
||||
doc = next(
|
||||
doc for doc in docs if doc.metadata["source"] == "tags_and_frontmatter.md"
|
||||
)
|
||||
|
||||
FRONTMATTER_FIELDS = {
|
||||
"aBool",
|
||||
"aFloat",
|
||||
"anInt",
|
||||
"anArray",
|
||||
"aString",
|
||||
"aDict",
|
||||
"tags",
|
||||
}
|
||||
DATAVIEW_FIELDS = {"dataview1", "dataview2", "dataview3"}
|
||||
assert (
|
||||
set(doc.metadata)
|
||||
== STANDARD_METADATA_FIELDS | FRONTMATTER_FIELDS | DATAVIEW_FIELDS
|
||||
)
|
||||
|
||||
|
||||
def test_tags_in_page_content() -> None:
|
||||
"""Verify a doc with tags are included in the metadata"""
|
||||
doc = next(doc for doc in docs if doc.metadata["source"] == "no_frontmatter.md")
|
||||
assert set(doc.metadata) == STANDARD_METADATA_FIELDS | {"tags"}
|
||||
|
||||
|
||||
def test_boolean_metadata() -> None:
|
||||
"""Verify boolean metadata is loaded correctly"""
|
||||
doc = next(
|
||||
doc for doc in docs if doc.metadata["source"] == "tags_and_frontmatter.md"
|
||||
)
|
||||
assert doc.metadata["aBool"]
|
||||
|
||||
|
||||
def test_float_metadata() -> None:
|
||||
"""Verify float metadata is loaded correctly"""
|
||||
doc = next(
|
||||
doc for doc in docs if doc.metadata["source"] == "tags_and_frontmatter.md"
|
||||
)
|
||||
assert doc.metadata["aFloat"] == 13.12345
|
||||
|
||||
|
||||
def test_int_metadata() -> None:
|
||||
"""Verify int metadata is loaded correctly"""
|
||||
doc = next(
|
||||
doc for doc in docs if doc.metadata["source"] == "tags_and_frontmatter.md"
|
||||
)
|
||||
assert doc.metadata["anInt"] == 15
|
||||
|
||||
|
||||
def test_string_metadata() -> None:
|
||||
"""Verify string metadata is loaded correctly"""
|
||||
doc = next(
|
||||
doc for doc in docs if doc.metadata["source"] == "tags_and_frontmatter.md"
|
||||
)
|
||||
assert doc.metadata["aString"] == "string value"
|
||||
|
||||
|
||||
def test_array_metadata() -> None:
|
||||
"""Verify array metadata is loaded as a string"""
|
||||
doc = next(
|
||||
doc for doc in docs if doc.metadata["source"] == "tags_and_frontmatter.md"
|
||||
)
|
||||
assert doc.metadata["anArray"] == "['one', 'two', 'three']"
|
||||
|
||||
|
||||
def test_dict_metadata() -> None:
|
||||
"""Verify dict metadata is stored as a string"""
|
||||
doc = next(
|
||||
doc for doc in docs if doc.metadata["source"] == "tags_and_frontmatter.md"
|
||||
)
|
||||
assert doc.metadata["aDict"] == "{'dictId1': '58417', 'dictId2': 1500}"
|
Loading…
Reference in New Issue
Block a user