diff --git a/docs/extras/modules/data_connection/document_loaders/integrations/recursive_url_loader.ipynb b/docs/extras/modules/data_connection/document_loaders/integrations/recursive_url_loader.ipynb index 2d402184ac9..b35b814af4b 100644 --- a/docs/extras/modules/data_connection/document_loaders/integrations/recursive_url_loader.ipynb +++ b/docs/extras/modules/data_connection/document_loaders/integrations/recursive_url_loader.ipynb @@ -1,6 +1,7 @@ { "cells": [ { + "attachments": {}, "cell_type": "markdown", "id": "5a7cc773", "metadata": {}, @@ -17,7 +18,7 @@ "\n", "But, the challenge is traversing the tree of child pages and actually assembling that list!\n", " \n", - "We do this using the `RecusiveUrlLoader`.\n", + "We do this using the `RecursiveUrlLoader`.\n", "\n", "This also gives us the flexibility to exclude some children (e.g., the `api` directory with > 800 child pages)." ] @@ -29,10 +30,11 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.document_loaders.recursive_url_loader import RecusiveUrlLoader" + "from langchain.document_loaders.recursive_url_loader import RecursiveUrlLoader" ] }, { + "attachments": {}, "cell_type": "markdown", "id": "6384c057", "metadata": {}, @@ -48,7 +50,7 @@ "outputs": [], "source": [ "url = 'https://js.langchain.com/docs/modules/memory/examples/'\n", - "loader=RecusiveUrlLoader(url=url)\n", + "loader=RecursiveUrlLoader(url=url)\n", "docs=loader.load()" ] }, @@ -119,6 +121,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "id": "40fc13ef", "metadata": {}, @@ -137,7 +140,7 @@ "source": [ "url = 'https://js.langchain.com/docs/'\n", "exclude_dirs=['https://js.langchain.com/docs/api/']\n", - "loader=RecusiveUrlLoader(url=url,exclude_dirs=exclude_dirs)\n", + "loader=RecursiveUrlLoader(url=url,exclude_dirs=exclude_dirs)\n", "docs=loader.load()" ] }, diff --git a/langchain/document_loaders/__init__.py b/langchain/document_loaders/__init__.py index 57e450c8844..2bf8f76e764 100644 --- a/langchain/document_loaders/__init__.py +++ b/langchain/document_loaders/__init__.py @@ -95,7 +95,7 @@ from langchain.document_loaders.psychic import PsychicLoader from langchain.document_loaders.pyspark_dataframe import PySparkDataFrameLoader from langchain.document_loaders.python import PythonLoader from langchain.document_loaders.readthedocs import ReadTheDocsLoader -from langchain.document_loaders.recursive_url_loader import RecusiveUrlLoader +from langchain.document_loaders.recursive_url_loader import RecursiveUrlLoader from langchain.document_loaders.reddit import RedditPostsLoader from langchain.document_loaders.roam import RoamLoader from langchain.document_loaders.rst import UnstructuredRSTLoader @@ -230,7 +230,7 @@ __all__ = [ "PySparkDataFrameLoader", "PythonLoader", "ReadTheDocsLoader", - "RecusiveUrlLoader", + "RecursiveUrlLoader", "RedditPostsLoader", "RoamLoader", "S3DirectoryLoader", diff --git a/langchain/document_loaders/recursive_url_loader.py b/langchain/document_loaders/recursive_url_loader.py index 7107f3a7345..b1a0250d74f 100644 --- a/langchain/document_loaders/recursive_url_loader.py +++ b/langchain/document_loaders/recursive_url_loader.py @@ -7,7 +7,7 @@ from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader -class RecusiveUrlLoader(BaseLoader): +class RecursiveUrlLoader(BaseLoader): """Loader that loads all child links from a given url.""" def __init__(self, url: str, exclude_dirs: Optional[str] = None) -> None: @@ -24,7 +24,7 @@ class RecusiveUrlLoader(BaseLoader): from bs4 import BeautifulSoup except ImportError: raise ImportError( - "The BeautifulSoup package is required for the RecusiveUrlLoader." + "The BeautifulSoup package is required for the RecursiveUrlLoader." ) # Construct the base and parent URLs