import logging import os import requests from urllib.parse import urlparse, urljoin from bs4 import BeautifulSoup from application.parser.remote.base import BaseRemote from application.parser.schema.base import Document from application.core.url_validation import validate_url, SSRFError from langchain_community.document_loaders import WebBaseLoader class CrawlerLoader(BaseRemote): def __init__(self, limit=10): self.loader = WebBaseLoader # Initialize the document loader self.limit = limit # Set the limit for the number of pages to scrape def load_data(self, inputs): url = inputs if isinstance(url, list) and url: url = url[0] # Validate URL to prevent SSRF attacks try: url = validate_url(url) except SSRFError as e: logging.error(f"URL validation failed: {e}") return [] visited_urls = set() base_url = urlparse(url).scheme + "://" + urlparse(url).hostname urls_to_visit = [url] loaded_content = [] while urls_to_visit: current_url = urls_to_visit.pop(0) visited_urls.add(current_url) try: # Validate each URL before making requests try: validate_url(current_url) except SSRFError as e: logging.warning(f"Skipping URL due to validation failure: {current_url} - {e}") continue response = requests.get(current_url, timeout=30) response.raise_for_status() loader = self.loader([current_url]) docs = loader.load() # Convert the loaded documents to your Document schema for doc in docs: metadata = dict(doc.metadata or {}) source_url = metadata.get("source") or current_url metadata["file_path"] = self._url_to_virtual_path(source_url) loaded_content.append( Document( doc.page_content, extra_info=metadata ) ) except Exception as e: logging.error(f"Error processing URL {current_url}: {e}", exc_info=True) continue # Parse the HTML content to extract all links soup = BeautifulSoup(response.text, 'html.parser') all_links = [ urljoin(current_url, a['href']) for a in soup.find_all('a', href=True) if base_url in urljoin(current_url, a['href']) ] # Add new links to the list of URLs to visit if they haven't been visited yet urls_to_visit.extend([link for link in all_links if link not in visited_urls]) urls_to_visit = list(set(urls_to_visit)) # Stop crawling if the limit of pages to scrape is reached if self.limit is not None and len(visited_urls) >= self.limit: break return loaded_content def _url_to_virtual_path(self, url): """ Convert a URL to a virtual file path ending with .md. Examples: https://docs.docsgpt.cloud/ -> index.md https://docs.docsgpt.cloud/guides/setup -> guides/setup.md https://docs.docsgpt.cloud/guides/setup/ -> guides/setup.md https://example.com/page.html -> page.md """ parsed = urlparse(url) path = parsed.path.strip("/") if not path: return "index.md" # Remove common file extensions and add .md base, ext = os.path.splitext(path) if ext.lower() in [".html", ".htm", ".php", ".asp", ".aspx", ".jsp"]: path = base if not path.endswith(".md"): path = f"{path}.md" return path