Compare commits
14 Commits
32ee43e77a
...
0.0.1
| Author | SHA1 | Date | |
|---|---|---|---|
| 6808538485 | |||
| fc29f49d41 | |||
| 3d6655084f | |||
| 5af411020c | |||
| 7f1b0d9c10 | |||
| b2173f3ef0 | |||
| de67c7b0b1 | |||
| 390eb22e1b | |||
| 358abc9b36 | |||
| 07293e4651 | |||
| fa14a79ec9 | |||
| 55245cf241 | |||
| 202329e190 | |||
| f59024ddd5 |
129
.drone.yml
Normal file
129
.drone.yml
Normal file
@@ -0,0 +1,129 @@
|
|||||||
|
---
|
||||||
|
kind: pipeline
|
||||||
|
type: docker
|
||||||
|
name: build-and-publish-pypi
|
||||||
|
|
||||||
|
platform:
|
||||||
|
os: linux
|
||||||
|
arch: arm64
|
||||||
|
|
||||||
|
workspace:
|
||||||
|
path: /drone/src
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: check-version
|
||||||
|
image: curlimages/curl:latest
|
||||||
|
environment:
|
||||||
|
PIP_REPO_URL:
|
||||||
|
from_secret: PIP_REPO_URL
|
||||||
|
PIP_USERNAME:
|
||||||
|
from_secret: PIP_USERNAME
|
||||||
|
PIP_PASSWORD:
|
||||||
|
from_secret: PIP_PASSWORD
|
||||||
|
commands:
|
||||||
|
- PACKAGE_NAME=$(grep -E '^name\s*=' pyproject.toml | head -1 | cut -d'"' -f2)
|
||||||
|
- VERSION=$(grep -E '^version\s*=' pyproject.toml | head -1 | cut -d'"' -f2)
|
||||||
|
- echo "🔍 Checking if $PACKAGE_NAME==$VERSION exists on $PIP_REPO_URL ..."
|
||||||
|
- |
|
||||||
|
if curl -fsSL -u "$PIP_USERNAME:$PIP_PASSWORD" "$PIP_REPO_URL/simple/$PACKAGE_NAME/" | grep -q "$VERSION"; then
|
||||||
|
echo "✅ $PACKAGE_NAME==$VERSION already exists — skipping build."
|
||||||
|
exit 78
|
||||||
|
else
|
||||||
|
echo "🆕 New version detected: $PACKAGE_NAME==$VERSION"
|
||||||
|
fi
|
||||||
|
|
||||||
|
- name: build-package
|
||||||
|
image: python:3.13-slim
|
||||||
|
commands:
|
||||||
|
- pip install --upgrade pip build
|
||||||
|
- echo "📦 Building Python package..."
|
||||||
|
- python -m build
|
||||||
|
- ls -l dist
|
||||||
|
|
||||||
|
- name: upload-to-private-pypi
|
||||||
|
image: python:3.13-slim
|
||||||
|
environment:
|
||||||
|
PIP_REPO_URL:
|
||||||
|
from_secret: PIP_REPO_URL
|
||||||
|
PIP_USERNAME:
|
||||||
|
from_secret: PIP_USERNAME
|
||||||
|
PIP_PASSWORD:
|
||||||
|
from_secret: PIP_PASSWORD
|
||||||
|
commands:
|
||||||
|
- pip install --upgrade twine
|
||||||
|
- echo "🚀 Uploading to private PyPI at $PIP_REPO_URL ..."
|
||||||
|
- |
|
||||||
|
twine upload \
|
||||||
|
--repository-url "$PIP_REPO_URL" \
|
||||||
|
-u "$PIP_USERNAME" \
|
||||||
|
-p "$PIP_PASSWORD" \
|
||||||
|
dist/*
|
||||||
|
|
||||||
|
trigger:
|
||||||
|
event:
|
||||||
|
- tag
|
||||||
|
|
||||||
|
---
|
||||||
|
kind: pipeline
|
||||||
|
type: docker
|
||||||
|
name: backfill-pypi-from-tags
|
||||||
|
|
||||||
|
platform:
|
||||||
|
os: linux
|
||||||
|
arch: arm64
|
||||||
|
|
||||||
|
workspace:
|
||||||
|
path: /drone/src
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: fetch-tags
|
||||||
|
image: alpine/git
|
||||||
|
commands:
|
||||||
|
- git fetch --tags --force
|
||||||
|
|
||||||
|
- name: build-and-upload-missing
|
||||||
|
image: python:3.13-slim
|
||||||
|
environment:
|
||||||
|
PIP_REPO_URL:
|
||||||
|
from_secret: PIP_REPO_URL
|
||||||
|
PIP_USERNAME:
|
||||||
|
from_secret: PIP_USERNAME
|
||||||
|
PIP_PASSWORD:
|
||||||
|
from_secret: PIP_PASSWORD
|
||||||
|
commands:
|
||||||
|
- apt-get update
|
||||||
|
- apt-get install -y git curl ca-certificates
|
||||||
|
- pip install --upgrade pip build twine
|
||||||
|
- |
|
||||||
|
set -e
|
||||||
|
|
||||||
|
PACKAGE_NAME=$(grep -E '^name\s*=' pyproject.toml | cut -d'"' -f2)
|
||||||
|
echo "📦 Package: $PACKAGE_NAME"
|
||||||
|
|
||||||
|
for TAG in $(git tag --sort=version:refname); do
|
||||||
|
VERSION="$TAG"
|
||||||
|
echo "🔁 Version: $VERSION"
|
||||||
|
|
||||||
|
if curl -fsSL -u "$PIP_USERNAME:$PIP_PASSWORD" \
|
||||||
|
"$PIP_REPO_URL/simple/$PACKAGE_NAME/" | grep -q "$VERSION"; then
|
||||||
|
echo "⏭️ Exists, skipping"
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
|
||||||
|
git checkout --force "$TAG"
|
||||||
|
|
||||||
|
echo "🏗️ Building $VERSION"
|
||||||
|
rm -rf dist
|
||||||
|
python -m build
|
||||||
|
|
||||||
|
echo "⬆️ Uploading $VERSION"
|
||||||
|
twine upload \
|
||||||
|
--repository-url "$PIP_REPO_URL" \
|
||||||
|
-u "$PIP_USERNAME" \
|
||||||
|
-p "$PIP_PASSWORD" \
|
||||||
|
dist/*
|
||||||
|
done
|
||||||
|
|
||||||
|
trigger:
|
||||||
|
event:
|
||||||
|
- custom
|
||||||
1
docs/core/content.md
Normal file
1
docs/core/content.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.core.content
|
||||||
1
docs/core/index.md
Normal file
1
docs/core/index.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.core
|
||||||
1
docs/core/parser.md
Normal file
1
docs/core/parser.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.core.parser
|
||||||
1
docs/core/scraper.md
Normal file
1
docs/core/scraper.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.core.scraper
|
||||||
1
docs/html/index.md
Normal file
1
docs/html/index.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.html
|
||||||
1
docs/html/parser.md
Normal file
1
docs/html/parser.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.html.parser
|
||||||
1
docs/html/scraper.md
Normal file
1
docs/html/scraper.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.html.scraper
|
||||||
1
docs/index.md
Normal file
1
docs/index.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread
|
||||||
1
docs/pdf/client.md
Normal file
1
docs/pdf/client.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.pdf.client
|
||||||
1
docs/pdf/index.md
Normal file
1
docs/pdf/index.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.pdf
|
||||||
1
docs/pdf/parser.md
Normal file
1
docs/pdf/parser.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.pdf.parser
|
||||||
1
docs/pdf/scraper.md
Normal file
1
docs/pdf/scraper.md
Normal file
@@ -0,0 +1 @@
|
|||||||
|
::: omniread.pdf.scraper
|
||||||
46
generate_docs.py
Normal file
46
generate_docs.py
Normal file
@@ -0,0 +1,46 @@
|
|||||||
|
"""
|
||||||
|
Programmatic MkDocs build script for OmniRead.
|
||||||
|
|
||||||
|
This script builds (or serves) the documentation by invoking MkDocs
|
||||||
|
*as a Python library*, not via shell commands.
|
||||||
|
|
||||||
|
Requirements:
|
||||||
|
- mkdocs
|
||||||
|
- mkdocs-material
|
||||||
|
- mkdocstrings[python]
|
||||||
|
|
||||||
|
Usage:
|
||||||
|
python generate_docs.py
|
||||||
|
python generate_docs.py --serve
|
||||||
|
"""
|
||||||
|
|
||||||
|
import sys
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
from mkdocs.commands import build as mkdocs_build
|
||||||
|
from mkdocs.commands import serve as mkdocs_serve
|
||||||
|
from mkdocs.config import load_config
|
||||||
|
|
||||||
|
|
||||||
|
PROJECT_ROOT = Path(__file__).resolve().parent
|
||||||
|
MKDOCS_YML = PROJECT_ROOT / "mkdocs.yml"
|
||||||
|
|
||||||
|
|
||||||
|
def main() -> None:
|
||||||
|
if not MKDOCS_YML.exists():
|
||||||
|
raise FileNotFoundError("mkdocs.yml not found at project root")
|
||||||
|
|
||||||
|
# Load MkDocs configuration programmatically
|
||||||
|
config = load_config(str(MKDOCS_YML))
|
||||||
|
|
||||||
|
# Decide mode
|
||||||
|
if "--serve" in sys.argv:
|
||||||
|
# Live-reload development server
|
||||||
|
mkdocs_serve.serve(config)
|
||||||
|
else:
|
||||||
|
# Static site build
|
||||||
|
mkdocs_build.build(config)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
53
mkdocs.yml
Normal file
53
mkdocs.yml
Normal file
@@ -0,0 +1,53 @@
|
|||||||
|
site_name: Aetoskia OmniRead
|
||||||
|
site_description: Format-agnostic document reading, parsing, and scraping framework
|
||||||
|
|
||||||
|
theme:
|
||||||
|
name: material
|
||||||
|
palette:
|
||||||
|
- scheme: slate
|
||||||
|
primary: deep purple
|
||||||
|
accent: cyan
|
||||||
|
font:
|
||||||
|
text: Inter
|
||||||
|
code: JetBrains Mono
|
||||||
|
features:
|
||||||
|
- navigation.tabs
|
||||||
|
- navigation.expand
|
||||||
|
- navigation.top
|
||||||
|
- navigation.instant
|
||||||
|
- content.code.copy
|
||||||
|
- content.code.annotate
|
||||||
|
|
||||||
|
plugins:
|
||||||
|
- search
|
||||||
|
- mkdocstrings:
|
||||||
|
handlers:
|
||||||
|
python:
|
||||||
|
paths: ["."]
|
||||||
|
options:
|
||||||
|
docstring_style: google
|
||||||
|
show_source: false
|
||||||
|
show_signature_annotations: true
|
||||||
|
separate_signature: true
|
||||||
|
merge_init_into_class: true
|
||||||
|
inherited_members: true
|
||||||
|
annotations_path: brief
|
||||||
|
show_root_heading: true
|
||||||
|
group_by_category: true
|
||||||
|
|
||||||
|
nav:
|
||||||
|
- Home: index.md
|
||||||
|
|
||||||
|
- Core (Contracts):
|
||||||
|
- Content Models: core/content.md
|
||||||
|
- Parsers: core/parser.md
|
||||||
|
- Scrapers: core/scraper.md
|
||||||
|
|
||||||
|
- HTML Implementation:
|
||||||
|
- HTML Parser: html/parser.md
|
||||||
|
- HTML Scraper: html/scraper.md
|
||||||
|
|
||||||
|
- PDF Implementation:
|
||||||
|
- PDF Client: pdf/client.md
|
||||||
|
- PDF Parser: pdf/parser.md
|
||||||
|
- PDF Scraper: pdf/scraper.md
|
||||||
@@ -0,0 +1,116 @@
|
|||||||
|
"""
|
||||||
|
OmniRead — format-agnostic content acquisition and parsing framework.
|
||||||
|
|
||||||
|
OmniRead provides a **cleanly layered architecture** for fetching, parsing,
|
||||||
|
and normalizing content from heterogeneous sources such as HTML documents
|
||||||
|
and PDF files.
|
||||||
|
|
||||||
|
The library is structured around three core concepts:
|
||||||
|
|
||||||
|
1. **Content**
|
||||||
|
A canonical, format-agnostic container representing raw content bytes
|
||||||
|
and minimal contextual metadata.
|
||||||
|
|
||||||
|
2. **Scrapers**
|
||||||
|
Components responsible for *acquiring* raw content from a source
|
||||||
|
(HTTP, filesystem, object storage, etc.). Scrapers never interpret
|
||||||
|
content.
|
||||||
|
|
||||||
|
3. **Parsers**
|
||||||
|
Components responsible for *interpreting* acquired content and
|
||||||
|
converting it into structured, typed representations.
|
||||||
|
|
||||||
|
OmniRead deliberately separates these responsibilities to ensure:
|
||||||
|
- Clear boundaries between IO and interpretation
|
||||||
|
- Replaceable implementations per format
|
||||||
|
- Predictable, testable behavior
|
||||||
|
|
||||||
|
----------------------------------------------------------------------
|
||||||
|
Installation
|
||||||
|
----------------------------------------------------------------------
|
||||||
|
|
||||||
|
Install OmniRead using pip:
|
||||||
|
|
||||||
|
pip install omniread
|
||||||
|
|
||||||
|
Or with Poetry:
|
||||||
|
|
||||||
|
poetry add omniread
|
||||||
|
|
||||||
|
----------------------------------------------------------------------
|
||||||
|
Basic Usage
|
||||||
|
----------------------------------------------------------------------
|
||||||
|
|
||||||
|
HTML example:
|
||||||
|
|
||||||
|
from omniread import HTMLScraper, HTMLParser
|
||||||
|
|
||||||
|
scraper = HTMLScraper()
|
||||||
|
content = scraper.fetch("https://example.com")
|
||||||
|
|
||||||
|
class TitleParser(HTMLParser[str]):
|
||||||
|
def parse(self) -> str:
|
||||||
|
return self._soup.title.string
|
||||||
|
|
||||||
|
parser = TitleParser(content)
|
||||||
|
title = parser.parse()
|
||||||
|
|
||||||
|
PDF example:
|
||||||
|
|
||||||
|
from omniread import FileSystemPDFClient, PDFScraper, PDFParser
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
client = FileSystemPDFClient()
|
||||||
|
scraper = PDFScraper(client=client)
|
||||||
|
content = scraper.fetch(Path("document.pdf"))
|
||||||
|
|
||||||
|
class TextPDFParser(PDFParser[str]):
|
||||||
|
def parse(self) -> str:
|
||||||
|
# implement PDF text extraction
|
||||||
|
...
|
||||||
|
|
||||||
|
parser = TextPDFParser(content)
|
||||||
|
result = parser.parse()
|
||||||
|
|
||||||
|
----------------------------------------------------------------------
|
||||||
|
Public API Surface
|
||||||
|
----------------------------------------------------------------------
|
||||||
|
|
||||||
|
This module re-exports the **recommended public entry points** of OmniRead.
|
||||||
|
|
||||||
|
Consumers are encouraged to import from this namespace rather than from
|
||||||
|
format-specific submodules directly, unless advanced customization is
|
||||||
|
required.
|
||||||
|
|
||||||
|
Core:
|
||||||
|
- Content
|
||||||
|
- ContentType
|
||||||
|
|
||||||
|
HTML:
|
||||||
|
- HTMLScraper
|
||||||
|
- HTMLParser
|
||||||
|
|
||||||
|
PDF:
|
||||||
|
- FileSystemPDFClient
|
||||||
|
- PDFScraper
|
||||||
|
- PDFParser
|
||||||
|
"""
|
||||||
|
|
||||||
|
from .core import Content, ContentType
|
||||||
|
from .html import HTMLScraper, HTMLParser
|
||||||
|
from .pdf import FileSystemPDFClient, PDFScraper, PDFParser
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
# core
|
||||||
|
"Content",
|
||||||
|
"ContentType",
|
||||||
|
|
||||||
|
# html
|
||||||
|
"HTMLScraper",
|
||||||
|
"HTMLParser",
|
||||||
|
|
||||||
|
# pdf
|
||||||
|
"FileSystemPDFClient",
|
||||||
|
"PDFScraper",
|
||||||
|
"PDFParser",
|
||||||
|
]
|
||||||
|
|||||||
@@ -0,0 +1,24 @@
|
|||||||
|
"""
|
||||||
|
Core domain contracts for OmniRead.
|
||||||
|
|
||||||
|
This package defines the **format-agnostic domain layer** of OmniRead.
|
||||||
|
It exposes canonical content models and abstract interfaces that are
|
||||||
|
implemented by format-specific modules (HTML, PDF, etc.).
|
||||||
|
|
||||||
|
Public exports from this package are considered **stable contracts** and
|
||||||
|
are safe for downstream consumers to depend on.
|
||||||
|
|
||||||
|
Submodules:
|
||||||
|
- content: Canonical content models and enums
|
||||||
|
- parser: Abstract parsing contracts
|
||||||
|
- scraper: Abstract scraping contracts
|
||||||
|
|
||||||
|
Format-specific behavior must not be introduced at this layer.
|
||||||
|
"""
|
||||||
|
|
||||||
|
from .content import Content, ContentType
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
"Content",
|
||||||
|
"ContentType",
|
||||||
|
]
|
||||||
|
|||||||
@@ -1,17 +1,62 @@
|
|||||||
|
"""
|
||||||
|
Canonical content models for OmniRead.
|
||||||
|
|
||||||
|
This module defines the **format-agnostic content representation** used across
|
||||||
|
all parsers and scrapers in OmniRead.
|
||||||
|
|
||||||
|
The models defined here represent *what* was extracted, not *how* it was
|
||||||
|
retrieved or parsed. Format-specific behavior and metadata must not alter
|
||||||
|
the semantic meaning of these models.
|
||||||
|
"""
|
||||||
|
|
||||||
from enum import Enum
|
from enum import Enum
|
||||||
from dataclasses import dataclass
|
from dataclasses import dataclass
|
||||||
from typing import Any, Mapping, Optional
|
from typing import Any, Mapping, Optional
|
||||||
|
|
||||||
|
|
||||||
class ContentType(str, Enum):
|
class ContentType(str, Enum):
|
||||||
|
"""
|
||||||
|
Supported MIME types for extracted content.
|
||||||
|
|
||||||
|
This enum represents the declared or inferred media type of the content
|
||||||
|
source. It is primarily used for routing content to the appropriate
|
||||||
|
parser or downstream consumer.
|
||||||
|
"""
|
||||||
|
|
||||||
HTML = "text/html"
|
HTML = "text/html"
|
||||||
|
"""HTML document content."""
|
||||||
|
|
||||||
PDF = "application/pdf"
|
PDF = "application/pdf"
|
||||||
|
"""PDF document content."""
|
||||||
|
|
||||||
JSON = "application/json"
|
JSON = "application/json"
|
||||||
|
"""JSON document content."""
|
||||||
|
|
||||||
XML = "application/xml"
|
XML = "application/xml"
|
||||||
|
"""XML document content."""
|
||||||
|
|
||||||
|
|
||||||
@dataclass(slots=True)
|
@dataclass(slots=True)
|
||||||
class Content:
|
class Content:
|
||||||
|
"""
|
||||||
|
Normalized representation of extracted content.
|
||||||
|
|
||||||
|
A `Content` instance represents a raw content payload along with minimal
|
||||||
|
contextual metadata describing its origin and type.
|
||||||
|
|
||||||
|
This class is the **primary exchange format** between:
|
||||||
|
- Scrapers
|
||||||
|
- Parsers
|
||||||
|
- Downstream consumers
|
||||||
|
|
||||||
|
Attributes:
|
||||||
|
raw: Raw content bytes as retrieved from the source.
|
||||||
|
source: Identifier of the content origin (URL, file path, or logical name).
|
||||||
|
content_type: Optional MIME type of the content, if known.
|
||||||
|
metadata: Optional, implementation-defined metadata associated with
|
||||||
|
the content (e.g., headers, encoding hints, extraction notes).
|
||||||
|
"""
|
||||||
|
|
||||||
raw: bytes
|
raw: bytes
|
||||||
source: str
|
source: str
|
||||||
content_type: Optional[ContentType] = None
|
content_type: Optional[ContentType] = None
|
||||||
|
|||||||
@@ -1,3 +1,20 @@
|
|||||||
|
"""
|
||||||
|
Abstract parsing contracts for OmniRead.
|
||||||
|
|
||||||
|
This module defines the **format-agnostic parser interface** used to transform
|
||||||
|
raw content into structured, typed representations.
|
||||||
|
|
||||||
|
Parsers are responsible for:
|
||||||
|
- Interpreting a single `Content` instance
|
||||||
|
- Validating compatibility with the content type
|
||||||
|
- Producing a structured output suitable for downstream consumers
|
||||||
|
|
||||||
|
Parsers are not responsible for:
|
||||||
|
- Fetching or acquiring content
|
||||||
|
- Performing retries or error recovery
|
||||||
|
- Managing multiple content sources
|
||||||
|
"""
|
||||||
|
|
||||||
from abc import ABC, abstractmethod
|
from abc import ABC, abstractmethod
|
||||||
from typing import Generic, TypeVar, Set
|
from typing import Generic, TypeVar, Set
|
||||||
|
|
||||||
@@ -12,11 +29,34 @@ class BaseParser(ABC, Generic[T]):
|
|||||||
|
|
||||||
A parser is a self-contained object that owns the Content
|
A parser is a self-contained object that owns the Content
|
||||||
it is responsible for interpreting.
|
it is responsible for interpreting.
|
||||||
|
|
||||||
|
Implementations must:
|
||||||
|
- Declare supported content types via `supported_types`
|
||||||
|
- Raise parsing-specific exceptions from `parse()`
|
||||||
|
- Remain deterministic for a given input
|
||||||
|
|
||||||
|
Consumers may rely on:
|
||||||
|
- Early validation of content compatibility
|
||||||
|
- Type-stable return values from `parse()`
|
||||||
"""
|
"""
|
||||||
|
|
||||||
supported_types: Set[ContentType] = set()
|
supported_types: Set[ContentType] = set()
|
||||||
|
"""Set of content types supported by this parser.
|
||||||
|
|
||||||
|
An empty set indicates that the parser is content-type agnostic.
|
||||||
|
"""
|
||||||
|
|
||||||
def __init__(self, content: Content):
|
def __init__(self, content: Content):
|
||||||
|
"""
|
||||||
|
Initialize the parser with content to be parsed.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
content: Content instance to be parsed.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
ValueError: If the content type is not supported by this parser.
|
||||||
|
"""
|
||||||
|
|
||||||
self.content = content
|
self.content = content
|
||||||
|
|
||||||
if not self.supports():
|
if not self.supports():
|
||||||
@@ -30,15 +70,25 @@ class BaseParser(ABC, Generic[T]):
|
|||||||
"""
|
"""
|
||||||
Parse the owned content into structured output.
|
Parse the owned content into structured output.
|
||||||
|
|
||||||
|
Implementations must fully consume the provided content and
|
||||||
|
return a deterministic, structured output.
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
Parsed, structured representation.
|
Parsed, structured representation.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
Exception: Parsing-specific errors as defined by the implementation.
|
||||||
"""
|
"""
|
||||||
raise NotImplementedError
|
raise NotImplementedError
|
||||||
|
|
||||||
def supports(self) -> bool:
|
def supports(self) -> bool:
|
||||||
"""
|
"""
|
||||||
Check whether this parser supports the content's type.
|
Check whether this parser supports the content's type.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
True if the content type is supported; False otherwise.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
if not self.supported_types:
|
if not self.supported_types:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|||||||
@@ -1,3 +1,22 @@
|
|||||||
|
"""
|
||||||
|
Abstract scraping contracts for OmniRead.
|
||||||
|
|
||||||
|
This module defines the **format-agnostic scraper interface** responsible for
|
||||||
|
acquiring raw content from external sources.
|
||||||
|
|
||||||
|
Scrapers are responsible for:
|
||||||
|
- Locating and retrieving raw content bytes
|
||||||
|
- Attaching minimal contextual metadata
|
||||||
|
- Returning normalized `Content` objects
|
||||||
|
|
||||||
|
Scrapers are explicitly NOT responsible for:
|
||||||
|
- Parsing or interpreting content
|
||||||
|
- Inferring structure or semantics
|
||||||
|
- Performing content-type specific processing
|
||||||
|
|
||||||
|
All interpretation must be delegated to parsers.
|
||||||
|
"""
|
||||||
|
|
||||||
from abc import ABC, abstractmethod
|
from abc import ABC, abstractmethod
|
||||||
from typing import Any, Mapping, Optional
|
from typing import Any, Mapping, Optional
|
||||||
|
|
||||||
@@ -10,6 +29,21 @@ class BaseScraper(ABC):
|
|||||||
|
|
||||||
A scraper is responsible ONLY for fetching raw content
|
A scraper is responsible ONLY for fetching raw content
|
||||||
(bytes) from a source. It must not interpret or parse it.
|
(bytes) from a source. It must not interpret or parse it.
|
||||||
|
|
||||||
|
A scraper is a **stateless acquisition component** that retrieves raw
|
||||||
|
content from a source and returns it as a `Content` object.
|
||||||
|
|
||||||
|
Scrapers define *how content is obtained*, not *what the content means*.
|
||||||
|
|
||||||
|
Implementations may vary in:
|
||||||
|
- Transport mechanism (HTTP, filesystem, cloud storage)
|
||||||
|
- Authentication strategy
|
||||||
|
- Retry and backoff behavior
|
||||||
|
|
||||||
|
Implementations must not:
|
||||||
|
- Parse content
|
||||||
|
- Modify content semantics
|
||||||
|
- Couple scraping logic to a specific parser
|
||||||
"""
|
"""
|
||||||
|
|
||||||
@abstractmethod
|
@abstractmethod
|
||||||
@@ -22,11 +56,20 @@ class BaseScraper(ABC):
|
|||||||
"""
|
"""
|
||||||
Fetch raw content from the given source.
|
Fetch raw content from the given source.
|
||||||
|
|
||||||
|
Implementations must retrieve the content referenced by `source`
|
||||||
|
and return it as raw bytes wrapped in a `Content` object.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
source: Location identifier (URL, file path, S3 URI, etc.)
|
source: Location identifier (URL, file path, S3 URI, etc.)
|
||||||
metadata: Optional hints for the scraper (headers, auth, etc.)
|
metadata: Optional hints for the scraper (headers, auth, etc.)
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
Content object containing raw bytes and metadata.
|
Content object containing raw bytes and metadata.
|
||||||
|
- Raw content bytes
|
||||||
|
- Source identifier
|
||||||
|
- Optional metadata
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
Exception: Retrieval-specific errors as defined by the implementation.
|
||||||
"""
|
"""
|
||||||
raise NotImplementedError
|
raise NotImplementedError
|
||||||
|
|||||||
@@ -0,0 +1,27 @@
|
|||||||
|
"""
|
||||||
|
HTML format implementation for OmniRead.
|
||||||
|
|
||||||
|
This package provides **HTML-specific implementations** of the core OmniRead
|
||||||
|
contracts defined in `omniread.core`.
|
||||||
|
|
||||||
|
It includes:
|
||||||
|
- HTML parsers that interpret HTML content
|
||||||
|
- HTML scrapers that retrieve HTML documents
|
||||||
|
|
||||||
|
This package:
|
||||||
|
- Implements, but does not redefine, core contracts
|
||||||
|
- May contain HTML-specific behavior and edge-case handling
|
||||||
|
- Produces canonical content models defined in `omniread.core.content`
|
||||||
|
|
||||||
|
Consumers should depend on `omniread.core` interfaces wherever possible and
|
||||||
|
use this package only when HTML-specific behavior is required.
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
from .scraper import HTMLScraper
|
||||||
|
from .parser import HTMLParser
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
"HTMLScraper",
|
||||||
|
"HTMLParser",
|
||||||
|
]
|
||||||
|
|||||||
@@ -1,6 +1,21 @@
|
|||||||
from typing import Any, Generic, TypeVar, Optional
|
"""
|
||||||
|
HTML parser base implementations for OmniRead.
|
||||||
|
|
||||||
|
This module provides reusable HTML parsing utilities built on top of
|
||||||
|
the abstract parser contracts defined in `omniread.core.parser`.
|
||||||
|
|
||||||
|
It supplies:
|
||||||
|
- Content-type enforcement for HTML inputs
|
||||||
|
- BeautifulSoup initialization and lifecycle management
|
||||||
|
- Common helper methods for extracting structured data from HTML elements
|
||||||
|
|
||||||
|
Concrete parsers must subclass `HTMLParser` and implement the `parse()` method
|
||||||
|
to return a structured representation appropriate for their use case.
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Any, Generic, TypeVar, Optional
|
||||||
from abc import abstractmethod
|
from abc import abstractmethod
|
||||||
|
|
||||||
from bs4 import BeautifulSoup, Tag
|
from bs4 import BeautifulSoup, Tag
|
||||||
|
|
||||||
from omniread.core.content import ContentType, Content
|
from omniread.core.content import ContentType, Content
|
||||||
@@ -13,13 +28,37 @@ class HTMLParser(BaseParser[T], Generic[T]):
|
|||||||
"""
|
"""
|
||||||
Base HTML parser.
|
Base HTML parser.
|
||||||
|
|
||||||
|
This class extends the core `BaseParser` with HTML-specific behavior,
|
||||||
|
including DOM parsing via BeautifulSoup and reusable extraction helpers.
|
||||||
|
|
||||||
Provides reusable helpers for HTML extraction.
|
Provides reusable helpers for HTML extraction.
|
||||||
Concrete parsers must explicitly define the return type.
|
Concrete parsers must explicitly define the return type.
|
||||||
|
|
||||||
|
Characteristics:
|
||||||
|
- Accepts only HTML content
|
||||||
|
- Owns a parsed BeautifulSoup DOM tree
|
||||||
|
- Provides pure helper utilities for common HTML structures
|
||||||
|
|
||||||
|
Concrete subclasses must:
|
||||||
|
- Define the output type `T`
|
||||||
|
- Implement the `parse()` method
|
||||||
"""
|
"""
|
||||||
|
|
||||||
supported_types = {ContentType.HTML}
|
supported_types = {ContentType.HTML}
|
||||||
|
"""Set of content types supported by this parser (HTML only)."""
|
||||||
|
|
||||||
def __init__(self, content: Content, features: str = "html.parser"):
|
def __init__(self, content: Content, features: str = "html.parser"):
|
||||||
|
"""
|
||||||
|
Initialize the HTML parser.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
content: HTML content to be parsed.
|
||||||
|
features: BeautifulSoup parser backend to use
|
||||||
|
(e.g., 'html.parser', 'lxml').
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
ValueError: If the content is empty or not valid HTML.
|
||||||
|
"""
|
||||||
super().__init__(content)
|
super().__init__(content)
|
||||||
self._features = features
|
self._features = features
|
||||||
self._soup = self._get_soup()
|
self._soup = self._get_soup()
|
||||||
@@ -32,6 +71,12 @@ class HTMLParser(BaseParser[T], Generic[T]):
|
|||||||
def parse(self) -> T:
|
def parse(self) -> T:
|
||||||
"""
|
"""
|
||||||
Fully parse the HTML content into structured output.
|
Fully parse the HTML content into structured output.
|
||||||
|
|
||||||
|
Implementations must fully interpret the HTML DOM and return
|
||||||
|
a deterministic, structured output.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Parsed representation of type `T`.
|
||||||
"""
|
"""
|
||||||
raise NotImplementedError
|
raise NotImplementedError
|
||||||
|
|
||||||
@@ -41,14 +86,42 @@ class HTMLParser(BaseParser[T], Generic[T]):
|
|||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def parse_div(div: Tag, *, separator: str = " ") -> str:
|
def parse_div(div: Tag, *, separator: str = " ") -> str:
|
||||||
|
"""
|
||||||
|
Extract normalized text from a `<div>` element.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
div: BeautifulSoup tag representing a `<div>`.
|
||||||
|
separator: String used to separate text nodes.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Flattened, whitespace-normalized text content.
|
||||||
|
"""
|
||||||
return div.get_text(separator=separator, strip=True)
|
return div.get_text(separator=separator, strip=True)
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def parse_link(a: Tag) -> Optional[str]:
|
def parse_link(a: Tag) -> Optional[str]:
|
||||||
|
"""
|
||||||
|
Extract the hyperlink reference from an `<a>` element.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
a: BeautifulSoup tag representing an anchor.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
The value of the `href` attribute, or None if absent.
|
||||||
|
"""
|
||||||
return a.get("href")
|
return a.get("href")
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def parse_table(table: Tag) -> list[list[str]]:
|
def parse_table(table: Tag) -> list[list[str]]:
|
||||||
|
"""
|
||||||
|
Parse an HTML table into a 2D list of strings.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
table: BeautifulSoup tag representing a `<table>`.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
A list of rows, where each row is a list of cell text values.
|
||||||
|
"""
|
||||||
rows: list[list[str]] = []
|
rows: list[list[str]] = []
|
||||||
for tr in table.find_all("tr"):
|
for tr in table.find_all("tr"):
|
||||||
cells = [
|
cells = [
|
||||||
@@ -64,11 +137,30 @@ class HTMLParser(BaseParser[T], Generic[T]):
|
|||||||
# ----------------------------
|
# ----------------------------
|
||||||
|
|
||||||
def _get_soup(self) -> BeautifulSoup:
|
def _get_soup(self) -> BeautifulSoup:
|
||||||
|
"""
|
||||||
|
Build a BeautifulSoup DOM tree from raw HTML content.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Parsed BeautifulSoup document tree.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
ValueError: If the content payload is empty.
|
||||||
|
"""
|
||||||
if not self.content.raw:
|
if not self.content.raw:
|
||||||
raise ValueError("Empty HTML content")
|
raise ValueError("Empty HTML content")
|
||||||
return BeautifulSoup(self.content.raw, features=self._features)
|
return BeautifulSoup(self.content.raw, features=self._features)
|
||||||
|
|
||||||
def parse_meta(self) -> dict[str, Any]:
|
def parse_meta(self) -> dict[str, Any]:
|
||||||
|
"""
|
||||||
|
Extract high-level metadata from the HTML document.
|
||||||
|
|
||||||
|
This includes:
|
||||||
|
- Document title
|
||||||
|
- `<meta>` tag name/property → content mappings
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Dictionary containing extracted metadata.
|
||||||
|
"""
|
||||||
soup = self._soup
|
soup = self._soup
|
||||||
|
|
||||||
title = soup.title.string.strip() if soup.title and soup.title.string else None
|
title = soup.title.string.strip() if soup.title and soup.title.string else None
|
||||||
|
|||||||
@@ -1,27 +1,97 @@
|
|||||||
|
"""
|
||||||
|
HTML scraping implementation for OmniRead.
|
||||||
|
|
||||||
|
This module provides an HTTP-based scraper for retrieving HTML documents.
|
||||||
|
It implements the core `BaseScraper` contract using `httpx` as the transport
|
||||||
|
layer.
|
||||||
|
|
||||||
|
This scraper is responsible for:
|
||||||
|
- Fetching raw HTML bytes over HTTP(S)
|
||||||
|
- Validating response content type
|
||||||
|
- Attaching HTTP metadata to the returned content
|
||||||
|
|
||||||
|
This scraper is not responsible for:
|
||||||
|
- Parsing or interpreting HTML
|
||||||
|
- Retrying failed requests
|
||||||
|
- Managing crawl policies or rate limiting
|
||||||
|
"""
|
||||||
|
|
||||||
import httpx
|
import httpx
|
||||||
from typing import Any, Mapping, Optional
|
from typing import Any, Mapping, Optional
|
||||||
|
|
||||||
from omniread.core.content import Content
|
from omniread.core.content import Content, ContentType
|
||||||
from omniread.core.scraper import BaseScraper
|
from omniread.core.scraper import BaseScraper
|
||||||
|
|
||||||
|
|
||||||
class HTMLScraper(BaseScraper):
|
class HTMLScraper(BaseScraper):
|
||||||
"""
|
"""
|
||||||
Base HTTP scraper using httpx.
|
Base HTML scraper using httpx.
|
||||||
|
|
||||||
|
This scraper retrieves HTML documents over HTTP(S) and returns them
|
||||||
|
as raw content wrapped in a `Content` object.
|
||||||
|
|
||||||
Fetches raw bytes and metadata only.
|
Fetches raw bytes and metadata only.
|
||||||
|
The scraper:
|
||||||
|
- Uses `httpx.Client` for HTTP requests
|
||||||
|
- Enforces an HTML content type
|
||||||
|
- Preserves HTTP response metadata
|
||||||
|
|
||||||
|
The scraper does not:
|
||||||
|
- Parse HTML
|
||||||
|
- Perform retries or backoff
|
||||||
|
- Handle non-HTML responses
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
*,
|
*,
|
||||||
|
client: httpx.Client | None = None,
|
||||||
timeout: float = 15.0,
|
timeout: float = 15.0,
|
||||||
headers: Optional[Mapping[str, str]] = None,
|
headers: Optional[Mapping[str, str]] = None,
|
||||||
follow_redirects: bool = True,
|
follow_redirects: bool = True,
|
||||||
):
|
):
|
||||||
self.timeout = timeout
|
"""
|
||||||
self.headers = dict(headers) if headers else {}
|
Initialize the HTML scraper.
|
||||||
self.follow_redirects = follow_redirects
|
|
||||||
|
Args:
|
||||||
|
client: Optional pre-configured `httpx.Client`. If omitted,
|
||||||
|
a client is created internally.
|
||||||
|
timeout: Request timeout in seconds.
|
||||||
|
headers: Optional default HTTP headers.
|
||||||
|
follow_redirects: Whether to follow HTTP redirects.
|
||||||
|
"""
|
||||||
|
|
||||||
|
self._client = client or httpx.Client(
|
||||||
|
timeout=timeout,
|
||||||
|
headers=headers,
|
||||||
|
follow_redirects=follow_redirects,
|
||||||
|
)
|
||||||
|
self.content_type = ContentType.HTML
|
||||||
|
|
||||||
|
def validate_content_type(
|
||||||
|
self,
|
||||||
|
response: httpx.Response,
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Validate that the HTTP response contains HTML content.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
response: HTTP response returned by `httpx`.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
ValueError: If the `Content-Type` header is missing or does not
|
||||||
|
indicate HTML content.
|
||||||
|
"""
|
||||||
|
|
||||||
|
raw_ct = response.headers.get("Content-Type")
|
||||||
|
if not raw_ct:
|
||||||
|
raise ValueError("Missing Content-Type header")
|
||||||
|
|
||||||
|
base_ct = raw_ct.split(";", 1)[0].strip().lower()
|
||||||
|
if base_ct != self.content_type.value:
|
||||||
|
raise ValueError(
|
||||||
|
f"Expected HTML content, got '{raw_ct}'"
|
||||||
|
)
|
||||||
|
|
||||||
def fetch(
|
def fetch(
|
||||||
self,
|
self,
|
||||||
@@ -29,20 +99,36 @@ class HTMLScraper(BaseScraper):
|
|||||||
*,
|
*,
|
||||||
metadata: Optional[Mapping[str, Any]] = None,
|
metadata: Optional[Mapping[str, Any]] = None,
|
||||||
) -> Content:
|
) -> Content:
|
||||||
with httpx.Client(
|
"""
|
||||||
timeout=self.timeout,
|
Fetch an HTML document from the given source.
|
||||||
headers=self.headers,
|
|
||||||
follow_redirects=self.follow_redirects,
|
Args:
|
||||||
) as client:
|
source: URL of the HTML document.
|
||||||
response = client.get(source)
|
metadata: Optional metadata to be merged into the returned content.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
A `Content` instance containing:
|
||||||
|
- Raw HTML bytes
|
||||||
|
- Source URL
|
||||||
|
- HTML content type
|
||||||
|
- HTTP response metadata
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
httpx.HTTPError: If the HTTP request fails.
|
||||||
|
ValueError: If the response is not valid HTML.
|
||||||
|
"""
|
||||||
|
|
||||||
|
response = self._client.get(source)
|
||||||
response.raise_for_status()
|
response.raise_for_status()
|
||||||
|
self.validate_content_type(response)
|
||||||
|
|
||||||
return Content(
|
return Content(
|
||||||
raw=response.content,
|
raw=response.content,
|
||||||
source=source,
|
source=source,
|
||||||
content_type=response.headers.get("Content-Type"),
|
content_type=self.content_type,
|
||||||
metadata={
|
metadata={
|
||||||
"status_code": response.status_code,
|
"status_code": response.status_code,
|
||||||
"headers": dict(response.headers),
|
"headers": dict(response.headers),
|
||||||
|
**(metadata or {}),
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
|
|||||||
25
omniread/pdf/__init__.py
Normal file
25
omniread/pdf/__init__.py
Normal file
@@ -0,0 +1,25 @@
|
|||||||
|
"""
|
||||||
|
PDF format implementation for OmniRead.
|
||||||
|
|
||||||
|
This package provides **PDF-specific implementations** of the core OmniRead
|
||||||
|
contracts defined in `omniread.core`.
|
||||||
|
|
||||||
|
Unlike HTML, PDF handling requires an explicit client layer for document
|
||||||
|
access. This package therefore includes:
|
||||||
|
- PDF clients for acquiring raw PDF data
|
||||||
|
- PDF scrapers that coordinate client access
|
||||||
|
- PDF parsers that extract structured content from PDF binaries
|
||||||
|
|
||||||
|
Public exports from this package represent the supported PDF pipeline
|
||||||
|
and are safe for consumers to import directly when working with PDFs.
|
||||||
|
"""
|
||||||
|
|
||||||
|
from .client import FileSystemPDFClient
|
||||||
|
from .scraper import PDFScraper
|
||||||
|
from .parser import PDFParser
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
"FileSystemPDFClient",
|
||||||
|
"PDFScraper",
|
||||||
|
"PDFParser",
|
||||||
|
]
|
||||||
79
omniread/pdf/client.py
Normal file
79
omniread/pdf/client.py
Normal file
@@ -0,0 +1,79 @@
|
|||||||
|
"""
|
||||||
|
PDF client abstractions for OmniRead.
|
||||||
|
|
||||||
|
This module defines the **client layer** responsible for retrieving raw PDF
|
||||||
|
bytes from a concrete backing store.
|
||||||
|
|
||||||
|
Clients provide low-level access to PDF binaries and are intentionally
|
||||||
|
decoupled from scraping and parsing logic. They do not perform validation,
|
||||||
|
interpretation, or content extraction.
|
||||||
|
|
||||||
|
Typical backing stores include:
|
||||||
|
- Local filesystems
|
||||||
|
- Object storage (S3, GCS, etc.)
|
||||||
|
- Network file systems
|
||||||
|
"""
|
||||||
|
|
||||||
|
from abc import ABC, abstractmethod
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
|
||||||
|
class BasePDFClient(ABC):
|
||||||
|
"""
|
||||||
|
Abstract client responsible for retrieving PDF bytes
|
||||||
|
from a specific backing store (filesystem, S3, FTP, etc.).
|
||||||
|
|
||||||
|
Implementations must:
|
||||||
|
- Accept a source identifier appropriate to the backing store
|
||||||
|
- Return the full PDF binary payload
|
||||||
|
- Raise retrieval-specific errors on failure
|
||||||
|
"""
|
||||||
|
|
||||||
|
@abstractmethod
|
||||||
|
def fetch(self, source: str) -> bytes:
|
||||||
|
"""
|
||||||
|
Fetch raw PDF bytes from the given source.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
source: Identifier of the PDF location, such as a file path,
|
||||||
|
object storage key, or remote reference.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Raw PDF bytes.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
Exception: Retrieval-specific errors defined by the implementation.
|
||||||
|
"""
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
|
||||||
|
class FileSystemPDFClient(BasePDFClient):
|
||||||
|
"""
|
||||||
|
PDF client that reads from the local filesystem.
|
||||||
|
|
||||||
|
This client reads PDF files directly from the disk and returns their raw
|
||||||
|
binary contents.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def fetch(self, path: Path) -> bytes:
|
||||||
|
"""
|
||||||
|
Read a PDF file from the local filesystem.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
path: Filesystem path to the PDF file.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Raw PDF bytes.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
FileNotFoundError: If the path does not exist.
|
||||||
|
ValueError: If the path exists but is not a file.
|
||||||
|
"""
|
||||||
|
|
||||||
|
if not path.exists():
|
||||||
|
raise FileNotFoundError(f"PDF not found: {path}")
|
||||||
|
|
||||||
|
if not path.is_file():
|
||||||
|
raise ValueError(f"Path is not a file: {path}")
|
||||||
|
|
||||||
|
return path.read_bytes()
|
||||||
49
omniread/pdf/parser.py
Normal file
49
omniread/pdf/parser.py
Normal file
@@ -0,0 +1,49 @@
|
|||||||
|
"""
|
||||||
|
PDF parser base implementations for OmniRead.
|
||||||
|
|
||||||
|
This module defines the **PDF-specific parser contract**, extending the
|
||||||
|
format-agnostic `BaseParser` with constraints appropriate for PDF content.
|
||||||
|
|
||||||
|
PDF parsers are responsible for interpreting binary PDF data and producing
|
||||||
|
structured representations suitable for downstream consumption.
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Generic, TypeVar
|
||||||
|
from abc import abstractmethod
|
||||||
|
|
||||||
|
from omniread.core.content import ContentType
|
||||||
|
from omniread.core.parser import BaseParser
|
||||||
|
|
||||||
|
T = TypeVar("T")
|
||||||
|
|
||||||
|
|
||||||
|
class PDFParser(BaseParser[T], Generic[T]):
|
||||||
|
"""
|
||||||
|
Base PDF parser.
|
||||||
|
|
||||||
|
This class enforces PDF content-type compatibility and provides the
|
||||||
|
extension point for implementing concrete PDF parsing strategies.
|
||||||
|
|
||||||
|
Concrete implementations must define:
|
||||||
|
- Define the output type `T`
|
||||||
|
- Implement the `parse()` method
|
||||||
|
"""
|
||||||
|
|
||||||
|
supported_types = {ContentType.PDF}
|
||||||
|
"""Set of content types supported by this parser (PDF only)."""
|
||||||
|
|
||||||
|
@abstractmethod
|
||||||
|
def parse(self) -> T:
|
||||||
|
"""
|
||||||
|
Parse PDF content into a structured output.
|
||||||
|
|
||||||
|
Implementations must fully interpret the PDF binary payload and
|
||||||
|
return a deterministic, structured output.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Parsed representation of type `T`.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
Exception: Parsing-specific errors as defined by the implementation.
|
||||||
|
"""
|
||||||
|
raise NotImplementedError
|
||||||
71
omniread/pdf/scraper.py
Normal file
71
omniread/pdf/scraper.py
Normal file
@@ -0,0 +1,71 @@
|
|||||||
|
"""
|
||||||
|
PDF scraping implementation for OmniRead.
|
||||||
|
|
||||||
|
This module provides a PDF-specific scraper that coordinates PDF byte
|
||||||
|
retrieval via a client and normalizes the result into a `Content` object.
|
||||||
|
|
||||||
|
The scraper implements the core `BaseScraper` contract while delegating
|
||||||
|
all storage and access concerns to a `BasePDFClient` implementation.
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Any, Mapping, Optional
|
||||||
|
|
||||||
|
from omniread.core.content import Content, ContentType
|
||||||
|
from omniread.core.scraper import BaseScraper
|
||||||
|
from omniread.pdf.client import BasePDFClient
|
||||||
|
|
||||||
|
|
||||||
|
class PDFScraper(BaseScraper):
|
||||||
|
"""
|
||||||
|
Scraper for PDF sources.
|
||||||
|
|
||||||
|
Delegates byte retrieval to a PDF client and normalizes
|
||||||
|
output into Content.
|
||||||
|
|
||||||
|
The scraper:
|
||||||
|
- Does not perform parsing or interpretation
|
||||||
|
- Does not assume a specific storage backend
|
||||||
|
- Preserves caller-provided metadata
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, *, client: BasePDFClient):
|
||||||
|
"""
|
||||||
|
Initialize the PDF scraper.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
client: PDF client responsible for retrieving raw PDF bytes.
|
||||||
|
"""
|
||||||
|
self._client = client
|
||||||
|
|
||||||
|
def fetch(
|
||||||
|
self,
|
||||||
|
source: str,
|
||||||
|
*,
|
||||||
|
metadata: Optional[Mapping[str, Any]] = None,
|
||||||
|
) -> Content:
|
||||||
|
"""
|
||||||
|
Fetch a PDF document from the given source.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
source: Identifier of the PDF source as understood by the
|
||||||
|
configured PDF client.
|
||||||
|
metadata: Optional metadata to attach to the returned content.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
A `Content` instance containing:
|
||||||
|
- Raw PDF bytes
|
||||||
|
- Source identifier
|
||||||
|
- PDF content type
|
||||||
|
- Optional metadata
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
Exception: Retrieval-specific errors raised by the PDF client.
|
||||||
|
"""
|
||||||
|
raw = self._client.fetch(source)
|
||||||
|
|
||||||
|
return Content(
|
||||||
|
raw=raw,
|
||||||
|
source=source,
|
||||||
|
content_type=ContentType.PDF,
|
||||||
|
metadata=dict(metadata) if metadata else None,
|
||||||
|
)
|
||||||
@@ -1,7 +1,19 @@
|
|||||||
httpx==0.27.0
|
httpx==0.27.0
|
||||||
beautifulsoup4==4.12.0
|
beautifulsoup4==4.12.0
|
||||||
|
pydantic==2.12.3
|
||||||
|
jinja2==3.1.6
|
||||||
# lxml==5.2.0
|
# lxml==5.2.0
|
||||||
|
|
||||||
|
# Test Packages
|
||||||
pytest==7.4.0
|
pytest==7.4.0
|
||||||
pytest-asyncio==0.21.0
|
pytest-asyncio==0.21.0
|
||||||
pytest-cov==4.1.0
|
pytest-cov==4.1.0
|
||||||
|
|
||||||
|
# Doc Packages
|
||||||
|
mkdocs==1.6.1
|
||||||
|
mkdocs-material==9.6.23
|
||||||
|
neoteroi-mkdocs==1.1.3
|
||||||
|
pymdown-extensions==10.16.1
|
||||||
|
mkdocs-swagger-ui-tag==0.7.2
|
||||||
|
mkdocstrings==1.0.0
|
||||||
|
mkdocstrings-python==2.0.1
|
||||||
|
|||||||
0
tests/__init__.py
Normal file
0
tests/__init__.py
Normal file
90
tests/conftest.py
Normal file
90
tests/conftest.py
Normal file
@@ -0,0 +1,90 @@
|
|||||||
|
import json
|
||||||
|
import pytest
|
||||||
|
import httpx
|
||||||
|
from pathlib import Path
|
||||||
|
from jinja2 import Environment, BaseLoader
|
||||||
|
|
||||||
|
from omniread import (
|
||||||
|
# core
|
||||||
|
ContentType,
|
||||||
|
|
||||||
|
# html
|
||||||
|
HTMLScraper,
|
||||||
|
|
||||||
|
# pdf
|
||||||
|
FileSystemPDFClient,
|
||||||
|
PDFScraper,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
MOCK_HTML_DIR = Path(__file__).parent / "mocks" / "html"
|
||||||
|
MOCK_PDF_DIR = Path(__file__).parent / "mocks" / "pdf"
|
||||||
|
|
||||||
|
|
||||||
|
def render_html(template_path, data_path) -> bytes:
|
||||||
|
template_text = Path(template_path).read_text(encoding="utf-8")
|
||||||
|
data = json.loads(Path(data_path).read_text(encoding="utf-8"))
|
||||||
|
|
||||||
|
env = Environment(
|
||||||
|
loader=BaseLoader(),
|
||||||
|
autoescape=False,
|
||||||
|
)
|
||||||
|
template = env.from_string(template_text)
|
||||||
|
|
||||||
|
rendered = template.render(**data)
|
||||||
|
return rendered.encode("utf-8")
|
||||||
|
|
||||||
|
|
||||||
|
def mock_transport(request: httpx.Request) -> httpx.Response:
|
||||||
|
"""
|
||||||
|
httpx MockTransport handler.
|
||||||
|
"""
|
||||||
|
path = request.url.path
|
||||||
|
if path not in ['/simple', '/table']:
|
||||||
|
return httpx.Response(
|
||||||
|
status_code=404,
|
||||||
|
content=b"Not Found",
|
||||||
|
request=request,
|
||||||
|
)
|
||||||
|
endpoint = path.split("/")[-1]
|
||||||
|
content = render_html(
|
||||||
|
MOCK_HTML_DIR / f"{endpoint}.html.jinja",
|
||||||
|
MOCK_HTML_DIR / f"{endpoint}.json",
|
||||||
|
)
|
||||||
|
|
||||||
|
return httpx.Response(
|
||||||
|
status_code=200,
|
||||||
|
headers={"Content-Type": ContentType.HTML.value},
|
||||||
|
content=content,
|
||||||
|
request=request,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.fixture
|
||||||
|
def http_scraper() -> HTMLScraper:
|
||||||
|
transport = httpx.MockTransport(mock_transport)
|
||||||
|
|
||||||
|
client = httpx.Client(transport=transport)
|
||||||
|
|
||||||
|
return HTMLScraper(client=client)
|
||||||
|
|
||||||
|
|
||||||
|
class MockPDFClient(FileSystemPDFClient):
|
||||||
|
"""
|
||||||
|
Test-only PDF client that routes logical identifiers
|
||||||
|
to fixture files.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def fetch(self, source: str) -> bytes:
|
||||||
|
if source in ["simple"]:
|
||||||
|
source = MOCK_PDF_DIR / f"{source}.pdf"
|
||||||
|
else:
|
||||||
|
raise FileNotFoundError(f"No mock PDF route for '{source}'")
|
||||||
|
|
||||||
|
return super().fetch(source)
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.fixture
|
||||||
|
def pdf_scraper() -> PDFScraper:
|
||||||
|
client = MockPDFClient()
|
||||||
|
return PDFScraper(client=client)
|
||||||
11
tests/mocks/html/simple.html.jinja
Normal file
11
tests/mocks/html/simple.html.jinja
Normal file
@@ -0,0 +1,11 @@
|
|||||||
|
<!DOCTYPE html>
|
||||||
|
<html>
|
||||||
|
<head>
|
||||||
|
<title>{{ title }}</title>
|
||||||
|
<meta name="description" content="{{ description }}">
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<div id="content">{{ content }}</div>
|
||||||
|
<a href="{{ link_url }}">{{ link_text }}</a>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
7
tests/mocks/html/simple.json
Normal file
7
tests/mocks/html/simple.json
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
{
|
||||||
|
"title": "Test Page",
|
||||||
|
"description": "Simple test page",
|
||||||
|
"content": "Hello World",
|
||||||
|
"link_url": "https://example.com",
|
||||||
|
"link_text": "Link"
|
||||||
|
}
|
||||||
31
tests/mocks/html/table.html.jinja
Normal file
31
tests/mocks/html/table.html.jinja
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
<!DOCTYPE html>
|
||||||
|
<html>
|
||||||
|
<head>
|
||||||
|
<title>{{ title }}</title>
|
||||||
|
<meta name="description" content="{{ description }}">
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<h1>{{ heading }}</h1>
|
||||||
|
|
||||||
|
<table id="{{ table_id }}">
|
||||||
|
<thead>
|
||||||
|
<tr>
|
||||||
|
{% for col in columns %}
|
||||||
|
<th>{{ col }}</th>
|
||||||
|
{% endfor %}
|
||||||
|
</tr>
|
||||||
|
</thead>
|
||||||
|
<tbody>
|
||||||
|
{% for row in rows %}
|
||||||
|
<tr>
|
||||||
|
{% for cell in row %}
|
||||||
|
<td>{{ cell }}</td>
|
||||||
|
{% endfor %}
|
||||||
|
</tr>
|
||||||
|
{% endfor %}
|
||||||
|
</tbody>
|
||||||
|
</table>
|
||||||
|
|
||||||
|
<a href="{{ link_url }}">{{ link_text }}</a>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
14
tests/mocks/html/table.json
Normal file
14
tests/mocks/html/table.json
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
{
|
||||||
|
"title": "Table Test Page",
|
||||||
|
"description": "HTML page with a table for parsing tests",
|
||||||
|
"heading": "Sample Table",
|
||||||
|
"table_id": "data-table",
|
||||||
|
"columns": ["Name", "Age", "City"],
|
||||||
|
"rows": [
|
||||||
|
["Alice", "30", "London"],
|
||||||
|
["Bob", "25", "New York"],
|
||||||
|
["Charlie", "35", "Berlin"]
|
||||||
|
],
|
||||||
|
"link_url": "https://example.org/details",
|
||||||
|
"link_text": "Details"
|
||||||
|
}
|
||||||
32
tests/mocks/pdf/simple.pdf
Normal file
32
tests/mocks/pdf/simple.pdf
Normal file
@@ -0,0 +1,32 @@
|
|||||||
|
%PDF-1.4
|
||||||
|
1 0 obj
|
||||||
|
<< /Type /Catalog /Pages 2 0 R >>
|
||||||
|
endobj
|
||||||
|
2 0 obj
|
||||||
|
<< /Type /Pages /Kids [3 0 R] /Count 1 >>
|
||||||
|
endobj
|
||||||
|
3 0 obj
|
||||||
|
<< /Type /Page /Parent 2 0 R /MediaBox [0 0 612 792] /Contents 4 0 R >>
|
||||||
|
endobj
|
||||||
|
4 0 obj
|
||||||
|
<< /Length 44 >>
|
||||||
|
stream
|
||||||
|
BT
|
||||||
|
/F1 12 Tf
|
||||||
|
72 720 Td
|
||||||
|
(Simple PDF Test) Tj
|
||||||
|
ET
|
||||||
|
endstream
|
||||||
|
endobj
|
||||||
|
xref
|
||||||
|
0 5
|
||||||
|
0000000000 65535 f
|
||||||
|
0000000010 00000 n
|
||||||
|
0000000061 00000 n
|
||||||
|
0000000116 00000 n
|
||||||
|
0000000203 00000 n
|
||||||
|
trailer
|
||||||
|
<< /Size 5 /Root 1 0 R >>
|
||||||
|
startxref
|
||||||
|
300
|
||||||
|
%%EOF
|
||||||
53
tests/test_html_simple.py
Normal file
53
tests/test_html_simple.py
Normal file
@@ -0,0 +1,53 @@
|
|||||||
|
from typing import Optional
|
||||||
|
|
||||||
|
from pydantic import BaseModel
|
||||||
|
from bs4 import Tag
|
||||||
|
|
||||||
|
from omniread import (
|
||||||
|
# core
|
||||||
|
Content,
|
||||||
|
|
||||||
|
# html
|
||||||
|
HTMLParser,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class ParsedSimpleHTML(BaseModel):
|
||||||
|
title: Optional[str]
|
||||||
|
description: Optional[str]
|
||||||
|
content: Optional[str]
|
||||||
|
link: Optional[str]
|
||||||
|
|
||||||
|
|
||||||
|
class SimpleHTMLParser(HTMLParser[ParsedSimpleHTML]):
|
||||||
|
"""
|
||||||
|
Parser focused on high-level page semantics.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def parse(self) -> ParsedSimpleHTML:
|
||||||
|
soup = self._soup
|
||||||
|
meta = self.parse_meta()
|
||||||
|
|
||||||
|
content_div = soup.find("div", id="content")
|
||||||
|
link_tag: Tag | None = soup.find("a")
|
||||||
|
|
||||||
|
return ParsedSimpleHTML(
|
||||||
|
title=meta["title"],
|
||||||
|
description=meta["meta"].get("description"),
|
||||||
|
content=self.parse_div(content_div) if content_div else None,
|
||||||
|
link=self.parse_link(link_tag) if link_tag else None,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def test_end_to_end_html_simple(http_scraper):
|
||||||
|
content: Content = http_scraper.fetch("https://test.local/simple")
|
||||||
|
|
||||||
|
parser = SimpleHTMLParser(content)
|
||||||
|
result = parser.parse()
|
||||||
|
|
||||||
|
assert isinstance(result, ParsedSimpleHTML)
|
||||||
|
|
||||||
|
assert result.title == "Test Page"
|
||||||
|
assert result.description == "Simple test page"
|
||||||
|
assert result.content == "Hello World"
|
||||||
|
assert result.link == "https://example.com"
|
||||||
49
tests/test_html_table.py
Normal file
49
tests/test_html_table.py
Normal file
@@ -0,0 +1,49 @@
|
|||||||
|
from typing import Optional
|
||||||
|
|
||||||
|
from pydantic import BaseModel
|
||||||
|
|
||||||
|
from omniread import (
|
||||||
|
# core
|
||||||
|
Content,
|
||||||
|
|
||||||
|
# html
|
||||||
|
HTMLParser,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class ParsedTableHTML(BaseModel):
|
||||||
|
title: Optional[str]
|
||||||
|
table: list[list[str]]
|
||||||
|
|
||||||
|
|
||||||
|
class TableHTMLParser(HTMLParser[ParsedTableHTML]):
|
||||||
|
"""
|
||||||
|
Parser focused on extracting tabular data.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def parse(self) -> ParsedTableHTML:
|
||||||
|
soup = self._soup
|
||||||
|
|
||||||
|
table_tag = soup.find("table")
|
||||||
|
|
||||||
|
return ParsedTableHTML(
|
||||||
|
title=soup.title.string.strip() if soup.title else None,
|
||||||
|
table=self.parse_table(table_tag) if table_tag else [],
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def test_end_to_end_html_table(http_scraper):
|
||||||
|
content: Content = http_scraper.fetch("https://test.local/table")
|
||||||
|
|
||||||
|
parser = TableHTMLParser(content)
|
||||||
|
result = parser.parse()
|
||||||
|
|
||||||
|
assert isinstance(result, ParsedTableHTML)
|
||||||
|
|
||||||
|
assert result.title == "Table Test Page"
|
||||||
|
assert result.table == [
|
||||||
|
["Name", "Age", "City"],
|
||||||
|
["Alice", "30", "London"],
|
||||||
|
["Bob", "25", "New York"],
|
||||||
|
["Charlie", "35", "Berlin"],
|
||||||
|
]
|
||||||
41
tests/test_pdf_simple.py
Normal file
41
tests/test_pdf_simple.py
Normal file
@@ -0,0 +1,41 @@
|
|||||||
|
from typing import Literal
|
||||||
|
from pydantic import BaseModel
|
||||||
|
|
||||||
|
from omniread import (
|
||||||
|
# core
|
||||||
|
Content,
|
||||||
|
|
||||||
|
# pdf
|
||||||
|
PDFParser,
|
||||||
|
)
|
||||||
|
|
||||||
|
class ParsedPDF(BaseModel):
|
||||||
|
size_bytes: int
|
||||||
|
magic: Literal[b"%PDF"]
|
||||||
|
|
||||||
|
|
||||||
|
class SimplePDFParser(PDFParser[ParsedPDF]):
|
||||||
|
def parse(self) -> ParsedPDF:
|
||||||
|
raw = self.content.raw
|
||||||
|
|
||||||
|
if not raw.startswith(b"%PDF"):
|
||||||
|
raise ValueError("Not a valid PDF")
|
||||||
|
|
||||||
|
return ParsedPDF(
|
||||||
|
size_bytes=len(raw),
|
||||||
|
magic=b"%PDF",
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def test_end_to_end_pdf_simple(pdf_scraper):
|
||||||
|
# --- Scrape (identifier-based, routed in conftest)
|
||||||
|
content: Content = pdf_scraper.fetch("simple")
|
||||||
|
|
||||||
|
assert content.raw.startswith(b"%PDF")
|
||||||
|
|
||||||
|
# --- Parse
|
||||||
|
parser = SimplePDFParser(content)
|
||||||
|
result = parser.parse()
|
||||||
|
|
||||||
|
assert result.magic == b"%PDF"
|
||||||
|
assert result.size_bytes > 100
|
||||||
Reference in New Issue
Block a user