mirror of
https://github.com/datalab-to/chandra.git
synced 2025-11-30 00:53:09 +00:00
Compare commits
9 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
089f987dfa | ||
|
|
c049e7524f | ||
|
|
b96eb84094 | ||
|
|
0f5f3d485c | ||
|
|
1bab4bf73a | ||
|
|
34f825351c | ||
|
|
068db0311e | ||
|
|
aafbb70ce8 | ||
|
|
910bcf100f |
@@ -1,4 +1,4 @@
|
||||
from typing import List
|
||||
from typing import List, Union, Optional
|
||||
import filetype
|
||||
from PIL import Image
|
||||
import pypdfium2 as pdfium
|
||||
@@ -13,26 +13,94 @@ def flatten(page, flag=pdfium_c.FLAT_NORMALDISPLAY):
|
||||
print(f"Failed to flatten annotations / form fields on page {page}.")
|
||||
|
||||
|
||||
def load_image(filepath: str):
|
||||
def load_image(
|
||||
filepath: str, min_image_dim: int = settings.MIN_IMAGE_DIM
|
||||
) -> Image.Image:
|
||||
image = Image.open(filepath).convert("RGB")
|
||||
if image.width < settings.MIN_IMAGE_DIM or image.height < settings.MIN_IMAGE_DIM:
|
||||
scale = settings.MIN_IMAGE_DIM / min(image.width, image.height)
|
||||
if image.width < min_image_dim or image.height < min_image_dim:
|
||||
scale = min_image_dim / min(image.width, image.height)
|
||||
new_size = (int(image.width * scale), int(image.height * scale))
|
||||
image = image.resize(new_size, Image.Resampling.LANCZOS)
|
||||
return image
|
||||
|
||||
|
||||
def load_pdf_images(filepath: str, page_range: List[int]):
|
||||
def load_pdf_images(
|
||||
filepath: str,
|
||||
page_range: List[int],
|
||||
image_dpi: Optional[Union[int, List[int]]] = None,
|
||||
min_pdf_image_dim: Optional[Union[int, List[int]]] = None,
|
||||
) -> List[Image.Image]:
|
||||
"""
|
||||
Load PDF pages as images with configurable DPI.
|
||||
|
||||
Args:
|
||||
filepath: Path to PDF file
|
||||
page_range: List of page indices to render
|
||||
image_dpi: Target DPI for rendering. Can be:
|
||||
- None: use settings.IMAGE_DPI for all pages (default)
|
||||
- int: use same DPI for all pages
|
||||
- List[int]: per-page DPI (must match length of page_range)
|
||||
min_pdf_image_dim: Minimum image dimension. Can be:
|
||||
- None: use settings.MIN_PDF_IMAGE_DIM for all pages (default)
|
||||
- int: use same value for all pages
|
||||
- List[int]: per-page value (must match length of page_range)
|
||||
|
||||
Returns:
|
||||
List of PIL Images, one per page in page_range
|
||||
"""
|
||||
doc = pdfium.PdfDocument(filepath)
|
||||
doc.init_forms()
|
||||
|
||||
# Determine default values
|
||||
default_image_dpi = image_dpi if isinstance(image_dpi, int) else settings.IMAGE_DPI
|
||||
default_min_pdf_image_dim = min_pdf_image_dim if isinstance(min_pdf_image_dim, int) else settings.MIN_PDF_IMAGE_DIM
|
||||
|
||||
# Handle per-page DPI lists
|
||||
is_per_page_dpi = isinstance(image_dpi, list)
|
||||
if not is_per_page_dpi and image_dpi is not None:
|
||||
# Convert single DPI value to list for all pages
|
||||
image_dpi = [image_dpi] * len(page_range)
|
||||
is_per_page_dpi = True
|
||||
|
||||
is_per_page_min_dim = isinstance(min_pdf_image_dim, list)
|
||||
if not is_per_page_min_dim and min_pdf_image_dim is not None:
|
||||
# Convert single min_dim value to list for all pages
|
||||
min_pdf_image_dim = [min_pdf_image_dim] * len(page_range)
|
||||
is_per_page_min_dim = True
|
||||
|
||||
if is_per_page_dpi and len(image_dpi) != len(page_range):
|
||||
raise ValueError(f"image_dpi list length ({len(image_dpi)}) must match page_range length ({len(page_range)})")
|
||||
if is_per_page_min_dim and len(min_pdf_image_dim) != len(page_range):
|
||||
raise ValueError(f"min_pdf_image_dim list length ({len(min_pdf_image_dim)}) must match page_range length ({len(page_range)})")
|
||||
|
||||
images = []
|
||||
page_idx_in_range = 0
|
||||
|
||||
for page in range(len(doc)):
|
||||
if not page_range or page in page_range:
|
||||
# Get DPI for this specific page
|
||||
if is_per_page_dpi:
|
||||
current_dpi = image_dpi[page_idx_in_range]
|
||||
elif image_dpi is None:
|
||||
current_dpi = settings.IMAGE_DPI
|
||||
else:
|
||||
current_dpi = default_image_dpi
|
||||
|
||||
# Get min_dim for this specific page
|
||||
if is_per_page_min_dim:
|
||||
current_min_dim = min_pdf_image_dim[page_idx_in_range]
|
||||
elif min_pdf_image_dim is None:
|
||||
current_min_dim = settings.MIN_PDF_IMAGE_DIM
|
||||
else:
|
||||
current_min_dim = default_min_pdf_image_dim
|
||||
|
||||
page_idx_in_range += 1
|
||||
|
||||
page_obj = doc[page]
|
||||
min_page_dim = min(page_obj.get_width(), page_obj.get_height())
|
||||
scale_dpi = (settings.MIN_PDF_IMAGE_DIM / min_page_dim) * 72
|
||||
scale_dpi = max(scale_dpi, settings.IMAGE_DPI)
|
||||
|
||||
scale_dpi = (current_min_dim / min_page_dim) * 72
|
||||
scale_dpi = max(scale_dpi, current_dpi)
|
||||
page_obj = doc[page]
|
||||
flatten(page_obj)
|
||||
page_obj = doc[page]
|
||||
|
||||
@@ -4,6 +4,7 @@ from chandra.model.hf import load_model, generate_hf
|
||||
from chandra.model.schema import BatchInputItem, BatchOutputItem
|
||||
from chandra.model.vllm import generate_vllm
|
||||
from chandra.output import parse_markdown, parse_html, parse_chunks, extract_images
|
||||
from chandra.settings import settings
|
||||
|
||||
|
||||
class InferenceManager:
|
||||
@@ -26,19 +27,29 @@ class InferenceManager:
|
||||
output_kwargs["include_headers_footers"] = kwargs.pop(
|
||||
"include_headers_footers"
|
||||
)
|
||||
bbox_scale = kwargs.pop("bbox_scale", settings.BBOX_SCALE)
|
||||
vllm_api_base = kwargs.pop("vllm_api_base", settings.VLLM_API_BASE)
|
||||
|
||||
if self.method == "vllm":
|
||||
results = generate_vllm(
|
||||
batch, max_output_tokens=max_output_tokens, **kwargs
|
||||
batch,
|
||||
max_output_tokens=max_output_tokens,
|
||||
bbox_scale=bbox_scale,
|
||||
vllm_api_base=vllm_api_base,
|
||||
**kwargs,
|
||||
)
|
||||
else:
|
||||
results = generate_hf(
|
||||
batch, self.model, max_output_tokens=max_output_tokens, **kwargs
|
||||
batch,
|
||||
self.model,
|
||||
max_output_tokens=max_output_tokens,
|
||||
bbox_scale=bbox_scale,
|
||||
**kwargs,
|
||||
)
|
||||
|
||||
output = []
|
||||
for result, input_item in zip(results, batch):
|
||||
chunks = parse_chunks(result.raw, input_item.image)
|
||||
chunks = parse_chunks(result.raw, input_item.image, bbox_scale=bbox_scale)
|
||||
output.append(
|
||||
BatchOutputItem(
|
||||
markdown=parse_markdown(result.raw, **output_kwargs),
|
||||
|
||||
@@ -1,8 +1,5 @@
|
||||
from typing import List
|
||||
|
||||
from qwen_vl_utils import process_vision_info
|
||||
from transformers import Qwen3VLForConditionalGeneration, Qwen3VLProcessor
|
||||
|
||||
from chandra.model.schema import BatchInputItem, GenerationResult
|
||||
from chandra.model.util import scale_to_fit
|
||||
from chandra.prompts import PROMPT_MAPPING
|
||||
@@ -10,12 +7,20 @@ from chandra.settings import settings
|
||||
|
||||
|
||||
def generate_hf(
|
||||
batch: List[BatchInputItem], model, max_output_tokens=None, **kwargs
|
||||
batch: List[BatchInputItem],
|
||||
model,
|
||||
max_output_tokens=None,
|
||||
bbox_scale: int = settings.BBOX_SCALE,
|
||||
**kwargs,
|
||||
) -> List[GenerationResult]:
|
||||
from qwen_vl_utils import process_vision_info
|
||||
|
||||
if max_output_tokens is None:
|
||||
max_output_tokens = settings.MAX_OUTPUT_TOKENS
|
||||
|
||||
messages = [process_batch_element(item, model.processor) for item in batch]
|
||||
messages = [
|
||||
process_batch_element(item, model.processor, bbox_scale) for item in batch
|
||||
]
|
||||
text = model.processor.apply_chat_template(
|
||||
messages, tokenize=False, add_generation_prompt=True
|
||||
)
|
||||
@@ -48,12 +53,12 @@ def generate_hf(
|
||||
return results
|
||||
|
||||
|
||||
def process_batch_element(item: BatchInputItem, processor):
|
||||
def process_batch_element(item: BatchInputItem, processor, bbox_scale: int):
|
||||
prompt = item.prompt
|
||||
prompt_type = item.prompt_type
|
||||
|
||||
if not prompt:
|
||||
prompt = PROMPT_MAPPING[prompt_type]
|
||||
prompt = PROMPT_MAPPING[prompt_type].replace("{bbox_scale}", str(bbox_scale))
|
||||
|
||||
content = []
|
||||
image = scale_to_fit(item.image) # Guarantee max size
|
||||
@@ -65,12 +70,15 @@ def process_batch_element(item: BatchInputItem, processor):
|
||||
|
||||
|
||||
def load_model():
|
||||
import torch
|
||||
from transformers import Qwen3VLForConditionalGeneration, Qwen3VLProcessor
|
||||
|
||||
device_map = "auto"
|
||||
if settings.TORCH_DEVICE:
|
||||
device_map = {"": settings.TORCH_DEVICE}
|
||||
|
||||
kwargs = {
|
||||
"dtype": settings.TORCH_DTYPE,
|
||||
"dtype": torch.bfloat16,
|
||||
"device_map": device_map,
|
||||
}
|
||||
if settings.TORCH_ATTN:
|
||||
|
||||
@@ -1,5 +1,6 @@
|
||||
import base64
|
||||
import io
|
||||
import time
|
||||
from concurrent.futures import ThreadPoolExecutor
|
||||
from itertools import repeat
|
||||
from typing import List
|
||||
@@ -27,10 +28,12 @@ def generate_vllm(
|
||||
max_workers: int | None = None,
|
||||
custom_headers: dict | None = None,
|
||||
max_failure_retries: int | None = None,
|
||||
bbox_scale: int = settings.BBOX_SCALE,
|
||||
vllm_api_base: str = settings.VLLM_API_BASE,
|
||||
) -> List[GenerationResult]:
|
||||
client = OpenAI(
|
||||
api_key=settings.VLLM_API_KEY,
|
||||
base_url=settings.VLLM_API_BASE,
|
||||
base_url=vllm_api_base,
|
||||
default_headers=custom_headers,
|
||||
)
|
||||
model_name = settings.VLLM_MODEL_NAME
|
||||
@@ -53,7 +56,9 @@ def generate_vllm(
|
||||
) -> GenerationResult:
|
||||
prompt = item.prompt
|
||||
if not prompt:
|
||||
prompt = PROMPT_MAPPING[item.prompt_type]
|
||||
prompt = PROMPT_MAPPING[item.prompt_type].replace(
|
||||
"{bbox_scale}", str(bbox_scale)
|
||||
)
|
||||
|
||||
content = []
|
||||
image = scale_to_fit(item.image)
|
||||
@@ -112,6 +117,7 @@ def generate_vllm(
|
||||
print(
|
||||
f"Detected vllm error, retrying generation (attempt {retries + 1})..."
|
||||
)
|
||||
time.sleep(2 * (retries + 1)) # Sleeping can help under load
|
||||
return True
|
||||
|
||||
if (
|
||||
@@ -122,6 +128,7 @@ def generate_vllm(
|
||||
print(
|
||||
f"Detected vllm error, retrying generation (attempt {retries + 1})..."
|
||||
)
|
||||
time.sleep(2 * (retries + 1)) # Sleeping can help under load
|
||||
return True
|
||||
|
||||
return False
|
||||
|
||||
@@ -6,7 +6,7 @@ from functools import lru_cache
|
||||
|
||||
import six
|
||||
from PIL import Image
|
||||
from bs4 import BeautifulSoup, NavigableString
|
||||
from bs4 import BeautifulSoup
|
||||
from markdownify import MarkdownConverter, re_whitespace
|
||||
|
||||
from chandra.settings import settings
|
||||
@@ -89,39 +89,6 @@ def parse_html(
|
||||
return out_html
|
||||
|
||||
|
||||
def escape_dollars(text):
|
||||
return text.replace("$", r"\$")
|
||||
|
||||
|
||||
def get_formatted_table_text(element):
|
||||
text = []
|
||||
for content in element.contents:
|
||||
if content is None:
|
||||
continue
|
||||
|
||||
if isinstance(content, NavigableString):
|
||||
stripped = content.strip()
|
||||
if stripped:
|
||||
text.append(escape_dollars(stripped))
|
||||
elif content.name == "br":
|
||||
text.append("<br>")
|
||||
elif content.name == "math":
|
||||
text.append("$" + content.text + "$")
|
||||
else:
|
||||
content_str = escape_dollars(str(content))
|
||||
text.append(content_str)
|
||||
|
||||
full_text = ""
|
||||
for i, t in enumerate(text):
|
||||
if t == "<br>":
|
||||
full_text += t
|
||||
elif i > 0 and text[i - 1] != "<br>":
|
||||
full_text += " " + t
|
||||
else:
|
||||
full_text += t
|
||||
return full_text
|
||||
|
||||
|
||||
class Markdownify(MarkdownConverter):
|
||||
def __init__(
|
||||
self,
|
||||
@@ -221,12 +188,12 @@ class LayoutBlock:
|
||||
content: str
|
||||
|
||||
|
||||
def parse_layout(html: str, image: Image.Image):
|
||||
def parse_layout(html: str, image: Image.Image, bbox_scale=settings.BBOX_SCALE):
|
||||
soup = BeautifulSoup(html, "html.parser")
|
||||
top_level_divs = soup.find_all("div", recursive=False)
|
||||
width, height = image.size
|
||||
width_scaler = width / settings.BBOX_SCALE
|
||||
height_scaler = height / settings.BBOX_SCALE
|
||||
width_scaler = width / bbox_scale
|
||||
height_scaler = height / bbox_scale
|
||||
layout_blocks = []
|
||||
for div in top_level_divs:
|
||||
bbox = div.get("data-bbox")
|
||||
@@ -255,7 +222,7 @@ def parse_layout(html: str, image: Image.Image):
|
||||
return layout_blocks
|
||||
|
||||
|
||||
def parse_chunks(html: str, image: Image.Image):
|
||||
layout = parse_layout(html, image)
|
||||
def parse_chunks(html: str, image: Image.Image, bbox_scale=settings.BBOX_SCALE):
|
||||
layout = parse_layout(html, image, bbox_scale=bbox_scale)
|
||||
chunks = [asdict(block) for block in layout]
|
||||
return chunks
|
||||
|
||||
@@ -1,5 +1,3 @@
|
||||
from chandra.settings import settings
|
||||
|
||||
ALLOWED_TAGS = [
|
||||
"math",
|
||||
"br",
|
||||
@@ -67,7 +65,7 @@ Guidelines:
|
||||
""".strip()
|
||||
|
||||
OCR_LAYOUT_PROMPT = f"""
|
||||
OCR this image to HTML, arranged as layout blocks. Each layout block should be a div with the data-bbox attribute representing the bounding box of the block in [x0, y0, x1, y1] format. Bboxes are normalized 0-{settings.BBOX_SCALE}. The data-label attribute is the label for the block.
|
||||
OCR this image to HTML, arranged as layout blocks. Each layout block should be a div with the data-bbox attribute representing the bounding box of the block in [x0, y0, x1, y1] format. Bboxes are normalized 0-{{bbox_scale}}. The data-label attribute is the label for the block.
|
||||
|
||||
Use the following labels:
|
||||
- Caption
|
||||
|
||||
@@ -1,7 +1,5 @@
|
||||
from dotenv import find_dotenv
|
||||
from pydantic import computed_field
|
||||
from pydantic_settings import BaseSettings
|
||||
import torch
|
||||
import os
|
||||
|
||||
|
||||
@@ -24,11 +22,6 @@ class Settings(BaseSettings):
|
||||
VLLM_GPUS: str = "0"
|
||||
MAX_VLLM_RETRIES: int = 6
|
||||
|
||||
@computed_field
|
||||
@property
|
||||
def TORCH_DTYPE(self) -> torch.dtype:
|
||||
return torch.bfloat16
|
||||
|
||||
class Config:
|
||||
env_file = find_dotenv("local.env")
|
||||
extra = "ignore"
|
||||
|
||||
Reference in New Issue
Block a user