From fe28f26fc237fbeb7612bb66a11a99602c6cf5e6 Mon Sep 17 00:00:00 2001 From: Vik Paruchuri Date: Fri, 7 Nov 2025 13:18:38 -0500 Subject: [PATCH 1/2] Adjust bbox format --- chandra/output.py | 3 +-- chandra/settings.py | 2 +- 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/chandra/output.py b/chandra/output.py index 9afe54a..8dc3ad9 100644 --- a/chandra/output.py +++ b/chandra/output.py @@ -1,5 +1,4 @@ import hashlib -import json import re from dataclasses import dataclass, asdict from functools import lru_cache @@ -241,7 +240,7 @@ def parse_layout(html: str, image: Image.Image): bbox = div.get("data-bbox") try: - bbox = json.loads(bbox) + bbox = bbox.split(" ") except Exception: bbox = [0, 0, 1, 1] diff --git a/chandra/settings.py b/chandra/settings.py index d0abd1b..5fa078a 100644 --- a/chandra/settings.py +++ b/chandra/settings.py @@ -15,7 +15,7 @@ class Settings(BaseSettings): TORCH_DEVICE: str | None = None MAX_OUTPUT_TOKENS: int = 12384 TORCH_ATTN: str | None = None - BBOX_SCALE: int = 1024 + BBOX_SCALE: int = 1000 # vLLM server settings VLLM_API_KEY: str = "EMPTY" From 3958707a80aa86d086b27464bb89dcb808516700 Mon Sep 17 00:00:00 2001 From: Vik Paruchuri Date: Mon, 10 Nov 2025 11:12:00 -0500 Subject: [PATCH 2/2] Support multiple formats --- chandra/model/hf.py | 3 +-- chandra/model/vllm.py | 2 -- chandra/output.py | 19 ++++++++----------- chandra/settings.py | 2 +- 4 files changed, 10 insertions(+), 16 deletions(-) diff --git a/chandra/model/hf.py b/chandra/model/hf.py index b88eb9e..50aa883 100644 --- a/chandra/model/hf.py +++ b/chandra/model/hf.py @@ -5,7 +5,6 @@ from transformers import Qwen3VLForConditionalGeneration, Qwen3VLProcessor from chandra.model.schema import BatchInputItem, GenerationResult from chandra.model.util import scale_to_fit -from chandra.output import fix_raw from chandra.prompts import PROMPT_MAPPING from chandra.settings import settings @@ -43,7 +42,7 @@ def generate_hf( clean_up_tokenization_spaces=False, ) results = [ - GenerationResult(raw=fix_raw(out), token_count=len(ids), error=False) + GenerationResult(raw=out, token_count=len(ids), error=False) for out, ids in zip(output_text, generated_ids_trimmed) ] return results diff --git a/chandra/model/vllm.py b/chandra/model/vllm.py index 5528571..1aabf69 100644 --- a/chandra/model/vllm.py +++ b/chandra/model/vllm.py @@ -9,7 +9,6 @@ from openai import OpenAI from chandra.model.schema import BatchInputItem, GenerationResult from chandra.model.util import scale_to_fit, detect_repeat_token -from chandra.output import fix_raw from chandra.prompts import PROMPT_MAPPING from chandra.settings import settings @@ -76,7 +75,6 @@ def generate_vllm( top_p=top_p, ) raw = completion.choices[0].message.content - raw = fix_raw(raw) result = GenerationResult( raw=raw, token_count=completion.usage.completion_tokens, diff --git a/chandra/output.py b/chandra/output.py index 8dc3ad9..47e9f98 100644 --- a/chandra/output.py +++ b/chandra/output.py @@ -1,4 +1,5 @@ import hashlib +import json import re from dataclasses import dataclass, asdict from functools import lru_cache @@ -21,15 +22,6 @@ def get_image_name(html: str, div_idx: int): return f"{html_hash}_{div_idx}_img.webp" -def fix_raw(html: str): - def replace_group(match): - numbers = re.findall(r"\d+", match.group(0)) - return "[" + ",".join(numbers) + "]" - - result = re.sub(r"(?:\|BBOX\d+\|){4}", replace_group, html) - return result - - def extract_images(html: str, chunks: dict, image: Image.Image): images = {} div_idx = 0 @@ -240,9 +232,14 @@ def parse_layout(html: str, image: Image.Image): bbox = div.get("data-bbox") try: - bbox = bbox.split(" ") + bbox = json.loads(bbox) + assert len(bbox) == 4, "Invalid bbox length" except Exception: - bbox = [0, 0, 1, 1] + try: + bbox = bbox.split(" ") + assert len(bbox) == 4, "Invalid bbox length" + except Exception: + bbox = [0, 0, 1, 1] bbox = list(map(int, bbox)) # Normalize bbox diff --git a/chandra/settings.py b/chandra/settings.py index 5fa078a..d0abd1b 100644 --- a/chandra/settings.py +++ b/chandra/settings.py @@ -15,7 +15,7 @@ class Settings(BaseSettings): TORCH_DEVICE: str | None = None MAX_OUTPUT_TOKENS: int = 12384 TORCH_ATTN: str | None = None - BBOX_SCALE: int = 1000 + BBOX_SCALE: int = 1024 # vLLM server settings VLLM_API_KEY: str = "EMPTY"