DeepSeek-OCR-vllm: update INPUT_PATH & Dockerfile envs

This commit is contained in:
kyy
2025-10-23 17:51:10 +09:00
parent 039842b465
commit b7f4a2537e
3 changed files with 27 additions and 22 deletions

9
.gitignore vendored Normal file
View File

@@ -0,0 +1,9 @@
# Cache directories
.cache/
__pycache__/
# Output directory
output/
# PDF files
*.pdf

View File

@@ -21,9 +21,7 @@ MODEL_PATH = 'deepseek-ai/DeepSeek-OCR' # change to your model path
# .jpg, .png, .jpeg: run_dpsk_ocr_image.py;
# Omnidocbench images path: run_dpsk_ocr_eval_batch.py
FILE_NAME='2025-27484-M21472.pdf'
INPUT_PATH = f'/workspace/2025-27484-M21472.pdf'
INPUT_PATH = f'/workspace/2018-0802140959-217049.pdf'
OUTPUT_PATH = '/workspace/output/'
PROMPT = '<image>\n<|grounding|>Convert the document to markdown.'

View File

@@ -1,41 +1,39 @@
# PyTorch 2.6.0 + CUDA 12.6 + cuDNN9
FROM pytorch/pytorch:2.6.0-cuda12.6-cudnn9-devel
# 기본 환경 변수
# 기본 환경 변수 설정
ENV DEBIAN_FRONTEND=noninteractive \
HF_HOME=/workspace/.cache/huggingface \
CUDA_HOME=/usr/local/cuda \
LD_LIBRARY_PATH=/usr/local/cuda/lib64:${LD_LIBRARY_PATH} \
PIP_DISABLE_PIP_VERSION_CHECK=1 \
PYTHONUNBUFFERED=1
PYTHONUNBUFFERED=1 \
TRITON_PTXAS_PATH=/usr/local/cuda/bin/ptxas \
TORCH_CUDA_ARCH_LIST="8.0"
WORKDIR /workspace
# 필수 빌드 도구
# 필수 빌드 도구 설치
RUN apt-get update && apt-get install -y --no-install-recommends \
git build-essential ninja-build \
&& rm -rf /var/lib/apt/lists/*
# pip
RUN python -m pip install -U pip setuptools wheel packaging ninja
# pip 업그레이드
RUN python -m pip install -U pip setuptools wheel
# (베이스 이미지에 torch 2.6.0이 이미 포함)
# torchvision/torchaudio 버전을 보장하려면 아래 라인을 사용하세요.
# cu126 인덱스가 없을 수 있으므로 실패해도 빌드가 계속되게 처리합니다.
RUN pip install --no-cache-dir \
torchvision==0.21.0 torchaudio==2.6.0 \
--extra-index-url https://download.pytorch.org/whl/cu126 || \
echo "Skip reinstalling torchvision/torchaudio (fallback to base image)"
# 기존 라이브러리 제거 및 특정 버전 재설치
RUN pip uninstall -y vllm torch torchvision torchaudio triton flash-attn || true
RUN pip install torch==2.6.0 torchvision==0.21.0 torchaudio==2.6.0
# 프로젝트 의존성 설치
COPY requirements.txt /tmp/requirements.txt
RUN pip install --no-cache-dir -r /tmp/requirements.txt
RUN pip install -r /tmp/requirements.txt
# vLLM 0.8.5 고정 (CUDA 12.x용 wheel 자동 선택)
RUN pip install --no-cache-dir vllm==0.8.5
# vLLM 특정 버전 설치
RUN pip install vllm==0.8.5
# FlashAttention (CUDA 12.6에서 빌드)
RUN pip install --no-cache-dir flash-attn==2.7.3 --no-build-isolation
# FlashAttention 소스에서 빌드하여 설치
RUN pip cache purge && \
pip install --no-cache-dir --no-build-isolation --no-binary=flash-attn flash-attn==2.7.3
# 작업 디렉터리 유지
WORKDIR /workspace
WORKDIR /workspace