# Copyright (c) 2025 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import argparse
import logging
import subprocess
import sys
import time
import warnings
from threading import Thread
import requests
from ._models import (
ChartParsing,
DocImgOrientationClassification,
DocVLM,
FormulaRecognition,
LayoutDetection,
SealTextDetection,
TableCellsDetection,
TableClassification,
TableStructureRecognition,
TextDetection,
TextImageUnwarping,
TextLineOrientationClassification,
TextRecognition,
)
from ._pipelines import (
DocPreprocessor,
DocUnderstanding,
FormulaRecognitionPipeline,
PaddleOCR,
PaddleOCRVL,
PPChatOCRv4Doc,
PPDocTranslation,
PPStructureV3,
SealRecognition,
TableRecognitionPipelineV2,
)
from ._version import version
from ._utils.deprecation import CLIDeprecationWarning
from ._utils.logging import logger
def _register_pipelines(subparsers):
for cls in [
DocPreprocessor,
DocUnderstanding,
FormulaRecognitionPipeline,
PaddleOCR,
PaddleOCRVL,
PPChatOCRv4Doc,
PPDocTranslation,
PPStructureV3,
SealRecognition,
TableRecognitionPipelineV2,
]:
subcommand_executor = cls.get_cli_subcommand_executor()
subparser = subcommand_executor.add_subparser(subparsers)
subparser.set_defaults(executor=subcommand_executor.execute_with_args)
def _register_models(subparsers):
for cls in [
ChartParsing,
DocImgOrientationClassification,
DocVLM,
FormulaRecognition,
LayoutDetection,
SealTextDetection,
TableCellsDetection,
TableClassification,
TableStructureRecognition,
TextDetection,
TextImageUnwarping,
TextLineOrientationClassification,
TextRecognition,
]:
subcommand_executor = cls.get_cli_subcommand_executor()
subparser = subcommand_executor.add_subparser(subparsers)
subparser.set_defaults(executor=subcommand_executor.execute_with_args)
def _register_install_hpi_deps_command(subparsers):
def _install_hpi_deps(args):
hpip = f"hpi-{args.variant}"
try:
subprocess.check_call(["paddlex", "--install", hpip])
subprocess.check_call(["paddlex", "--install", "paddle2onnx"])
except subprocess.CalledProcessError:
sys.exit("Failed to install dependencies")
subparser = subparsers.add_parser("install_hpi_deps")
subparser.add_argument("variant", type=str, choices=["cpu", "gpu", "npu"])
subparser.set_defaults(executor=_install_hpi_deps)
def _register_install_genai_server_deps_command(subparsers):
def _install_genai_server_deps(args):
try:
subprocess.check_call(
["paddlex", "--install", f"genai-{args.variant}-server"]
)
except subprocess.CalledProcessError:
sys.exit("Failed to install dependencies")
subparser = subparsers.add_parser("install_genai_server_deps")
subparser.add_argument(
"variant", type=str, choices=["vllm", "sglang", "fastdeploy"]
)
subparser.set_defaults(executor=_install_genai_server_deps)
def _register_genai_server_command(subparsers):
# TODO: Register the subparser whether the plugin is installed or not
try:
from paddlex.inference.genai.server import get_arg_parser, run_genai_server
except RuntimeError:
return
def _show_prompt_when_server_is_running(host, port, backend):
if host == "0.0.0.0":
host = "localhost"
while True:
try:
resp = requests.get(f"http://{host}:{port}/health", timeout=1)
if resp.status_code == 200:
break
except (requests.exceptions.ConnectionError, requests.exceptions.Timeout):
pass
time.sleep(1)
prompt = f"""The PaddleOCR GenAI server has been started. You can either:
1. Set the server URL in the module or pipeline configuration and call the PaddleOCR CLI or Python API. For example:
paddleocr doc_parser --input demo.png --vl_rec_backend {backend}-server --vl_rec_server_url http://{host}:{port}/v1
2. Make HTTP requests directly, or using the OpenAI client library."""
logger.info(prompt)
def _run_genai_server(args):
Thread(
target=_show_prompt_when_server_is_running,
args=(args.host, args.port, args.backend),
daemon=True,
).start()
try:
run_genai_server(args)
except subprocess.CalledProcessError:
sys.exit("Failed to run the server")
paddlex_parser = get_arg_parser()
subparser = subparsers.add_parser(
"genai_server", parents=[paddlex_parser], conflict_handler="resolve"
)
subparser.set_defaults(executor=_run_genai_server)
def _get_parser():
parser = argparse.ArgumentParser(prog="paddleocr")
parser.add_argument(
"-v", "--version", action="version", version=f"%(prog)s {version}"
)
subparsers = parser.add_subparsers(dest="subcommand")
_register_pipelines(subparsers)
_register_models(subparsers)
_register_install_hpi_deps_command(subparsers)
_register_install_genai_server_deps_command(subparsers)
_register_genai_server_command(subparsers)
return parser
def _execute(args):
args.executor(args)
def main():
logger.setLevel(logging.INFO)
warnings.filterwarnings("default", category=CLIDeprecationWarning)
parser = _get_parser()
args = parser.parse_args()
if args.subcommand is None:
parser.print_usage(sys.stderr)
sys.exit(2)
_execute(args)