Skip to main content
Glama
_cli.py6.3 kB
# Copyright (c) 2025 PaddlePaddle Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import logging import subprocess import sys import time import warnings from threading import Thread import requests from ._models import ( ChartParsing, DocImgOrientationClassification, DocVLM, FormulaRecognition, LayoutDetection, SealTextDetection, TableCellsDetection, TableClassification, TableStructureRecognition, TextDetection, TextImageUnwarping, TextLineOrientationClassification, TextRecognition, ) from ._pipelines import ( DocPreprocessor, DocUnderstanding, FormulaRecognitionPipeline, PaddleOCR, PaddleOCRVL, PPChatOCRv4Doc, PPDocTranslation, PPStructureV3, SealRecognition, TableRecognitionPipelineV2, ) from ._version import version from ._utils.deprecation import CLIDeprecationWarning from ._utils.logging import logger def _register_pipelines(subparsers): for cls in [ DocPreprocessor, DocUnderstanding, FormulaRecognitionPipeline, PaddleOCR, PaddleOCRVL, PPChatOCRv4Doc, PPDocTranslation, PPStructureV3, SealRecognition, TableRecognitionPipelineV2, ]: subcommand_executor = cls.get_cli_subcommand_executor() subparser = subcommand_executor.add_subparser(subparsers) subparser.set_defaults(executor=subcommand_executor.execute_with_args) def _register_models(subparsers): for cls in [ ChartParsing, DocImgOrientationClassification, DocVLM, FormulaRecognition, LayoutDetection, SealTextDetection, TableCellsDetection, TableClassification, TableStructureRecognition, TextDetection, TextImageUnwarping, TextLineOrientationClassification, TextRecognition, ]: subcommand_executor = cls.get_cli_subcommand_executor() subparser = subcommand_executor.add_subparser(subparsers) subparser.set_defaults(executor=subcommand_executor.execute_with_args) def _register_install_hpi_deps_command(subparsers): def _install_hpi_deps(args): hpip = f"hpi-{args.variant}" try: subprocess.check_call(["paddlex", "--install", hpip]) subprocess.check_call(["paddlex", "--install", "paddle2onnx"]) except subprocess.CalledProcessError: sys.exit("Failed to install dependencies") subparser = subparsers.add_parser("install_hpi_deps") subparser.add_argument("variant", type=str, choices=["cpu", "gpu", "npu"]) subparser.set_defaults(executor=_install_hpi_deps) def _register_install_genai_server_deps_command(subparsers): def _install_genai_server_deps(args): try: subprocess.check_call( ["paddlex", "--install", f"genai-{args.variant}-server"] ) except subprocess.CalledProcessError: sys.exit("Failed to install dependencies") subparser = subparsers.add_parser("install_genai_server_deps") subparser.add_argument( "variant", type=str, choices=["vllm", "sglang", "fastdeploy"] ) subparser.set_defaults(executor=_install_genai_server_deps) def _register_genai_server_command(subparsers): # TODO: Register the subparser whether the plugin is installed or not try: from paddlex.inference.genai.server import get_arg_parser, run_genai_server except RuntimeError: return def _show_prompt_when_server_is_running(host, port, backend): if host == "0.0.0.0": host = "localhost" while True: try: resp = requests.get(f"http://{host}:{port}/health", timeout=1) if resp.status_code == 200: break except (requests.exceptions.ConnectionError, requests.exceptions.Timeout): pass time.sleep(1) prompt = f"""The PaddleOCR GenAI server has been started. You can either: 1. Set the server URL in the module or pipeline configuration and call the PaddleOCR CLI or Python API. For example: paddleocr doc_parser --input demo.png --vl_rec_backend {backend}-server --vl_rec_server_url http://{host}:{port}/v1 2. Make HTTP requests directly, or using the OpenAI client library.""" logger.info(prompt) def _run_genai_server(args): Thread( target=_show_prompt_when_server_is_running, args=(args.host, args.port, args.backend), daemon=True, ).start() try: run_genai_server(args) except subprocess.CalledProcessError: sys.exit("Failed to run the server") paddlex_parser = get_arg_parser() subparser = subparsers.add_parser( "genai_server", parents=[paddlex_parser], conflict_handler="resolve" ) subparser.set_defaults(executor=_run_genai_server) def _get_parser(): parser = argparse.ArgumentParser(prog="paddleocr") parser.add_argument( "-v", "--version", action="version", version=f"%(prog)s {version}" ) subparsers = parser.add_subparsers(dest="subcommand") _register_pipelines(subparsers) _register_models(subparsers) _register_install_hpi_deps_command(subparsers) _register_install_genai_server_deps_command(subparsers) _register_genai_server_command(subparsers) return parser def _execute(args): args.executor(args) def main(): logger.setLevel(logging.INFO) warnings.filterwarnings("default", category=CLIDeprecationWarning) parser = _get_parser() args = parser.parse_args() if args.subcommand is None: parser.print_usage(sys.stderr) sys.exit(2) _execute(args)

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/PaddlePaddle/PaddleOCR'

If you have feedback or need assistance with the MCP directory API, please join our Discord server