Differences between Gradio and code output

#10
by SimoneG98 - opened

Hello everyone,
I wanted to test MinerU on my server, so I installed everything inside a Docker starting from the latest slang image, and I got everything working. I copied the parameters that your demo dashboard on Gradio is sending, and I'm using the vlm_slang. However, I get very different results loading the same document in your dashboard or in my code. Am I missing something? is the open-source code a different version from the demo one?

image.png

I call in this way, with method "auto", language "ch", and backend "vlm-sglang-engine"

Thank you!

OpenDataLab org

Please commit more information about the different.

myhloli changed discussion status to closed

Hi,
I am having the same issue with the docker implementation and the demo. The demo produces the correct results but the docker implementation is off e.g it is capturing the header but the row data is missing.
I know it is not the OCR model as I updated to MinerU2.5-Pro-2604-1.2B in docker.

These are the parameters that I am using:
"return_middle_json": "false",
"return_model_output": "false",
"return_md": "true",
"return_images": "false",
"end_page_id": "99999",
"parse_method": "auto",
"start_page_id": "0",
"lang_list": "en",
"server_url": "",
"return_content_list": "false",
"backend": "vlm-auto-engine",
"table_enable": "true",
"response_format_zip": "false",
"return_original_file": "false",
"formula_enable": "true",

OpenDataLab org

@dalow24 We have just released version 3.1.0 to support the 2.5 Pro model. You need to update MinerU to the latest version for compatibility.

Thanks. I realize I had to update to the latest model opendatalab/MinerU2.5-Pro-2604-1.2B on mineru-api.
I was updating model on the wrong instance on mineru-openai-server
I am now getting the same results in my docker deploy versus the online demo

Sign up or log in to comment