compose.yaml.bak 3.0 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687
  1. services:
  2. mineru-openai-server:
  3. image: mineru:latest
  4. container_name: mineru-openai-server
  5. restart: always
  6. profiles: ["openai-server"]
  7. ports:
  8. - 30000:30000
  9. environment:
  10. MINERU_MODEL_SOURCE: local
  11. entrypoint: mineru-openai-server
  12. command:
  13. --host 0.0.0.0
  14. --port 30000
  15. # --data-parallel-size 2 # If using multiple GPUs, increase throughput using vllm's multi-GPU parallel mode
  16. # --gpu-memory-utilization 0.5 # If running on a single GPU and encountering VRAM shortage, reduce the KV cache size by this parameter, if VRAM issues persist, try lowering it further to `0.4` or below.
  17. ulimits:
  18. memlock: -1
  19. stack: 67108864
  20. ipc: host
  21. healthcheck:
  22. test: ["CMD-SHELL", "curl -f http://localhost:30000/health || exit 1"]
  23. deploy:
  24. resources:
  25. reservations:
  26. devices:
  27. - driver: nvidia
  28. device_ids: ["0"] # Modify for multiple GPUs: ["0", "1"]
  29. capabilities: [gpu]
  30. mineru-api:
  31. image: mineru:latest
  32. container_name: mineru-api
  33. restart: always
  34. profiles: ["api"]
  35. ports:
  36. - 23424:8000
  37. environment:
  38. MINERU_MODEL_SOURCE: local
  39. entrypoint: mineru-api
  40. command:
  41. --host 0.0.0.0
  42. --port 8000
  43. # parameters for vllm-engine
  44. # --data-parallel-size 2 # If using multiple GPUs, increase throughput using vllm's multi-GPU parallel mode
  45. # --gpu-memory-utilization 0.5 # If running on a single GPU and encountering VRAM shortage, reduce the KV cache size by this parameter, if VRAM issues persist, try lowering it further to `0.4` or below.
  46. ulimits:
  47. memlock: -1
  48. stack: 67108864
  49. ipc: host
  50. deploy:
  51. resources:
  52. reservations:
  53. devices:
  54. - driver: nvidia
  55. device_ids: ["0"] # Modify for multiple GPUs: ["0", "1"]
  56. capabilities: [gpu]
  57. mineru-gradio:
  58. image: mineru:latest
  59. container_name: mineru-gradio
  60. restart: always
  61. profiles: ["gradio"]
  62. ports:
  63. - 23425:7860
  64. environment:
  65. MINERU_MODEL_SOURCE: local
  66. entrypoint: mineru-gradio
  67. command:
  68. --server-name 0.0.0.0
  69. --server-port 7860
  70. # --enable-api false # If you want to disable the API, set this to false
  71. # --max-convert-pages 20 # If you want to limit the number of pages for conversion, set this to a specific number
  72. # parameters for vllm-engine
  73. # --data-parallel-size 2 # If using multiple GPUs, increase throughput using vllm's multi-GPU parallel mode
  74. # --gpu-memory-utilization 0.5 # If running on a single GPU and encountering VRAM shortage, reduce the KV cache size by this parameter, if VRAM issues persist, try lowering it further to `0.4` or below.
  75. ulimits:
  76. memlock: -1
  77. stack: 67108864
  78. ipc: host
  79. deploy:
  80. resources:
  81. reservations:
  82. devices:
  83. - driver: nvidia
  84. device_ids: ["0"] # Modify for multiple GPUs: ["0", "1"]
  85. capabilities: [gpu]