docker-compose.yml 1.6 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465
  1. version: "3.8"
  2. services:
  3. # vLLM - Qwen3 0.6B
  4. vllm-0.6b:
  5. image: swr.cn-north-4.myhuaweicloud.com/ddn-k8s/docker.io/vllm/vllm-openai:latest
  6. container_name: vllm-qwen3-0.6b
  7. ports:
  8. - "8100:8000"
  9. environment:
  10. - HF_ENDPOINT=https://hf-mirror.com
  11. command: >
  12. --model Qwen/Qwen3-Embedding-0.6B
  13. --dtype float16
  14. --port 8000
  15. volumes:
  16. - ./models:/root/.cache/huggingface
  17. deploy:
  18. resources:
  19. reservations:
  20. devices:
  21. - driver: nvidia
  22. count: all
  23. capabilities: [ gpu ]
  24. # # vLLM - Qwen3 4B
  25. # vllm-4b:
  26. # image: vllm/vllm-openai:latest
  27. # container_name: vllm-qwen3-4b
  28. # ports:
  29. # - "8200:8000"
  30. # command: >
  31. # --model Qwen/Qwen3-Embedding-4B
  32. # --dtype float16
  33. # --api-port 8000
  34. # volumes:
  35. # - ./models:/root/.cache/huggingface
  36. #
  37. # # vLLM - Qwen3 8B
  38. # vllm-8b:
  39. # image: vllm/vllm-openai:latest
  40. # container_name: vllm-qwen3-8b
  41. # ports:
  42. # - "8300:8000"
  43. # command: >
  44. # --model Qwen/Qwen3-Embedding-8B
  45. # --dtype float16
  46. # --api-port 8000
  47. # volumes:
  48. # - ./models:/root/.cache/huggingface
  49. # Milvus 向量数据库
  50. # milvus:
  51. # image: swr.cn-north-4.myhuaweicloud.com/ddn-k8s/docker.io/milvusdb/milvus:v2.4.5
  52. ## image: milvusdb/milvus:v2.4.4-standalone
  53. # container_name: milvus
  54. # ports:
  55. # - "19530:19530"
  56. # - "9091:9091"
  57. # environment:
  58. # - ETCD_USE_EMBED=true
  59. # - MINIO_USE_EMBED=true
  60. # - PULSAR_USE_EMBED=true
  61. # volumes:
  62. # - ./milvus_data:/var/lib/milvus
  63. # command: ["milvus", "run", "standalone"]