011、AI大模型,GPU调度方法 2024-07-21 字数统计: 117字 | 阅读时长≈ 1分 本人使用的显卡是: A100使用AI大模型参数级别是: 100b Ai大模型,GPU调度方法12345678910111213141516171819202122232425262728293031323334353637383940414243444546kubectl label node gpu=a100apiVersion: apps/v1kind: StatefulSetmetadata: name: ollama namespace: llamaspec: serviceName: "ollama" replicas: 1 selector: matchLabels: app: ollama template: metadata: labels: app: ollama spec: nodeSelector: gpu: a100 containers: - name: ollama image: ollama/ollama:latest ports: - containerPort: 11434 resources: requests: cpu: "6000m" memory: "6Gi" nvidia.com/gpu: "6" limits: cpu: "30000m" memory: "30Gi" volumeMounts: - name: ollama-volume mountPath: /root/.ollama tty: true volumeClaimTemplates: - metadata: name: ollama-volume spec: accessModes: ["ReadWriteOnce"] resources: requests: storage: 2000Gi AI