-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Kubeshare prototyping and compute sharing deep dive #88
Comments
Comparing with GaiaGPU solution, I think Kubeshare's architecture has some good designs to monitor/manage the compute resources. |
To do: Run a pod with 50% utilization request with KubeShare, and monitor GPU utilization fluctuation. |
Gemini algorithm major features:
|
Gemini algorithm is deserved to be further explored and tested. I like it very much. The next step analysis involves with a lot of coding and testing. two main questions: 1 Will GPU be idle if one pod runs out of quota and there is no other pod/thread asking for GPU? answer: normally it doesn't run into situation. even it runs into that unique situation, it should be avoidable by changing parameters. 2. does cuda do context switching when the pod is changed? It is vGPU driver's question, not Kubeshare's question. In my research, vGPU driver should use one context among all sharing threads. |
No description provided.
The text was updated successfully, but these errors were encountered: