-
Couldn't load subscription status.
- Fork 21
Update Jetstream catalog to support MaxText and Pytorch #30
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: rlakhtakia The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
|
Welcome @rlakhtakia! |
|
Thanks for improving this! Let's remove the llama and gemma examples under the Then also comment validating the yamls deploy and the model servers start correctly. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Pull Request Overview
Adds support for JetStream-PyTorch (gemma-7b-it) and JetStream-MaxText (llama3-8b) by creating dedicated gke overlays and base patches for both models.
- Introduces new
gkeoverlay directories with job, deployment, HPA, and README for PyTorch and MaxText. - Adds base kustomization and patch files for service, job, and deployment under each model.
- Removes unused
kagglemounts from the shared base job and cleans up legacy definitions.
Reviewed Changes
Copilot reviewed 30 out of 30 changed files in this pull request and generated no comments.
Show a summary per file
| File | Description |
|---|---|
| serving-catalog/core/deployment/jetstream/pytorch/gemma-7b-it/gke/job.patch.yaml | New TPU job overlay |
| serving-catalog/core/deployment/jetstream/pytorch/gemma-7b-it/gke/hpa.patch.yaml | HPA patch for inference server |
| serving-catalog/core/deployment/jetstream/pytorch/gemma-7b-it/gke/deployment.patch.yaml | Deployment labels for gemma-7b-it |
| serving-catalog/core/deployment/jetstream/pytorch/gemma-7b-it/gke/README.md | Usage instructions |
| serving-catalog/core/deployment/jetstream/pytorch/gemma-7b-it/base/* | Base patches and kustomization for gemma-7b-it |
| serving-catalog/core/deployment/jetstream/maxtext/llama3-8b/gke/* | New GKE overlay for llama3-8b |
| serving-catalog/core/deployment/jetstream/maxtext/llama3-8b/base/* | Base patches and kustomization for llama3-8b |
| serving-catalog/core/deployment/jetstream/base/job.yaml | Removed legacy kaggle mounts |
|
The Kubernetes project currently lacks enough contributors to adequately respond to all PRs. This bot triages PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
|
The Kubernetes project currently lacks enough active contributors to adequately respond to all PRs. This bot triages PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle rotten |
|
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs. This bot triages PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /close |
|
@k8s-triage-robot: Closed this PR. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
Uh oh!
There was an error while loading. Please reload this page.