-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Finalize generation script #33
Finalize generation script #33
Conversation
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
Signed-off-by: thepetk <[email protected]>
WIP: because I'm working on an issue with the MODEL_ENDPOINT_BEARER env var and the update of the deployment done by the tekton pipeline |
Signed-off-by: thepetk <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Just some review of the README's with some minor wording suggestions
I'm booting up a cluster now and will checkout this branch and given a go later today @thepetk - thanks
Testing the 3 charts worked for me @thepetk , including an update to the app getting rolled out. I'm guessing until we have an update to |
Yeah it should be this way and pretty much this is the main reason I added the workflow just to be sure every time we are up-to-date with the ai-lab-app dependency. For the |
Co-authored-by: Gabe Montero <[email protected]>
Co-authored-by: Gabe Montero <[email protected]>
Co-authored-by: Gabe Montero <[email protected]>
Sure all that is good. I'm just saying we can't fully validate until a change is made to those repo(s). Before you move your Jira to closed, were you planning on making some sort of change in those repo(s) that would get caught by your github workflow, or get seen if someone ran |
I think we are getting new changes right? Because we are importing the changes from the linter merged in the |
I've managed to make the bearer auth work. In order to achieve that I've opened a PR in the rhdh-pipelines side and I made an update to the app-config so it can stay permanently as a configmap so we can fetch each time the necessary variables. @gabemontero PTAL :) |
Example screenshot using existing model server, bearer auth and the updated pipeline introduced by redhat-ai-dev/rhdh-pipelines#11 Btw the whole experience with an existing model server is way faster, the application is up and running very fast, so the git repo and the pipelines. |
Either I'm mis-interpreting what you are saying, or I'm not making myself sufficiently clear, or something along those lines @thepetk Let me try this way: At some point we should be able to submit at PR against this repo, and see the print at https://github.com/redhat-ai-dev/ai-lab-helm-charts/pull/33/files#diff-119dd431be9907b10ad9714895fb89d998ccd55bf25f3cc6bb4f1573ed7ca7e2R21 occur because on of those other repositories have been updated, correct? Ideally we test that flow with perhaps a simple change, non effecting change in one of the other repos, merge in one of those repos, and then see your new workflow flag that Correct? |
Yeap! That's correct! Also apologies if my message was not clear. I meant that in the current PR we can confirm that the Now to test also the opposite, e.g that the workflow fails when "a change is introduced in workflow but not included in the PR", I've added a quick change in my fork of the In my |
If that occurred, I am unable to determine that. Is it possibly one of the 18 commits that pulled in that change that you can point me to?
Perfect :-) .... your verification via fork changes is even better then the post mortem "temp change" I was suggesting. |
So is there a situation where we need to merge the rhdh-pipelines change BEFORE this one, or are you just citing the relationship, but they can merge independently @thepetk ? |
I'm quite on the fence, but honestly I don't see a blocker but as you say there's a relationship. My plan is:
|
I think that one is a good example: ad618d8 |
What does this PR do?:
The PR finalizes the automatic conversion of the shared gitops resources between the
ai-lab-app
and theai-lab-helm-chart
.More specifically it introduces the following updates:
generate.sh
is added and is the main point to update all the shared resources. Along with the script a workflow has been added to check in every new PR that we are aligned with theai-lab-app
repo. This way we guarantee that bothai-lab-app
&ai-lab-samples
are always up-to-date. Finally, a quick readme has been added to capture all this information. I find the best location to be theai-software-templates
directory which contains all the related charts.chatbot-ai-sample
chart, all the updates brought from Add partially-automatic conversion flow #31 are now tested and captured in the readme tmpl (and ofc in the readme) of the chart.values.dbRequired
is removed for now from the new resource content as it is only related with the RAG case. I think the best approach here is to create a new chart that will cover this case and not use the chatbot chart for this.Apart from the above, also some other details are:
existingModelServer
will remove themodel-server
deployment if set, the conversion script now adds on the top and bottom of all*model-server*
files a condition.Which issue(s) this PR fixes:
Fixes RHDHPAI-379
PR acceptance criteria:
Testing and documentation do not need to be complete in order for this PR to be approved. We just need to ensure tracking issues are opened and linked to this PR, if they are not in the PR scope due to various constraints.
How to test changes / Special notes to the reviewer:
In regards to testing, I've tested the following cases:
vLLM
support. IfvllmSelected
is set the model server deployed is based on this technology. To test this case you need to provision a cluster with GPU support: