Replies: 2 comments
-
I believe @booxter is proposing something as well. |
Beta Was this translation helpful? Give feedback.
0 replies
-
I posted #1238 that is relevant here. I'm focusing on API modeling and provider interfaces to Jobs API layer. Particular backends are an interesting discussion; I would need to explore the space a bit more to make any judgement one way or another. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
In order for Llama Stack to be deployable either as a standalone application and or as distributed, service oriented architecture, some sort of abstraction over computation-heavy resources is needed. This can be custom or it can leverage existing, mature solutions.
Use cases include at least training, synthetic data generation, and document processing.
It seams reasonable to say that basic requirements would be that such an abstraction should be able to function in a single-node deployment, ideally without requiring a message queue in that case, as a distributed cluster, or on Kubernetes.
Options include:
It seems that Airflow would be the most general solution that subsumes at least most of the other major options.
Has this been discussed before for Llama Stack?
Beta Was this translation helpful? Give feedback.
All reactions