Skip to content
This repository has been archived by the owner on Jun 25, 2023. It is now read-only.

Electriclizard triton solution #36

Open
wants to merge 34 commits into
base: main
Choose a base branch
from

Conversation

electriclizard
Copy link
Contributor

@electriclizard electriclizard commented Jun 11, 2023

Hi @rsolovev, i've made a new solution, based on nvidia-triton, metrics raise x2 on local tests, can you please explain me how can i correctly deploy an application using docker-compose? Now it starts with docker-compose up
telegram-cloud-photo-size-2-5201986564564044257-y

@electriclizard
Copy link
Contributor Author

electriclizard commented Jun 15, 2023

Hi @rsolovev, finally i've decided to avoid from docker-compose back to a single dockerfile, now it builds in ci and metrics raises again, hope to see a test a results on your environment
now service starts with docker run --gpus 0 --shm-size 16G -p 8080:8080 image-name --shm-size is recuired for tritonserver

local test results
telegram-cloud-photo-size-2-5213390626152892225-y

@electriclizard electriclizard force-pushed the Electriclizard-triton-solution branch from 78b0ec2 to bd49505 Compare June 15, 2023 10:05
Copy link
Collaborator

@rsolovev rsolovev left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

hey @electriclizard, thank you for this iteration, unfortunately this didnt work out:

shm-size is not configurable in k8s [1]. There are some workarounds with mounting an empty volume, but in k8s 1.24 they removed dockershim support completely [2] so this workarounds kinda go against our infrastructure "standards" (not that it is impossible to configure, just "feels unstable" in this form). Also image size is quite large (25GB) -- it failed to build on github runner and I had to build it locally.

I'm sure that with more time we could make this approach feasible and stable, but right now it is inconvenient to follow

@electriclizard
Copy link
Contributor Author

but right now it is inconvenient to follow

hey @rsolovev, thank you for a try anyway, yes image is quite big, thats true. I just wanted to test it, because i am sure and local tests confirm my theory, that it might be impossible to run models on python with the same productivity

@electriclizard
Copy link
Contributor Author

electriclizard commented Jun 19, 2023

Hi @rsolovev, i can decrease image size to fix bulding on a github runner, but i cannot change infrastructure, may be can i try to add these lines to helm?

volumes:
        - name: dshm
          emptyDir:
            medium: Memory

Because i've checked it on 1.24 api k8s version and it works

@rsolovev
Copy link
Collaborator

Because i've checked it on 1.24 api k8s version and it works

that sounds great, lets try reducing the image size first, and then I'll tweak the volume confs backstage

@electriclizard
Copy link
Contributor Author

Because i've checked it on 1.24 api k8s version and it works

that sounds great, lets try reducing the image size first, and then I'll tweak the volume confs backstage

great, i'll push the commit later

@electriclizard
Copy link
Contributor Author

hey @rsolovev i decreased the image size to 20.1gb, by pulling models before the app start, not while building the image, it takes a few minutes before start now

Copy link
Collaborator

@rsolovev rsolovev left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@electriclizard impressive! I take my words back, no issues with dshm volume from the first try. An absolute peak rps and total requests record (121 rps, 26850 total) -- full stats

p.s added shm mount as a reference

@electriclizard
Copy link
Contributor Author

electriclizard commented Jun 19, 2023

@electriclizard impressive! I take my words back, no issues with dshm volume from the first try. An absolute peak rps and total requests record (121 rps, 26850 total) -- full stats

That's great!!! Thank you for a run, it was my latest commit to a challenge))

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants