Trying to deploy Llava-Mistral using a simple Docker image, receive both success & error msgs
I am using a simple Docker script to deploy Llava-Mistral. In the system logs, it creates the container successfully. In the container logs, I get the following:
Script:
The system logs spam me with "start container" as well.
I made sure to use absolute paths to make certain that this is pointed at the right spot. I also tested this in Docker Desktop and it worked flawlessly. My question is what am I doing wrong here? Why am I unable to get a connection to the endpoint?
I'd also like to know what a typical request would look like to an exposed port on the https /run endpoint. Typically reverse proxies don't use ports so I'd like to know what the norm is for that.
22 Replies
You need to use
sleep infinity
to keep your container alive. I also have a LLaVA template that you can use that is working.yes that'd be awesome. any tips on getting mistral in particular working?
was i on the right track with my container or?
Mistral 7B is the default model in my green template
so where is the template
regular or instruct
"LLaVA 1.6" under the "Communtiy" section of "Explore".
im trying to implement this on serverless though
this is the serverless support section
I don't see a runpod hander in your Dockerfile
i'm new to runpod
can you show me what i need to change?
Here are some resources for getting started with RunPod serverless:
https://blog.runpod.io/serverless-create-a-basic-api/
https://www.youtube.com/@generativelabs/videos
https://trapdoor.cloud/getting-started-with-runpod-serverless/
well you just told me what was missing, so how about just telling me directly so i don't have to sift through all that
It is not my job to hold your hand and do everything for you. I told you your handler was missing, use your brain and follow the resources I sent you otherwise I will gladly help you for $100 per hour of my time.
i asked for a courtesy, you respond with sass?
you said in the article yourself you aren't an expert with implementing llava. your time is not worth $100 per hour
Then stuggle with it yourself
you are a childish man
Nope, I told you want to do but you are too lazy and expect everyone to do everything for you. That is not how life works. I offered to help for my hourly rate then you insult me, when I am one of the most experienced people on RunPod. YOU are chilidhs and a comple fucking idiot.
i don't know a single noteworthy person who yells their credentials when somebody upsets them
imagine going into a help section and calling someone the r slur
@ashleyk Lets chill and let me handle this. No need to make another argue 🙂
@B1llstar have you tried to put
as docker command. Also do you use network/volume storage?
I would also change way you store models in image:
though note it will work on pods as for serverlles you need have handler file that will process job requests
nice, i will look at this today. thank you for the level-headed response
Though like askleyk said have look at the links he send they are good examples how to start with serverless
i don't know if that guy represents you but it's probably not a good idea to have someone yelling obscenities like that
ashleyk is person that creates many templates and he is always willing to help. Though not except that we are ChatGPT and we will give you working solution cause you want to.
i didn't quite understand that second sentence, but i think i understand what you're getting at?
i honestly mainly asked for direct help because i figured the fix was a single line of code or something that i was missing in the docker file lol
i'll be looking into the hander today though. i did glance at the articles and they were well-written. i can "separate the artist from their work"