Monster
Explore posts from serversRRunPod
•Created by Monster on 7/13/2024 in #⚡|serverless
not enough GPUs free
Hi there,
wish you a good day today. I have a serverless endpoint running on runpod, it is created on top of the network storage belongs to US-OR-1 data center. it was running well for somedays, but 20 mins before, I have encountered the issue that no worker is able to be created because no GPU resource. the system throws a log like this repeatedly.
2024-07-13T06:32:22Z create container USERNAME/ENDPOINT
2024-07-13T06:32:22Z error creating container: not enough GPUs free
how can I make sure there are GPU resources whenever the request comes, should I change the endpoint and the network volume to other region which has more GPU resoures? how often this shortage will be happening. it post a risk on the stability and quality of service which is critical in most scenarios.
thank you.
39 replies
deploy react+fastapi+postgresql+celery on railway
Hi, I have a github repo composed of fastapi as backend, and react as the frontend, and the postgreSQL as DB, all of them are dockerized, all of the dockers are configured to be running on local machine perfectly. now I am trying to deploy the full stack on railway, but I did not find much information to guide me through. my friend recommended me to railway instead of aws, it supposed to be straightforward to deploy such a typical stack. I am confused, am I misunderstood the way how railway is supposed to be used.
67 replies
RRunPod
•Created by Monster on 6/13/2024 in #⚡|serverless
is there example code to access the runpod-worker-comfy serverless endpoint
Hi, I have managed to run the runpod-worker-comfy serverless endpoint. and I know it supports for 5 entries: RUN, RUNSYNC, STATUS, CANCEL, HEALTH. but I do not exactly know how to access the service from my python code. like how to prepare the api-key, the worker id, how to prepare the request for RUN, and how to check the status until it is finished, and download the generated image. anywhere exists a example code to do these basic operation from my python code?
Previously I have python code to communicate directly with the comfyUI server, which will create a websocket, send the workflow with http post, keep checking the history, once the work is done, read the image from the output which passed through the websocket connection. but when wrapped with runpod-worker-comfy, indeed, the interface is more easy, and there is input validation which is great. but I do not know how to use it from my code, and did not find any example code to access it, sorry for my ignorance.
28 replies
RRunPod
•Created by Monster on 6/12/2024 in #⚡|serverless
can 3 different serverless workers running from same network volume?
Hi @digigoblin
I have checked your answer about symbol linking network volume dir to serverless dir and run worker from the network volume as it did as if it was a separate pod instance.
https://github.com/ashleykleynhans/runpod-worker-comfyui/blob/main/start.sh#L5-L7
I fully understand it and thank you for the instruction. Nevertheless, if I am running 3 serverless workers, all of them will launch comfyui main script from the same network volume folder and read and write the ./input and ./output directory or the big model files simultaneously from the same volume. is this a problem in terms of the performance, or do you have a best practice for such a typical scenario where multiple workers sharing same source directory because we do not want to copy big models and working directory for each of these workers.
thank you so much and looking forward to your kindly response.
64 replies
can I deploy flask, celery, redis, postgreSQL on runpod?
Hi,
as you know the pod only persist data under /workspace folder. for all python related packages I can use venv to put all the data and configuration under /workspace.
while if I need to install all the tools like flask, celery, redis, postgreSQL they are not python installation, the configuration files will be scattered here and there. all these file and configuration will disappear after pod restart.
So, what is the best practice to do so ? or runpod is not designed to be used in this scenario.
thank you !
42 replies
Apply a fix public ip and attach to the running pod, Attach a network volume to the same pod.
Hi, I am new user of runpod. I have one pod running. but I cannot find and place to apply for a fix public ip and attach to the running pod.
also I need to put the data on a persistent storage, that is why I have created a network volume. but I cannot find any where to attach it to my pod.
I think this is very basic requirements that majority of the user will need. it must be somewhere in the document. but unfortunately I did not find the answer in the document either.
please help, thx
4 replies