R
RunPod12mo ago
J.

How to WebSocket to Serverless Pods

@Merrell / @flash-singh Wondering if I can get code-pointers on how to use the API to expose ports on serverless programatically, so that I can do stuff like web sockets?
86 Replies
Jason
Jason12mo ago
Wow if this is possible then it's great But I think they'll allow some ports for you to use rather than allowing you to control the open ports on serverless ..
J.
J.OP12mo ago
they used to be able to have in the UI the ability to open up ports just like GPU Pod, but they said they removed it to not cause customer confusion / etc. But supposedly, since it's just a GPU Pod under the hood of serverless, anything should be possible on serverless that is on GPU Pod. Well either way, even if its just a dedicated port on serverless for web sockets, that still nice, to be able to stream data / files
Jason
Jason12mo ago
Ooh that's weird customer confusion? So, Now they simplify it by automating the assignment of ports ye Yeah I'm interested in this too 👌
J.
J.OP12mo ago
I think there is no ports exposed at all right now, just a simple handler function; except for certain customers i guess of theirs are using it, but not widely known knowledge
Jason
Jason12mo ago
Oh yea for gpu pods* I mean tcp ports are randomly assigned right
flash-singh
flash-singh11mo ago
you would need to use our graphql api to add ports to a serverless template, from there the workers will have ports, by default gui will not let you add ports tcp ports are randomly assigned but we do expose them in env variables, so technically you can expose that port back to your infra using progress updates within a job heres the flow that someone else used - run a job - worker will launch and env is populated, in progress of the job share back port / ip details - catch port / ip details in your app using /status or /stream, do whatever you want with the port / ip - when your done using the port / ip, send message through that port / ip service to mark the job done on worker side - once job is done, the worker will turn off automatically if there are no more jobs
J.
J.OP11mo ago
Thank you! ❤️ Appreciate it! ❤️
Jason
Jason11mo ago
Oh is it already possible by now and documented in the reference? Which env variable BTW?
flash-singh
flash-singh11mo ago
on any pod, you can get env variables and ports are exposed as part of env, ill have to check where we document it
Builderman
Builderman11mo ago
is it possible to stream with this? websockets are definitely more efficient
Jason
Jason11mo ago
Yeah oh really? We sockets more efficient?
Builderman
Builderman11mo ago
considering my method of streaming yes we just spam the request to get all chunks
flash-singh
flash-singh11mo ago
websockets should work the same since your using ports, also you can use server-sent events, SSE https://medium.com/deliveryherotechhub/what-is-server-sent-events-sse-and-how-to-implement-it-904938bffd73
Jason
Jason11mo ago
Hey @flash-singh have you got the env variable yet? Please send the env name for exposed ports if there is
flash-singh
flash-singh11mo ago
heres an example, if you expose 22 port in your template, env will show this RUNPOD_TCP_PORT_22=22097 that is saying port 22 is mapped to 22097 externally to users
Jason
Jason11mo ago
Ohh it works on pods and serverless? ok thanks Oh and will i be able to get the RUNPOD_PUBLIC_IP everytime? for serverless
flash-singh
flash-singh11mo ago
yes, serverless runs on secure cloud and all secure cloud pods have public ip, just keep in mind ip and port can be different for every job you run in serverless
Jason
Jason11mo ago
port_22 = os.environ.get('RUNPOD_TCP_PORT_22')
public_ip = os.environ.get('RUNPOD_PUBLIC_IP')

#ITs all none when i return them
return {
'message': "public ip: {}, port_ref: {}".format(public_ip, port_8188) # None, None
}
port_22 = os.environ.get('RUNPOD_TCP_PORT_22')
public_ip = os.environ.get('RUNPOD_PUBLIC_IP')

#ITs all none when i return them
return {
'message': "public ip: {}, port_ref: {}".format(public_ip, port_8188) # None, None
}
i did this on dockerfile: EXPOSE 22/tcp I did the graphql port assignment too on tcp
flash-singh
flash-singh11mo ago
did it work?
Jason
Jason11mo ago
Nah as I said it returned only none object on python Why is that? Do you wanna check out my template to see if the ports is there?
flash-singh
flash-singh11mo ago
yeah sure
Jason
Jason11mo ago
@flash-singh 3pgp4kglgq heres the template id sorry for the wait lol can you make sure if its already right and any updates on the public ip env or the port env? or any alternatives for websockets in serverless? ( it needs ports right? which defaults to http or https ) i should've used the stream instead for now yeah
briefPeach
briefPeach9mo ago
hi @nerdylive did the public port work for you in serverless machine? I'm trying to do the same thing!
Jason
Jason9mo ago
no its not supported yet right @flash-singh ? Another alternative is to start ws conection to your own server
briefPeach
briefPeach9mo ago
oh really? why is that? I thought serverless pods and pods are the same secure cloud infra under the hood
Jason
Jason9mo ago
from serverless -> your server yeah like different platform
briefPeach
briefPeach9mo ago
umm i actually need to expose public ports from serverless machine specifically.. i really wonder why it cannot be achieved..
Jason
Jason9mo ago
well their limitations
briefPeach
briefPeach9mo ago
will ngrok work in this case? local tunneling?
Jason
Jason9mo ago
Hmm yes ngrok, cloudflared works but you have to pass it programatically, since it'll be ineffecient to run it manually from a terminal right
briefPeach
briefPeach9mo ago
ok great! i'll try ngrok or cloudflare tunneling!
Jason
Jason9mo ago
btw what are you trying to do with that?
briefPeach
briefPeach9mo ago
cuz i need to run a ui on serverless machine, (comfyui)
Jason
Jason9mo ago
oh
briefPeach
briefPeach9mo ago
pod starts up too slow
Jason
Jason9mo ago
damn hahahah you're gonna make a proxy for comfyui ?
briefPeach
briefPeach9mo ago
serverless is way faster cold start time than pods
Jason
Jason9mo ago
true
briefPeach
briefPeach9mo ago
i think they cache the container image for serverless i don't understand why they don't support port expose in serverless, it will be so much more flexible
Jason
Jason9mo ago
hmm this feels abit like abusing the power of serverless lol yeah they say they usually allow them but because of its complexity for users they decided to remove it... from what i remember
briefPeach
briefPeach9mo ago
umm but low startup time is a technical need (a very decent need)...they should aim to optimize for that technically
Jason
Jason9mo ago
True Can be done in many ways too ( from their side ofc )
briefPeach
briefPeach9mo ago
yeah..i just hope runpod can be more flexible and powerful on their serverless offering
Jason
Jason9mo ago
Flexible like what I thought it's already powerful enough 😅
briefPeach
briefPeach9mo ago
the serverless is really amazing in cold starts. beats every other cloud service (EC2 is too slow..) its the major reason i choose runpod oh really, i guess i just want the ports hahah
Jason
Jason9mo ago
Oh rlly
briefPeach
briefPeach9mo ago
apart from local tunneling like ngrok, what other way to do this?
Jason
Jason9mo ago
Ec2 serverless? Hmm Idk, never thought about this.. Reverse tunneling is the way to bypass it
briefPeach
briefPeach9mo ago
oh i see, is it different from local tunneling?
Jason
Jason9mo ago
I'm not sure what's local tunneling
briefPeach
briefPeach9mo ago
ok no worries. cloudflare tunnel is local tunneling i think
flash-singh
flash-singh9mo ago
we do support but haven't exposed it as a feature, its something we can do if needed
Jason
Jason9mo ago
hmm how can i get it then?
flash-singh
flash-singh9mo ago
be nice and ping me about it and your reasoning for needing it it might be possible using our graphql api, you need to set ports on the template which we hide in GUI for serverless comfyui is likely a cool feature to expose to your users, spin up a job, that jobs lasts until your user quits comfy ui, and you can expose a port / ip per job
Jason
Jason9mo ago
@briefPeach @flash-singh Ohk then, I don't think I need it now since I have used stream and it works kinda well for now
Encyrption
Encyrption9mo ago
The trick is to have your serverless worker establish a web socket connection to server. Once that happens you have 2 way communication over the web socket. The serverless worker needs to act as a web socket client, it cannot be a web socket server.
briefPeach
briefPeach9mo ago
yes i finally made it work with cloudflare tunneling, to spin up a serverless job and start tunneling until the job is canceled. similar to this code: https://colab.research.google.com/github/ltdrdata/ComfyUI-Manager/blob/main/notebooks/comfyui_colab_with_manager.ipynb#scrollTo=jjjjjjjjjjjjjj people in comfyui community is gonna love this sooo much.
Jason
Jason9mo ago
Not like per request? I thought comfyui is like stateless Is it not?
Encyrption
Encyrption9mo ago
What are you building? Hope you don't mind me asking.
Jason
Jason9mo ago
Yeah that's what I said to b4
briefPeach
briefPeach9mo ago
starting comfy in seconds is so coool 💖 thanks for the great infra at runpod @flash-singh i love the runpod serverless so much, the cold start time is amazing
Jason
Jason9mo ago
I was talking about comfyui
briefPeach
briefPeach9mo ago
comfyui is sort of stateful because frontend is not fully decoupled from backend code, comfyui relies on the server's file system to serve some static js files (it is really a flaw of design 😕 making serverless comfyui so hard to achieve)
Encyrption
Encyrption9mo ago
a remote comfyui would be cool.. seperate the front end from the back end.
Jason
Jason9mo ago
Oh yeah sessions
briefPeach
briefPeach9mo ago
Hi im building a comfyui serverless machines platform to let you config your own "comfyui machines" (custom nodes versions, comfyui versions, etc.) and you can share your customized "comfyui machine" with others as serverless app, so people can reuse your comfyui environment
Jason
Jason9mo ago
But it can be stateless right By keeping the session ids
briefPeach
briefPeach9mo ago
yes this is exactly what I'm building. serverless comfyui! let me know if you are interested to test out or colaborate! @Encyrption
Encyrption
Encyrption9mo ago
Yes, for sure! this could work with the marketplace I am building.
briefPeach
briefPeach9mo ago
the problem is to keep the model files, input files etc... oh wow what kind of marketplace are you building? would love to hear about it!
Jason
Jason9mo ago
Woo what marketplace? Yeah could use some front-end storage and backend storage for that You can add custom logic for that if you wanna store them in serverless btw
briefPeach
briefPeach9mo ago
@flash-singh thanks for the advice. I tried to use graphql for it, but when mutate saveTemplate, if isServerless: true is in the payload then it will ignore the ports: '8080/http' that i set, and still set ports to null so i can only bypass isServerless: true i wonder is this flag required for serverless template? can i just not set it?
Jason
Jason9mo ago
Last time I tried on graphql it didn't work well, I guess you need Runpod's assistance to activate that on an endpoint
flash-singh
flash-singh9mo ago
we will need to update this to allow users this behavious even for serverless, so far one of the reasons we haven't made serverless too flexible is to provide a better experience, slowly we will introduce more and more features
briefPeach
briefPeach9mo ago
Omg this sounds great! I’m super happy to hear you’re investing more in serverless.❤️❤️ I think this is the future of ai ml service! The magic fast boot time 🚀
Brandon
Brandon7mo ago
Does anyone know if this is possible now?
Encyrption
Encyrption7mo ago
You can use the RunPod proxy URL if you add /ws onto the end of it and it will work with websockets with HTTP port. There is a bit of a delay before it kicks in though.
Brandon
Brandon7mo ago
Does that also work for serverless, thought it only worked for Pods?
Encyrption
Encyrption7mo ago
Yes, it works for serverless It is a bit tricky to pull off the timing though. Your handler has to get your POD_ID from the ENV variable (RUNPOD_POD_ID) and send it out via an update like this:
runpod.serverless.progress_update(job, {'POD_ID': POD_ID})
runpod.serverless.progress_update(job, {'POD_ID': POD_ID})
You can get this info by polling the STATUS endpoint. With the POD_ID you can create the proxy URL with /ws appended. This works with the RUN endpoint. I am not sure if it is possible when using the RUNSYNC endpoint. WS_URL = f'wss://{POD_ID}-{WS_PORT}.proxy.runpod.net/ws'
Brandon
Brandon7mo ago
Ahh that makes sense, thank you so much
kingtut AI
kingtut AI6mo ago
There are a lot of video on doing this that I've used on Youtube.
Syed Muhammad Asad
GitHub
GitHub - asdkazmi/runpod-serverless-socketio
Contribute to asdkazmi/runpod-serverless-socketio development by creating an account on GitHub.
yhlong00000
yhlong000002mo ago
GitHub
GitHub - runpod-workers/worker-websocket
Contribute to runpod-workers/worker-websocket development by creating an account on GitHub.
lxe
lxe2mo ago
is everyone in this thread building a "persistent comfyui but you pay only when workflow runs?"
Slothify⚡Daily Gmove
how do you do this? would like to know. i want to run comfyui on the cloud because my local machine can't handle it beyond some casual light use
Vlad
Vlad4w ago
Related to this -- I just came across this blog post: https://blog.runpod.io/introduction-to-websocket-streaming-with-runpod-serverless/ which is excitingly-titled, but the content has nothing to do with websockets 🙂
RunPod Blog
Introduction to Websocket Streaming with RunPod Serverless
In this followup to our 'Hello World' tutorial, we'll create a serverless endpoint that processes base64-encoded files and streams back the results. This will demonstrate how you can work with file input/output over our serverless environment by encoding the file as data within a JSON payload. Prerequisites * Complete our

Did you find this page helpful?