Postgres Volume
I use n8n with postgres, doing some tests I downloaded the same file several times and it filled up postgres.
My question is what do I lose if I clean the volume, do I lose my credentials and workflows too?
I can no longer access it to make a backup.
44 Replies
Project ID:
N/A
N/A
All your data will be deleted and won't be possible to restore it
Is it possible to access /var/lib/postgresql/data to delete the downloaded files?
I would like to increase the volume, but it is not enabling it.
There has to be a way out, I can't lose everything.
You can use the file browser template, but I don't suggest navigating through DB data.
I will silenty tag @Brody here and see if it's possible to expand your volume size.
Project ID: 166d3751-42cb-4032-8849-ca8f64aac2f7
what is the current size of your volume?
I can resize the volume.
However, it fails when opening the n8n application.
please answer the question
It was 500 MB, I changed it to 1GB
please confirm that for me
okay great, please explain what "it fails" means
n8n application does not open
is n8n running?
no
that would be the problem 😆
yes
I can't log into the application
I already did redploy but I can't access it
what errors are you getting in the deployment logs
I'll get
esta caindo o serviço
English please
inattention
why have they crashed
not sure why you removed the image?
Solution
you need to redeploy the crashed services
yes, I removed the postgres image, I returned the image with :last.
now it works.
I'm a beginner.
thanks
I would still like to delete the files I downloaded and that filled up the memory.
How do I do this?
please dont swap image tags if you dont know what you are doing, it can very easily lead to data loss, railway cant recover data if the user was at fault, luckily it didnt cause issues this time
ok
images are not stored in postgres
I'm saying about this
files are not stored in postgres
being a hobby user you can grow your volume to 5GB
but that suddenly filled the postgres memory
ok. tanks
I did more testing by running the n8n workflow by downloading the PDF file from Google Drive and Postgres went from 534MB to 545MB. I would like to know how to do it for which file is saved in Postgre. Please
See that every time I download a PDF file from Google Drive to n8n my Postgres database increases in size. I would like to understand if this was supposed to happen. I have 2000 PDFs to process and this increase that occurred was only with 2 PDFs.
as brody said, files are not stored in postgres. use an object/cloud bucket like amazon’s S3
What would this increase in volume be with each execution of the workflow then?
are you storing any data about the pdf / accessing the database in any way? If not, it may just be coincidental
Sorry for so many questions, I'm a beginner.
What I did was install N8N as shown in the image and I didn't do any configuration, I just accepted the default one to be able to use N8N.
It's no coincidence that I uploaded 393 PDFs, see the increase in volume in the image.
Can you tell me what I lose if I clean the volume?
If I clean it and redistribute Postgres, will it work normally again?
Uploaded the pdfs to where? Where are they stored?
pinecone, but it has nothing to do with uploading, because the volume increases when downloading the pdf binary from google drive to n8n.
I’m sorry, but you’re really not making your problem clear. You’ve said that the volume increases while uploading, but now you’re saying it only happens when downloading?
What data does this postgres db store? How does your script interact with it? If everything is done correctly and your script just pulls the pdf into Google Drive without interacting with the db, the db’s volume size should not increase.