I am trying to make a backup of a very big postgresql(around 165 gb) for two days now. and at last I find out my mistake. never take a dump at same disk. it eats a lot of IO and kill all services which depends to dumped database.
first try I am dumping database to same disk and after around 6 hours web server started to giving timeout and lovely sitescope mails :) and I had to kill that process.
then I read a lot and started to dumping the database to another machine and it was smooth took around 4 hours to dump and no web server gave any timeout.
example commands
pg_dump -Fc dbname > db.backup
and I started restore like this
pg_restore -d dbname db.backup
before that I needed to recreate the db from psql. and this have not finished I must tell that pg_restore has -j parameter which gives more thread to read the dump file and you can give cpu number to -j which will work faster.
Subscribe to:
Post Comments (Atom)
diablo immortal
strangely I start liking diablo immortal in mobile, and now decided to install in windows :)
-
at the end of any product there will be a reporting interfaces for counts. let say you build a advertisement site which gives people to publ...
-
In order to use jhipster user system I needed to find out how to call jhipster, I am planning to do this with curl because it is easiest wa...
-
if you have wl11 ejb server and in that environment generated ejb client and deploy to wl12 it will give you this exception java.lang.NoSu...
No comments:
Post a Comment