I am trying to make a backup of a very big postgresql(around 165 gb) for two days now. and at last I find out my mistake. never take a dump at same disk. it eats a lot of IO and kill all services which depends to dumped database.
first try I am dumping database to same disk and after around 6 hours web server started to giving timeout and lovely sitescope mails :) and I had to kill that process.
then I read a lot and started to dumping the database to another machine and it was smooth took around 4 hours to dump and no web server gave any timeout.
pg_dump -Fc dbname > db.backup
and I started restore like this
pg_restore -d dbname db.backup
before that I needed to recreate the db from psql. and this have not finished I must tell that pg_restore has -j parameter which gives more thread to read the dump file and you can give cpu number to -j which will work faster.
Wednesday, June 01, 2011
pg_dump and slowness
Subscribe to: Post Comments (Atom)
strangely I start liking diablo immortal in mobile, and now decided to install in windows :)
at the end of any product there will be a reporting interfaces for counts. let say you build a advertisement site which gives people to publ...
In order to use jhipster user system I needed to find out how to call jhipster, I am planning to do this with curl because it is easiest wa...
if you have wl11 ejb server and in that environment generated ejb client and deploy to wl12 it will give you this exception java.lang.NoSu...
Post a Comment