Backup Type :: Daily, Nightly, Incremental?
May 13, 2008What type of backup do you use with your host? (daily, nightly, incremental, etc)
View 10 RepliesWhat type of backup do you use with your host? (daily, nightly, incremental, etc)
View 10 RepliesI realize that this is probably wishful thinking, but if someone could point me to the closest existing technology to do what I need to do I would really appreciate it. I run two forums from an SDX account at JaguarPC. For those of you that don't know, an SDX account is basically a large shared hosting account with more available resources and less accounts perserver. Anyway, they've been great, but I need to set up some sort of backups and I'd like to do them to my home pc. What would be ideal....and I doubt it exists....would be something like robocopy on windows which syncs up changed/added files on a regular basis. The problem is, this is a Linux account and my home pc is an XP machine. Is anything even remotely similar to this possible?
Just as an aside, I'm already creating nightly DB Dump backups but they are stored in the account? This won't help much in the case of a catastrophic failure. The backup, if possible, would backup these DB backups as well since they are in the account.
I current do some rsync backups with a command like so every day
rsync -az -e ssh --stats --delete --exclude "stuff" / user@server:/home/user/
What I want to do is have some incremental backups in there in subdirectories. So, for example, something like this on the remote server
/home/user/something.tuesday
/home/user/something.friday
I thought the --backup --backup-dir Switches were used to store just the files that had changed in seperate directories, am I wrong on that?
I've read everything I could find, including the big rsnapshot scripts, but I'm not able to do what I want, it seems so simple but something's not right, am I wrong that subdirs should have just files that are new or have changed. I tried various things like this, but had no luck
rsync -az -e ssh --stats --delete --backup --backup-dir=/home.Thursday --exclude "stuff" / user@server:/home/user/
Does anyone have experience in backing up mysql incrementally? From what I've read in the docs it's possible using the binary logs but I haven't been able to find a good resource on how to make this work.
I have a database that is over 5GB. There are a few Myisam tables that are insert/select only and one innodb table that receives updates/deletes/selects/inserts.
Ideally I wouldn't have to backup the 5GB every night, I'd prefer to only get the items that have changed. If I could make this work, then I could also get backups more often rather than once a night.
I was wondering if I would be better to use incremental backup in cpanel instead of disabling the incremental feature and having tar.gz backup.
Im doing ssh overnigth on an offsite server.
Gzip is -rsync tagged when activated, as per this ccron command:
Quote:
0 1 * * * export GZIP="--rsyncable" ; /scripts/cpbackup
so they should be rsync friendly?
Whats your tougth on this?
do incremental backup on my local hard disk but i would use backup compression, is there any way to tar each account and use incremental system? i have read something on GZIP --rsyncable
View 1 Replies View RelatedI want to back-up accounts managed by cPanel/WebHostManager to a remote server. I want to be able to restore to any of the last 7 days. I believe doing an incremental back-up each night can accomplish this best, saving both bandwidth and disk space on the remote server.
Using the back-up functionality built into WHM is appealing because WHM knows what files need to be backed up to do a full restore. However, WHM can't do exactly what I'm looking for either. I've been looking into using rsync as described in this document:
[url]
It looks becaue of the "hard link" capabilities of linux, rsync can be used to create what functions like a full back-up every night but requires a full backup the first day and then each day after that only the incremental changes have to be transfered. However, because I don't know what directories need to be backed up, I'm not sure how to implement an rsyc solution either. (BTW, I want to back-up everything necessary to be able to restore each account except the log files. So, that means MySQL, mail, config info, and whatever the user has uploaded)
i want to run /scripts/cpbackup every day begain with 12:00 AM and i put this line in /tmp/crontab.XXXXwuxGUI File
0 0 * * * /scripts/cpbackup
but the backup didn't work and do the job
I have my WHM/cPanel installation configured with daily and weekly backups. I checked at what time of the day the server load was at the minimum and configured the cPanel backup cron to run then.
The problem now is: Backing up a few hundred accounts results in a high server load. My server configuration:
Dual Processor Quad Core Xeon 5335 2.0GHz with 4GB RAM and 2 x 250GB SATA HDD hosted at SoftLayer.
The accounts are located on the first HDD and the backup archives are placed on the second HDD.
What can I do about this? I'd like to take daily backups of all accounts but not if my server load increases up to 10... That kind of renders the cPanel backup feature useless if it doesn't even work on a powerful server like this one...
Would it help if I use an application such as Auto Nice Daemon to give the backup process a lower priority? But then again that won't work on the MySQL dumps? And I think it's not a CPU problem but an I/O wait problem? Other processes have to wait for disk access because the disk-intensive backup process is running?
I have seen resellerzoom provides daily full backup to their customer. How should i configure my WHM so that it create daily backup and delete old backup.
View 4 Replies View RelatedI want to know how can backup my linux server data's to windows server? due high number of files and daily updates i cant use FTP
View 3 Replies View Relatedto decrease the load in server when daily backup start ,, the load in server before backup start from 0.80 to 1.20 after daily backup started i see very high load from 16.00 to 32 and 40
any solve for decrease load when backup start from 3 to 7 alot
i just wana know is it safe to do remote daily backup for about 70,000 files?
file sizes is about 200kb and every day i have about 1000 new file, so rsync first should check old files becouse i am deleting about 30-50 of them daily and them backup new 1000 files ,
so how much it will take every time to compare that 70,000 files?
i have 2 option now:
1-using second hdd and raid 1
2-using rsync and backuping to my second server , so i can save about $70 each month.
Is there any way to create a cron to backup mysql data daily ( or weekly )? I mean an "auto script" to run this command daily
mysqldump -u usernam -p password dataname > file.sql
I'm new to Parallels Panel. I use version 11.0.9. I want to backup mysql database daily. First of all, what is the best way for daily database backup in plesk. I'm trying to do this in Scheduled Tasks and I use mysqldump command although I'm not sure.
I chose the time and day first and then I switched on the task. I typed the following command to Command line.
mysqldump -uusername -ppassword databasename | gzip -9 >/httpdocs/mydatabase.sql.gz
or
mysqldump -uusername -ppassword databasename >/httpdocs/mydatabase.sql.gz
This created only a blank file. When I use this without gzip, nothing changes.
1- Is mysqldump right command for database backup?
2- Should I define full path for mysqldump, gzip and database? If so, how can I find out the full path of mysqldump, gzip and my database? Because I can't see their locations in panel.
3- I can't see any error message. There is not any log file in httpdocs folder. Where does the log file exist?
4- It is weird but should the username be "database user" or should I write "root" ?
When i try to install BotNET 1.0 on my dedicated, i got this error :
root@leet [~/botnet/BotNET-1.0]# . install.sh
Compiling source code . . .
In file included from src/main.c:9:
src/../include/bot.h:43: error: array type has incomplete element type
src/../include/bot.h:57: error: array type has incomplete element type
src/../include/bot.h:89: error: array type has incomplete element type
src/main.c: In function:
src/main.c:146: error: type of formal parameter 1 is incomplete
Here is my install.sh file:
Code:
#!/bin/bash
# BotNET installation script.
# If this script causes problems, try "make all" instead.
# Usage: . install.sh
cc="/usr/bin/gcc"
echo "Compiling source code . . ."
bot=`$cc src/main.c src/launch.c src/memo.c src/seen.c src/parse.c src/help.c src/log.c src/info.c -o bin/bot -pthread || (echo 1)`
botnet=`$cc src/botserv.c -o bin/botserv || (echo 1)`
if [ "$bot" != "1" ]; then
echo "Installation complete."
echo "Executables will be found in bin/"
else
echo "Errors encountered during compilation!"
fi
My OS is centOs 5.x
Kernel : Linux 2.6.18-53.el5 #1 SMP Mon Nov 12 02:22:48 EST 2007 i686 i686 i386 GNU/Linux
* I have tried all other way to install (make all) and other *
I am starting a software as a service company and have a dedicated MySQL server running on Linux. It has RAID mirroring but I need to take nightly backups as well and store them either remotely or on a locally connected drive.
There will be 100, and hopefully eventually 1000’s of databases.
I am thinking of using the built in MySQL tools backup script in the GUI tools, which seems to work well. But I have also read about rsync and other disk mirroring backups. What do you guys recommend?
Also, do you guys think a simple external USB connected drive 500GB to the MySQl server is good idea, or is it worth the money to look at TAPE or remote backup on a geographically different location? What exactly is the benefit of TAPE anyway? Seems to me like TAPE is so freaking old, and expensive, I don’t understand why anybody would use it over a simple USB external drive.
I have two folders:
/misc
/misc-backup
how do I go about making /misc-backup a mirror of /misc correctly?
I was wondering which of the backup method has the fastest restoration time - is it incremental uncompressed backup or compressed backup?
I am making 2 daily backup which is both uncompressed and compressed. However in case my server broke down, I want to choose which one is the fastest?
We have incremental backups running in our server. We want to move the cPanel accounts to a new cPanel server and restore in that server. The Backup Restore option in WHM will recognize only compressed backups (like .tar.gz) when it is placed in the /home directory. So how do I do about restoring a incremental backup through the WHM of the new server?
View 4 Replies View Related