I have a tiny MySQL database (10MB) and I'm looking for some remote backup service. Someting that will backup the database and let me recover it if the host is in trouble.
I strongly prefer something simple and fully automated.
I have a 1GB MySQL database (compresses down to 300MB) and would like an automated method of backing it up to a remote server. Both accounts are shared hosting accounts (if it matters, both are running CPanel, no shell access on either).
I've just purchased a second server of which I wish to use a database server for a number of my databases and I'd like manage it remotely (from the web server) via phpmyadmin.
I've configured my firewall and granted access to the web server on root@webserverip and also username@webserverip (username being just one of the database accounts I wish to grant access in order to test this). Accessing mysql remotely from the webserver now works on both accounts (via ssh and php).
However, having configured the phpmyadmin config file to include the ip (network ip) of the 2nd server I am unable to see the 2nd server when logged in to phpmyadmin. I just have the usual root@localhost and username@localhost.
I have one server setup to power my forum and its database. It has been running from this server for a while now and as the site gets bigger i have decided i want to move the database to a seperate server.
I have another server at Ecatel and wanted to know if its possible to do this, as they are in seperate datacenters.. or is it only possible between two servers on a private network in the same DC?
I am creating a software as a service product and I am going back and forth on a database design issue.
Should I create a separate database for each new customer, or create one big database and store all customers and do the standard relational database design idea.
The clear choice for me is create a new database for each new customer, but the only problem is rolling out schema updates and bug fixes. If I have 5000 customers and then 5000 databases, how to manage creating a new table, editing a column name or data type, editing a stored procedure, or creating a new trigger on 5000 separate customer databases?
So, there are many benefits for making a new database for each customer that I can think of:
With separate databases, customer’s data are each separate entities, it is the whole encapsulation/modular design idea. It is impossible to get data out of a different customer databases. With one large database a query that forgets to include the WHERE clause selecting a customer_id can cause havoc.
With one large database all customers are located in one, actually maybe a few, files. So many of the selects, inserts, deletes will scan all customers data, when all they ever need or care about is their own data. Even with indexes, inserts and deletes become slower with one large database as the size of the database increases.
What happens if you have one large database with 100,000 customers and say a table with 30 million records? Then you need to spilt the database across multiple servers, a cluster, because one machine inst cutting it. With a single database design, cluttering becomes a headache. With multiple databases, no problem, just setup a new server and start adding new databases.
Also, last thing is database integrity. With one large database if you database gets corrupt or accidentally deleted, or modified in a way which causes unwanted results all your customers are screwed. With multiple databases problems only effect the one single customer, again each customer is a separate entity.
The only problem of course with multiple databases, and the root of my question; how do I manage database schema maintenance? How do I effectively roll out changes to each customer database schema?
PRODUCT, VERSION, MICROUPDATE, OPERATING SYSTEM, ARCHITECTURE Plesk 11.5, 11.5.30 Update #44, last updated at April 30, 2014 03:25 AM, Centos 6.5 (Final), Linux 64bit
PROBLEM DESCRIPTION When I try to add an extra MySQL server, running on a different server, I always get an error that the user has no admin access. But the user I created does have admin access to the database. The remote database is a MySQL 5.5 on a Linux Ubuntu 12.04 LTS server
STEPS TO REPRODUCE Tools & Settings -> Database Servers -> Add Databases server
ACTUAL RESULT Error: Test connection to the database server has failed because the supplied account does not possess administrative privileges: Access denied for user 'admin'@'%' to database 'd536cedf97037e'
EXPECTED RESULT A working database connection.
ANY ADDITIONAL INFORMATION These queries are logged by MySQL:
If I just ignore this message, go to a subscription and create a database, then the database gets created just fine. It's just Plesk that thinks that something went wrong.
For the last two months the forum has been going down every couple of days because the database service has been continuously stoped running on the server with no reason.
What can cause such a frequent database failures?
Can someone attack the forum/database/server in a way that will cause the above problem?
(there is a good reason I consider this possibility very seriously)
I have a linux dedicated server and use BMU (BackMeUp) for create cpanel full backup. But this cant create remote backup for me. BMU use "ncftp" for this. and i see following error about this problem:
Time: Fri May 16 03:38:17 2008
The following list of files have FAILED the md5sum comparision test. This means that the file has been changed in some way. This could be a result of an OS update or application upgrade. If the change is unexpected it should be investigated:
/usr/local/bin/ncftp: FAILED So, im install "ncftp" manually, but issue not resolved and i havent remote ftp backup too!
I have already developed shell scripts to automate account management, add/delete users, change quota, passwords etc. I am now working on web interface which will be used for managing the accounts, testing some free scripts and writing tutorials.
So far, my scripts can create chrooted (jailed) accounts with some basic shell commands.
The same disk space is also configured in samba and rsync daemon. Customers can use ssh/sftp or simply mount it as a network drive. Free utilities for Windows users (like deltacopy) work just fine with rsync daemon. Script can also modify iptables-based firewall and allow only certain IPs to access the server but I am still working on that part.
It takes about 10 seconds to create a new account and I have tested it on CentOS server.
I'm thinking about buying a shared hosting plan with lots of space. Anyone do backups to remote ftp server instead of rsync? I was wondering if it was possible to configure ftp in such a way to only transfer files that are NOT present on the remote ftp. This way, it doesn't overwrite existing files and thus save on bandwidth. Is it possible to transfer an entire folder with a ftp command, I know the mput command is for uploading multiple files to another location. Lastly, what are the advantages of rsync over ftp besides saving banddwidth and trannsferring only files that were changed?
I noticed that when "WHM Remote Backup" backs up my websites via ftp, the tar.fz archives are incomplete, meaning the backup is pretty much useless. I have tried with both PASV on and off and it still persists.
Any ideas why?
Also, what is the command to force the whm remote backup to run?
What exactly do you do when you take a remote backup through resync?
For instance, if i signed up with say bqbackup and started resyncing and taking backups, would i be copying all the data from my server to the remote server or is it possible only to copy the cpanel backups made daily/monthly/weekly to the remote space?
Also, when the resync runs everyday, does it replace the old files or does it create a new copy of everything that has changed?