I've seen this mentioned before, that you can run SQL on a separate server and lot of people do it for SQL heavy applications.
Only thing I wonder, how well does it work bandwidth wise? Is there a noticeable slowdown in query speed? What about bandwidth cost, does it use a lot, enough to run over a typical dedicated server monthly limit?
Also what about security, are queries/username/password sent out in clear text?
I have a Ubuntu Lamp server setup in my building which is used for development purposes.
So I also have phpmyadmin setup and am using a Mysql Query browser and Mysql Administrator desktop programs to administor the database on the server. However I am not able to connect to the server from my desktop. I am able to login to my myphpadmin but not able to login via the Mysql Query Browser program.
Hopwing someone can help. I have read about port forwarding and also read a post about editing the my.cnf file on the server and changing the Bind-Address= 127.0.0.1 to the actual hostname but before I start editing the actual configuration files on the server I want to see if their is another way as I am worried that if I change the me.conf file that the phpmyadmin may stop working?
I see option to setup remote Mysql server in WHTM but wonder can i connect 2 servers through private network then fill in Remote server address something like 192.168.1.2. If it is possible it would save me a lot of bandwidth for both servers.
I'm facing some odd issues with 2 servers located in the same datacenter.
One server is the HTTP server, and every 5 minutes it works, then for 5 minutes Server One and Server Two (MySQL) hate each other and refuse to talk to each other. Not only can neither not connect to each others MySQL, but also neither can ping each other.
The odd thing is, Server Three can however connect to the Server Two MySQL with no problems during this period. It seems to be just the communication between Server One and Server Two that keeps getting halted.
I've tried "service iptables stop" to disable the firewall, and changing SELinux to disabled, but neither have fixed it.
i am getting mysql down issues with my current dedicated server. I would like to move my database to a dedicated server where only mysql will be running. Just wondering if this will be faster than the localhost? I have planned to buy dedicated server from another hosting company so definitely there will be a good enough connectivity distance.
i have two server one for running apache and second is sql server.Now come to problem.
I have a very strange type of ddos attacks on my httpd server and due to these attacks my apche server failed many times (10 to 15 times)in a day.
When i checked max http connection use by a single ip then i am surprised to see that my remote sql server ip is using maximum of http connections and behind this there is some other ips whose doing the same.
I installed media layer it seems to work fine for me but suddenly he blocked my remote sql server ip due to using maximum http connection lolz its very strange any body please help me out of this problem.
my site is working fine for somedays and suddenly its stoped and hanged and also my PPP is not working , and file manager in PPP its give me this error VZAgent returns error #422: "Cannot initialize connection to remote host" and the suppord admin said I have problem in failcnt
I have a few users that have a hosting account with me and are trying to connect to their own MySQL servers elsewhere. Both said that they are able to connect to the MySQL server with other hosting accounts but not the ones hosted with me. I asked if they had setup permissions to allow remote connections from my server and they confirmed that they did.
What am I missing? Do I need to configure my firewall somehow? I figured the connection is just using port 80 since it's from an PHP script right?
We have got 200 customers, and we will to change to cpanel.
The MYSQL server is on a solaris zone. It is possible to make the remote connecting with it. I will import the 200 users MYSQL databases to cpanel, with privileges (user settings). The mail/web dir a will import manuali to cpanel server(copy-paste-setuser:group).
I have a 1GB MySQL database (compresses down to 300MB) and would like an automated method of backing it up to a remote server. Both accounts are shared hosting accounts (if it matters, both are running CPanel, no shell access on either).
I have create a remote mysql database in whm for another server on my network. Now my problem is that the remote mysql in fact works as the data is still being pulled from it with mysql disabled on the main server. I did however put in the root password during the remote mysql server setup. Now the only problem I seem to be having is the databases dont show up in any cpanel accounts and phpmyadmin doesnt work. but all sites using a database made before this change still work.
Why so many sleeping processes from server 2, and none or almost none from server 1 ? Is it normal that every query from server 2 is open on a different port ? Should they be grouped by virtual host, or is there one different port for each script running concurrently?
What could be a reason for server 2 having problems to open connections on server 1, and how to solve it?
Right now I'm about to deport more users from server 1 to new virtual hosts on server 2, hoping it will lower the load on the first machine and overall accelerate the response times. I just thought I'd post here first because this all seems very weird to me and i thought someone with an external look might tell me what I'm doing wrong. And also if network would be the problem maybe that's not the clever thing to do to make more users remote.
I have 2 identical Fedora8 boxes with Apache 2.2.8, MySQL 5.0.45 and PhP 5.2.4. I use one as the Web+DB server (say box A), and the other (box B) just as the Web server connecting to the DB server on box A. I use this 2 machine configuration to test a LAMP based-Web application. I have a client program on other machines that can emulate a massive web workload to these 2 servers; it can emulate hundreds to thousands of users using the application simultaneously.
Before running the test, the connection to the DB server from both boxes seem fine. Using the mysql client program on either box A or B trying to connect to the DB, the connection goes through instantly. In the /etc/my.cnf file, I have max_connections=4096 and max_user_connections=4096. Note that the web app uses one single db user to connect to the db. To allow remote connection, I inserted one record into the user table of the mysql db whose the host field's value is '%' (allow connection from all remote hosts).
After running the test (which I found out that many requests sent to box B failed), mysql client program on box A is still able to connect to the db instantly; but the one on box B has a problem: it takes extremely long (5 - 10 minutes or even more) to establish the connection, it doesn't time out, just takes that long. I believe that's the cause for the failure of requests to box B.