Basically mysql is behaving very very intermittently. Crashes were every 4 hours, I've brought them down to once every 8 or so hours but mysql keeps dying.
the error log will show the same routine each time.
on mysql start:
Quote:
091101 21:58:03 [Warning] option 'open_files_limit': unsigned value 120000 adjusted to 65535
091101 21:58:03 [Warning] Could not increase number of max_open_files to more than 65535 (request: 200110)
091101 21:58:03 [Note] /usr/sbin/mysqld: ready for connections.
Then we'll see errors due to crashed databases:
Quote:
091102 0:33:07 [ERROR] /usr/sbin/mysqld: Incorrect information in file: './<nameofdatabase.frm>'
following this a heap of:
Quote:
091102 0:36:35 [ERROR] /usr/sbin/mysqld: Can't open file: '>another database here.frm>
091102 0:36:36 [ERROR] /usr/sbin/mysqld: Sort aborted
091102 0:36:52 [ERROR] /usr/sbin/mysqld: Sort aborted
091102 0:43:00 [ERROR] Error in accept: Too many open files
I been having a heck of a time just with this one cpanel server and open files limit. At first using open_files_limit did not work so I changed it to open-files-limit that seemed to work but now it rejects the value and sets its down to 65535.
Then system open files limit is 500000. I try to set it to any value about 65535 in my.cnf and here is usual error:
090630 9:32:07 [Warning] option 'open_files_limit': unsigned value 120510 adjusted to 65535 090630 9:32:07 [Warning] option 'open_files_limit': unsigned value 120510 adjusted to 65535
When I run something like the tuning-primer it shows:
Current open_files_limit = 120510 files
The open_files_limit should typically be set to at least 2x-3x that of table_cache if you have heavy MyISAM usage.
Your open_files_limit value seems to be fine
But Im not sure if it is just reading my.cnf or something. I am still getting complaints from users about lost connections and I see the errors in error log. Ive looked everywhere and cant seem to find a solution to this.
We have three virtual hosts on our Apache 2.2 installation on Windows Server 2003. For some reason, I'm unable to open log files (error.log and each virtual hosts-specific log), even though I have full administrator rights. (The log folder is full access to admins.) Every time I try to open the file or even copy it to another location, it just says "Access Denied." I temporarily solved the issue for one of the logs by adding BufferedLogs On
How to increase the Open Files limits descriptor in Apache. In the earlier version of Cpanel, we had an option of Raise FD Size Limit to 16384, but the option no longer appear while rebuilding Apache. What is the way to do it and make the change permanent?
i have Made a VPs on my Own dedicated Server Which i use to run TorrentFlux for Personal Use. I am facing a few problems and dont know where to askf or help.
when i start more than about 12, i get errors in SSh (if i llogin) or th4e Apache Restarts killing all the Transfers.
I ahve 2 Gb Ram, Dual Core CPU.
the Error Via SSh is: sh: pipe error: Too many open files in system
and i ahve attached a Errors Log From Apache.
i am a Noob in Servers so i ahve Lxadmin Contorl Panel Installed and the Log is generated by it.
I am trying to increase my open file limit on my CentOS release 4.6 VPS.
I have tried modifying my /etc/security/limits.conf with the following: @mysql soft nofile 4096 @mysql hard nofile 4096 @mysql soft sigpending 4096 @mysql hard sigpending 4096
Even is a mannual set it with "ulimit -n 4096 mysql" it will not stick.
Any ideas what I can do to get this limit increased permanently?
I recently had a harddrive failure and luckliy I can still access certain directories on this failed drive. I can still access the /var/lib/mysql/ directory which holds all the users databases and have backed all these up separately using tar.
Now what I need to know is how do you restore these database files to another server? I tried simply untar'ing one of these to the new servers /var/lib/mysql/ direcotry and it stuffed Mysql up - it went offline. I had to get a cpanel tech to bring Mysql back online.
how can I get these database files to fully work on a new server?
I did the absolutely most stupid thing yesterday, I emptied the wrong mysql table via phpmyadmin. Fortunately my host managed to grab a the table in question from a few days ago.
So I have:
auctions.frm, auctions.MYD, and auctions.MYI
My assumption, is that I can just overwrite the current files of the same name in the mysql folder, and everyone will be happy. Is there some more to this process?
I have following problem-i have over 20 sites on server and each site has own database.Is there a way to speed up backup and transfer or other server?
Method which i use right now it's following:first i archive entire directory using command tar -pczf name.tar.gz public_html,and then repeat with each directory.But i think i could simply archive all requied directories,but that will took toomuch time,so if i drop connection during archiving it wont be archived at all then.So i think best solution will be to create some kind of batch command which can load in background,so that means command wont stop if client lost connection.
So let's say i have 2 sites and two directories located and different places.
One is at home/site1 and other at home/site2 So i think i would need to put command into batch file tar -pczf site.tar.gz. /home/site1 and tar -pczf site2.tar.gz /home/site2
Will that work? Also second part,mysql databases,i founded if i login into phpmyadmin as root i can see all databases.I managed to export all databases,but question will import again to phpmyadmin work.I think phpmyadmin create command for each database "if there is no db sitename_mysqlbase,create it",but howmuch i know phpmyadmin have limit with importing size of mysql database.Could that be done with import/export mysql command?
I couldn't keep my mouth shut (technically fingers). A customer wanted to upgrade servers and he needed a way to move the data across. Since I don't allow hard drives to be swapped, they have to do it manually all by themselves. I generally allow up-to 4 days for them to transfer data and make DNS changes, etc. But this time, I offered help! I agreed to move the data (darn me) and it just came out of me, involuntarily.
God knows what just happened... but in a positive way, customer is extremely happy!
So...
Both servers are on cPanel - with root access (duh)
200 odd files which total to 25 GB
1 database about 100 MB in size (no biggie)
I was planning on using one of my Windows 2003 servers (via remote desktop) to download the 25 GB and upload the 25 GB, but that sounds like a waste of resources and time.
Is there a way to exploit openssl? someone came to me asking me for a job on hostparlor.com and i said no. They then uploaded a remote view script to exploit perl and gain access. I patched that. Then he again said that he can exploit through openssl. I thaught open ssl was security itself? He then told me that he can gain access through php sites to root? Is that possible? We have phpsafemode set to off because we run whmcs and it requires it to be off. He said he can hack us through that aswell. Is this guy just bluffing or can this seriously be done? Like i said this is why i have 2 server admins...
I installed the MySQL binary packages in /usr/local/mysql/ after removing the MySQL RPM package. MySQL is functioning when I executed /usr/local/mysql/bin/safe_mysqld. I reinstalled MySQL before I installed PHP. When I used a PHP script to access a MySQL database, it outputs an error:
Code: Warning: mysqli::mysqli() [function.mysqli-mysqli]: (HY000/2002): can't connect to local mysql server through socket /var/lib/mysql/mysql.sock in index.php on line 2 However, I installed MySQL in /usr/local/mysql, not in /var/lib/mysql. How do I fix MySQL?
i would be interested in knowing like what would you prefer for your dedicated server? i.e Behind a firewall or on public network. Can anyone experienced provide pros and cons on this one?
I bought two Dell R200 because they was extremely cheap.
My target is Open Source SAN with active/passive setup.
Now i was wondering what RAID-level I should go for.
I will use 1TB SATA-II disks.
1) RAID-1 in both servers and mirror each other with DRBD. With this setup i have like double RAID-1 so i lose lots of disk space. 4 disks -> 1TB
2) No-raid at all and i will mirror data with DRBD. 4 disks -> 2TB
Then there is change to go for RAID-5. Theres enough space to put third disk in that case
3) RAID-5 in both servers. Mirrors data with DRBD. 6 disks -> 2TB but more performance.
But in every setup i lose space more then i would like to.
Actually i dont need space more then 1TB, but i would like to get best possible redundancy and most space available i could.
Also need to remember that those servers does not support hot-plug HDs so in case there is disk-failure i need to be able to shutdown one server and iSCSI should still be up and running.
Of course theres change to make it active/active.
I was thinking that if I go for the RAID-5 I will make LVM-VG on both nodes (san-vg1 and san-vg2) and use it 50/50.
In case of fail both VG's will be used from working node.
Im going to use Debian Lenny I guess..
I was looking for OpenFiler because of the GUI, but im familiar with Debian and have always done everything from cmd.
i have a linux machine with centos5 64 bit. I have a firewall and would like to know how to open and close ports? I have tried several different things and cant seem to open a port?
Is there a command line i can use in ssh to open a port? Or is there something i can addon to cpanel to control ports?
i have to get rid of an open relay state on my server, and i cant do it! i have antirelayd on my WHM but it seems its not working, any idea? please or at least wich line in /scripts/antirelayd i have to check to see if its everything ok.
i saw an older post, tried everything but nothing works x_X