I have one server where cPanel WHM displays the load averages at the top right, but another server where it doesn't. How do I enable this? And what do the three different numbers shown (e.g. "Load Averages: 0.99 1.17 1.43") stand for?
way to get MySQL stats such as queries per day and things like that?
I understand how the load averages are calculated and what the numbers represent but what would be considered a high load average and at what point would the server need upgraded?
That would be impossible as the LOAD as given in linux/unix has nothing to do with the cpu load of a server but the total of processes queued (waiting for their turn). If their main node has such a high queue count it should be fairly easy to see for them which process is causing that. No idea why some hosts keep sending their customers to fairytale land every time when it comes to linux load information. Anyways; Exim with that many processes doesnt sound to good either.
As for one VPS bringing down the main node; this is really not that strange as you keep in concideration that a VPS is still a node with many shared customers. Where some things can be regulated, like cpu cycles and memory, but if your node goes completely balistic on its disk IO it will bring a VPS-main node to its knees.
I've never fully understood load averages, I've read articles about what it means but I still don't fully grasp what they are and their meaning.
For example, if I'm to understand correctly, if the amount of processors is greater than the load, then the server is not behind on processing. For example, if a server has a load average of 3.4 and is quad core, this would mean it's not 'under strain', correct?
I have a dedicated server, 2 3GHZ CPUs and the following load averages:
Quote:
Load Averages3.25 3.61 3.69
Does this equate to my server being strained, or handling fine?
the Load averages in Linux I tend to find confusing, I've read that its the system load averaged over 1, 5 and 15 minutes but beyond that I never seem to find a clear cut answer as to what is considered a normal system load.
For instance I have a p4 D 3Ghz with 2gb ram and a single 250gb SATA HDD
the 15min average sits around 0.85 during the busiest time Is that good bad or indifferent?
The server hosts my sites and that of a few friends. [url]shows what MRTG is outputting about the load averages.
It also says the CPU is 520% although I believe it multiplies it by 10? (i changed the title to say that) so its around 52%
been checking out this site for a while and finally decided to register because I have a problem. Also hope this is the correct forum for this topic, sorry if it isn't.
So I have a problem with Apache. One of the sites that I run/host has a moderately large vBulletin board, and Apache just seems to eat up the CPU. Load averages have shot up between 20-30 and I've seen it as high as 80. Apache and MySQL are optimized already, I'm using suPHP for security because there are other sites on this box.
The funny thing about this is that it only started happening about a week ago. After checking for rootkits and all that garbage, I reinstalled the OS just to be on the safe side. Everything comes back clean still. I also got fed up and hired Platinum Server Management for a month, to see if they could find a solution (and I've been interested in reselling their services, but that's not relevant). So far the only thing they can come up with is disable suPHP, which isn't an option. I do realize that suPHP is ~20-25 times slower than mod_php, but what totally baffles me is that it worked beforehand and started going all crazy like this. I did try running the site using an dso configuration, the load did drop, but nothing to be proud of.
This site, and the server overall hasn't had any increase in load, I've held off putting new accounts on it until I get this fixed.
In the meantime, I have said forums running on lighttpd, which lowered the load. (Also writing a tutorial on having lighty work with cPanel)
I have a webserver that our main two company websites reside on and this box also hosts ~50 other misc. websites. None of the websites on the server are very CPU intensive and bandwidth usage on the box is next to none.
Every morning around 4-6am the server takes forever to load anything. Seeing as this is when I do most of my work, I quickly become frustrated when trying to load our admin area, webmail, etc. Note that things do load, just very slowly...
I've contacted my tech a couple of times but he says this is normal as it's backups and stats running. I guess my question is, IS this normal? If so, is there a way to lessen the load at all?
I look after a number of sites and monitor their stats. We use both webalizer and AWSTATS so we can have a comparison. Up until recently the stats for both of them where relatively the same, webalizer usually showing higher numbers as it doesn't filter bots but the progression of increase and decrease in stats was proportional. However, over the last few months a curious trend is appearing, the stats in AWSTATS are decreasing every month whereas the stats in Webalizer are increasing, the gap between them now is huge.
I am just a beginner and it is about a month that the mysql loads on my server is so high , I have checked the mysql processlist via cpanel and there is no account that takes high mysql and make it high , and even I have checked for any possible rotten databases. but still the server load is so high and I can not control it , even when I restart it just after a minute it goes up again.
just the spamd command takes a high CPU usage sometimes. what it should be from? the spamd for a special account!
I need to optimize mysql , and need help. please do not tell me to hire an expert , just help me thanks
and another question is that how can I check which account is sending spam and stop it?
I have a dual xeon server that host around 15 small website, but these days I see that the load generated from mysql is very high as you can see below is consumes 32.6 CPU
There is a problem with mysql and server load. Some how mysql usage is going to over 100, but server load is not increasing, but it is crashing more often.
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 5931 mysql 16 0 111m 20m 3888 S 71 2.1 10488:45 /usr/sbin/mysqld --basedir=/ --datadir=/var
What could cause the problem? I did mysqladmin pr and only 2 database connections are existing.
it was very fast until mysql upgraded to 5.0.45 it was 4.. i can’t even turn my forum if it is a busy time cos it is so slow i get page not found after a while but when it is quiet it is not too bad... but it was alot more faster with mysql 4 i don't really want to downgrade please give me some ideas to fix this issue ...........
MySQL the last few days seems to be constantly the most demanding process in top, which it never was before. As far as I can tell, nothing has substantially changed with regards to traffic to MySQL driven sites on the server. Is there anything that might be wrong with the databases, etc., that might throw MySQL into a tizzy?
I have a cheap managed server and support is of little help BUT they will do what I tell them, where should I tell them to start to fix high mysql load?
The server currently has a large IPB forum, a large 4images gallery, and a couple popular Wordpress blogs, and this combo keeps taking mysql down.
What can I do server side to help this? And what shouldn't I do?
My server very slow and the is very loaded. How can I fix this problem? My server specialities and loaded softwares which are below. And sometimes there is "mysql connect failed", "Lost connection to MySQL server during query" errors. Thanks for kinds...
-- MYSQL PERFORMANCE TUNING PRIMER -- - By: Matthew Montgomery -
MySQL Version 5.0.41-log i686
Uptime = 3 days 4 hrs 34 min 17 sec Avg. qps = 92 Total Questions = 25406284 Threads Connected = 1
Server has been running for over 48hrs. It should be safe to follow these recommendations
To find out more information on how each of these runtime variables effects performance visit: .............
SLOW QUERIES Current long_query_time = 5 sec. You have 115 out of 25406302 that take longer than 5 sec. to complete The slow query log is enabled. Your long_query_time seems to be fine
WORKER THREADS Current thread_cache_size = 8 Current threads_cached = 7 Current threads_per_sec = 0 Historic threads_per_sec = 0 Your thread_cache_size is fine
MAX CONNECTIONS Current max_connections = 500 Current threads_connected = 2 Historic max_used_connections = 110 The number of used connections is 22% of the configured maximum. Your max_connections variable seems to be fine.
MEMORY USAGE Max Memory Ever Allocated : 835 M Configured Max Per-thread Buffers : 3 G Configured Max Global Buffers : 46 M Configured Max Memory Limit : 3 G Total System Memory : 3.96 G
Max memory limit exceeds 85% of total system memory
KEY BUFFER Current MyISAM index space = 35 M Current key_buffer_size = 32 M Key cache miss rate is 1 : 73658 Key buffer fill ratio = 34.00 % Your key_buffer_size seems to be too high. Perhaps you can use these resources elsewhere
QUERY CACHE Query cache is enabled Current query_cache_size = 4 M Current query_cache_used = 1 M Current query_cach_limit = 1 M Current Query cache fill ratio = 29.83 % MySQL won't cache query results that are larger than query_cache_limit in size
SORT OPERATIONS Current sort_buffer_size = 2 M Current record/read_rnd_buffer_size = 1 M Sort buffer seems to be fine
JOINS Current join_buffer_size = 1.00 M You have had 14127 queries where a join could not use an index properly You should enable "log-queries-not-using-indexes" Then look for non indexed joins in the slow query log. If you are unable to optimize your queries you may want to increase your join_buffer_size to accommodate larger joins in one pass.
Note! This script will still suggest raising the join_buffer_size when ANY joins not using indexes are found.
OPEN FILES LIMIT Current open_files_limit = 2500 files The open_files_limit should typically be set to at least 2x-3x that of table_cache if you have heavy MyISAM usage. Your open_files_limit value seems to be fine
TABLE CACHE Current table_cache value = 256 tables You have a total of 636 tables You have 256 open tables. Current table_cache hit rate is 1% , while 100% of your table cache is in use You should probably increase your table_cache
TEMP TABLES Current max_heap_table_size = 16 M Current tmp_table_size = 32 M Of 2271787 temp tables, 3% were created on disk Effective in-memory tmp_table_size is limited to max_heap_table_size. Created disk tmp tables ratio seems fine
TABLE SCANS Current read_buffer_size = 1 M Current table scan ratio = 28 : 1 read_buffer_size seems to be fine
TABLE LOCKING Current Lock Wait ratio =1 : 112 You may benefit from selective use of InnoDB. If you have long running SELECT's against MyISAM tables and perform frequent updates consider setting 'low_priority_updates=1' If you have a high concurrentcy of inserts on Dynamic row-length tables consider setting 'concurrent_insert=2'.
Just wondering if mysql load balancing is possible in a shared environment.
Example:
I have 3 shared accounts. On one server, I have the write and the 2nd and 3rd server I have the read (select). Or vice versa .. Would it be possible to create a php script to perform this function?
Quad Core server, 4GB ram. MySQL runs at all times between 200 - 300% CPU. Server does only 5K unique per day, and runs zen cart.
I am at a loss, I have experiece with tracking dowen reasons for this but this one has stumpped me. So I was hoping to get new eyes on this and see if anyone had any ideas.
# The following directives should be commented out # but included as they are things that get added # very frequently on tickets. These are more in a # need-this-feature basis.
# The below 2 cannot be set on the fly. If the customer already has # InnoDB tables and wants to change the size of the InnoDB tablespace # and InnoDB logs, then: # 1. Run a full backup with mysqldump # 2. Stop MySQL # 3. Move current ibdata and ib_logfiles out of /var/lib/mysql # 4. Uncomment the below innodb_data_file_path and innodb_log_file_size # 5. Start MySQL (it will recreate new InnoDB files) # 6. Restore data from backup #innodb_data_file_path = ibdata1:2000M;ibdata2:10M:autoextend #innodb_log_file_size = 100M
So the site got featured on [url]and now the server is drowning...
The Coppermine Gallery usually hovers around 30~50 users daily and now, 1800, and im at a lost as how I should configure mysql to take on such a load. right now it takes about 10 secs or more to load a page and sometimes it would time out. Because it si coppermine, all pages are dynamic and can't be cached -_-"
Here's the my.cnf right now after i played around with the numbers
server spec Opteron 170 (2ghz) 2gb ram 250 7200rpm
# # KeepAlive: Whether or not to allow persistent connections (more than # one request per connection). Set to "Off" to deactivate. # KeepAlive Off
# # MaxKeepAliveRequests: The maximum number of requests to allow # during a persistent connection. Set to 0 to allow an unlimited amount. # We recommend you leave this number high, for maximum performance. # MaxKeepAliveRequests 100
# # KeepAliveTimeout: Number of seconds to wait for the next request from the # same client on the same connection. # KeepAliveTimeout 15
# prefork MPM # StartServers: number of server processes to start # MinSpareServers: minimum number of server processes which are kept spare # MaxSpareServers: maximum number of server processes which are kept spare # ServerLimit: maximum value for MaxClients for the lifetime of the server # MaxClients: maximum number of server processes allowed to start # MaxRequestsPerChild: maximum number of requests a server process serves <IfModule prefork.c> StartServers 8 MinSpareServers 5 MaxSpareServers 20 ServerLimit 200 MaxClients 200 MaxRequestsPerChild 1500 </IfModule>
# worker MPM # StartServers: initial number of server processes to start # MaxClients: maximum number of simultaneous client connections # MinSpareThreads: minimum number of worker threads which are kept spare # MaxSpareThreads: maximum number of worker threads which are kept spare # ThreadsPerChild: constant number of worker threads in each server process # MaxRequestsPerChild: maximum number of requests a server process serves <IfModule worker.c> StartServers 2 MaxClients 150 MinSpareThreads 25 MaxSpareThreads 75 ThreadsPerChild 25 MaxRequestsPerChild 0 </IfModule>
if it was possible to load vhosts from a MySQL database and the only solutions i found didnt work for me. Im running Apache on Windows. Has anyone seen anything that would work with Windows? The only alternative i can think of is to read the database info using PHP and output it to a vhosts config file. I dont want to do that though because it means restarting the server to load the new config and i hear that a lot of vhosts slows the server down
I'm hosting about 6 private websites on my server, and some have problems.
My biggest website is a vb forum with 30.000 members and 1.5 million posts.
When i browse with firefox i see that vb forum loads fast, but phpads which is integrated on the top and footer loads very slow. When i browse with IE, it takes like 10 seconds before the whole site is loaded and sometimes IE even crashes.
When i try to login on another website / admin panel (oscommerce), i get a mysql gone away error, that means timeout. But why does it show that? I ran mysqlreport and can't find anything why it should display mysql gone away. # ./mysqlreport MySQL 5.0.15-standard uptime 3 19:36:47 Mon Feb 5 11:08:10 2007
__ Key _________________________________________________________________ Buffer usage 78.94M of 96.00M %Used: 82.23 Write ratio 0.60 Read ratio 0.00
How can i check (using SSH) which databases/users cause server load to mysql ?
I've tried "mysqladmin proc stat" but it shows just the current. How can i get stats of the last 24 hours for example ?
I've also seen slow connections stats. What is the command to check more detailed report of the slow connections; which databases caused it etc', in the last 24 hours for example as well.
Anyone know a good script that I can run with cron or something? Mysql seems to be the #1 problem with a lot of my web sites, a restart usually fixes it right away for me, but I can't keep restarting my servers everyday manually.
I currently have one server a Dual Xeon 5130 2GHZ (woodcrest) 2GB Ram. Running cPanel/WHM
Now I run a website that is VERY PHP & MySQL Intensive and MySQL is ALWAYS the top of the process list, hogging a ton of usage. It's getting to the point where the site is needed a second server and I know there's a few options; but I'm not sure which one would be the best.
They way I see it my two options are getting a 2nd server and setting up the two to do load balancing, or getting a 2nd server and setting one up for just Apache and the Other for MySQL and using the 2nd as a remote SQL server.
If I do the Apache on a seperate server would I need such a powerful server? And if I also would want to upgrade this server along with getting a 2nd server would I be better of upgrading to 4GB of RAM or upgrading the processors?