I have server in LT, the network is so good, but i got problem with the apache, it keep going down because the process too high ( i think)
when i list the process it show me :
nobody
0.000.250.0 Top Process%CPU 98.9/usr/local/apache/bin/httpd -DSSL Top Process%CPU 98.8/usr/local/apache/bin/httpd -DSSL Top Process%CPU 2.9
Other than that all domain only use not more than 10% process, I use Cpanel X and Cpanel XP skin, please help
Now my domain , will get blank page when try to access, i dont know why, even though all service is running and i already reboot the server
So my server is "unresponsive" for abour 18 hours, burst net didnt answer my tickets and I dont know what to do. Ive been with this setup for almost 5 months with no problems, No changes have been made to hardware or software.
CentOS 4.5 final Apache 2.0.59 php 4.3.11 MySQL 4.1.20
I have a user who has been importing large sql databases and this causes the cpu to run at 80% - 90% this in turn causes high loads and sometimes causing the server to crash.
Is there a way to automatically limit or control how much the process can use or stop the process to allow the server to recover.
Just trying to put some sort of safe guard in place
top output from my server shows a this user was creating a lot of proccesses, even this process has been running for about 30 mins, how can I limit the time a process can stay alive? I suspect it was causing a high load on the server when started.
I have a single mysql process that never stops and running with 10-200% CPU load: URL....
I restart mysql => process coming back I restart server => process coming back I kill process => process coming back
I have run: # mysqladmin -uadmin -p`cat /etc/psa/.psa.shadow` -i 1 processlist but there was noting that runs >10min.
Code: > SHOW FULL PROCESSLIST; +------+-------+-----------+------+---------+------+-------+-----------------------+ | Id | User | Host | db | Command | Time | State | Info | +------+-------+-----------+------+---------+------+-------+-----------------------+ | 328 | admin | localhost | psa | Sleep | 56 | | NULL | | 8110 | admin | localhost | NULL | Query | 0 | init | SHOW FULL PROCESSLIST | +------+-------+-----------+------+---------+------+-------+-----------------------+ 2 rows in set (0.00 sec) I have strace the PID
Code: # timeout 1m strace -f -c -p 5873 Process 5873 attached with 30 threads - interrupt to quit Process 10499 attached (waiting for parent) Process 10499 resumed (parent 5873 ready) Process 10502 attached (waiting for parent) Process 10502 resumed (parent 5873 ready) Process 10503 attached (waiting for parent)
[code]....
I found with google a hint for high cpu URL....I have deinstalled "health monitor" module, but that was not the reason.I use plesk 12.0.18 Update #38 with CentOS 6.6 (Final).
We have flash files site hosted on dedicated server in that Apache is utilizing the more CPU. When the server reached more than 1000 connections server get down automatically.
I have a fairly high end server in which I have installed SIM. SIM is restarting Apache up to 10 times a day, presumably due to high load causing un-availability.
On restart, Apache / MYSQL is stable until the load / mem usage begins to climb then it is restarted again. Here are my 'load' stats for today:
Load for today High (2:18am): 4.63 Low (3:30am): 1.20 Mean: 1.84 Latest: 1.61
Mem usage for today High (1:36am): 9,192.9 MB Low (1:48am): 7,995.7 MB Mean: 8,683.1 MB Latest: 8,781.7 MB
I have seen it using 20GB RAM before.
I have tried to follow various optimisation guides but these seem tailored to less powerful servers.
The web application I run on this server is almost entirely MYSQL based, with thousands of DB calls a day. Across the entire system I probably get 200,000 bot hits per day or even more. At peak times search engine bots are literally hammering the server.
I deal with a server that gets positivey slammed once a week for a few months per year. I'd tell you how many hits we got tonight, but I'm still waiting for AWSTATS to chew through the 2gb access_log file...
Tonight, I made some changes that SEEM to work, but I"mn not sure what the long-term effects could be. If we have any apache experts on the forums, I'd really like to bend your ear for a few to see what you know.
Obviously, with PHP, we're limited to prefork MPM.
First of all, I dropped Timeout from 300 to 120. That should be MORE than enough time to know that we've timed out. Then I dropped KeepAliveTimeout to 5 from 15.
Here's the radical one. Watching the process list and the load, it seemed that load spiked when the processes hit their end of useful life and respawned. Duh. This was happening every four seconds at the load we were under. MaxRequestsPerChild was set to 10,000. I upped this to 80,000 over a period of hours that we were under the load. I didn't see any significant memory leakage, but it's the change I'm worried about the most. I've seen Apache do some bad things when people allow this to go unlimited, and had always used the relatively low default as a guide.
Besides not loading a bunch of dynamic modules (also done, I usually do this so I'm not worried about it), what else can I do tuning-wise to keep load down? Please note that caching and load-balancing aren't acceptable solutions; I have one server to work with (for now) and the boss says no to caching because of how frequently our information updates. We also have extensive .htaccess files, so there's no LHTTPD in my future.
I am in a bind with Apache's multi process limit. Let me explain what I am doing. There's this website which has career details of all the football players since the beginning of professional football. They have a simple web form which allows you to look at a player's profile by entering his name or his 7 digit numeric id number (on that website).
One of my client wants a list of all the players with a certain "flag" in their profile. So I created an automatic form submission and HTML parsing script to get details of all the players with that "flag" in their profile. Let me not go into too much details and tell you that after applying a few pattern rules to the id number, the number of possible id numbers comes to about 1 million (instead of 10^7; each field can have {0,1,2,3,4,5,6,7,8,9}=10 digits, so net combinations = 10*10*10*10*10*10*10).
Therefore, to completely automate this process I wrote a script which would generate an id number, submit the form with that id number, and parse the resulting HTML profile for the "flag". If the script finds a hit on the flag, it stores all the fields of that player in a database. This script is working absolutely fine but the speed I was getting was about one check per second which means that I would have to leave the script running for about 11 days (to process all of about 1 million checks).
So i came up with this idea to divide the check into ten parts and i created separate scripts for each part. Now basically the first script checks for the first 100 thousand combinations, the second checks for another 100 thousand combinations, and so on.
The problem is that I am able to get only two of these scripts running at the same time. So it would still take me at least 5 days to get all the results. The rest of the scripts just sit there in the server's backlog. This is definitely due to Apache's limitation to handle multiple processes. The server I am using to run this script as well the target webserver both run on Apache2. I am sure it's not a problem with the receiving server. It has to be my Apache web server which is running the scripts. I have tried using mpm_winnt (on a windows server) as well as the prefork and worker modules (on a linux server) without any luck. Has any of you ever faced the same situation?
For those concerned about the legitimacy of this work, rest assured, this is absolutely legit. There's nothing in the website's use policy which restricts somebody from doing this. Moreover, my client hired me to do this only because the website owners were not able to hand over the data he required. They gave the stupid reason that they are helpless in providing the data because they don't have a system in place which would allow them to do a search restriction!
I recently switched from XAMPP to a self-setup WAMP installation using Apache Lounge's 2.4.9 package in order to improve performance, to have more tweaking possibilities and to solve some problems that I was not able to solve in XAMPP (e.g. creating large PDF files using TCPDF) on my local Windows 7 development PC.PHP is running as FCGI module, which I thought was a good idea to achieve good performance, but I am afraid I have messed up the configuration a bit.
What happens is that sometimes, PHP script requests do not seem to terminate(?), but I do neither know why, nor do I know what is happening. It probably has something to do with scripts that process kind of a larger
I've got a quite difficult problem which I don't know hoe to solve. We use a self created ISAPI module which is a "business server" running behind an Apache.
From time to time a client app crashes the server (the circumstances are not quite clear) - and because there is only one server process, all other clients crash as well. Here is an exemplary log of a crash:
Exception code: 0xc00000fd means stack overflow as I learnt. Therefore we configured mpm_winnt_module to use a 8M ThreadStackSize but this didn't work.
A solution might be that Apache starts for every client its own server process with the module ABCServer.dll. Because it is quite small and there are not thousands of customers this sounds like the perfect solution. No other clients/customers would be affected by a crash.
Unfortunately mpm_winnt_module supports only 1 process AFAIK. Worker and Prefork MPM are not available in Apache's Windows version what I read.
How can we configure Apache to start a new module process per client?
It should be a straight forward change. [URL] .... does not redirect to [URL]...... It simply tries to load /denver-cars/ and denver-cars is in the URL. Am I missing something here? I have tried moving it up and down the list of rules and have tried numerous types of flags to no avail. Everything else in the htacess works fine with out the line:
Options +FollowSymlinks RewriteEngine on RewriteBase / # Force www # Redirect google index dir's to new dir RewriteRule ^/(.*)-cars/ /newcars-in-$1/ [NC,R=301,L]
I'm looking to pass the entries to a web form, via Apache, to an external process (listening on a port say 4321) running on the same host as Apache.Is there a way to "coerce" Apache into doing this?
I have a site that is eating up my server resources and need to know what the best solution for this is. I'm thinking of getting another server just for mysql but do not know what specs the server should be to handle the current traffic/database load and have the site run smoothly without slowing down to a snail's pace.
An alternative is to get another server just for the videos being served and leave the database and html on the current server. This is where I'm stuck and don't know what route to take with this.
I've attached screenshots of top and bandwidth usage per day. Hopefully with this information you could tell me if I need another server or if there are any things I can do to the current server to help things move faster.
Right now I am using the built-in cpanel backup feature to backup all the accounts on my server to an offsite FTP backup account I have on a weekly basis. The problem I am having is that everytime the backup runs, the load (thats what I'm guessing is the cause) is crashing apache/bind, causing sites to be unreachable until I can restart the services. What options do I have to resolve this issue? Currently I already have spri installed, and it does help out a bit, but its still not enough.
At the time I took this particular snapshot, it's not near its peak... it's not uncommon for the CPU load approach 60%. Reading around, it seems the CPU load should normally be under 1% (such as 0.0139% or what-not). Is this true?
The weird thing though... I have no idea where that number is coming from, because according to "top", the CPU is actually 90% idle.
I actually just raised the MaxClients from 512 to 1024 because I was hitting a constant cap of 40 requests/sec... and I was worried it was going to bottleneck. When I raised that value, the max requests per second now seem to be freed up.
If the actual CPU of the server is 90% idle... am I okay? Anyone know where Apache's getting the CPU Load info from?
if anyone has had problems with high cpu load after upgrading APACHE to 2.2.8. We were running 1.3.5 with comfortable CPU loads of 2-10 on dual Xeon 2.8's. Now the loads are 20-70 with most of the CPU being used by many many httpd processes. I've heard that APACHE v2 does consume more resources, but am wondering if there was a problem with the build, or is it that much more demanding.
OTHER INFO: WHM:11.15 CPANEL: 11.18 OS: CENTOS 4.6 KERNAL: 2.6.20.4-ts.grh.mh.i386
They don't appear for some time if I kill them (a day or more). But it repeats again and again. One day there were 8 similar processes in total in max which used all 4 cores at 100% (and even ssh console was extremely slow to do something there).
I think that somebody is trying to make a small attack of some sort but I need to check it first. I tried to look at apache logs but there were too many posting requests from different IPs and no dublicates for little period of time so I had no success.
Anyway, that script worked for us for 4 years already and we didn't have any problems with it even on our old single core P4 2.8 ghz.
way to make sure is this an attack of some sort or just this script doesn't work correctly on our new machine?
Are there any ways to get IPs of visitors who are running posting.php with CPU overloading?
been checking out this site for a while and finally decided to register because I have a problem. Also hope this is the correct forum for this topic, sorry if it isn't.
So I have a problem with Apache. One of the sites that I run/host has a moderately large vBulletin board, and Apache just seems to eat up the CPU. Load averages have shot up between 20-30 and I've seen it as high as 80. Apache and MySQL are optimized already, I'm using suPHP for security because there are other sites on this box.
The funny thing about this is that it only started happening about a week ago. After checking for rootkits and all that garbage, I reinstalled the OS just to be on the safe side. Everything comes back clean still. I also got fed up and hired Platinum Server Management for a month, to see if they could find a solution (and I've been interested in reselling their services, but that's not relevant). So far the only thing they can come up with is disable suPHP, which isn't an option. I do realize that suPHP is ~20-25 times slower than mod_php, but what totally baffles me is that it worked beforehand and started going all crazy like this. I did try running the site using an dso configuration, the load did drop, but nothing to be proud of.
This site, and the server overall hasn't had any increase in load, I've held off putting new accounts on it until I get this fixed.
In the meantime, I have said forums running on lighttpd, which lowered the load. (Also writing a tutorial on having lighty work with cPanel)
In the last day or two I've been having issues with memory on one of by boxes.
Something eats it all up, so the OS starts swapping, the I/O wait shoots up, and soon the load is up in the hundreds and the thing is totally useless.
During the day today I've tracked the something to occational apache processes. It seems that occationally a thread is started which uses upwards of 150M of memory. These threads are obvioudly doing something heavyweight and take a while to complete. When I get a few of them running together it soons finishes off all my available memory. Below is an extract from top when a couple of these threads are running.
Code: PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 10230 nobody 25 0 197m 27m 9392 R 45.5 2.9 0:03.27 /usr/local/apache/bin/httpd -DSSL 10231 nobody 25 0 197m 25m 8376 R 52.8 2.7 0:01.60 /usr/local/apache/bin/httpd -DSSL I was wondering if anyone could help out by suggesting some ways of tracking down what page/script it which uses so much memory. It's a cPanel server so it's not really practical to tail -f the apache logs (not knowing which account it is means I don't know which file to watch).
I currently have a vps server from GIGEServers, running at 512mb of dedicated and 1024mb spike it seems like apache is using alot of resources, does anyone know of a less hog, or tips on tweaking it, and the same for mysql server any tips on tweaking.
We are currently running plesk 11.5 and over the last few months, the apache memory usage has been very high. On investigation we have gone through the logs and we can't find a cause.
I can issue a service httpd restart and the memory drops for about 30 minutes and then we see the apache memory grow to almost 1.5 - 2GB. Why this keeps happening?