How Do I Fix "excessive Apache Connections"
May 2, 2008
I just received an email saying
We were forced to suspend your account due to excessive apache connections causing ftp to go down. Here is a snapshot of the activity:
242-0 22189 0/57/57 W 1.24 10 0 0.0 0.09 0.09 xxxxxx xxxxxx.com
243-0 22194 0/1/1 W 0.01 0 0 0.0 0.00 0.00 xxxxxxx xxxxxx.com GET
244-0 22214 0/4/4 W 0.15 4 0 0.0 0.03 0.03 xxxxxx xxxxx.com GET
And when I was on the phone with the support, they told me my site had opened up 400 apache connections + using the GET command and was causing FTP slow downs.
It just happens so that a file I posted made its way to a popular forum and all the sudden, everyone was downloading the file.
What is the best way to correct this problem? and GET? Isn't that a FTP Command? but people were downloading via [url]
View 2 Replies
ADVERTISEMENT
Jul 12, 2007
I have the impression that I am being affected from a kind of DDOS or email worm attack. Is there a way I can track the sources of the connections?
The control panel I am using is Plesk 8.2 on Linux CentOS 4.2
View 2 Replies
View Related
Nov 4, 2013
What was causing this:
I installed a url shortner script but the link that the script creates takes you to a server error page.
I viewed the logs and I get this error over and over again.
Request exceeded the limit of 10 internal redirects due to probable configuration error. Use 'LimitInternalRecursion' to increase the limit if necessary. Use 'LogLevel debug' to get a backtrace.
this is what is in my htaccess file
RewriteEngine On
RewriteCond %{REQUEST_FILENAME} !-d
RewriteCond %{REQUEST_FILENAME} !-f
RewriteCond %{REQUEST_FILENAME} !-l
RewriteRule ^(.+)$ index.php?url=$1 [QSA,L]
View 1 Replies
View Related
Sep 15, 2008
Just logged in my cPanel, and Apache Server Status shows
Parent Server Generation: 7
Server uptime: 2 hours 52 minutes 5 seconds
Total accesses: 701666 - Total Traffic: 63.7 GB
CPU Usage: u1610.22 s255.4 cu0 cs0 - 18.1% CPU load
68 requests/sec - 6.3 MB/second - 95.2 kB/request
400 requests currently being processed, 0 idle workers
I told customer service and said my website (a big forum) have 4000 people now, I felt very slow, could the slowness caused by this max apache connection setting?
I got reply: "400 seems to be as high as Apache can go. Your httpd.conf settings currently show 500 max connections enabled. If Apache is stopping at 400 then this is it's hard limit for maximum connections. Also If it was able to go even higher you would eventually run into memory issues on the server that would cause the server to crash."
Can anyone tells me if "400 requests currently being processed, 0 idle workers " is a problem or could it be the cause of the slowness. I imagin if more people request connection, and apache can't deal with that much, it has to let those request wait in the queue, therefore caused slowness or time-out.
The seem server could deal with 8000 people online before, no any problem at all and speed was quite fast. I don't know what i should do now.
View 14 Replies
View Related
Jan 19, 2007
is a way to understand whats the bext max apache settings for me?
Maybe to look over httpd-status requests currently being processed and the number of idle servers or the number of strokes?
Or probably the best way is to use some benchmark application but then i am not sure how to test my config?
View 4 Replies
View Related
Sep 24, 2007
I currently have a dedicated server with the following specs:
Celeron 2.6ghz
2gb ram
100mbit connection
the cpu load is always under 0.7, and I always have at least 400mb of free ram.
The site takes about 5-10 seconds to load a test page with just a single word on it. When I type netstat -n | grep :80 | wc -l
I get around 1100-1200 connections. I get about 130k page views per day.
My site is about 95% static html, it has about 150 images per page. How can I speed up my site?
Here are my apache httpd.conf settings:
Timeout - 100
KeepAlive - on
KeepAliveTimeout - 12
StartServers - 48
MinSpareServers - 32
MaxSpareServers - 64
MaxClients - 1500
MaxRequestsPerChild - 1000000
View 6 Replies
View Related
Feb 22, 2008
whats the easiest way to monitor the number of apache connections?
View 9 Replies
View Related
Jul 11, 2009
Our web server encounter a problem lately, that an IP address in India repetitively sent requests that uses up all connections available in Apache. All connections appear to be in W state. The connections are not terminated by Apache even though timeout has been set to lower 30 seconds. Similarly, MySQL connections also are not dropped until Apache is restarted.
View 7 Replies
View Related
Feb 24, 2007
I would like to know what is the maximum number of Apache connections a Server can handle? Does this depends on the Config of the Server? Is it Possible for a server to handle more than 2500 Active Apache connections without timeout / connection failure / slowness?
View 6 Replies
View Related
Apr 19, 2007
I've got an older "Super Server P4" series server @ ThePlanet - it's been a great box for years. We recently upgraded php4 to php5 and did a mysql upgrade as well. Ever since, randomly, without warning, Apache stops terminating connections, so the max connections fills up, and httpd won't respond. Apache doesn't stop running, it just max's out and stops accepting new connections, so customers assume the server is "down", although email/FTP work fine.
The server never crashes, the loads stay down, but httpd just fills up and won't accept more connections. We can't increase max connections (we actually DECRESED IT), because they'll just keep piling up, never terminating, and then ultimately, it will crash the box.
My admin has worked for nearly 2 weeks trying to figure it out, and Scott (AtomicRocketTurtle) and his team have been evaluating it for about a week - it's happened 3x in 2 days ... last night, httpd quit responding for about 7+ hours and since we didn't have httpd monitoring, we never knew until the office opened this AM and I had two very angry customers. Scott suspects it may be some rogue application that triggers it that didn't affect it prior to the php5/mysql upgrades.
Both Scott and Parm, my admin, have about thrown in the towel and are recommending we retire the server and migrate to a new box.
It's older - much older .. I'm pasting specs below, RHEL3, Plesk 7.5 .. but just wonder, before retiring an old server that was RUNNING GREAT prior to the upgrade, maybe someone has seen this happen before?
If so - PLEASE LET US KNOW before I spend the $$$ on a new box.
Current box's specs:
CPU GenuineIntel, Intel(R) Pentium(R) 4 CPU 2.80GHz
Version psa v7.5.4_build75060118.18 os_RedHat el3
OS Linux 2.4.21-47.0.1.ELsmp
RAM: 1GB, with 2GB swap
apache 1.3.31
bind 9.2.4-20.EL3
coldfusion Component was not installed
coldfusion-support 7.3-2.96.128
courier-imap 3.0.8-rhel3.build75050824.12
drweb 4.32.2-rh7_psa
drweb-qmail 4.32-rhel3.build75050824.12
frontpage 5.0.2.2634
httpd 2.0.46-61.ent
jdk 1.4.2
logrotate 3.7
mailman 2.1.5.1-25.rhel3.7
mod_perl 1.99_09-10.ent
mod_python 3.0.3-5.ent
mysql 4.1.21-2.rhel3.art
perl-Apache-ASP 2.57-rhel3.build75050824.12
php 5.0.4-13.rhel3.art
phpmyadmin 2.5.3
phppgadmin 2.4.2
postgresql 7.3.6-7
postgresql-server 7.3.18-1
proftpd 1.2.9
psa 7.5.4-rhel3.build75050824.12
psa-agent 1.3.2-2_psa7.1
psa-api-rpc 7.5.4-rhel3.build75050930.11
psa-bu 7.5.4-rhel3.build75050926.17
psa-horde 3.0.5-rhel3.build75050824.12
psa-imp 4.0.3-rhel3.build75050824.12
psa-logrotate 3.7-rhel3.build75050824.12
psa-manual-custom-skin-guide 7.5.4-rhel3.build75050824.12
psa-migration-manager 7.5.4-rhel3.build75050930.11
psa-proftpd 1.2.10-rhel3.build75050824.12
psa-qmail 1.03-rhel3.build75050824.12
psa-qmail-rblsmtpd 0.70-rhel3.build75050824.12
psa-spamassassin Component was not installed
psa-tomcat-configurator 7.5.4-rhel3.build75050824.12
psa-turba 2.0.3-rhel3.build75050926.17
qmail 1.03
rblsmtpd 0.70
samba 3.0.9-1.3E.12
spamassassin 3.1.3-1
SSHTerm 0.2.2-rhel3.build75050824.12
stunnel 4.04-4
tomcat 4.1.24-full.2jpp
webalizer 2.01_10-15.ent
View 4 Replies
View Related
Apr 11, 2013
I am trying to figure out how Apache is working on windows 7 - so far so good - but how do I set restrictions and limitations on bandwidth usage and max number of IP-connections?
View 1 Replies
View Related
Oct 29, 2013
Webserver is restarted all TCP connections randomly. we are not able to get any error into errorlog file.
Webserver close all TCP connection and start created new TCP connection for new request.
AcceptFilter http none
AcceptFilter https none
Also, conf file is saying that
Defaults Values are:
EnableMMAP On
EnableSendfile Off
Not sure what's wrong with server.
We have installed "Apache 2.4.6-x86 server" on Win 2008 R2 Standard server (64-Bit).
Also, we are using mod_jk module for connecting to tomcat server "tomcat-connectors-1.2.37-windows-i386-httpd-2.4.x.zip".
View 4 Replies
View Related
Nov 13, 2008
I am facing to this trouble: apache max out the number of connections. When viewing status, all I get here
Code:
SrvPIDAccMCPU SSReqConnChildSlotClientVHostRequest
0-028240/6/65R 0.1414290.00.020.32 ??..reading..
1-028170/17/65R 0.111510.00.030.22 ??..reading..
2-028350/3/57R 0.20112210.00.000.24 ??..reading..
3-027570/33/62R 1.431412930.00.130.21 ??..reading..
4-026430/75/75R 2.12153190.00.310.31 ??..reading..
5-026440/52/52R 2.061610.00.240.24 ??..reading..
6-028180/13/60R 0.621600.00.070.22 ??..reading..
7-026490/63/63R 2.84143070.00.370.37 ??..reading..
8-028190/7/43R 0.271610.00.020.27 ??..reading..
9-028080/27/76R 0.361410.00.040.29 ??..reading..
10-026750/50/52R 2.9617630.00.260.26 ??..reading..
11-027720/23/49R 3.641600.00.100.19 ??..reading..
12-026620/64/64R 2.201520.00.290.29 ??..reading..
13-027730/25/64R 1.121500.00.140.19 ??..reading..
which account on server is the reason please? I am on CentOS box with DirectAdmin.
View 2 Replies
View Related
Apr 12, 2007
any good rule to limit Apache (port 80) connections from 1 IP to 15 with iptables/csf?
And total connections to the box to 100?
View 6 Replies
View Related
Feb 5, 2007
As you can see my apache is full of Reading connections..... they are filling up my server dening legitimate users to browse trought the websites hosted there... I think this is what is happening to me:
http://mail-archives.apache.org/mod_...l.gmail.com%3E
Im using apache 1.3.3.7 on RHES 3 with latest patches and kernel.
930 requests currently being processed, 6 idle servers
RRRRRRRRWRRRRRWRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR
RRRRRRRRRRRRRRRRRRWRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR
RRRWRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRWRRRRRRRRRRRRR
RRRRRRRRRWRRRWRRRRRRRRRRRRRRRWRRRRRRRRRRRRRRRRRRRWRRRRRRRRRRRRRR
RRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR_RRRRRRRRRRRRR_RRRRRRR
RRRRRRRRRRRRRRRRRRRRRRWRRRWRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR
R_WRRRRRRRRRRRRRRRRRRRRWWRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR
RRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRWRRRRRRR
RRRRRRRRRRRRRWRR_RRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR_
RRRRRRRRRRRRRRRRRRRRRRWRRRRRRRRRRRRWRRRRRRRRRRRRRRRRRRRRRRRRRRRR
RRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRWRRRRRRRRRRRRRRRRRRRRRRRRRRRR
RRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRR
RRRRRRRRRRRRRRRRRRRRRRRRR_RRRRRRRRWRRRRRRRRRRRRRRRRRRRRRRRRRRRRR
RRRRRRRRRRRRRRWRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRWRRRRR
RRRRRRRRRRWRRRRRRRRRRRRRRRRRRRRRRRRRR.RRR.......................
1-044940/40/40R 0.33340.00.060.06 ??..reading..
2-044950/55/55R 0.47330.00.110.11 ??..reading..
3-044960/35/35R 0.31330.00.050.05 ??..reading..
4-044970/38/38R 0.13210.00.090.09 ??..reading..
5-045410/40/40R 0.16360.00.130.13 ??..reading..
6-046190/28/28R 0.12110.00.030.03 ??..reading..
7-046220/34/34R 0.35320.00.100.10 ??..reading..
8-051640/34/34R 0.135100.00.300.30 ??..reading..
Also, here is the top output of the top command:
top - 12:46:21 up 10 min, 1 user, load average: 1.40, 2.03, 1.06
Tasks: 1063 total, 2 running, 1060 sleeping, 0 stopped, 1 zombie
Cpu(s): 4.9% us, 1.6% sy, 0.0% ni, 93.1% id, 0.3% wa, 0.0% hi, 0.0% si
Mem: 2073516k total, 2001984k used, 71532k free, 42384k buffers
Swap: 2048276k total, 0k used, 2048276k free, 178096k cached
Server seems to be fine, but total tasks are always between 1060 and 1124... thats very rare too..
how to avoid this attack?
View 14 Replies
View Related
Jan 9, 2007
I've been having trouble the past few days with someone who's been "attacking" my site so to speak by continuously downloading very large files with as many connections as (he) can open. I operate a large downloads site for computer games, this person has selected the largest files (like 400-500MB). Not sure of the real intent other than to clog up my bandwidth capacity. Also he appears to be using proxies since as soon as I ban one, another shows up seeminly from China.
Anyway, I have mod_bw and I've limited the number of connections in the downloads area to 2. While that works ok, his tool uses threads like a download manager would and he's using up 30-40 child threads for his 2 file downloads.
So 2 questions,
Is there anyway to not only limit file downloads to 2, but limit the number of connections per request? Many of my visitors do use download managers and I'd like for them to continue using them but use a reasonable number of threads like 6 or 8, but not 30.
Also, is there a way to restrict access to someone using a proxy?
View 2 Replies
View Related
Apr 7, 2014
I am interested to know number of current connections in my Apache server
Server version: Apache/2.2.15 (UNIX)
OS : Solaris 5.10
View 5 Replies
View Related
May 31, 2008
I keep getting emails from LFD saying this user is using to many resources and it is because of their shoutcast is their a way to take care of this problem?
View 0 Replies
View Related
Oct 9, 2009
Resource: Virtual Memory Size
Exceeded: 149 > 100 (MB)
Executable: /usr/bin/php
I've been receiving e-mails about this. May I know how to fix this?
View 6 Replies
View Related
Apr 5, 2007
I'm having a problem that I've never run across before, and was wondering if anyone might have any ideas as to what may be causing this.
Basically, on 3 of 5 new servers on a brand new private rack from The Planet, we're having what we've narrowed down to be a problem with PHP or Apache. Loading any sort of PHP page with a larger output (even such as a simple 'phpinfo' call) results in, depending on the computer or browser in use:
- The page loading for a split second then reverting to a DNS Server Not Found page (observed in IE)
- The page loading, but filling the source code with vast amounts of extra blank spaces, making a simple phpinfo call download 5+mb of HTML (observed in both IE and Firefox)
- The page loading part way, then hanging (observed in Firefox)
- Occasionally the page will reload over and over again all by itself until it ultimately goes to a DNS error page (observed in IE)
Pages not including PHP, including very long .HTML and .SHTML pages, load just fine.
Here's a link to a page calling a simple phpinfo string, and nothing else (as this is my first post, I can't link directly to URLs, sorry):
View 1 Replies
View Related
Nov 20, 2008
Excessive resource usage: dbus (2015)
I get below alarm from lfd
Quote:
Time: Sun Sep 28 12:16:06 2008 +0200
Account: dbus
Resource: Process Time
Exceeded: 134303 > 1800 (seconds)
Executable: /bin/dbus-daemon
The file system shows that this executable file that the process is running has been deleted. This typically happens if the original file has been replaced by a new file when the application is updated. To prevent this being reported again, restart the process that runs this excecutable file.
Command Line: dbus-daemon --system
PID: 2015
Killed: No
How can I find which process runs this excecutable file ?
View 1 Replies
View Related
Jun 21, 2014
I fixed that there is a process that loads the server and much why it does not.
Name process: md2_resync
View 2 Replies
View Related
Feb 17, 2007
Mountain View (CA) - As a company with one of the world's largest IT infrastructures, Google has an opportunity to do more than just search the Internet. From time to time, the company publishes the results of internal research. The most recent project one is sure to spark interest in exploring how and under what circumstances hard drives work - or not.
There is a rule of thumb for replacing hard drives, which taught customers to move data from one drive to another at least every five years. But especially the mechanical nature of hard drives makes these mass storage devices prone to error and some drives may fail and die long before that five-year-mark is reached. Traditionally, extreme environmental conditions are cited as the main reasons for hard drive failure, extreme temperatures and excessive activity being the most prominent ones.
A Google study presented at the currently held Conference on File and Storage Technologies questions these traditional failure explanations and concludes that there are many more factors impacting the life expectancy of a hard drive and that failure predictions are much more complex than previously thought. What makes this study interesting is the fact that Google's server infrastructure is estimated to exceed a number of 450,000 fairly mainstream systems that, in a large number, use consumer-grade devices with capacities ranging from 80 to 400 GB in capacity. According to the company, the project covered "more than 100,000" drives that were put into production in or after 2001. The drives ran at a platter rotation speed of 5400 and 7200 rpm, came from "many of the largest disk drive manufacturers and from at least nine different models."
Google said that it is collecting "vital information" about all of its systems every few minutes and stores the data for further analysis. For example, this information includes environmental factors (such as temperatures), activity levels and SMART parameters (Self-Monitoring Analysis and Reporting Technology) that are commonly considered to be good indicators to describe the health of disk drives.
In general, Google's hard drive population saw a failure rate that was increasing with the age of the drive. Within the group of hard drives up to one year old, 1.7% of the devices had to be replaced due to failure. The rate jumps to 8% in year 2 and 8.6% in year 3. The failure rate levels out thereafter, but Google believes that the reliability of drives older than 4 years is influenced more by "the particular models in that vintage than by disk drive aging effects."
Breaking out different levels of utilization, the Google study shows an interesting result. Only drives with an age of six months or younger show a decidedly higher probability of failure when put into a high activity environment. Once the drive survives its first months, the probability of failure due to high usage decreases in year 1, 2, 3 and 4 - and increases significantly in year 5. Google's temperature research found an equally surprising result: "Failures do not increase when the average temperature increases. In fact, there is a clear trend showing that lower temperatures are associated with higher failure rates. Only at very high temperatures is there a slight reversal of this trend," the authors of the study found.
In contrast the company discovered that certain SMART parameters apparently do have an effect drive failures. For example, drives typically scan the disk surface in the background and report errors as they discover them. Significant scan errors can hint to surface errors and Google reports that fewer than 2% of its drives show scan errors. However, drives with scan errors turned out to be ten times more likely to fail than drives without scan errors. About 70% of Google's drives with scan errors survived the first eight months after the first scan error was reported.
Similarly, reallocation counts, a number that results from the remapping of faulty sectors to a new physical sector, can have a dramatic impact on a hard drive's life: Google said that drives with one or more reallocations fail more often than those with none. The observed average impact on the average fail rate came in at a factor of 3-6, while about 85% of the drives survive past eight months after the first reallocation.
Google discovered similar effects on hard drives in other SMART categories, but them bottom line revealed that 56% of all failed drives had no count in either one of these categories - which means that more than half of all failed drives were put out of operation by factors other than scan errors, reallocation count, offline reallocation and probational counts.
In the end, Google's research does not solve the problem of predicting when hard drives are likely to fail. However, it shows that temperature and high usage alone are not responsible for failures by default. Also, the researcher pointed towards a trend they call "infant mortality phase" - a time frame early in a hard drive's life that shows increased probabilities of failure under certain circumstances. The report lacks a clear cut conclusion, but the authors indicate that there is no promising approach at this time than can predict failures of hard drives: "Powerful predictive models need to make use of signals beyond those provided by SMART."
View 6 Replies
View Related
Dec 22, 2008
Sometimes my server surcharge load average increase at 60 , and all my configuration are OK
when i type :
netstat -plan|grep :80|awk {'print $5'}|cut -d: -f 1|sort|uniq -c|sort -nk 1
i have : ...
View 8 Replies
View Related
Dec 21, 2008
I tried to update a plugin at my blog its a wordpress blog, as soon as the update was started that site on the server stopped working, (later on i closed the upgradation window), after few minutes website start working automatically, Now in my opinion I think that update process is still running in background thats why connections are creating continuously to that website IP.
[root@server ~]# netstat -alpn | grep :80 | awk '{print $4}' | cut -d: -f1 |sort |uniq -c
1001 serverIPhere
its even touching 1500, I tried to contact my server support but unfortunately they can investigate the issue, instead they told me to check with the following command.
netstat -plan |grep :80 | awk '{print $5}' | cut -d: -f1 | sort | uniq -c
which is not an answer to my question. Can anybody please tell me why those connections are making to that website's IP? I don't think its a Ddos attack, because it was just started when i updated the plugin.
View 10 Replies
View Related
May 19, 2008
Could someone comment on the kind of load a VPS service can handle? If I were to run an HTTP server how many connection/sec would be realistic.
View 3 Replies
View Related
Mar 6, 2007
How many simulteanous connections to the site do alot of webhosting company usually allow with shared hosting packages. I was wondering because4 some companies say pay $$ a month get 300gb of bandwith a month. Can they limit the bandwith by limiting your simulteanous connections? I am asking because I just found out my host only allows 50 per hosting package that is on a shared server. To me that seems to be very little.
View 1 Replies
View Related
Apr 19, 2007
WARNING: One or more of your DNS servers does not accept TCP connections. Although rarely used, TCP connections are occasionally used instead of UDP connections. When firewalls block the TCP DNS connections, it can cause hard-to-diagnose problems. The problem servers are:
Error [No response to TCP packets].
APF is installed on the server, how do I allow TCP DNS connections? I already added port 53 to ingress/egress for TCP and UDP.
View 7 Replies
View Related