My server was running a cronjob backup, and my apache always fail
root@b# service httpd restart
/etc/init.d/httpd restart: httpd not running, trying to start
/etc/init.d/httpd restart: httpd started
root@b# service httpd restart
/etc/init.d/httpd restart: httpd not running, trying to start
/etc/init.d/httpd restart: httpd started
root@b# service httpd restart
/etc/init.d/httpd restart: httpd not running, trying to start
/etc/init.d/httpd restart: httpd started
I can only restart them on my CPanel WHM
Attempting to restart httpd
Waiting for httpd to restart.... . . . . . . . . . . finished.
httpd status
root 5805 0.8 0.9 21092 19504 ? S 22:39 0:00 /usr/local/cpanel/whostmgr/bin/whostmgr ./reshttpd
root 1012 4.0 0.6 23348 13828 ? Ss 22:39 0:00 /usr/local/apache/bin/httpd -DSSL
httpd started ok
root@b# service httpd restart
/etc/init.d/httpd restart: httpd restarted
I have changed to mod_filter from AddOutputFilterByType and now cannot rewrite the body content of a 302 page. I can rewrite the Location header using mod_headers, but I need to modify the informational content of the 302 page (usual a href "page has gone here") message as the source app puts internal details on the page.
At times as I'm developing, due to some coding error in PHP on my part, particularly calling a COM object, the apache server crashes. I'm delighted that it recovers, but in so doing it always tries to rerun the query that crashed it, which just causes another crash, and so on. Is there some way of getting round this, so that it recovers but the problematic code is not rerun?
XP SP3 (still!) Apache/2.4.3 (Win32) mod_fcgid/2.3.7 PHP/5.4.9 Firefox (Aurora)
I am trying to setup a pass through on our apache proxy server, typically this is not any trouble but our developers have started using a product call DEV EXPRESS and I cannot figure out how to get my pass through to work. Here's what my pass through looks like
ProxyPass /tmsdev http://tmsdev.dot.missouri/ ProxyPassReverse /tmsdev http://tmsdev.dot.missouri/ The initial url is http://tmsdev.dot.missouri/RealEstate/AAH/LitterPickup.aspx
I am getting 404 errors on everything that has /RealEstate/DXR.axd......... The DXR.axd is something from DEVEXPRESS here's the error I'm getting.
<!DOCTYPE HTML PUBLIC "-//IETF//DTD HTML 2.0//EN"> <html><head> <title>404 Not Found</title> </head><body> <h1>Not Found</h1> <p>The requested URL /RealEstate/DXR.axd was not found on this server.</p> <hr /> <address>IBM_HTTP_Server/6.0.2.35 Apache/2.0.47 (Win32) Server at ghweb04 Port 80</address> </body></html>
I've installed apache on my windows 2008 vps . It's run and when I browse localhost , it's works fine . But when I enter the vps ip address , it says Chrome could not connect to ... , I'm sure that httpd is running on port 80 (I checked it via netstat -ao and task manage ) ....
I'm not an admin, I can't make apache run cgi's, I have followed several documents in different pages without luck, I can run normal html pages but not cgi's. I only get the code displayed in screen.
I have small issue with load speed of pages when number of apache running processes goes over +-310, during peak traffic of day. IE: They load really slow.
Here's pic of what I mean:
=> [url]
Only thing that fixes it temporarily (at least until peak traffic ends) is to stop either httpd or mysql for several seconds, as everything cools down, then start it back up. Stop/starting mysql usually has longer terms temporary fix.
Anyway, would you have any ideas on what to configure or change within the system in order to keep the processes under 300?
The load is normal. Even at load 40, pages are blistering fast, as long as processes are under 300. So it's got nothing to do with load.
Code: [root@serwer /]# httpd restart /usr/sbin/httpd restart: httpd not running, trying to start /usr/sbin/httpd restart: httpd could not be started
Code: [root@serwer /]# httpd status Looking up localhost Making HTTP connection to localhost Alert!: Unable to connect to remote host. lynx: Can't access startfile [url]
How do I configure Lighttpd to serve images/javascript/css files that are littered about my websites? Can I just input the paths of all the directories that I want Lighttpd to serve from?
For example, I have 4 separate websites with each having a message board, main page etc... Could I have lighty serve the files from:
We run httpd-2.2.3 on CentOS 5.8 with caching enable. The server has been running for 4 years, but htcacheclean has never been run. The cache folder is about 10GB now (a "du" on the folder takes a very long time to complete). Should we expect any problem when running htcacheclean for a first time such as server slowing or freezing, ... Furthermore, in case of freezing, if we disable caching and restart the server, do we expect any problems? We did a few test on our test servers, and everything was fine, but our test servers do not have a large cache folder.
I have been a web developer using PHP for years now, and my work laptop runs CentOS with PHP5 as an Apache Module. This suits me just fine for development, but now I want to setup a web server that runs PHP 4.4.8 and PHP 5.2.5 and offer hosting to people. I know PHP4 is very old, but I am looking to offer my users the freedom of choice, as there are still third party applications out there that may need to be run on PHP4.
I will want PHP 5 to be the default, and allow PHP 4 to be used if either the file extension is .php4, or they have a line in their .htaccess file. I have seen various tutorials in search engine results saying to run both, I can either do one of the following:
1. Install PHP5 as a module, and run PHP4 using FastCGI 2. Install PHP5 and PHP4 and run both using FastCGI 3. Install PHP4 as a module, and run PHP5 using FastCGI
In the future I will also be looking to support PHP6 once a stable version has been released, though that will probably be optional to begin with and require a line in the .htaccess file too - like PHP4. I will be using Apache 2.2.8 on CentOS 5.1. I am also looking to install Ruby on Rails and Django too, which I think use FastCGI.
What's the difference between running as an Apache Module, or using FastCGI? This will be for a shared hosting environment so performance over lots of connections, stability and security are my concerns. Should I run everything using FastCGI, if not, would the default PHP version be better off installed as an Apache Module?
I am trying to run a simple script which brings up a gnome terminal on the user's display. I have created a perl cgi script which calls a shell script using a system call. If I run the perl cgi script from the command line, it brings up the gnome terminal correctly. If I try running the perl cgi script via the Apache web server, I get the following error in /var/log/httpd/error_log:
I'm not sure I understand the server-status page enough to know if this is a problem or not, but I have several processes that seem to run forever, or until I restart Apache. e.g.
Code: 13-1 21045 0/697/4264 W 59.45 19641 0 0.0 43.28 274.97 66.249.66.133 www.example.com GET /wp/2005/01/ HTTP/1.1 19-1 408 1/834/1831 C 83.52 32463 0 14.8 149.66 263.48 66.249.66.133 www.example.com GET /wp/ HTTP/1.1 30-1 14416 0/430/431 W 35.19 13347 0 0.0 37.42 37.44 66.249.66.133 www.example.com GET /wp/category/issues/ HTTP/1.1
They are almost always on a single domain (there's about 100 on the server) that's a Wordpress site. These processes are also almost always a search engine.
On the rare case I see them running on other domains on the same server they're always on Wordpress sites.
The longer the processes run, the more processor/memory they use, the more they slow the server down.
It seems to have just started in the past few weeks, I've had the site there for a couple years.
I'm using rsync to backup my cpanel/whm VPS data to a remote backup server.
Is there any inherent problem in bypassing the built in cpanel/whm backup feature, the cpbackup script, and just usign a cron with rysnc to sync the /home directory of my VPS to the backup server?
I want to make sure that there's not a security issue doing it this way, and that there is no later on restore problem by doing it this way.
Running programs named Perl with Heavy CPU usage, with the ownership of user apache.
We found the problem on Fedora 3 and Fedora 6.
In our case, it was the result of a Trojan activity.
Quick Solution
Check the cron jobs of user apache crontab -u apache -e */1 * * * * perl /tmp/.tmp/tmpfile delete the cronjob entry. Also delete the file /tmp/.tmp/tmpfile also added "apache" to the file /etc/cron.deny
Kind of a newb question but when I go into my cpbackup folder, then for example, into my daily folder..I see a number of accounts in there. Some are shown as only "accountname" and then some are shown as "accountname.0" What is the difference between them? Are the ones that end in .0 currently active accounts on the server? Thanks
My server load on my server is very high, but my server is not slow, my memory usage is fine, my CPUs are handling everything fine so it's not really a problem... except I've got a lot of "sleeping processes" which seem to increase the server load but not affect the server really.
My problem is, cpbackup refuses to run when it decides the server load is too high. So, is there a way to make it run no matter what the server load is? Because I do want backups to run again...
The server load never goes down because these things are always running, so running it at early hours of the morning doesn't make a difference.