Apache :: 2.4.3 Not Generating Correct URLs?
Aug 22, 2013
We're running Apache 2.4.3 on Slackware 14:
LCMlinux ~> uname -a
Linux LCMlinux 3.2.29-smp #2 SMP Mon Sep 17 13:16:43 CDT 2012 i686
LCMlinux ~> httpd -v
Server version: Apache/2.4.3 (Unix)
Server built: Aug 23 2012 11:07:26
LCMlinux ~>
We are using this both for the Trac issue-tracking application and for a small, simple internal mirror web site. Trac is working perfectly; the web site works if exact URLs are provided (as in <a href=...>
View 7 Replies
ADVERTISEMENT
Jun 28, 2008
Recently I use suphp with fcgi on my cPanel Server. When I apply RLimitCPU for each vhost, I see that some scripts that potentially overloading the server is killed. I think this is a good way to control load on server.
But each time a php process killed, apache create a coredump files under users's directory that is large enough to fill user's space limit.
How to disable apache from creating core dump files?
I have try:
- set "ulimit -c 0" on users and root
- set "/proc/sys/fs/suid_dumpable" to 0
- set /etc/security/limits.conf with 0 limit for core parameter
- set CoreDumpDirectory to specific directory ...
View 12 Replies
View Related
Feb 8, 2015
So I've got a problem where a small percentage of incoming requests are resulting in "400 bad request" errors and I could really use some input. At first I thought they were just caused by malicious spiders, scrapers, etc. but they seem to be legitimate requests.
I'm running Apache 2.2.15 and mod_perl2.
The first thing I did was turn on mod_logio and interestingly enough, for every request where this happens the request headers are between 8000-9000 bytes, whereas with most requests it's under 1000. Hmm.
There are a lot of cookies being set, and it's happening across all browsers and operating systems, so I assumed it had to be related to bad or "corrupted" cookies somehow - but it's not.
I added "%{Cookie}i" to my LogFormat directive hoping that would provide some clues, but as it turns out half the time the 400 error is returned the client doesn't even have a cookie. Darn.
Next I fired up mod_log_forensic hoping to be able to see ALL the request headers, but as luck would have it nothing is logged when it happens. I guess Apache is returning the 400 error before the forensic module gets to do its logging?
By the way, when this happens I see this in the error log:
request failed: error reading the headers
To me this says Apache doesn't like something about the raw incoming request, rather than a problem with our rewriting, etc. Or am I misunderstanding the error?
I'm at a loss where to go from here. Is there some other way that I can easily see all the request headers? I feel like that's the only thing that will possibly provide a clue as to what's going on.
View 1 Replies
View Related
Nov 30, 2013
1.Find .htaccess file in root folder
2.Open .htaccess file
3.Delete all content
4.Type in this code(using your domain ):
<IfModule mod_rewrite.c>RewriteEngine onRewriteBase /RewriteCond %{HTTP_HOST} !^www.example.co.uk$ [NC]RewriteRule ^(.*)$ http://www.example.co.uk/$1 [L,R=301]RewriteRule .* - [E=HTTP_AUTHORIZATION:%{HTTP:Authorization},L]</IfModule>
View 3 Replies
View Related
Feb 23, 2013
I need htaccess file for one of my sites. I've searched everywhere, and I can't find a solution for my problem.
System Info:
Easy Apache v3.18.1
CentOS 6.3 (x86_64)
So I'm trying to redirect all URLs in a certain query range to a new website. i.e.,
mydomain.com/?p=35000
mydomain.com/?p=35001
mydomain.com/?p=35002
mydomain.com/?p=35003
etc.
all need to 301 redirect to myotherdomain.com
I don't need to append each query to the new domain. All the URLs simply need to redirect to the naked domain of the other site. So
mydomain.com/?p=35000 is not redirecting to myotherdomain.com/?p=35000
It is simply redirecting to myotherdomain.com
Also, I need to redirect the specific series ([35000-35999]) to the new domain. I want to redirect the www and the non-www URLs.
I've attempted to write the code myself, but when I upload the htaccess file, the URLs are not redirecting. I'm doing something wrong.
Here's what I've tried:
Code:
RewriteEngine on
RewriteCond %{HTTP_HOST} ^mydomain.com$ [OR]
RewriteCond %{HTTP_HOST} ^www.mydomain.com$
RewriteCond %{QUERY_STRING} ^p=([35000-35999]*)$
RewriteRule ^(.*)$ "http://myotherdomain.com/" [R=301,L]
View 2 Replies
View Related
Jan 28, 2014
This redirect redirects all subpages of a domain to another domain while not redirecting to index.php but to same page, only at another domain:
RewriteEngine On
RewriteBase /
RewriteCond %{HTTP_HOST} !^cz.hq-scenes.com$ [NC]
RewriteRule ^(.*)$ http://cz.hq-scenes.com/$1 [R=301,L]
I want to achieve that only pages like /viewtopic.php...............
will be redirected (............. means any other characters)
How to redirect all viewtopic.php pages only? How to modify it?
View 2 Replies
View Related
Apr 24, 2013
I need to accomplish the following:
1. User hits my new 2.4 reverse proxy at [URL] ....
2. I proxy the request through to my "real" app server at [URL] ....
3. I also use a re-write rule to add a querystring to the URL: ?Parameter=Foo
4. So, client's request arrives at the my app server as [URL] .....
5. When my app server responds, it is including the Parameter=Foo key/value combination. I don't want this.
6. I want my reverse proxy (somebox.com) to strip "Parameter=Foo" from the string which gets returned to the client.
I have steps 1 & 2 working nicely, but it looks like I can't handle the last bit with with mod_rewrite. I found mod_filter and mod_substitute, but it appears that this stuff is used for re-writing strings IN the document. Can these libs be used to maybe modify (I'm guessing here) the headers so that the "?Parameter=Foo" string can't be seen on the client if they're running something like fiddler?
View 3 Replies
View Related
Jun 3, 2013
I've taken over a site that caters for client access. They all access there own folder, and in the folder the files have an include with a relative path as below.
/core - contains all the actual files
/client/file.php -
<? include "../core/file.php";?>
but with the growing number of clients I want to go a level deeper and separate them better...
/uk/client/file.php -
<? include "../../core/file.php";?>
This is fine but when the files are included, they too have there own relative includes and this is where it breaks.
There are so many files I can't easily go through them to change all the include paths so I would like to maybe do a rewrite to fake the path? I've tried this...
RewriteRule^uk/$ /
But that doesn't work.
View 2 Replies
View Related
Mar 11, 2015
I used a little vServer with ubuntu (turnkey) and use logwatch to be informed by email about any errors. I'm confused about the following errors from Apache:
--------------------- httpd Begin ------------------------
Requests with error response codes
404 Not Found
http://translate.google.com/gen204: 1 Time(s)
http://www.teddybrinkofski.com/ip_json.php: 1 Time(s)
503 Service Unavailable
http://www.google.com/: 1 Time(s)
---------------------- httpd End -------------------------
These errors are definetly not from my own code. I have checked that mod_proxy is disabled and i disabled also CONNECT like here described: [URL] ....
What does these errors mean and how can i disabled this?
View 4 Replies
View Related
Nov 15, 2013
My website is on 1and1 shared hosting and to enable PHP I needed to put in my .htaccess the following
AddType x-mapp-php5 .php .html .htm .shtml
View 11 Replies
View Related
Oct 23, 2014
I am attempting to add a header to a number of requested urls i.e. domain/feeds/chicago, domain/feeds/*.I understand that this can be achieved by
•using mod_rewrite to set an environment variable
•using mod_headers to add a header based on the existence of an environment variable.
So far I can add the a new header thus Header add RSS_FEED_URL "Akamai-Edge-Control"
View 1 Replies
View Related
Nov 7, 2012
I have a WP online shop using WP E-commerce plugin 3.8.9 together with the SEO Yoast plugin.My problem in that when exploring the product URLs ending with / in google webmaster tools, it displays 404. But the same URL without / is found and ok. I must day that both URLs show up correctly in browsers and the non / version is redirected to the one ending in /.Here is my .htaccess:
<IfModule mod_rewrite.c>
RewriteEngine On
RewriteBase /
RewriteRule ^index.php$ - [L]
RewriteCond %{REQUEST_FILENAME} !-f
RewriteCond %{REQUEST_FILENAME} !-d
RewriteRule . /index.php [L]
</IfModule>
I could solve this problem by setting a redirect to non / URLs instead.
View 3 Replies
View Related
Apr 16, 2015
I have two problems regarding Mailman Mailing Lists.
1) The Mailman Interface is usually located under lists.domain.tld/mailman (I changed the config from /cgi-bin/mailman/ to just /mailman/), but it seems the apache confs for this subdomain are not applied – I only see the servers default page when visiting this URL. domain.tld/mailman works, though. It would be great if lists.domain.tld/mailman would actually work and domain.tld/mailman wouldn't work.
How do I change/repair the configurations properly? I've installed the newest MU and already to reinstall mailman.
2) I get an Internal Server Error when visiting the domain.tld/mailman Interface. Reason is mod_suexec, which I need to disable for domain.tld/mailman, but where and how? All the vhost configurations are created automatically.
I could imagine this is related to my first problem and the mod_suexec thing is properly configured in the lists-subdomain config.
View 3 Replies
View Related
Aug 22, 2014
A dynamic php+mySQL site was converted to a static html by mirroring (wget). The result was that each page became an html file named some-page.html
View 4 Replies
View Related
Aug 28, 2014
I am trying to setup dynamic content configuration in apache 2.2 httpd conf. Here is the configuration and couple test cases:
CONF:
<FilesMatch ".(js|css|gif|jpg|jpeg|png)$">
Header set Cache-Control "no-cache, no-store, max-age=0, must-revalidate"
</FilesMatch>
TEST 1:
Without <FilesMatch> directive, the headers are set to all requested content as it should. This includes content served from apache server and content from a separate application server.
TEST 2:
If I use the above configuration example, I can see that static content comes with correct headers only
View 1 Replies
View Related
May 20, 2015
I have question: How to redirect specific useragent on specific URLs to specific URLs in .htaccess [Question]
E.g.:
I want to redirect 301 with conditional:
Code:
useragent: Firefox
from my url1: domain[dot]com/old-url1/
from my url2: domain[dot]com/old-url2/
to
Code:
to new url1: in my url1: domain[dot]com/new-url1/
to new url2: in my url1: domain[dot]com/new-url2/
I create this in my .htaccess but not work
Code:
RewriteEngine On
RewriteCond %{HTTP_USER_AGENT} googlebot [NC]
RewriteRule ^/?this-is-url/?$ domain[dot]com [L,R,NC]
View 1 Replies
View Related
Feb 11, 2008
On my server, i have one user ho create load on my server.
user 29508 22.0 0.0 0 0 ? Z 15:18 0:00 [php] <defunct>
That user has more site added with addons from cpanel. How can I found witch site is generating that high load ? Also some time, I have php index.php ( and that don't help me very much )
The server run php as cgi module.
View 1 Replies
View Related
Jul 22, 2008
Is there a way you can just disable the Generate Full Backup link in the Backup section of CPanel, and not disable the Backup section completely?
Meaning, I just want to disable that one function a client has the ability to do, but I don't want to stop them from being able to download their home directory, mysql databases,etc.
View 2 Replies
View Related
Aug 17, 2014
I am using the plesk firewall and trying to set up SSH rule which only allows from my IP but deny from everywhere else. In previous versions this worked fine by adding an ip selecting Allow from selected sources, deny from others and the icon in the rules would be orange with the lines
allow incoming from xxx.xxx.xxx.xx
Deny incoming from all others
However this no longer works as the deny from all others is not appearing and is not being generated in the iptables by plesk.
View 1 Replies
View Related
Jul 13, 2009
how can i provide temporary urls for users on my server like [url]until the actual domain resolves? ive seen this done in with cpanel but i dont know how its done exactly. my current server does not have cpanel.
View 11 Replies
View Related
May 23, 2009
Are there any scripts out there that can protect URLs? For an example I am trying to protect a megaupload.com URL with a masking URL and making sure that the masking URL is only access by a referral site. Can this be done?
View 1 Replies
View Related
Jul 19, 2008
Let's say you want to protect againts hacking,and using method with simply blocking loading url.So let's say someone hacked your index.html and changed links to lead to his domain.com.Is it possible to block what would be loaded on site ?(to prevent possible future hacking intrusions)
View 6 Replies
View Related
May 4, 2008
I'm testing scripts on new server now, and server has 2 problems.
1. I can not enter domain name as "get" parameter. For example, if I'm requesting URL like domain.com/file.php?url=[url] - it does not work. If I'm requesting URL like domain.com/file.php?url=[url](please note it has INVALID extension for TLD) - it works!
2. fsockopen and file_get_contents does not work. I added these settings into php.ini:
allow_url_fopen = On
allow_url_include = On
...and nothing works. I get just blank pages when using these functions.
Server is running cpanel + apache 2.2 + php 5 + APF firewall
View 6 Replies
View Related
Aug 15, 2008
I just found hundreds of rubbish urls in awstats for a particular domain. Is this referrer spam or something more serious and can I do something about this?
I have attached a screenshot.
View 4 Replies
View Related
Sep 27, 2007
One of the sites I have, is placed on a non-Apache server (the others are). Phpinfo() gives this:
Server API CGI
I'd like to make search engine-friendly URLs for all my sites. All of them will do fine with mod_rewrite, but that's not possible on this server, it seems. Anybody here knows how I can do this for this particular server?
View 1 Replies
View Related
Jan 28, 2015
I'm trying to use nginx php-fpm on my forum, and to set friendly url's, I am told to add this to nginx config:
Code:
location / {
try_files $uri $uri/ /index.php?$uri&$args;
index index.php index.html;
[Code] ....
I try adding this using the Additional Nginx Directives, and get this error:
Code:
Invalid nginx configuration: nginx: [emerg] duplicate location "/" in /var/www/vhosts/system/domain.com/conf/vhost_nginx.conf:1 nginx: configuration file /etc/nginx/nginx.conf test failed
I think if my domain was in a folder, this wouldn't be a problem. I really don't want to change that, though.
View 3 Replies
View Related
May 28, 2014
I'm migrating some websites from old server with virtualmin, some websites have files with special characters as à ,ö,ç etc...
On the other server the files (images for example) are served well but on the new server with plesk 11.5 error 404 appears. (Nginx reverse proxy is activated)...
View 2 Replies
View Related
Oct 29, 2014
My Linux (CentOS) server with Plesk 12 is giving HTTP 414 errors ("URL too long") in response to URLs which are over 256 characters in length. They happen to include a GET variable in the query string which accounts for most of this length, and if I shorten it manually, it works. But I can't change the script to submit a shorter URL or send it by POST, because it comes from an external payment processing server which I don't control.
Adding the following lines to my /etc/httpd/conf/httpd.conf file and restarting Apache does not work:
LimitRequestLine 8190
LimitRequestFieldSize 8190
The URLs I'm trying to use are well short of 8190 bytes; they are around 800 characters long.
Is this something that Plesk affects / can control? Is there a way to see what the current maximum setting for URL length is, and to change it?
View 14 Replies
View Related