<VirtualHost xxx.xxx.xxx.xxx:80>
ServerAdmin cs@reflexnetworks.net
DocumentRoot /home/reflextest/public_html
<Directory "/home/reflextest/public_html">
Options Indexes FollowSymLinks
AllowOverride None
Order allow,deny
Allow from all
</Directory>
ServerName test.reflexnetworks.net
ServerAlias www.test.reflexnetworks.net
ErrorLog logs/test.reflexnetworks.net-error_log
CustomLog logs/test.reflexnetworks.net-access_log common
</VirtualHost>
( xxx.xxx.xxx.xxx is the correct IP). The <Directory "/home/reflextest/public_html">...</Directory> part does not make any difference
Permissions:
Code:
ls -lR reflextest/
reflextest/:
total 4
drwxr-xr-x 2 reflextest apache 4096 Mar 25 04:50 public_html
reflextest/public_html:
total 4
-rw-r--r-- 1 reflextest apache 22 Mar 25 04:50 index.html
This is a source compile of apache. See for yourself: [url]
I've been having some trouble with Apache 2.0 recently. Whenever I load somewhere between 200 and 230 virtual hosts into Apache's config (as per my hosting control panel), Apache will fail to restart when the command "apache2ctl restart" is run. The command "/etc/init.d/apache2 restart" fails as well. All it does is signal all child processes of Apache to be terminated and fail to start again.
Here's what the log has to say:
Code: [Tue Nov 13 19:04:18 2007] [notice] SIGHUP received. Attempting to restart [Tue Nov 13 19:04:18 2007] [notice] seg fault or similar nasty error detected in the parent process Is it that Apache is too bloated with vhosts to restart or is it something else?
if it was possible to load vhosts from a MySQL database and the only solutions i found didnt work for me. Im running Apache on Windows. Has anyone seen anything that would work with Windows? The only alternative i can think of is to read the database info using PHP and output it to a vhosts config file. I dont want to do that though because it means restarting the server to load the new config and i hear that a lot of vhosts slows the server down
I work for a start up as well as do independent consulting. I had a friend set up my apache server initially for the start up. It's been running fine. Now I want to host a client's site on the same server. I tired just adding a vhost section at the bottom of the conf file, but it said that I was running two sites off the same port. I tried manipulating the stuff my buddy did for me for the first site, then putting two vhosts on the bottom, but that just didn't work.
I'm trying to configure a host to use vhosts but the server keeps breaking when I uncomment the Include vhosts file line. The apache process starts but none of the hostnames work. When the line is commented out, only the main hostname (defined in httpd.conf) works.
I tried searching the forums a bit and the Apache manual, but I was unable to find much related to my specific problem.
We have a system such that each user has a 'www' folder in their home directory so that example.com/~user will point to that folder. The user information, however, is not on the www server and instead we use a network file system to access the files in the home folder.
This all works fine and dandy with mod_rewrite. However, we want to allow each user to have access to the error logs, etc. for their own site. The best way to do this seems to simply store it in their home directory.
The problem is that how do you split these files apart. While it would be possible to have a separate process do it, the best way seems to just have each user be their own vhost. mod_vhost_alias seems to be applicable in this situation.
However, mod_vhost_alias relies on either the Host: field or the ip to split queries into their respective vhosts. As said before, our system uses the directory.
So the question is, is there a way to either use mod_rewrite to tell mod_vhost_alias what the vhost's name is. Or, is there a way to split the logs another way.
I have hundreds of sites on a server running Plesk. When I try to add more, Apache refuses to start, because it is out of file handles.
It is out of file handles because Plesk includes a custom ErrorLog and CustomLog in every vhost config file it makes. I tried overriding them with a vhost.conf file, but am not sure how to unset set options. I also tried pointing all the same ErrorLogs to the same file, but that didn't seem to do anything.
Does anyone know how I can either log everything to a single file, or just shut off logging? I really never use logging, and it seems like it would save me progressing power and storage (not to mention file handles!) if I could just disable it.
I have a Problem with my VHosts on one of two IP addresses. One IP is dedicated and works well
The other IP is shared. Since two days every call to one of the websites of this share ip goes to the Plesk Default Website. I tried to reconfigure single sites and all sites (/usr/local/psa/admin/bin/httpdmng --reconfigure-all), nothing works and restarting apache shows some warnings.
apache2ctl -S shows the following lines:
Code:
[Mon Dec 01 19:56:08 2014] [warn] VirtualHost 85.214.41.141:7080 overlaps with VirtualHost 85.214.41.141:7080, the first has precedence, perhaps you need a NameVirtualHost directive [Mon Dec 01 19:56:08 2014] [warn] VirtualHost 85.214.41.141:7081 overlaps with VirtualHost 85.214.41.141:7081, the first has precedence, perhaps you need a NameVirtualHost directive
[Code] ....
In an other thread I found a possible solution for the overlapping vhosts. The problem there seems to be a missing ssl-certificate and that matches with the situation on my server. After copying any other certificate-file as the one with the missing filename the apache starts without any warnings, but after that all websites (the sites from the dedicated ip too) only shows the apache default site ("it works"). I Tried to reconfigure single sites and all sites (/usr/local/psa/admin/bin/httpdmng --reconfigure-all), nothing works.
The server is hosted by STRATO, so there are daily backups. I turned back one week (the problem exists since two days). Then for some hours all works fine but suddenly all the sites on the ip 85.214.41.141 get broken again and shows only plesks default page.
We are using a server with Ubuntu 14.04 LTS and Plesk 12. When installing plesk, the vhosts directory is automatically created in /var/www/vhosts . Our problem is that /var is on a 120 GB SSD drive. We would like to move it to /data, which is a 2 TB drive. Whenever we used:
This KB-Article [URL] .... (exits with errors) this third-party documentation [URL] .... (Which is from a user who has the exact same problem with the same paths) all we get is ERROR 403 on our demo site. The site was created before the move and we also tried to create it after the move. The Plesk installation is fresh out of the box. There were absolutely no changes to the installation. Even the domain default or server default pages will not load anymore.
From what I can tell the vhosts on my Plesk Panel keep resetting (I think around once a day), all my webstatistics (AWStats) begin to 404 and I have a Wordpress installed in the root of a domain which is now displaying the "Web Server's Default Page" (any sub-directory of the Wordpress install gives a 404 such as the wp-admin directory). The way I've fixed this temporarily is to run "%plesk_cli%
epair.exe" --update-vhosts-structure. I do not see any errors or warnings in the event viewer or log files.
On another note is there a way to increase the logging for event viewer to include more than just errors (such as informational) to make it easier to track down issues?
We run a high traffic server and the access logs get filled up very quick. I know we could implement rotation, but I would also like to prevent performance loss by having an access log, doesnt matter how marginal that would be.
I've only ever had a shared hosting account with Hostgator, plus a few freebie hosts. However, I'm now pulling some heavy traffic and I'm concerned that Hostgator is going to suspend me soon.
My traffic on Saturday for example was ~2600 unique visitors and ~5000 page views. All of this traffic was from WordPress blogs and a small SMF forum. I've since converted one of the blogs to a static site to limit my CPU usage and I've setup caching for my other WordPress blogs. Advice I've heard on the Hostgator forums is that 7000 page views per day for a database driven site is around the time you should be upgrading and based on my traffic from Saturday (which admittedly was a bit of a spike) I could potentially be receiving 150,000 page views/month, so about 20x the point at which they recommend upgrading at.
Anyhows, in a nutshell I need to upgrade, or risk Hostgator throwing a tantrum at me ... but I don't have a lot of cash to pay for an upgrade Due to my lack of cashflow I've been considering moving to a VPS. The company which has interested me the most is HostV.com who offer a 256 MB (with 1000 MB 'burst' RAM) for only US$39.99 which seems quite reasonable to me.
They say that their 256 MB plan should be able to handle over 5000 page views per day for a WordPress run site, but I'm a little suspect. Do any of you know if this is a reasonable expectation from a 256 MB chunk of a virtual server? I have no idea and am always wary of believing the sales pitch of a random company across the other side of the world.
PHPAuction GPL Enhanced V2.51 Auction Software seems to be the perfect solution for my client, but the PHP requirements are very specific and the host we currently are using doesn't allow anyone to play with .htacess which is the usual workaround. Can anyone recommend a host that they know has the following setup?
Minimal server requirements are as follows: - Apache web server - PHP 4.0.6 or later (see below) with safe_mode=Off - register_globals=on - no open_basedir restriction - MySQL Database - 3.0 or higher - "Cookies" MUST be enabled on your computer!
Alternatively, does anyone know of any auction software that restricts sellers to only the admin?
I have a client that asked me to educate myself about web hosting and make a recommendation to him about where he should be. He currently has a shared hosting server at Network Solutions and finds unexplained slow downs and disk corruption reports in his forums DB unacceptable.
I'm glad I found this site-lots of good info but nothing like throwing up some stats and seeing what people recommend. The client told me he wanted to move to a dedicated server but I'm thinking a VPS might do the trick. Especially if upgraded with dedicated Core as well as RAM such as wiredtree is offering.
Looking for a managed, Unix based server that in a typical month serves 100k unique visitors 230k page views 500Gb of downloads
But needs to be easily upgradeable to handle his expected traffic levels in the next year of monthly visits in the order of: 250k unique visitors 600k page views 1.1Tb of throughput As far as features:
*Currently they use about 15 gigs of disk space. Some of that is inefficient disk management but the bulk is them supporting previous software releases.
*needs to be fully managed
*US datacenter with all the features you guys would expect to have as far as backbone access, security, power backups, etc..
*Backups by provider. Let's say 5 gigs worth since the old software versions don't really need to be backed up.(I'll recommend his own backups as well)
*Either plesk or cpanel
*15 minute hardware SLA is what the client is asking for but i'd like to present some comparisons to 1 hour SLA companies to see how much he'd save.
And finally, i tried to search for the answer to this but the keywords kept bringing up lots of hits without good info. The client sells software so the bandwidth needed is pretty consistent until they release a new version. Then it skyrockets to the point they may have 1500 people trying to download a 50Meg file simultaneously. What is the right way to handle that? Use a CDN or negotiate with the hosting provider to provide burstable bandwidth as needed. As a side note while looking at many offerings I was most surprised that bandwidth seems to sold in large chunks with overage costs hidden.
I'm not quite sure how much of Ram I need for my vps, But I'm going to get 1GB Vps from wiredtree.com
Anyone can tell me what kind of website I would be able to run on such an VPS? If it's just wordpress driven website...
Maybe anyone can share how much traffic your site have and how much ram it's using?
At the moment I have website with about 40k uniques/day and ~100k pageloads per day hosted on shared hosting but they have gave me 3days to find another hosting because they say I use to much of their traffic...
I use zoneedit to point my domain to the server, and a few times their servers don't respond for a few minutes that causes my site to be unaccessible. I was wondering if there was any better way of doing this? Please give me suggestions on what to do to have proper dns.
I have mixed feelings on the mail() function in php. I've seen it send out amazing numbers of emails and yet I keep reading that it is prone to problems. I'm wondering if someone can put an approximation on what you could/would send out in say 60 seconds. I know that there are tonnes of variables, so i'll try and remove a few:
- server hardware is typical of a cpanel/WHM shared hosting environment - running apache/linux - email is say 20kb HTML format - mail is sent in a php script loop, one by one; alternatively, mail() could be called in clusters of x emails - like send 5 emails at once - there are 400 recipients
I wanted to host multiple domains on the same hosting a/c and so I bought GoDaddy's economy hosting which supported multiple domains. But now when I try to add multiple domains, it displays this message -
"The domain path cannot be modified for economy hosting accounts."
That means all domains point to the same folder/content which is not what I wanted. So is there way to put a .htacess file or make an index.php file which can read where the request comes from and automatically redirect them to appropriate folders???
How well does ThePlanet.com handle DMCA complaints?
Recently got shafted by a client for over $30K (3 months of work plus our own out of pocket expenses). To make matters worse, this crook took the PHP source code which my company offers as a hosted solution and installed it on a dedicated server at Theplanet.com.
I have sent DMCA take down notices to theplanet, following their procedure (which is the standard legal procedure) but they have not done anything to the perpetrator. He has managed to spring up 2 websites already, and is no doubt planning to launch more.
Not sure if anyone else has been in this type of situation but I need to get something done about this. I have no doubt in my mind that even if theplanet did shut his sites down, or at least wiped the infringing software from the disks, he would jump over to another host and do the same thing again. My company doesn't have time to waste chasing him around, however the software he has contains a lot of proprietary code that we created and was never intended to be public.
So far both of the domains that he is using have Network Solutions as the registrar. Does Netsol assist with this kind of thing? Hopefully someone with similar experience can chime in with some advice.
I have VPS's with two companies that have managed/semi-managed support (depending on how you define it) and rely on them for a fair amount.
Whenever submitting a support request, I have to submit my root and cPanel passwords. Do people in my situation leave their root password as they would normally and just changing it however often they would if it wasn't given to support? Or, do you change your root/cpanel passwords before making a support request, and then change it back after the ticket is closed?
No offense intended to either of the VPS companies or their personell (that monitor WHT), both have been great. But, the reality is that I take it everyone at the company that has access to submitted tickets now have access to the root password, and since as a customer, I don't know when there has been employee turnover, that seems a security risk.
So, I am curious how others handle this. Not really sure if this belongs here or in the VPS forum, but since it could apply to any type of server/hosting account, I figured it belonged here.
if upgrading to that new server that I'll mention will probably solve my problems. Whatever help you can provide would be greatly appreciated. Below are the details:
In the GMT evenings and nights my current server gets so loaded that every page load takes 10 - 30 seconds. Even the pure html pages will be so slow to load. It seems that after a certain treshold it just suddenly becomes that much slower. Not much middleground there. I have high MaxClients and ServerLimit values now and the error log doesn't say that they are exceeded anymore but that didn't help enough.
I have a high traffic website that is using latest version of apache (2.2.x) with the prefork MPM and apache is optimized, PHP 5.2.5 and APC 3.0.15.
I get 160,000 - 210,000 pageloads per day. 32,000 - 45,000 visits per day.
Most of its pages are PHP but shouldn't be too CPU or databes intensive. Mysql isn't used and I mostly used sharedmem (php's shm functions) for databases. 2 semaphores are quite heavily used but that can't explain how a few more users would make the server serve pages so much slower.
Swap usage is practically 0 and CPU user % usage is like 1 - 2 % and CPU system % is also about the same even during peak times. However the Average Load or whatever that "top" reports is 6 - 9.
My current server scecs: 1 GB Ram, Pentium D 3 ghz, CentOS 5 32bit fully updated.
I load all pictures and even the stylesheet from a secondary server by using href="$secondaryserverIP..." in the html code, so the main server practically just serves the pages.
My new server will have apache with the worker MPM and latest versions of every software. Also its specs are: 2 GB of RAM, Intel Dual Core Xeon 2.40GHz, CentOS 5.1 32bit fully updated.
I have a sophisticated netstat based ddos script that is an improved version of DDoS Deflate and while some of these slowdowns seem to have been caused by attacks that it then was able to defend me from, most of them are not. I am even protected from users who constantly have 7+ connections to my site and if someone has a way too high number of connections, the script won't even check if it constantly has it and the script just bans that user outright. It probably is banning a bunch of innocent proxy users too but that is a small price to pay.