Dramatically Increase Your Performance Of Web Server By Disk Cache
Jun 22, 2009
it took me one year to develop the disk cache tool which can dramatically boost your host and save your harddisk. it is like supercache,but more cheap and better speed.
check picture to see what it can do.
i will offer free download to test the tool by first 10 people. if u host huge traffic website, do not hesitate to try it. i already test it for half a year. it is time to publish it. pm me or post here to get free download.
I use apache with CentOS VPS hosting for my blog. I only host one blog in this VPS account. I have 1.5GB RAM and I have 7, 500 page preview per day. My page loading time is 2-3 seconds (according to the pingdom tool).
I want to know what is the best performance (faster web page loading) W3 Total cache option for VPS hosting blog. Currently I use Disk to enhance for page cache and database cache for disk.
I'm considering upgrading our company's server. We run an ecommerce site php/mysql based, we have about 12,000 visitors per day and we use up about 1500 gig per month. Simple ecommerce site, nothing fancy. We had some complains from people that our site is "slow" (despite it running smoothly for us, but these people claimed that only our site was "slow" and that they use fast connections) so that's why now I'm thinking maybe its time to upgrade.
We are using softlayer, for an extra $50 per month they recommended that we upgrade from our Opteron 1216 to "Xeon 3220 - 2.40GHz (Kentsfield) - 2 x 4MB cache"
Are we going to see a serious increase in performance? Will the increase be enough that most visitors will actually be able to feel it?
By default ModSecurity relies on the regular expression library built into Apache for pattern matching. This works well with Apache 2.x but not so much with Apache 1.x. The Apache 1.x regular expression engine is several times slower. Since 1.9.2 it is possible to compile ModSecurity for Apache 1.x against an external regular expression library (PCRE, [url] the same library used in Apache 2.x) and achieve significant performance increase. This is achieved with the USE_PCRE compile-time flag.
If you have PCRE already installed on your system it may be sufficient to compile ModSecurity like this:
Code: [root@servidor1 customapache]# apxs -DUSE_PCRE -cia mod_security.c gcc -DLINUX=22 -DHAVE_SET_DUMPABLE -I/usr/include/gdbm -DMOD_SSL=208128 -DUSE_HSREGEX -DEAPI -fpic -DSHARED_MODULE -I/usr/include/apache -DUS E_PCRE -c mod_security.c mod_security.c:37:18: pcre.h: No such file or directory mod_security.c: In function `my_pregcomp': mod_security.c:929: error: `PCRE_CASELESS' undeclared (first use in this function) mod_security.c:929: error: (Each undeclared identifier is reported only once mod_security.c:929: error: for each function it appears in.) mod_security.c:929: warning: assignment makes pointer from integer without a cast mod_security.c: In function `my_regexec': mod_security.c:940: error: `PCRE_ERROR_NOMATCH' undeclared (first use in this function) apxs:Break: Command failed with rc=1 I double checked if I have PCRE installed:
Code: [root@servidor1 customapache]# yum search pcre pcre.i386 4.5-3.2.RHEL4 installed Matched from: pcre Perl-compatible regular expression library. PCRE has its own native API, but a set of "wrapper" functions that are based on the POSIX API are also supplied in the library libpcreposix. Note that this just provides a POSIX calling interface to PCRE: the regular expressions themselves still follow Perl syntax and semantics. The header file for the POSIX-style functions is called pcreposix.h. [url] And I even located the /lib/libpcre.so.0.0.1 and copied it to /usr/lib/apache/libpcre.so, with no sucess. Got the same error.
Has anyone sucessfully compiled mod_security with PCRE and can help me?
One of the things people seem to bring up a lot is disk IO Performance.
Why? Because theres little you can do about a customer being stupid and creating a disk swap nightmare.
There is however something you can do to reduce the impact across your clients, Have a separate raid array for swap space.
This does 2 things, it splits some of the Disk IO across 2 arrays, but more important it reduces the affect someone overusing there swap will have on the ones that are not.
does mod_cache use real disk space or real memory to locate "cache"? I mean if i use mod_cache, does my vps/server will use more disk space/memory?
Quote:
mod_cache implements an RFC 2616 compliant HTTP content cache that can be used to cache either local or proxied content. mod_cache requires the services of one or more storage management modules. Two storage management modules are included in the base Apache distribution:
Today we are going to conduct a detailed study of RAIDability of contemporary 400GB hard drives on a new level. We will take two "professional" drives from Seagate and Western Digital and four ordinary "desktop" drives for our investigation. The detailed performance analysis and some useful hints on building RAID arrays are in our new detailed article.
I have a pages with high load - load avg is about 10-20.
and in the error_log of the apache message:
[Mon Mar 17 18:10:19 2008] [info] server seems busy, (you may need to increase StartServers, or Min/MaxSpareServers), spawning 8 children, there are 6 idle, and 32 total children
Upon reading http://www.theregister.co.uk/2008/08...sky_black_hat/ it appears those who use network address translation may be vulnerable to DNS cache poisoning even after patching their DNS servers.
"another 15 per cent are still vulnerable to some extent because they use network address translation gear that prevents the patch from working."
I am using dreamhost host 3 of my web sites and 1 blog. Dreamhost is great, offers alot space and bandwidth.
but I think they are oversellling their space, sometimes it gets really slow. (overselling ? ok, I dont really know, but sometimes its really slow, and most my asian readers said need to refresh to load the page. I am wondering if theres a way to check if they are overselling or not.)
I am thinking about buying vps, even tho, I still got 5 month left with dreamhost.
I found 2 vps companies are highly recommanded on this forum, JaguarPC and LiquidWeb.
theres already a post compared both companies in terms of price and service. I say I will pick JagarPc, cuz, its basic plan just 20 USD, and htey got promotion now, its even cheaper. and basic Liquidweb vps plan is 60 bucks.
I am wondering why Jagarpc is so cheap , are they overselling? how can we check if they are overselling.
I found a few posts saying how good jaguarPc is. and they are not overselling, but those members just signed up this month, and only have 1-3 posts. I cannot really trust those new members.
Can someone share their experience with JaguarPC? compare JaguarPc performance and liquidweb performance. antoher question is switch from dreamhost to JaguarPC basic vPS plan, will performance gets better?
last question: VPS account allows 3 IP, 3ip = 3 domains? if not, how many domains can I have?
We are getting into VPS hosting and wanted to get some opinions and feedback as we're quite unsure on what to expect as for performance and how many clients we can generally keep on a box.
For now we've bought 3 dell R710 with dual Xeon L5520, 72GB ram and 8 x 2.5" SAS drives.
We are thinking of a base offering of 512 megabytes of ram and was hoping to get about 40-50 onto a server.
With 40 there should be -plenty- free ram and plenty drivecache.
Then a next offering of 1 gig ram and next one of 2 gigs.
Even if we do the biggest 2 gig offering with 25 on a server we should have free ram to spare.
The software would be virtuozzo.
Any thoughts on this, am I expecting too much, or am I being fairly realistic?
Hi"Optimization of computing resources has long been an important management issue. One of its aspects concerns server scalability and the question of whether an organization should scale-up or scale out.Assume that the computing performance of the servers can be measured by variable 0 <=p, that their total cost is given by "c" and the relationship between server performance and cost is defined by c=αp^β"
a. What is the cost-performance elasticity(ђ), precisely?
b. What would be the range of values for ђ that would be expected by moore's law and what are its implications?
c. What would be the range of values for ђ that would lead managers to scale-out? Draw a graph and throughly explain the implications.
We are a web-based Yacht Charter company, with offices scattered around the world: www.boatbookings.com
Currently, both our web site and our back-office business management system are hosted on a single server in the UK, with an automatic fail-over to a server in Dallas, TX, USA.
The problem we are having is that our sales office in Singapore is having really slow response times and this is very frustrating for them.
Using an application called "JustPing" we see that response times From Singapore are much slower than other parts of the world. (the cities closer to London are fastest, the ones further away are slowest) JustPing Results
Is there anyway to improve this or is hosting our applications on multiple servers the only way to improve performance. What's the best cost-effective method of multiple server hosting?
(Incidentally, if I JutPing Google, response times are fantastic worldwide, but we know they're hosted on many, very large servers)
I was just playing around with litespeed and I thought I would switch back to apache for a few min to see how the server reacts. The load with litespeed was 1.00 - 3.00. I switched to apache the load jumped to 28.00 - 35.00. Its amazing how litespeed is handling connections.
Is there a site which will enable me to enter url of my website and it will simulate visitors from multiple locations. It needs to open the page completely, and run for example 10 minutes. Two things I found are host-tracker, but it just gets headers from multiple locations, and does it only once. Another thing is Paessler software which can test exactly what I want (number of visitors for some period of time with full page download) but it must be run from one (my) PC, so I can not test bandwidth from multiple locations. I need combination of these two, anyone knows for something like that on the net?
My dedicated server is sometimes sluggishly slow. I would like to get a grasp of its performance during a day, to get a better understanding of what's going on.
Therefore I am looking for a server performance monitoring service. All service I found so far were simply monitoring uptime (server is down / server is up). I need something more - a service that checks every 30 seconds or so the loading time of the main page.
Then it would allow me to download the data in CSV or draw a response time graph.
I've recently found (last 6 weeks) that the performance of my client's websites on my trusted host's servers isn't how it used to be and/or how it should be. The download seems to take way much longer than before. So far it hasn't resulted in a significant drop in average pageviews per unique visitor but it might do if things continue. The host claims a number of attacks on the mailserver and an unusually high load at one point in time. Above all it's just annoying - isn't reliable server performance what can be expected from a reputable host?
Here's the question: what's the best way of monitoring online how the web and mail servers are performing so that I can take this report to my host and urge them to take (more) action? Ideally I'd like to compare this with a seperate web server that I use for another client. I don't mind spending a little bit of money but high subscription fees are not in my budget.
[mysqld] datadir=/var/lib/mysql socket=/var/lib/mysql/mysql.sock # Default to using old password format for compatibility with mysql 3.x # clients (those using the mysqlclient10 compatibility package). old_passwords=1
Now-a-days server is having too much load due to http and in access logs we see following message : ======================================== 127.0.0.1 - - [11/Oct/2008:01:40:02 -0700] "OPTIONS * HTTP/1.0" 200 - 127.0.0.1 - - [11/Oct/2008:01:40:03 -0700] "OPTIONS * HTTP/1.0" 200 - 127.0.0.1 - - [11/Oct/2008:01:40:02 -0700] "OPTIONS * HTTP/1.0" 200 -=============================================
And due to this there is load on server. We are not able to understand why this is happening and how to stop this. So please suggest with some solution.
I have LAMP -server running and I was wondering how to test it's performance. So is there any good tools for that? I'm interest how many http queries my server could handle etc.
I am working on a busy and popular website which has a large amount of database activity - and requires hourly backups of all database data.
At the moment the site is hosted on two servers - one for the front end web server, one for the database.
Both servers are running a RAID HDD system which allows quick swaps of faulty HDDs without data loss. An hourly full backup of database tables is running which is killing the server when it runs.
ISP has suggested installing a third server to run as a slave to the existing DB server, and hence always hold a duplicated of the live database.
I have a feeling however that this is basically just like having RAID mirroring, but on a different machine - so to solve the problem of a potential dodgy SQL statement wiping out ALL copies of the live database, we'd STILL need hourly backups to run, and hence would still see the major system speed drop each hour at the time of backup.