Hi"Optimization of computing resources has long been an important management issue. One of its aspects concerns server scalability and the question of whether an organization should scale-up or scale out.Assume that the computing performance of the servers can be measured by variable 0 <=p, that their total cost is given by "c" and the relationship between server performance and cost is defined by c=αp^β"
a. What is the cost-performance elasticity(ђ), precisely?
b. What would be the range of values for ђ that would be expected by moore's law and what are its implications?
c. What would be the range of values for ђ that would lead managers to scale-out? Draw a graph and throughly explain the implications.
i am having quite problem with my server i cant get it to optimize it. During peak times the load goes beyond 10 sometimes 20 to 50 sometimes its just under 1. I have my webserver running on P4 Core 2 6300, 1 GB ram, 250 HDD, on CentOS 4.4 (kernel 2.6.9-42.0.10) etc. It has cpanel/whm running Apache 1.3.7, php 4.4, mysql 4.1, mailscanner, eaccelerator, etc i dont know really what other info to show you.
I host about 200 accounts and there are 4 large php forums (not like WHT) running on it. Sometimes pages stop loading especially these forums have to refresh to get page load.
During day there are approx 700-900 connections.
Previously i had phpsuexec and the load never went beyond 4-5, i have removed it to install eaccelerator.
I hope some of you folks could give me some pointers here. I have a server that is used for a gaming clan's purposes. Nothing major as far as games go, but what I got I want running at peak efficiency.
The most important is the Counter-Strike source. I really need the kernel frequency to be tweaked for this. The number one question is can I merely modify the existing config file and recompile that or do I need to start all over? If I could do that is it something like make old or something to that matter?
I am running a couple other servers as well, but there is never a time where all the servers have anyone in them. I keep everything running. After installing some weird stuffs here and there I got the memory usage to go down. I think. Regarding a scenario of having more than one instance of a game, is there any way to share memory between them?
My Swap. Is there a way to force idling games to move to it instead of using physical memory?
The kernel is what I really need to take care of. I'm still somewhat of a n00b and have just switched from centos to debian to top it off. So any tips would make my day. Thanks for reading me.
I'm on a dedicated server and the httpd service of the server keeps going down which results in downtime for the number of hours until i dont realise that the site is down.
So when I realise the site is down, I ask the hosting support to look into the issue, they tell me the server loads are high and ask me to upgrade even when I just upgraded the server last month.
Generally the forum has around (250 - 300 members) + guests visitors online and is in the alexa top 20k sites.
The first thing I would want to do is hire a person who specializes in server optimization but then I have heard a lot of bad things like the server admin stealing databases and selling it in open market.
The second thing I wanted to do is contact vbulletin for server optimization but they need me to give them server specs and I dont know how to obtain them.
I have purchased a proxy website using phproxy script hosted at dedicated server. Its now about 4 months and everything is fine but i want to move my site to another dedicated server provider because current one has some issues.
When i purchased that site seller told me that he has optimized server and installed some scripts {server scripts} to get best out of proxy website. Now seller is out of my contact.
I am very much pleased with current server settings but i need to move. So i want to make sure that i can optimize new server and install same scripts to keep going with current performance.
I only know about http.conf settings and i have noted it ... plz help me about copying every other settings which that guy had done to optimize the server and how can i know which and where those scripts are installed which has made server quite efficient?
Can anyone recommend a good company or someone that would be able to optimize Apache and MySQL on my server? I've slightly configured Apache so that it works well with my server so I'm mainly looking for someone to optimize MySQL (since it's using a lot of cpu).
The server load is pretty high so I'm hoping to reduce it before having to upgrade to a new server.
i run a fedora server with plesk 8 on it i was wondering if some one can help me to find the best way to optimize the databases on the server at once my servers contains many clients with unicode data in the database.
I'm sure by now most of you have read the Web Server Optimization Guide by Shaw Networks. It came in handy for me when I was first starting out, I reduced my load & memory usage by tonnes. I thought I would make a new thread with an updated How-To.
[mysqlhotcopy] interactive-timeout To Save: CTRL-X Restart Service: "service mysqld restart" or "service mysql restart" or "/etc/rc.d/init.d/mysql restart"
Use Pico (pico /usr/local/apache/conf/httpd.conf) or Download via WinSCP for editing, Change the following settings in your httpd.conf...
Set "Timeout" value to "Timeout 60" Set "KeepAlive" to "KeepAlive on" Set "KeepAliveTimeout" to "KeepAliveTimeout 3" Set "MinSpareServers" to "MinSpareServers 16" Set "MaxSpareServers" to "MaxSpareServers 32" Set "MaxRequestsPerChild" to "MaxRequestsPerChild 256" Set "HostnameLookups" to "HostnameLookups Off"
Note: These settings will not work under all server environments its recommended that you tweak around with the numbers until your web server is running 100% please read Apache documentation before changing any settings so you know what you are changing [url]
To Save: CTRL-X Restart Service: "service httpd restart" or "/etc/rc.d/init.d/httpd restart"
Installing eAccelerator: eAccelerator is a further development from mmcache PHP Accelerator & Encoder. It increases performance of PHP scripts by caching them in compiled state, so that the overhead of compiling is almost completely eliminated. [url]
If you run CPanel please visit [url] for an auto-installer which will do all the hard work for you.
Run the following in SSH.
Code: cd /
mkdir ea
cd ea
wget [url]
bunzip2 eaccelerator-0.9.5.tar.bz2 tar -xvf eaccelerator-0.9.5.tar
So the site got featured on [url]and now the server is drowning...
The Coppermine Gallery usually hovers around 30~50 users daily and now, 1800, and im at a lost as how I should configure mysql to take on such a load. right now it takes about 10 secs or more to load a page and sometimes it would time out. Because it si coppermine, all pages are dynamic and can't be cached -_-"
Here's the my.cnf right now after i played around with the numbers
server spec Opteron 170 (2ghz) 2gb ram 250 7200rpm
# # KeepAlive: Whether or not to allow persistent connections (more than # one request per connection). Set to "Off" to deactivate. # KeepAlive Off
# # MaxKeepAliveRequests: The maximum number of requests to allow # during a persistent connection. Set to 0 to allow an unlimited amount. # We recommend you leave this number high, for maximum performance. # MaxKeepAliveRequests 100
# # KeepAliveTimeout: Number of seconds to wait for the next request from the # same client on the same connection. # KeepAliveTimeout 15
# prefork MPM # StartServers: number of server processes to start # MinSpareServers: minimum number of server processes which are kept spare # MaxSpareServers: maximum number of server processes which are kept spare # ServerLimit: maximum value for MaxClients for the lifetime of the server # MaxClients: maximum number of server processes allowed to start # MaxRequestsPerChild: maximum number of requests a server process serves <IfModule prefork.c> StartServers 8 MinSpareServers 5 MaxSpareServers 20 ServerLimit 200 MaxClients 200 MaxRequestsPerChild 1500 </IfModule>
# worker MPM # StartServers: initial number of server processes to start # MaxClients: maximum number of simultaneous client connections # MinSpareThreads: minimum number of worker threads which are kept spare # MaxSpareThreads: maximum number of worker threads which are kept spare # ThreadsPerChild: constant number of worker threads in each server process # MaxRequestsPerChild: maximum number of requests a server process serves <IfModule worker.c> StartServers 2 MaxClients 150 MinSpareThreads 25 MaxSpareThreads 75 ThreadsPerChild 25 MaxRequestsPerChild 0 </IfModule>
I have a fairly high end server in which I have installed SIM. SIM is restarting Apache up to 10 times a day, presumably due to high load causing un-availability.
On restart, Apache / MYSQL is stable until the load / mem usage begins to climb then it is restarted again. Here are my 'load' stats for today:
Load for today High (2:18am): 4.63 Low (3:30am): 1.20 Mean: 1.84 Latest: 1.61
Mem usage for today High (1:36am): 9,192.9 MB Low (1:48am): 7,995.7 MB Mean: 8,683.1 MB Latest: 8,781.7 MB
I have seen it using 20GB RAM before.
I have tried to follow various optimisation guides but these seem tailored to less powerful servers.
The web application I run on this server is almost entirely MYSQL based, with thousands of DB calls a day. Across the entire system I probably get 200,000 bot hits per day or even more. At peak times search engine bots are literally hammering the server.
I'm currently in the process of installing a mysql database server on my dedicated game server box. I was using shared hosting before, but i got a complaint from them that i was using too many connections and memory so i have to host the database myself.
Here's the stats of the server:
Dual Core Dual CPU Intel Xeon 3.0Ghz (2 core per cpu & 2 cpu's) 4GB RAM 120GB SATAII Hard Drive Windows 2003 Enterprise Edition SP2 (180 Day Trial)
Currently, im hosting around 10 game servers on the machine with around 30% CPU load. I'm running with the page file disabled (its using 2 GB of RAM instead). Server runs fine like this (i heard its a performance boost, and it is). Over 2GB of ram are free at all times even when most of the game servers are full.
I want the database server to be able to as many connections are need from the game servers but at the same time i dont want to affect performance of the game servers. Any suggestions on what settings i should use?
P.S. The SQL will have 3 databases (AMXBans, SourceBans, & a RP character db). Everytime a player connects to one of my servers, their information is checked against the database if they are an admin. So in total i have around 126 slots for players to connect on.
Can anyone recommend a Apached and MYSQL optimization expert? I just some optimization done on mysql and apache upgraded... but pache wouldn't restart with the following entries:
if i remove above then it starts... but my CPU spikes and I am getting a ton of errors emailed to me:
Invalid SQL: SELECT COUNT(DISTINCT(userid)) AS count FROM vb_session WHERE vb_session.userid>0 AND vb_session.lastactivity>1245294346;
MySQL Error : Out of memory (Needed 8388580 bytes) Error Number : 5 Request Date : Wednesday, June 17th 2009 @ 09:39:06 PM Error Date : Wednesday, June 17th 2009 @ 09:39:06 PM
Its constantly complaining about this same query and i thought i removed all of the vbulletin options...
i am pulling my hair out... cuz my server isn't swapping, ther are 0ver 8GBs in cache.... but my CPU goes up
Here is original problem thread...
[url]
2nd post has recommended config... and my subsequent probs
We just upgraded to VBulletin 3.6.5, and are experiencing strange behavior. My forums are on a dedicated server, 27,000 users.
We'll be running along fine with loads of .5-2.00, etc. Then suddenly, the loads start climbing to 50, 70, 85+. We've been trying to figure out why. We even went on a different (much more robust) server, still the same result. This only started happening after the upgrade. Restarting the server corrects the problem, but only until it happens again. It can happen at anytime - during peak or off peak. The server may run for an hour or two until this happens, or it can happen 10 minutes after a restart.
Late last night when loads were normal we benched the server hard, and it ran just fine. We just can't figure out where this load spike is coming from.
Server info below:
Servers: Mysql: server version: 5.0.27-log PHP: PHP 5.1.6 (cli) Zend Engine v2.1.0, Apache Server version: Apache/2.2.3 Kernel: 2.6.19-1.2911.fc6 #1 SMP
Server: Pentium III with 1 gig of ram 15k RPM SCSI Raid 5 1 Gig RAM
I am using dreamhost host 3 of my web sites and 1 blog. Dreamhost is great, offers alot space and bandwidth.
but I think they are oversellling their space, sometimes it gets really slow. (overselling ? ok, I dont really know, but sometimes its really slow, and most my asian readers said need to refresh to load the page. I am wondering if theres a way to check if they are overselling or not.)
I am thinking about buying vps, even tho, I still got 5 month left with dreamhost.
I found 2 vps companies are highly recommanded on this forum, JaguarPC and LiquidWeb.
theres already a post compared both companies in terms of price and service. I say I will pick JagarPc, cuz, its basic plan just 20 USD, and htey got promotion now, its even cheaper. and basic Liquidweb vps plan is 60 bucks.
I am wondering why Jagarpc is so cheap , are they overselling? how can we check if they are overselling.
I found a few posts saying how good jaguarPc is. and they are not overselling, but those members just signed up this month, and only have 1-3 posts. I cannot really trust those new members.
Can someone share their experience with JaguarPC? compare JaguarPc performance and liquidweb performance. antoher question is switch from dreamhost to JaguarPC basic vPS plan, will performance gets better?
last question: VPS account allows 3 IP, 3ip = 3 domains? if not, how many domains can I have?
We are getting into VPS hosting and wanted to get some opinions and feedback as we're quite unsure on what to expect as for performance and how many clients we can generally keep on a box.
For now we've bought 3 dell R710 with dual Xeon L5520, 72GB ram and 8 x 2.5" SAS drives.
We are thinking of a base offering of 512 megabytes of ram and was hoping to get about 40-50 onto a server.
With 40 there should be -plenty- free ram and plenty drivecache.
Then a next offering of 1 gig ram and next one of 2 gigs.
Even if we do the biggest 2 gig offering with 25 on a server we should have free ram to spare.
The software would be virtuozzo.
Any thoughts on this, am I expecting too much, or am I being fairly realistic?
We are a web-based Yacht Charter company, with offices scattered around the world: www.boatbookings.com
Currently, both our web site and our back-office business management system are hosted on a single server in the UK, with an automatic fail-over to a server in Dallas, TX, USA.
The problem we are having is that our sales office in Singapore is having really slow response times and this is very frustrating for them.
Using an application called "JustPing" we see that response times From Singapore are much slower than other parts of the world. (the cities closer to London are fastest, the ones further away are slowest) JustPing Results
Is there anyway to improve this or is hosting our applications on multiple servers the only way to improve performance. What's the best cost-effective method of multiple server hosting?
(Incidentally, if I JutPing Google, response times are fantastic worldwide, but we know they're hosted on many, very large servers)
I was just playing around with litespeed and I thought I would switch back to apache for a few min to see how the server reacts. The load with litespeed was 1.00 - 3.00. I switched to apache the load jumped to 28.00 - 35.00. Its amazing how litespeed is handling connections.
Is there a site which will enable me to enter url of my website and it will simulate visitors from multiple locations. It needs to open the page completely, and run for example 10 minutes. Two things I found are host-tracker, but it just gets headers from multiple locations, and does it only once. Another thing is Paessler software which can test exactly what I want (number of visitors for some period of time with full page download) but it must be run from one (my) PC, so I can not test bandwidth from multiple locations. I need combination of these two, anyone knows for something like that on the net?
My dedicated server is sometimes sluggishly slow. I would like to get a grasp of its performance during a day, to get a better understanding of what's going on.
Therefore I am looking for a server performance monitoring service. All service I found so far were simply monitoring uptime (server is down / server is up). I need something more - a service that checks every 30 seconds or so the loading time of the main page.
Then it would allow me to download the data in CSV or draw a response time graph.
I've recently found (last 6 weeks) that the performance of my client's websites on my trusted host's servers isn't how it used to be and/or how it should be. The download seems to take way much longer than before. So far it hasn't resulted in a significant drop in average pageviews per unique visitor but it might do if things continue. The host claims a number of attacks on the mailserver and an unusually high load at one point in time. Above all it's just annoying - isn't reliable server performance what can be expected from a reputable host?
Here's the question: what's the best way of monitoring online how the web and mail servers are performing so that I can take this report to my host and urge them to take (more) action? Ideally I'd like to compare this with a seperate web server that I use for another client. I don't mind spending a little bit of money but high subscription fees are not in my budget.
[mysqld] datadir=/var/lib/mysql socket=/var/lib/mysql/mysql.sock # Default to using old password format for compatibility with mysql 3.x # clients (those using the mysqlclient10 compatibility package). old_passwords=1
Now-a-days server is having too much load due to http and in access logs we see following message : ======================================== 127.0.0.1 - - [11/Oct/2008:01:40:02 -0700] "OPTIONS * HTTP/1.0" 200 - 127.0.0.1 - - [11/Oct/2008:01:40:03 -0700] "OPTIONS * HTTP/1.0" 200 - 127.0.0.1 - - [11/Oct/2008:01:40:02 -0700] "OPTIONS * HTTP/1.0" 200 -=============================================
And due to this there is load on server. We are not able to understand why this is happening and how to stop this. So please suggest with some solution.
I have LAMP -server running and I was wondering how to test it's performance. So is there any good tools for that? I'm interest how many http queries my server could handle etc.
I am working on a busy and popular website which has a large amount of database activity - and requires hourly backups of all database data.
At the moment the site is hosted on two servers - one for the front end web server, one for the database.
Both servers are running a RAID HDD system which allows quick swaps of faulty HDDs without data loss. An hourly full backup of database tables is running which is killing the server when it runs.
ISP has suggested installing a third server to run as a slave to the existing DB server, and hence always hold a duplicated of the live database.
I have a feeling however that this is basically just like having RAID mirroring, but on a different machine - so to solve the problem of a potential dodgy SQL statement wiping out ALL copies of the live database, we'd STILL need hourly backups to run, and hence would still see the major system speed drop each hour at the time of backup.
I am currently hosting my website on one server with the specs:
2.8ghz Dual Quad-Core processor + 8 gigs of ram + two 500 hard drives with a 50 mbps unmetered bandwidth package.
My current problem lies in high server loads and very slow server performance throughout the day.
I am considering migrating over to The Planet onto server with the specs:
3.0ghz Dual Quad-Core + 18 gigs of ram + two 50gb hard drives with 2TB of monthly bandwidth transfer.
In an attempt to have great bandwidth pricing and server performance, I plan on downgrading my current server with my current host to a lowe-end server and keeping it only to host my VIDEO and MUSIC files with the 50mbps unmetered package. The Planet will then host my database and all other web related files on their new server.
Is this a good idea as an attempt to save money in bandwidth costs and eliminating my server lag issues?
I was offered a setup of a separate web and database server at my current host but from what I have read, no one touches the performance and reliability The Planet has to offer.
I have a Windows 2000 Advanced Server where there's a performance issue with some of the .asp pages that retrieve data from Access databases, (I know Access databases aren't ideal for data). These pages will just get stuck/freeze, and then either suddenly spring back to life, or give a script timeout error 0113.
The largest Access database I've seen is 136MB (is that way too large?)
I will probably move some of the large Access databases onto a different server but before I do:
- Are there any tools you can recommend to diagnose exactly what files / databases are causing the problem. I don't think the Win 2000 performance monitor tools even work.
- Can anyone explain more about the technicalities behind this issue. I expect it has something to do with processes, threads, memory, Access drivers being loaded into memory etc etc. Can anyone tell me what they know to put me in the picture better?
I have centos -7 and apache 2.4.6. I have made one small php scripts and measuring the performance of server.
Average time is around 15 ms for per request. But when adding concurrent requests the average time increase too much high for the request see the below output
ab -n 10000 http://127.0.0.1/1.php
Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.0 0 0 Processing: 14 17 2.6 15 35 Waiting: 0 1 0.2 1 10 Total: 14 17 2.6 15 35
With Concurrent connections
ab -n 1000 -c 100 http://127.0.0.1/1.php See the processing the mean request time its around 150 ms.
Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 1.1 0 5 Processing: 25 150 24.7 153 213 Waiting: 2 134 25.2 138 194 Total: 29 150 24.1 153 216
ab -n 1000 -c 100 http://127.0.0.1/1.php See the processing the mean request time its around 150 ms when total request is 1000 only.
Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 1.1 0 5 Processing: 25 150 24.7 153 213 Waiting: 2 134 25.2 138 194 Total: 29 150 24.1 153 216
How to configure the Apache so the request time decreas to low ms ???
In Centos -7 Default configuration of MPM Directory also not found ??? checked in httpd.conf and other file also.