if there is any specific way (maybe logs) to see what is using the swap memory of a dedicated server..
its a server with 8GB ram.. it has 60% of memory used, and a constant swap memory usage of 30%.. i thought that in normal conditions swap memory was not used..
I have a server that I have had for a little while now that was runnning perfectly fine. All of the sudden it started just using up a ton of the memory and started using up swap. This has caused the server to slow down and e-mail to stop working. I would upgrade to more memory but I suspect it is someone on the server doing something they shouldn't be. My only reasoning for this is due to the fact that this problem just suddenly arised. I cannot for the life of me determine where this ram usage is coming from.
Here are a few of the errors I have recently got just while logging in..
RIGHT after logging in... id: cannot find name for group ID 0 id: cannot find name for user ID 0 [root@www root]# top top: Unknown terminal "xterm" in $TERM [root@www root]# killall -9 spamd bash: /usr/bin/killall: Too many open files in system
I have added 2 screen shots as well. First one (top1) is what happens once I can get the command "top" to actually run (normally takes quite a few tries). Second attachement (top2) is after I sorted by mem usage...
I'm having an issue with what I believe is Apache. Things will be running nicely, then all of a sudden things will just 'blow up'. Apache seems to fork a bunch of new processes, and it starts ripping into the memory. I've had it essentially eat through all the free memory and swap, and the system will then freeze up as there's no more memory. This is a CentOS server running CPanel, the configuration had been up and running for a long time without problems then this suddenly started happening.
I have a linux server that is only hosting one site , and although the httpd and mysql are working well , then the server load is high and when I restart the server the server load is high till the swap accelerates to 100% and then it uses the memory alot.
I got 5 wordpress and 5 statics website on this server and 100 visitors by 24H00 each day.
Question 1 : Why the memory is so low and the swap so high ? Question 2 : Why i don't find high usage process in top command ? Question 3 How can i resolve this problem ?
I am running a website with a huge traffic on a single server with 2GB Ram and its working fine from last 3 years but from last 1 week the website is being unstable, not all the time but sometimes,
Following are the conditions when the problem occures:-
- When we enter the URL in browser and hit enter, it takes bit long on connecting...before reaching to the website, once the title appear at the top of the browser the page loads normally and while browsing the website, sometimes it gives page cannot be displayed error and we have to refresh the page.
- at that time when I check the result from [url]it gives bit high Average Response Time.
- On the other everything seems to be normal on SSH, load average: 0.90, 0.96, 0.79 Mem: 3116360k total, 2841948k used, 274412k free,
i've recently ran into a problem that i do not have power to solve as it requires my datacenter to fix.
I did like to know how could i verify, monitor and come up with enough data to prove where and that there is a problem within the network so they can at least see and think if they want to keep the problem or solve it at some point ?
Tools that i could use, methods everything is welcome.
I also belive that many that do not have this knowledge will like this topic if any experts could come by and share a little of knowledge.
Is there a way to determine from monitoring the packets coming in to my IP address what domain on my server is being attacked? Something like Tcpdump maybe can tell me? Having DDoS trouble and I'm trying to identify the domain being hit.
Does anyone have experience with the cheap $1 web hosts?
I don't mind paying for hosting, but I only need one script installed on a server that runs a cron file once per day and I don't want to pay $5 to $10 for that.
My question is basically this: is my credit card safe with these $1 web hosts? Can they be trusted to keep my identity safe?
I need a quicker way to find spammers. I've found a decent way to find the scripts, but I want to find heavy offenders by a simple command line or something to identify all scripts sending e-mail in let's say a text document or something.
I have a VPS. And have had an issue both when it was 1Gig and now I recently downgraded it to 768m, because I am moving some sites to a dedicated.
However, the part I am having trouble grasping is that when I look at graphs from Munin, it will typically always show 200-400MB free memory (and free -m and top agrees with munin), but Munin shows 'committed' memory that is above the total Ram on the VPS and once the 'committed' ram exceeds the VPS limit, processes start failing.
So, why is 'committed' memory exceeding the RAM on my VPS, when Munin, free -m and top all show there is free memory available?
Code: root@server [~]# free -m total used free shared buffers cached Mem: 768 449 318 0 0 0 -/+ buffers/cache: 449 318 Swap: 0 0 0 Here's a graph that munin produces that shows the 'committed' memory exceeding the total memory. [url]
What would cause a Linux server to run out of swap? Would it be a memory leak? This happened today to my server and it had to be forcefully rebooted by the data center.
Opcode caches have stability problems with some PHP apps:
[url]
I've been trying to get XCache working with gallery2, right now apache has to restart every 2-4 hours but it's been worse at times and it'd be nice to know what's going on
I noticed that both XCache and eAccel use a small amount of swap space all the time, it grows slowly and if I'm lucky enough to go without a segfault it will reach maybe 4MB. That's not a lot but it doesn't happen at all without a PHP cache (swap stays locked at 72kb unless RAM runs out and the server thrashes).
Here's the error and my settings:
[Tue Feb 19 06:19:55 2008] [error] PHP Warning: Invalid argument supplied for foreach() in /home/cityv4/public_html/gallery/modules/customfield/classes/CustomFieldHelper.class on line 233 [Tue Feb 19 06:19:55 2008] [error] PHP Warning: Invalid argument supplied for foreach() in /home/cityv4/public_html/gallery/modules/customfield/classes/CustomFieldHelper.class on line 233 [Tue Feb 19 06:19:56 2008] [notice] child pid 31028 exit signal Segmentation fault (11) [Tue Feb 19 06:19:56 2008] [notice] child pid 10638 exit signal Segmentation fault (11)
I have purchased several VPS from a provider and found they do not provide swap space with VPS, and even with 256MB ram, I get 'out of memory' trying to compile a perl library... Creation of a swap file by myself doesn't work (operation is not permitted). Hosting providers runs HyperVM.
So the question. Is it common or is it a misconfiguration? For now I got just 'checking on this for you' and three days of silence from their support.
I'm not opening hosting provider name, but I will if they say "You must pay for more RAM", just because other 5 VPS providers support my VPS servers with swap space.
I have talked with the moderators and they have agreed that I can start a new thread to clarify the issue of Xen and swapping as long the discussion remains technical.
For people who are curious, I would first like to explain why this is important. What we have here is someone making a specific technical accusation against Xen, and if it is indeed crucial, it needs to be solved, or otherwise people have to know it before they get into Xen.
Claim Number 1: The original claim is that users can create arbitrarily large swap and this can lead to the equivalent of overselling.
Fact: Arbitrarily large swap has absolutely zero effect on a normal system, since Linux treats swap as an auxiliary storage, and will not use it unnecessarily. Linux will always use the RAM to the full, but swap is used only when all the buffers have been cleared. If you run free on a dedicated hardware, you will see that swap usage is most of the times zero, even though you have assigned a very large swap to the system. In fact, you can try this by simply increasing the swap to very huge value, you will see that Linux will ignore it completely.
For large swap to cause a problem, the user not only has to assign a large swap, he has to run a really huge workload on his limited RAM, and this being physically impossible, will cripple his vps long before it has any serious impact on the host.
I think the above has been agreed to by the person who made the initial claim.
Claim Number 2 Merely having a swap can lead to vps getting slowed down.
Fact ; The use of swap does not mean thrashing .
Thrashing is a technical term, and it means that the application is registering a swap hit every few instructions. This is rare in normal systems, because of a property of programs that execution is always localized. That is, at any particular time, a certain portion of the program will be continuously being executed, and other portions would be idle. Linux has algorithms that will swap out the Least Recently Used page, and this will mean that the system will not run into too many swap hits.
Now again, like in the earlier case, there will be trouble if the vps customer is trying to run a 1GB workload in a 64MB RAM. This will actually cripple his vps, which is what's the right thing to happen. So normal usage of Swap will not lead to disk I/O, since Linux has explicit algorithms to reduce swap usage.
So for minor OverUsage of memory:
For virtuozzo: the application will crash
For Xen: there will be a very minor degradation of quality. And sometimes it won't have any affect at all, since as I said, just because swap drive is non-empty doesn't mean that linux is constantly swapping out.
So summarizing:
a) Inordinately large swap has zero effect on Linux
b) Non-empty swap doesn't mean that the system is registering swap hits. For normal workloads, the swap hits will be very minimal.
And an extra advice is that if you are using Xen, don't ever use snapshoting, as it will double the disk I/O, and the worst thing is that the vps that's causing this will not even be penalized. The overhead will be completely borne by the system.
I've come across a XEN based plan that offers additional 256MB swap space, and a few others Virtuozzo based plans that offer bursting capability (some even upto 8GB).
Which would be the better option?
Also, in Virtuozzo, will SLM allocation be significantly more beneficial than UBC for a small VPS?
Another question ... DirectAdmin officially claims to be able to run on a minimum of 64MB RAM. How would it perform on a 128 MB VPS ? I'm not really looking to do much, only to host a few (5-6) small sites along with some other non-webhosting applications (which is the reason for getting a VPS instead of a reseller).
I've seen most people suggest atleast 256MB for DirectAdmin, but that is beyond my budget. (I'm also seriously considering the option of employing vi as a control panel to further conserve my limited resources)
I have a handful of machines running Rails on CentOS 5, each with 8GB RAM. They're monitored with Nagios, so I get paged if we exceed 80% memory usage usage or 2% swap. FWIW, they're primarily dual x. dual-core Opterons, and we're using PAE.
The problem is that I keep getting paged in the middle of the night for situations where we're using 2-20% swap yet there are still several gigs of RAM free.
On some level, this makes sense -- if nothing is using a big chunk of RAM, we might as well move it out to disk. On the other hand, when there's still several gigs of RAM free, I don't see any point in bothering, especially when it causes me to get paged at 2am.
So my question is threefold:
Is there any easy way to see what's using swap? top can be made to show a swap column, but it's not what you'd think it is. (For example, most processes show as using lots of swap, and the sum vastly exceeds the actual size of our swap partition.) Can I get a list of what processes are actually swapped out to disk?
Is this a problem other people have? Are others just less aggressive in getting alerted at swap usage, or does this not happen to other people? Our setup is probably not too common (Opterons, PAE, adn 8GB on an older CentOS release?), which makes me think there's a small chance a bona fide bug plays into this somehow.
Is this a case for playing with the swappiness sysctl to make the machine less obsessed with swapping things out to disk? And if so, is there any good documentation on what the 0-100 value actually means, beyond the one sentence explaining that 100 is most likely to swap and 0 is least? Does the current setting of 60 "mean" anything, for example? 98% of the stuff online is just a mirror of the e-mail thread in which Andrew Morton and others go back and forth about what setting is ideal, which doesn't help and is tiring to read over and over.
i have a problem with my server it is a Xeon with 2Gb ram, i have a igh swap usage and when it reach the size of 4gb that i have set it go in kernel panic, this is the actual value