I try to search and not found the solutions to limit concurrent of db users to connect to MySQL in same time for example:
Limit user root to connect to MySQL maximum 5 connections in same time
There have max_connections setting but it limit for all users in all connection to MySQL.
But I found that in version 4.1.x there have max_connections property in user table (Users and global privileges) of mysql database, this setting is limit connection in 60 minutes.
I would like to know if there any setting or code modify for this solution that I want.
I'm on a vps using Dell PowerEdge Servers and 750 GB of transfer.
I have concurrent users settings to where i now have them at "50" and I'm wondering how much higher i can go without troubles.
My app is a perl cgi script in flat files and 95% of visitors will pass through my domain without seeing my webpage. They are only there approx 1 second, maybe 2 as they are quickly redirected from this platform.
My host has a custom solution installed to check for apache stalls every few minutes and it will restart apache when it sticks - and it will stick from time to time.
Just now i do about 75000 per day but plan to double to 150,000 next week and soon after will triple to 225,000 visitors per day.
Should i increase concurrent users now in prep for the increase to 150k or at what level.
Please give some suggestions for concurrent settings at future traffic levels. This host has a multiple server set up in case one fails - i should be good, i think.
Like the topic suggested, anybody have a rough estimate of how many concurrent users game websites such as wildtangent and instantaction have? Also, what kind of server configuration would be necessary to support those kind of users assuming you have 1 server for web and 1 for database.
I have tried forking anywhere from 50 to 300 to 1000 fastcgi children with PHP_FCGI_CHILDREN (through lighttpd, I'll attempt to do with same with spawn-fastcgi and fpm)
I am using to lighttpd's lighttpd-status to estimate concurrent connections.
When I refresh the panel, it shows that there are around 100-150 connections and around 150 requests/s in the last 5 seconds.
My vmstats show that CPU is 98% idle. Blocks written/read is neglible. MySql key_buffer set to 2gb and I'm pretty sure it's not mysql. The overwhelming majority of requests do not access mysql.
EDIT: Uh oh, I just realized that tcp_mem could be a huge bottle neck.
I just set it to:
net.ipv4.tcp_mem = 4096000 87380000 4194304000
It was previously: net.ipv4.tcp_mem = somenumber somenumber 393,216 <<<--- WTF!
x1000 for my read values (it's an access server only). I can't benchmark the server right now so let me know if you have any suggestions besides this. I do think that this was the problem. When under load images could not be accessed either.
Is there any option to limit users only to upload (No other permissions like permission change, directory access etc, only upload) in FTP.. I am using PureFTP..
I have seen options to do it in ProFTP, like below,
e.g. I create 5 users and each of them uses the ssh server as a proxy. I want each user to create only 5GB traffic a month. How can I do this? I've already tried iptables but they can only limit the traffic created by the user who started the ssh server
i dont want clients taking the servers I/O and server load over 4.00 when they do major update ect... querys on sql. is there a way to limit the ammount they can do?
I been having a heck of a time just with this one cpanel server and open files limit. At first using open_files_limit did not work so I changed it to open-files-limit that seemed to work but now it rejects the value and sets its down to 65535.
Then system open files limit is 500000. I try to set it to any value about 65535 in my.cnf and here is usual error:
090630 9:32:07 [Warning] option 'open_files_limit': unsigned value 120510 adjusted to 65535 090630 9:32:07 [Warning] option 'open_files_limit': unsigned value 120510 adjusted to 65535
When I run something like the tuning-primer it shows:
Current open_files_limit = 120510 files
The open_files_limit should typically be set to at least 2x-3x that of table_cache if you have heavy MyISAM usage.
Your open_files_limit value seems to be fine
But Im not sure if it is just reading my.cnf or something. I am still getting complaints from users about lost connections and I see the errors in error log. Ive looked everywhere and cant seem to find a solution to this.
I am trying to increase my open file limit on my CentOS release 4.6 VPS.
I have tried modifying my /etc/security/limits.conf with the following: @mysql soft nofile 4096 @mysql hard nofile 4096 @mysql soft sigpending 4096 @mysql hard sigpending 4096
Even is a mannual set it with "ulimit -n 4096 mysql" it will not stick.
Any ideas what I can do to get this limit increased permanently?
I have a few users that have a hosting account with me and are trying to connect to their own MySQL servers elsewhere. Both said that they are able to connect to the MySQL server with other hosting accounts but not the ones hosted with me. I asked if they had setup permissions to allow remote connections from my server and they confirmed that they did.
What am I missing? Do I need to configure my firewall somehow? I figured the connection is just using port 80 since it's from an PHP script right?
The upgrade has an error when manage the users database.
PRODUCT, VERSION, VERSION OF MICROUPDATE, OPERATING SYSTEM, ARCHITECTURE OS Microsoft Windows Server 2008 R2 Service Pack 1 x64 Panel version 11.5.30 Update #13, last updated at Sept 1, 2013 03:30 PM
PROBLEM DESCRIPTION In a costumer panel have a one database MSSQL, and assign to this DB 3 users, but the tab option "Users" don't work fot his costumer and show this error:
Error Javascript: TypeError: template is null this.template = template.toString(); in protototype.js 8472831 (lÃnea 807)
ACTUAL RESULT Error Javascript: TypeError: template is null this.template = template.toString(); in protototype.js 8472831 (lÃnea 807)
EXPECTED RESULT Show users in the tab users for database.
i have a vps, and im current use lighttpd, but i want move to litespeed Standard. And i see they limit Max Concurrent Connections =150 on standard version.
but what is Max Concurrent Connections? where can i find it.
and is it the number connection via port 80 (netstat -nt | grep :80 | wc -l)?
On my server, users can connect to any database as long as they have the database user and password. This makes it easier to hack any database on the server. What I want to do is to make the users can only connect to their own databases and not other's.
I tried changing the localhost ip address but it didn't work ( I assume I didn't do it the right way)
I was looking at some load balancers hosting companies offer and some of the load balancer specs say they can handle up to 15million concurrent sessions(users online at the same time), so does this mean if i had a site like wikipedia that had 15 million users online at the same time, would i be able to do this with only 2 dedicated servers, or will the Cpu's not be enough?
Is it possible to run two concurrent VPN connections using the Linux Cisco VPN client? We need to have one server be able to communicate over two different vendors gateways.
Do we need to possibly add an additional NIC to the server first?
When we try to connect to another profile we get the error:
Initializing the VPN connection.
Secure VPN Connection terminated locally by the Client
Currently I use WHM/cPanel latest on a few boxes, I have (at least to my knowledge) copied the exact compile settings between servers, yet some of my servers allow me to do concurrent file downloads (using DownThemAll or another "accelerator") whilst others only let me make 1 connection. These servers also don't show me the final filesize while downloading. So my question is, what did I set/unset to cause this to happen? It's been bugging me for a month or so now and no amount of recompiling has fixed the issue (aside from a clean install of the OS/cPanel). I can provide a list of modules and other config settings if required.
I am planning to get a Juniper firewall, but due to SSG140 has a maximum of 48,000 concurrent sessions per second, so it triggers me how do I measure the concurrent session of a linux server of the total throughput instead of just port 80?