where do you go host HUGE websites, youtube like sites, with HUGE bandwidth usage?
I don't believe people go on host like rackspace, with their 150GB / month packages, unless they want to pay an absurd amount of $$.... so where do these guys go to host? What kind of hosts are these?
I have three domains that I want to host myself on my own computer. Two are temporary sites that I only put up on Memorial Day weekend for the Detroit Electronic Music Festival or Movement 09 to direct people to my after parties and the other site is a two page bio site for myself. I have a very tight budget this year and I can't afford to pay for web hosting all year round when these are only temporary sites even though I own the domain names all year round. I do know some of the promoters for the festival do the hosting themselves.
I currently host my own website with windows server 2003, it is a long ip address. I use a DNS redirector service to make it a real domain name. I was wondering if I can host more than one website?
Does any web host give free dedicated server/cluster hosting to non-profit or famous websites? I'm just wondering coz if I made a site and it got 1 million visits per day, who do you think would give me free dedicated cluster server hosting? Or Would only a rich company like google or microsoft be able to do that?
I'm seriously considering the LVPS-2 plan at SWVPS: 10GB space, 200GB bandwidth, 512MB RAM (burstable to 768MB). I have four sites hosted with MediaTemple but I want to give VPS another shot since it provides full root access. MediaTemple is hosting me fine, but I keep running into roadblocks with things that involve custom compiling. I'd really like some more flexibility.
So how easy would it be to add 4 domains to one VPS plan and have them all in their own subdirectory? Also, two of these domains would have one e-mail address, so I'd need a mail server along with a FTP server. I know these two to be slight resource hogs from my previous VPS experiences, as they both caused me to go over my limit (which was, I think, 256MB).
I sent SWVPS an email but all I got was a suggestion to get a dedicated server. These days it's hard to tell a genuine recommendation from someone trying to milk another $40/mo out of you. I really don't think I need a dedicated server for these sites, as I've hosted much more than that on shared servers (granted, that's a completely different area). I also can't afford dedicated hosting. This $20 price point for LVPS-2 is convenient, because it's the same price at MT, which I'm barely skimming by on.
Site #1 is a personal blog with one email account, powered by PHP and MySQL.
Site #2 is a company website and blog, powered by RoR, but not much database load. It would, however, run Subversion and browse it with Warehouse, a RoR SVN browser. This would probably be the main offender as everything SVN would be managed and run from here.
Site #3 is a software website in RoR, all downloads would be from Site #2's SVN server. It has forums, a bug tracker, a plugin/theme browser, and a development blog, and everything else is just (R)HTML.
Site #4 is another personal site, with one email account.
Just moved to a new server, and of course, 10GB doesn't seem that large for a server but for some reason wget is not able to handle the transfer of that backup for me... it transfers about 1MB then tells me "successful transfer..."
The old server is using cPanel, and the new server is just a plain old server that I haven't loaded up yet.
how I can get this full backup over to the new server?
I'm sure this question has been asked before, but I'm looking for a nice and simply way of breaking up log files into smaller chunks.
I've been running apache2 on a VPS for the past few months and one of the access.log files is now 700mb big... bit of a waste of space. I'm currently just doing:
I've been using mod_security for a long time, but apparently I accidentally enabled some kind of log or something that uses mysql. I don't remember it being there before.. but the point is; the database is like 145100k!
I had several user accounts that were pushing their quota. I was digging around in SSH and found that the INBOX file in /home/username/mail was huge even though the user does not keep messages on the server. I deleted this file to free up space and all seems file. A couple seconds later I did check and the file was recreated with new incoming mail.
My question is how do I keep this file from growing out of control? One of the users I had for almost 2 years had an INBOX file of almost 2GB!
Server Details: VPS running WHM 11.23.2 cPanel 11.23.3-R25623 Redhat 9
I'm just curious as to what kind of things the huge sites--Youtube, Myspace, etc.--are doing to try to keep scalable. What sites do you guys just hate for failing in this regard, and perhaps most importantly, what are some ways we can prevent downtime?
I have been receivig a huge logwatch report, seems that logwatch is not parsing the /var/log/secure file, but sending the log entries instead of any resume of it. I got thousands of lines like
Cp-Wrap: Pushing "47 GETDISKUSED pvargas lights.com.co" to '/usr/local/cpanel/bin/eximadmin' for UID: 47 : 25 Time(s) Cp-Wrap: Pushing "47 GETDISKUSED r.perez konecrans.com" to '/usr/local/cpanel/bin/eximadmin' for UID: 47 : 69 Time(s) Cp-Wrap: Pushing "47 GETDISKUSED r.rodriguez konecrans.com" to '/usr/local/cpanel/bin/eximadmin' for UID: 47 : 114 Time(s)
I have upgraded to the most recent version of Logwatch with default configuration. Any ideas on what could be wrong?
ways to improve the database performance in the situation when I have to modify a large table (several million rows), by e.g. adding a column. Currently this would take several hours which is too slow. The bottleneck is disk I/O. I am considering either partitioning the table over several innodb files on several disks, or going to a RAID-5 or RAID-10, it this will give me better write performance.
The database is 130GB large, and the problem table (which I make period changes to) is the largest table on the server. I cannot have downtime of 3 hours each time I make a change and adding blank fields (to be used later, when a new field is needed) is not an option.
Each time I add a column, the cpu goes into high (80%) io wait state for about 3 hours.
I have a hack which would allow me to split the large table into multiple smaller tables based on some criteria (for example, forumID or such). Here are a couple of things but would like to know which is best, and am open to new ideas. The ideas so far:
1. Split the table into 3 or 5 smaller tables each on it's own disk. The disk IO would then not be so bad, and it might only take 1 hour to perform the table change. But this might not work because the changes to the database (as in adding a column) might be serial, meaning only 1 disk is being written to at a time. (Then again, maybe it will work if I launch 3 different scripts, one to update each table at once).
2. Do RAID 5 or 10, and have 3 or 5 disks. This again might not help at all because of the above issue with MySQL writing serially.
I am using latest MySQL 5.0.45 with InnoDB engine on Debian etch Linux
I have one domain where is hosted a lot of subdomains,and for some reason it constantly have 4% cpu usage and 33% mem usage.Since that domain is inactive,could be that usage beacuse of addon domains but it simply not presented correctly in whm?
I had 18GB bandwidth.log file at /etc/log/ directory? What is the meaning of bandwidth.log file? And what may be reason increasing file size to 18GB, especially in one night.
I have done my research, befriend a few super proxy webmasters, and learned everything I need to know about being successful in the proxy business. So I am selling almost all my websites to fund this huge project. I will also be flipping proxies from time to time to fund the project even more. This will be a year long project and will be my full time job sooner or later. My goal is to have 1,000 proxy sites.
So with this knowledge, my questions are the following;
1) Which hosting plan should I get right now "Reseller" or "VPS"?
2) Which one would be more profitable in the short term?
A While back I found a great deal for SSL certficates so I purchased a bulk package of about 10 of them and used several of them at the time. Now when I went back to use the rest of my pre-purchased SSL certificates (more than a year later), the "contracts" have apparently EXPIRED and the money that was put into those contracts has been frozen along with the contracts! WHAT THE F#$@!
That is such BS! When you pay money for something you should get something in return.
What have I learned... That to me seems extremely manipulative of RapidSSL and Geotrust...
I WILL NEVER PURCHASE AN SSL FROM Rapid SSL or Geo Trust AGAIN! and I hope this post inspires others to select one of the many other certificate sellers out there that are more upfront about their business.
I have contacted both of them and both are telling me that they cannot help me.
Now that I am looking for a new SSL provider can someone give me a good respectable company.
server has huge serverloads of 25+ at random. When I login as root and type the top -s command, the highest cpu usage is less than 5%. The total is less than 50%. Yet my serverload can reach as high as 80.
I also get the "lfd: High 5 minute load average alert " email, but that also does not show what process uses such high resources.
i am getting a huge DDoS attack in one of my servers they are botnets attacks came from Turkey's ip block where the computers have dynamic ips and every ip sends 1 packet 48 Byte and closing the connection To 80 22 110 25 ports so the machine became unaccessiable because of the syn attack what would you advice do you advice cisco pix series or layeredtechs ddos protection PIX 501 Cisco PIX 501 Cisco PIX 501 - 1 Server Only - $99 Monthly Charge - $49 Set Up 99.0 i can buy this there are 1834 banned ips by the software firewall i am thinking is this cisco pix can handle a such attack
Has anyone else been having big latency issues with ThePlanet (EV1 Houston 1 or 2) datacenter?
I called and they said they are having issues which was causing slow connection. If this was the case there would be at least some threads going on in here about it. Anyone else can confirm?
transfer a client's site files (over 220 MB) to my server. The client does not use cPanel or have SSH access.
FTP is horribly tedious. I have created the account on my server and have SSH enabled. I have a feeling I can use wget to download the files to the account's home directory, but I am not sure of the correct syntax to recursively download all the directories and the files.
I have a request to built a standard 32 bit Windows 2003 server as big as possible using standard parts. I am thinking if i use 750GB x 4 Raid 5, that will give me 2.1TB of usable space. Is there any limitations or bottlenecks I should be wary about?
I am currently moving from my current dedicated servers because they simply cannot handle the load. I have a site which frequently makes it onto radio, digg and other similar sites.
I need a dedicated server that can take a beating from Digg and offline Media. For most of the month the server load is really low, the site hardly uses up anything. However, when it hits those sites, it suffers.
I am OK with using Shell, just basic tars/logs/sqldumping/httpd.conf editing/rebooting etc.. anything beyond that like installing and configuring software I cant really do.
I guess I am looking at a dedicated option (linux based) with a host that'll setup software/modules modules for me when I ask, but doesn't really need to hold my hand all the time.
How are ThePlanet.com's servers? Do they manage the servers?
Let's suppose you have a site on a shared hosting plan, and all of the sudden it gets a huge surge in traffic as a result of being featured in the news or something like that. What would be a good plan of action to deal with the surge quickly?
(ex. maybe your hosts takes the site offline from bandwidth overuse)