A While back I found a great deal for SSL certficates so I purchased a bulk package of about 10 of them and used several of them at the time. Now when I went back to use the rest of my pre-purchased SSL certificates (more than a year later), the "contracts" have apparently EXPIRED and the money that was put into those contracts has been frozen along with the contracts! WHAT THE F#$@!
That is such BS! When you pay money for something you should get something in return.
What have I learned... That to me seems extremely manipulative of RapidSSL and Geotrust...
I WILL NEVER PURCHASE AN SSL FROM Rapid SSL or Geo Trust AGAIN! and I hope this post inspires others to select one of the many other certificate sellers out there that are more upfront about their business.
I have contacted both of them and both are telling me that they cannot help me.
Now that I am looking for a new SSL provider can someone give me a good respectable company.
These two hosts are scams. i was with gigenoc for a few and then there website shutdown with a lovely blank white index and today i just got a email from a employe from gigenoc. heres the email:
Quote:
Gigenoc is a ****ing **** Service, Tim wanted to scamm you over he made a new site www.servermoment.com ALL SPREAD THE WORD
TIM IS AT SERVERMOMENT UNDER THE NAME MARK
Tell people not to sign up its HIM!
So basically, Tim likes to scam people i guess. bad idea because now i can dispute the orders i had with them and so are all my friends that have gotten hosting from them. good thing webhostingtalk is viewed by thousands of people everyday
does anyone know if RapidSSL is accepted by most used browsers (IE, Firefox, Opera and Safari)?
also, if I want to protect "secure.domain.com", do I need to issue the certificate exactly for "secure.domain.com" and it probably wouldn't match "www.secure.domain.com"? Can't I get a certificate for "*.domain.com"?
I entered the domain name in wrong - domain.com instead of www .domain.com - and decided to not approve it when you get to the final step as rapidssl said to do.
So how long do i need to wait before namecheap will update the status so that I can start the process again ?
I signed up with namecheap.com for RapidSSL. I had my hosting provider generate a CSR for me and I put it in namecheap. Then I receive an email from namecheap that says:
========== Thank you for your RapidSSL order.
Supplier Order Id: ###### An email will be sent to the designated approver with instructions on how to approve your certificate request for xxx.domain.com.
==========
This was sent about two hours ago. I need this SSL fast as it is the backbone of my website. Namecheap doesn't have phone support and RapidSSL is closed on weekends. Any insight on when I'll get instructions to move onto the next step?
Just moved to a new server, and of course, 10GB doesn't seem that large for a server but for some reason wget is not able to handle the transfer of that backup for me... it transfers about 1MB then tells me "successful transfer..."
The old server is using cPanel, and the new server is just a plain old server that I haven't loaded up yet.
how I can get this full backup over to the new server?
I'm sure this question has been asked before, but I'm looking for a nice and simply way of breaking up log files into smaller chunks.
I've been running apache2 on a VPS for the past few months and one of the access.log files is now 700mb big... bit of a waste of space. I'm currently just doing:
I've been using mod_security for a long time, but apparently I accidentally enabled some kind of log or something that uses mysql. I don't remember it being there before.. but the point is; the database is like 145100k!
I had several user accounts that were pushing their quota. I was digging around in SSH and found that the INBOX file in /home/username/mail was huge even though the user does not keep messages on the server. I deleted this file to free up space and all seems file. A couple seconds later I did check and the file was recreated with new incoming mail.
My question is how do I keep this file from growing out of control? One of the users I had for almost 2 years had an INBOX file of almost 2GB!
Server Details: VPS running WHM 11.23.2 cPanel 11.23.3-R25623 Redhat 9
I'm just curious as to what kind of things the huge sites--Youtube, Myspace, etc.--are doing to try to keep scalable. What sites do you guys just hate for failing in this regard, and perhaps most importantly, what are some ways we can prevent downtime?
where do you go host HUGE websites, youtube like sites, with HUGE bandwidth usage?
I don't believe people go on host like rackspace, with their 150GB / month packages, unless they want to pay an absurd amount of $$.... so where do these guys go to host? What kind of hosts are these?
I have been receivig a huge logwatch report, seems that logwatch is not parsing the /var/log/secure file, but sending the log entries instead of any resume of it. I got thousands of lines like
Cp-Wrap: Pushing "47 GETDISKUSED pvargas lights.com.co" to '/usr/local/cpanel/bin/eximadmin' for UID: 47 : 25 Time(s) Cp-Wrap: Pushing "47 GETDISKUSED r.perez konecrans.com" to '/usr/local/cpanel/bin/eximadmin' for UID: 47 : 69 Time(s) Cp-Wrap: Pushing "47 GETDISKUSED r.rodriguez konecrans.com" to '/usr/local/cpanel/bin/eximadmin' for UID: 47 : 114 Time(s)
I have upgraded to the most recent version of Logwatch with default configuration. Any ideas on what could be wrong?
ways to improve the database performance in the situation when I have to modify a large table (several million rows), by e.g. adding a column. Currently this would take several hours which is too slow. The bottleneck is disk I/O. I am considering either partitioning the table over several innodb files on several disks, or going to a RAID-5 or RAID-10, it this will give me better write performance.
The database is 130GB large, and the problem table (which I make period changes to) is the largest table on the server. I cannot have downtime of 3 hours each time I make a change and adding blank fields (to be used later, when a new field is needed) is not an option.
Each time I add a column, the cpu goes into high (80%) io wait state for about 3 hours.
I have a hack which would allow me to split the large table into multiple smaller tables based on some criteria (for example, forumID or such). Here are a couple of things but would like to know which is best, and am open to new ideas. The ideas so far:
1. Split the table into 3 or 5 smaller tables each on it's own disk. The disk IO would then not be so bad, and it might only take 1 hour to perform the table change. But this might not work because the changes to the database (as in adding a column) might be serial, meaning only 1 disk is being written to at a time. (Then again, maybe it will work if I launch 3 different scripts, one to update each table at once).
2. Do RAID 5 or 10, and have 3 or 5 disks. This again might not help at all because of the above issue with MySQL writing serially.
I am using latest MySQL 5.0.45 with InnoDB engine on Debian etch Linux
I have one domain where is hosted a lot of subdomains,and for some reason it constantly have 4% cpu usage and 33% mem usage.Since that domain is inactive,could be that usage beacuse of addon domains but it simply not presented correctly in whm?
I had 18GB bandwidth.log file at /etc/log/ directory? What is the meaning of bandwidth.log file? And what may be reason increasing file size to 18GB, especially in one night.
I have done my research, befriend a few super proxy webmasters, and learned everything I need to know about being successful in the proxy business. So I am selling almost all my websites to fund this huge project. I will also be flipping proxies from time to time to fund the project even more. This will be a year long project and will be my full time job sooner or later. My goal is to have 1,000 proxy sites.
So with this knowledge, my questions are the following;
1) Which hosting plan should I get right now "Reseller" or "VPS"?
2) Which one would be more profitable in the short term?
server has huge serverloads of 25+ at random. When I login as root and type the top -s command, the highest cpu usage is less than 5%. The total is less than 50%. Yet my serverload can reach as high as 80.
I also get the "lfd: High 5 minute load average alert " email, but that also does not show what process uses such high resources.
i am getting a huge DDoS attack in one of my servers they are botnets attacks came from Turkey's ip block where the computers have dynamic ips and every ip sends 1 packet 48 Byte and closing the connection To 80 22 110 25 ports so the machine became unaccessiable because of the syn attack what would you advice do you advice cisco pix series or layeredtechs ddos protection PIX 501 Cisco PIX 501 Cisco PIX 501 - 1 Server Only - $99 Monthly Charge - $49 Set Up 99.0 i can buy this there are 1834 banned ips by the software firewall i am thinking is this cisco pix can handle a such attack
Has anyone else been having big latency issues with ThePlanet (EV1 Houston 1 or 2) datacenter?
I called and they said they are having issues which was causing slow connection. If this was the case there would be at least some threads going on in here about it. Anyone else can confirm?
transfer a client's site files (over 220 MB) to my server. The client does not use cPanel or have SSH access.
FTP is horribly tedious. I have created the account on my server and have SSH enabled. I have a feeling I can use wget to download the files to the account's home directory, but I am not sure of the correct syntax to recursively download all the directories and the files.
I have a request to built a standard 32 bit Windows 2003 server as big as possible using standard parts. I am thinking if i use 750GB x 4 Raid 5, that will give me 2.1TB of usable space. Is there any limitations or bottlenecks I should be wary about?
I am currently moving from my current dedicated servers because they simply cannot handle the load. I have a site which frequently makes it onto radio, digg and other similar sites.
I need a dedicated server that can take a beating from Digg and offline Media. For most of the month the server load is really low, the site hardly uses up anything. However, when it hits those sites, it suffers.
I am OK with using Shell, just basic tars/logs/sqldumping/httpd.conf editing/rebooting etc.. anything beyond that like installing and configuring software I cant really do.
I guess I am looking at a dedicated option (linux based) with a host that'll setup software/modules modules for me when I ask, but doesn't really need to hold my hand all the time.
How are ThePlanet.com's servers? Do they manage the servers?