I just recently had a hard drive failure that produced the following error
root@re:/# mount /dev/hda3 /mnt
mount: wrong fs type, bad option, bad superblock on /dev/hda3,
missing codepage or other error
In some cases useful info is found in syslog - try
dmesg | tail or so
The host is going to mount this HD on the same machine after adding a new hard drive and fresh install... Does anyone have any recommendations for how I can go about recovering data? Specifically mysql databases?
As I have never used cPanel neither have whatsoever experience with control panels, in the process of ordering a couple of new server I am wondering what's the average recovery time after a server failure which involves data (ie. disk failure).
I am interested to understand this, because I need to choose between hardware raid or backup disk.
For example, at LiquidWeb (this is just one of the many managed providers i am evaluating) you can pay the same server 264$ with 2x250 hardware raid 1, and 219$ with 2x250 simple hard drives, where the second one is supposed to be a backup drive.
I am yet to ask, but on this servers most likely (at this prices) they do not have hot swappable disk drives where they can rebuild the array "live", so also if i choose raid hardware, in case of a disk failure, i face a downtime to bring down the server, rebuild the array, bring on the server, which I guess is something not less than 30-60 minutes downtime.
The advantage of this solution is that a disk failure keeps the system running, so you can schedule a maintenance window to replace the disk and rebuild the array, I guess.
On the other side, relying on a backup disk (and of course rsyncing data to an offsite server, i would do this anyway) you save 45$ each month, and disks after all do not fail every month.
If the main disk fails, they need to replace the disk, install again the server with what I suppose being a standard image (so let's figure a couple of hours, as they have a 30min hardware replacement sla), then you need to restore backups (which in my case are something like 20gigs). So I guess, on an average, it would take something like 4 hours downtime.
Am I correct ?
What would you reccomend as a solution with cpanel, taking into account the huge price difference ? Backup drive or Hardware raid ?
This would be for a single big website, not a hosting-company with resellers and customers environment, so I would value more the monthly saving, rather than the high availability, but of course I am interested to know what is the average time to recover a cpanel server after a drive failure.
Besides I have also another question, as I have never had a colocated server with raid hardware, so i do not know anything about this.
When a drive fails in a raid1 hardware environment, how the hosting provider gets notified ? I suppose it's not a led on the server to blink, as no one would see it.
So if you have software raid (like my computer at home) you get an email from mdadm with "warning, array degraded", so you know it quick and you can check it anytime doing a ' cat /proc/mdstat '. What about hardware raid?
My fear is that nobody notices my raid1 drive failure, the server keeps going on just a drive, then maybe the other drive fails, and it would be unpleasant: while a single drive failure would be obviously very easy to spot for me.
[cPanel smartcheck] Possible Hard Drive Failure Soon
ATA Error Count: 1512 (device log contains only the most recent five errors) Error 1512 occurred at disk power-on lifetime: 11736 hours (489 days + 0 hours) Error 1511 occurred at disk power-on lifetime: 11736 hours (489 days + 0 hours) Error 1510 occurred at disk power-on lifetime: 11736 hours (489 days + 0 hours) Error 1509 occurred at disk power-on lifetime: 11736 hours (489 days + 0 hours) Error 1508 occurred at disk power-on lifetime: 11736 hours (489 days + 0 hours) ----END /dev/sda--
What do you advice me to do? ask the DC to change the Hard Drive or wait till damaged?
I'm currently hosted on a VPS server with Galaxy Solutions, but just last night I was informed of a hard drive failure.
This morning they said the DC couldn't recover the data, and now they are trying to recover the data themselves. Right now I'm assuming that if the DC can't recover it, then there isn't much of a chance of it being recovered at all.
I've put in countless hours and dedicated so much of my time to my sites, something which certainly cannot be compensated for. I would like to exhaust any possible way to recover the data.
Can anyone recommend what I should do in a situation like this? Would it be advisable to consult a data recovery specialist? It would be great if you could also recommend one.
The cost doesn't matter. I'm extremely frustrated, annoyed, and confused because of all this. Just like that, all my work is gone.
I did have backups, but they were on the same server.
S.M.A.R.T Errors on /dev/sdb From Command: /usr/sbin/smartctl -q errorsonly -H -l selftest -l error /dev/sdb Please note the following marginal Attributes: ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE 190 Unknown_Attribute 0x0022 054 037 045 Old_age Always In_the_past 46 ----END /dev/sdb-- i have done cpanel update in the log i have this error
Running Sanity Checks & Notifications...Using smartcheck config 5.32 for smartctl(5.33) Checking /dev/sda....Ok Checking /dev/sdb.... Errors: Please note the following marginal Attributes: ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE 190 Unknown_Attribute 0x0022 052 037 045 Old_age Always In_the_past 48
I received this a month ago and now ive just done a cpanel update and in the log i noticed this:
Running Sanity Checks & Notifications...Using smartcheck config 5.32 for smartctl(5.33) Checking /dev/sda....Ok Checking /dev/sdb.... Errors: ATA Error Count: 213 (device log contains only the most recent five errors) Error 213 occurred at disk power-on lifetime: 7 hours (0 days + 7 hours) Error 212 occurred at disk power-on lifetime: 7 hours (0 days + 7 hours) Error 211 occurred at disk power-on lifetime: 7 hours (0 days + 7 hours) Error 210 occurred at disk power-on lifetime: 7 hours (0 days + 7 hours) Error 209 occurred at disk power-on lifetime: 7 hours (0 days + 7 hours)
My server hard disk is crashed badly. The rescue function in the server cannot take part and so I've tried using some recovery software to get back my data.
I've tried using Easy Recovery Professional. It sort all the files by it file type into different folder. I found a folder named .DB, there are also some .ado and . ldb folder too. I guess one of it is my database. The problem now is, i dunno how to read the file.
Do you have any idea to read it? I've tried many recovery system. eg. DiskInternals Linux Recovery, Disk Doctor Linux Recovery.
I have 100+ sites on this hard drive, and one site in particular that meant the world to me.
My host sent the drive to Gillware first, but they failed saying that the file system was so severely damaged that they could not recover anything.
Then shortly after, my host sent it to DriveSavers, a very well-known company, but they also FAILED.
I'm extremely depressed because of this. Please don't post if you're going to say "make sure you do backups next time" because I've heard it 504329504395 times now, and while I do realize my mistake, saying that does NOT help me.
I am willing to spend ALOT to get my sites back. I still have hope. Are there any other companies out there BETTER than DriveSavers? Assuming that you'd still have hope even after two companies failed, where you would you go or what would you do?
so this is the second time I got an email like this from my server:
This email was generated by the smartd daemon running on:
host name: xxxxxxxxxxxxxxxxxxxx DNS domain: xxxxxxxxxxxxxxxxxxx NIS domain: (none)
The following warning/error was logged by the smartd daemon:
Device: /dev/sdb, 6 Offline uncorrectable sectors
For details see host's SYSLOG (default: /var/log/messages).
You can also use the smartctl utility for further investigation. No additional email messages about this problem will be sent.
What can I do about this?
I actually COLO a box.. I sent them the server NEW, and the drive failed a week after I sent them a new one. This SECOND drive is now having errors apparently.. what can I tell my tech to do about this? Please advise me on what I can do.. (There is a chance that this could be my primary drive)
I asked FDC to check the controller and drive, and they said there was nothing they could do.
Now iam useing 320Gig SATA harddrive as my primary hard drive,i dont use 2rd harddrive,iam haveing pure download site,in TOP command 4.5%wa is this bit high? or can i add 2rd harddisk and move some data to there to reduce wa,but my load of the server is fine or any way to reduce wa?
1- i did Os Reload with new Hard drive for "home" 2- data "backups" drive lost
3- replaced the old home drive as "/old drive"
4- " /old drive " is now the secondry drive in my server and it has the all sites usres and evry thing
5- what i need , to trasfer, copy this sites from " /old drive " to "home"
but data center said
The /olddrive/home directory contains the contents that were previously in the /home directory. You can copy files from this directory to any other directory on your server.
The command to copy files in the UNIX environment is the "cp" command.
The user directories in /olddrive/home directory contain the web page files for the users. However, simply copying the contents over will not recreate the users or domain entries in DNS/httpd. If you wish these back you will need to recreate them manually or restore them from backups.
I have a terminal server with a bunch of applications on it, among which is a database driven app. There have been complaints that access to the db is sluggish. right now the server is on just one 7.2k drive.
I am guessing its a hard drive bottleneck because memory and cpu usage seems okay. I have a few ideas. Please tell me which of these you would recommend.
1. Upgrade entire server to raid 10 system.
2. Upgrade the entire server to a 5-disk RAID 5 system.
3. Create two raid 1 setups. One for the OS and regular apps, the second one to host the DB.
4. Create a raid 1 setup for the OS and regular apps, and a RAID 5 setup for the DB.
Ideally I would like to see improved read/write speeds on both regular files/apps and on the database. The RAID 10 system was what I was leaning towards at first because the stripping increases throughput, but then I realized I may see better performance by keeping the regular files and the DB on independent setups so that OS and file read/writes won't affect the DB read/writes.
I have never had a hard drive fail on me, i dont think the I/O on my servers would ever warrant it but im looking to ask people that have had a hard drive die one them the following:
1. Were you ever cautious before your hard drive failure or did you lose data and if you did lose data, did it make you buck up your ideas (i.e do you now have a solid backup plan or do you still play maverick)
2. If you have had a hard drive failure, has it made you, do you think, overly-cautious?
3. Has a hard-drive failure ever swayed you towards the more expensive raptor drives? and if so, why?
I just purchased a 500GB ATA hard drive (meant to get SATA but must have misordered, but I don't think it makes that much difference) to replace a 80GB SATA hard drive on a Windows Server 2003 server. There are two partitions on the 80GB drive, and rather than add the new HD as a third partition, I would like to clone the data on the 80GB drive to the 500GB drive, then increase the partition sizes and remove the 80GB completely. Is there a specific method or software I can use to accomplish this?
I am going to mount my new hard drive to /home/user/public_html but I have a question before I do so. If I upload stuff to /home/user/public_html will it go to my new hard drive that I just mounted there or will it still go in the old hard drive?
The reason why I want to mount it there is because my other hard drive ran out of space, and I would like to continue this user account's space without having to move this user.