1) I use DNSMadeEasy for a couple of my important domains so I can utilize their failover service.
2) I use my own nameservers for everyone else.
At my register (GoDaddy) I've added host entries to my domain (let's call it host.com) for ALL of my nameservers: DNSMadeEasy and mine. For example here are my host entries:
At the register I've then configured host.com to use the first five nameservers for itself, the DNSMadeEasy nameservers.
For less critical sites that I host I simply point them to ns1.host.com and ns2.host.com, my nameservers.
Now, here's the twist. If I use dig to look up www.host.com I get:
[root@lax1 ~]# dig +trace www.host.com
; <<>> DiG 9.3.3rc2 <<>> +trace www.host.com ;; global options: printcmd . 220048 IN NS D.ROOT-SERVERS.NET. ........................................... . 220048 IN NS K.ROOT-SERVERS.NET. ;; Received 228 bytes from 66.63.160.2#53(66.63.160.2) in 1 ms
net. 172800 IN NS J.GTLD-SERVERS.net. ........................................... net. 172800 IN NS G.GTLD-SERVERS.net. ;; Received 497 bytes from 128.8.10.90#53(D.ROOT-SERVERS.NET) in 74 ms
host.com. 172800 IN NS nsdme0.host.com. host.com. 172800 IN NS nsdme1.host.com. host.com. 172800 IN NS nsdme2.host.com. host.com. 172800 IN NS nsdme3.host.com. host.com. 172800 IN NS nsdme4.host.com. ;; Received 225 bytes from 192.48.79.30#53(J.GTLD-SERVERS.net) in 125 ms
www.host.com. 1800 IN CNAME host.com. host.com. 75 IN A 60.55.55.55 host.com. 86400 IN NS nsdme2.host.com. host.com. 86400 IN NS nsdme1.host.com. host.com. 86400 IN NS nsdme5.host.com. host.com. 86400 IN NS nsdme0.host.com. host.com. 86400 IN NS nsdme4.host.com. host.com. 86400 IN NS nsdme3.host.com. ;; Received 276 bytes from 123.123.123.123#53(nsdme0.host.com) in 68 ms BUT, if I lookup the nameserver (ns1.host.com) I get:
Code: [root@lax1 ~]# dig +trace ns1.host.com
; <<>> DiG 9.3.3rc2 <<>> +trace ns1.host.com ;; global options: printcmd . 218964 IN NS M.ROOT-SERVERS.NET. ........................................... . 218964 IN NS K.ROOT-SERVERS.NET. ;; Received 228 bytes from 66.63.160.2#53(66.63.160.2) in 1 ms
net. 172800 IN NS H.GTLD-SERVERS.net. ........................................... net. 172800 IN NS G.GTLD-SERVERS.net. ;; Received 497 bytes from 202.12.27.33#53(M.ROOT-SERVERS.NET) in 115 ms
ns1.host.com. 172800 IN A 60.55.55.55 host.com. 172800 IN NS nsdme0.host.com. host.com. 172800 IN NS nsdme1.host.com. host.com. 172800 IN NS nsdme2.host.com. host.com. 172800 IN NS nsdme3.host.com. host.com. 172800 IN NS nsdme4.host.com. ;; Received 241 bytes from 192.54.112.30#53(H.GTLD-SERVERS.net) in 151 ms
What I've realized is that the actual IP addresses for nameserver host entries come from a higher level server than my own, in this case H.GTLD-SERVERS.net. I guess this makes sense but I just hadn't realized it before. It looks like I don't even need to have record entries in my DNS records for the host nameservers.
Now for the question. Can I:
1) Remove my custom host nameserver entries from my register.
2) Add entries in my DNSMadeEasy records to specify the location of ns1.host.com and ns2.host.com.
3) Use the failover provided by DNSMadeEasy to also fail-over my DNS entries for my nameservers?
I know this would require one more hop if it works but it would allow me to provide failover ability to fifty domains without having to purchase the extra domains at DNSMadeEasy.
I have server for testing UBC an SLM memory management (Virtuozzo 4). I use SLM memory limits and setup 2 VPS's with my hosting plan (1024MB guaranteed and 4096 MB dynamic).
Host machine have 8 GB of RAM. My problem is, that host machine has have 100 MB of memory free when these two VPS's are running.
I was using apache on my old xp machine but recently got a new computer, vista 64 bit which comes with IIS server.
I am familiar with html/css and starting to learn php, asp etc. Am I correct in assuming that it would be good to setup both Apache and IIS on my machine so that I can test database driven sites on my system, because I will want to work on many client websites who some will use windows/IIS hosts and others apache.
Or is it that with just IIS, I can test all sites in my dev environment since IIS supports everything apache does and more?
I am not sure if when developing a site for a client with a linux/apache host there is a lot of apache configuration that I would not be able to test on my local IIS server?
In fact, I am not even sure if a web programmer would need to do anything different at all dependant on which server type their code was running on.
I have a dedicated machine with Xen configured... Dom0 stuff runs great.
I have a pre-made image from jailtime.org, with an ubuntu.7-04.img, ubuntu-7-05.xen3.cfg, and ubuntu.swap -- obviously, the actual image, the config file, and the swap file.
Starting it fails:
Code: $ sudo xm create ./ubuntu.7-04.xen3.cfg Using config file "././ubuntu.7-04.xen3.cfg". The config file:
- Are the /boot and /dev/sda1 literal? That is, do they relate to Dom0 names (/boot on the server, and /dev/sda1, my "real" disk), or are they telling the new DomU what to call them?
- I don't want DHCP.... Do I change dhcp to "0" / "false," or do I specify an IP?
I've found a zillion guides out there, and they cover everything about setting up Xen, except for this one part, it seems?
I recently got a 2nd dedicated server to run MySQL for me. I host game servers that require MySQL. I usually ran it locally on my dual xeon machine until it started using a lot of resources, more then all of the game servers combined. So I moved my MySQL to a dual core machine. The MySQL machine is in the same datacenter as my dual xeon machine (I rent from softlayer). So I am using the private network ip to communicate between the 2 servers. However the queries are lagging out a lot, and causing my game servers to freeze up. I never had this problem when it ran locally, both servers are 100mbit and the mysql machine has a 15K RPM hardrive. The queries seem to not go through as fast as they did when it ran locally (of course), but not as much as I thought it would. Is there anything I can do to make it operate faster? Also, both servers are Windows 2003 Server.
I feel like I'm making this much harder than it is. I have one server with multiple IPs. I list my ns1 as (example) 1.2.3.4 and ns2 as 1.2.3.5. All of that's squared away, and it all resolves properly.
Except that tinydns only listens on 1.2.3.4, and I can't for the life of me figure out how to make it listen on the second IP too. Consequently, queries to ns2 fail.
What I ended up doing was just starting a second session with /etc/tinydns2 (and /service/tinydns2)... This is surely not the right solution, but it's made even worse because my "cp -R /service/tinydns /service/tinydns2" command doesn't do anything.
to move data (a lot) from one server to another. The thing is that the old server's host will not allow SSH access, not even just for a few hours. The new server is a dedicated, so I will of course have SSH there, but how should I handle this situation?
The data in question is massive..Much too much to download to the PC via FTP and upload to the new server. I'm not too familiar with FTP on linux. Could I use SSH on the new machine to FTP into the old machine and recursively grab everything (IIRC, the FTP protocol doesn't allow recursive gets...although it's been awhile since I've used CLI FTP)
I have a small linux box that I use as a router (CentOS 4.4 on OpenVZ).
I have quite a few clients connecting to it and using it as a gateway.
I would like to monitor their bandwidth usage if possible, I have iptables installed and am using iptables -L -v -n, which shows me the data transferred on specific ports that I am forwarding to them.
So, is there an bit of software out there that will monitor each IP for all UDP and TCP traffic, and wont be lost if I restart iptables.
I have looked at Cacti, but have never managed to get it to work...?
For some reason, every time I setup the SSL cert on my Windows box, I receive a error 400 when I try to visit the site. When I take the SSL cert out then the site loads again. Anyone know what the problem is here?
Currently running a server with a little less than 100 accounts, running 32 bit CentOS 4.6 on a single chip running a current version of Cpanel and we're going to upgrade to a two quad core chips, and I figured I'd take this time to upgrade to a 64 bit version of CentOS 5.1. Here's my question: We're running an old version of Apache, and PHP
and while all we have on the server now are wordpress blogs and a few forums, I have to confess a level of uninformed-ness when it comes to upgrading Apache and then transferring all the accounts.
These will be different machines, as I hope to make the backup/restore fairly seemless.
Should I upgrade Apache to the Apache2.* before I make the transfer, or does it matter? Same with PHP to PHP5.* There shouldn't be any conflicts, but I'm posting this because I haven't had to deal with it as yet and thought someone knew of an large issue I could be on the lookout.
Also, the RPM/Perl modules's I've installed over time. Uh, is there a "differential" list or am I going to have to create a list of the RPMs I've got now, and then check after I make the build on the new machine? There's probably a dozen or so ones specific to some applications on the server that aren't required for the core operation, and damned if I can remember which they were...
I have more worries, but most of it involves hand holding issues I think I'll work on privately, but if I do have any more questions, I'll add it to this thread.
I am currently researching the options open to me for Virtualisation, the two main ones I have seen are Xen or KVM.
I mainly use CentOS (RHEL), but have read that the version of Xen with it is very old, broken and unstable. KVM isn't included in the kernel that ships with CentOS, as it is too old, apparently it was first featured in Kernel v2.6.20. There isn't likely to be an update till RHEL6, which is due for release first quarter of 2010. I can't wait over a year, so need to find another Distro for use as the Host OS/Hypervisor.
I have built a pretty powerful server, it has an Intel Xeon 3230 which has VT - so I might be better off using KVM over Xen. I am going to collocate this server, so realistically I can make this decision only once - as it would be a PITA to re-install a host Linux distro remotely.
I did a search on distrowatch for distros with the latest version of the kernel, and Slackware came up as being just one minor version behind the most current (v2.6.27.7).
Now this distro is very mature, so should be a fairly safe bet, but it is a 32bit version and can't host 64bit VMs. I have 8GB of ram so want to be able to use it all, and offer the choice of 32/64bit VMs. So that's that out of the window.
I have used Arch Linux on and off for a couple of years as a workstation OS, but because it is so bleeding edge, when pacman updates it can break itself. But I suppose if I just use it as the Host OS, and never let it update/reboot, then it won't break. It should be fairly lightweight and stable, as I will be installing the bare minimum packages. I have a management card, so if the server fails to boot, then I can still remote in to fix it.
If I do want to update the kernel, is it possible to update without rebooting? I think it is somehow... unless I can just reboot during an unused time at 3am or something.
As you can tell I am leaning towards KVM on Arch Linux (x86_64). Is this a good plan?
I'm running apache to serve PHP files on (/home/www/) and thttpd to serve images on (/home/www/images)
thttpd runs on a different IP than apache, apache only listens to its IP.
After doing this, the number of apache processes decreased significantly, however performance has gone down, and apache is starting to crash very frequently (swapping).
Could there be a file-locking issue? Do I have to separate the images folder from the www folder?
Basically I have installed asp.net and mssql on my machine at home and I was wondering what steps I need to do so that I can have my machine act as a web server?
I've set up a website using a no ip-account, nothing fancy or business-oriented, and I have it working this way:
no-ip(port 80 redirect to port 6500)->router(redirect port 6500 to pc2 with ip 192.168.1.3)->pc2(vmware server redirects port 6500 to virtual machine with ip 192.168.60.100, which is running an asp .net server serving on port 6500)
The router does not forward any port other than that one, but I would like to know if this has any risk for the other machines on the lan.
I was wondering if anyone can help me out or point me to a guide that will allow me to to create a Windows 2003 x64 Virtual MAchine using VMWare server on a CentOS or Redhat installation.
I am planning to get a Juniper firewall, but due to SSG140 has a maximum of 48,000 concurrent sessions per second, so it triggers me how do I measure the concurrent session of a linux server of the total throughput instead of just port 80?
Can someone suggest me good Free virtual machine software? To be able to run on linux platform, and able to run windows, linux, unix and bsd systems from ISO files.
I want to be able to clone a existing virtual machine image(whether it's openvz or xen) and deploy it.
It's sort of like Amazon EC2.
I want to set up a virtual machine image once. Then deploy it(it doesn't matter whether it's on the same physical machine or not) to a new machine or overriding existing image on a existing machine. It's fine that the VPS provider charge me a monthly charge for every virtual machine I deploy.
This way, I only need to maintain one master virtual machine. It'd be even better if I can export a image to my own machine, modify it locally, then upload it for deployment.
I run mdaemon mail server and IIS6. Recently the machine started acting real freaking slow and I noticed that mdaemon was taking up 25-50% cpu almost all the time. When browsing through my computer/explorer, it takes a number of seconds (about 5-10) to show all the folders as you drill down any directory tree. Also, when you right click on any folder/set of folders/set of files/individual file, it takes a good 5 seconds for the right click menu to show up.
The entire machine is just going butt slow for some reason and I'm at a total loss as to why. Here are a few bits of info regarding how the machine responds to things. If I reboot the machine, its totally fine for a short time... mail takes up the usual 1-5% cpu and all folder browsing is normal and speedy. When the machine does become slow as crap, if I turn off the mail server, the machine still runs super slow, nothing changes really. The folder browsing takes forever still and so does right clicking anything. I've done a full virus scan, and nothing. I'm not sure what else to do, but there are no processes running that are taking up a lot of cpu or ram that might be programs causing problems. Bandwidth is not being eaten up either, its on a T1 and on average does about 5-10GB of total IN/OUT bandwidth and that has not changed. I'm not stupid when it comes to computers, but I'm at a total loss as to what could be causing this.