I was using apache on my old xp machine but recently got a new computer, vista 64 bit which comes with IIS server.
I am familiar with html/css and starting to learn php, asp etc. Am I correct in assuming that it would be good to setup both Apache and IIS on my machine so that I can test database driven sites on my system, because I will want to work on many client websites who some will use windows/IIS hosts and others apache.
Or is it that with just IIS, I can test all sites in my dev environment since IIS supports everything apache does and more?
I am not sure if when developing a site for a client with a linux/apache host there is a lot of apache configuration that I would not be able to test on my local IIS server?
In fact, I am not even sure if a web programmer would need to do anything different at all dependant on which server type their code was running on.
I'm running apache to serve PHP files on (/home/www/) and thttpd to serve images on (/home/www/images)
thttpd runs on a different IP than apache, apache only listens to its IP.
After doing this, the number of apache processes decreased significantly, however performance has gone down, and apache is starting to crash very frequently (swapping).
Could there be a file-locking issue? Do I have to separate the images folder from the www folder?
I have been using IIS 7.5 on a Win7 32-bit computer. We moved to a Win7 64-bit machine and 1 app does not work. I am thinking of trying Apache 32-bit to see if that works. The script that is causing problems calls Office Word to create a document. Here is a simplified version of script.
I have build windows xp machine. This machine is running OTRS ticket system. All of our users login using [URL] .... and agent can login with [URL] ....
I have also build ubuntu 9.10 machine and installed OTRS ticket system. I like redirect the [URL] ..... in windows xp to Ubuntu 9.10 with [URL] ....
I searched httpd.conf file and tried to change virtual host but it did not worked.
I am 3 days new to figuring out how to get Perl scripts to run on my Windows XP box. I downloaded and installed the Apache installation file "httpd-2.0.65-win32-x86-openssl-0.9.8y.msi" and the "strawberry-perl-5.18.2.1-32bit.msi" from the perl.org site in hopes that I could get a feedback form to work for a web site that I am working on.
Out of the many pages that I have viewed online of how to configure the Apache Server, nothing has given any favorable results with their explanations.
My last attempt was [URL] ..... where I could not get the example to work. I did the changes to the Apache file "Edit the Apache httpd.conf Configuration File" fairly easily but I must be having problems with the test.pl because I can't get it to work.
I used a different version of Perl (Strawbery from perl.org because it installed without giving me an error pop up after installation) and after copy pasting the script, in an attempt to get it working, I ended up changing it in hopes that I could get it working, shown below.
where I assumed that "#!" meant the "C:" drive and substituted the first "/perl" with the folder the Strawbery Perl had installed itself to and left the second "/perl" in the first line thinking that it was referring to the executable in the "C:strawberryperlbin" folder.
This is the error I get when trying to get the script to run when typing "localhost/test.pl" in the address bar.
"Internal Server Error The server encountered an internal error or misconfiguration and was unable to complete your request. ........"
Know how I can change the title of this post to read "Configuring Apache HTTP Server 2.0 to run Perl in Windows"?
I have server for testing UBC an SLM memory management (Virtuozzo 4). I use SLM memory limits and setup 2 VPS's with my hosting plan (1024MB guaranteed and 4096 MB dynamic).
Host machine have 8 GB of RAM. My problem is, that host machine has have 100 MB of memory free when these two VPS's are running.
I have a dedicated machine with Xen configured... Dom0 stuff runs great.
I have a pre-made image from jailtime.org, with an ubuntu.7-04.img, ubuntu-7-05.xen3.cfg, and ubuntu.swap -- obviously, the actual image, the config file, and the swap file.
Starting it fails:
Code: $ sudo xm create ./ubuntu.7-04.xen3.cfg Using config file "././ubuntu.7-04.xen3.cfg". The config file:
- Are the /boot and /dev/sda1 literal? That is, do they relate to Dom0 names (/boot on the server, and /dev/sda1, my "real" disk), or are they telling the new DomU what to call them?
- I don't want DHCP.... Do I change dhcp to "0" / "false," or do I specify an IP?
I've found a zillion guides out there, and they cover everything about setting up Xen, except for this one part, it seems?
I recently got a 2nd dedicated server to run MySQL for me. I host game servers that require MySQL. I usually ran it locally on my dual xeon machine until it started using a lot of resources, more then all of the game servers combined. So I moved my MySQL to a dual core machine. The MySQL machine is in the same datacenter as my dual xeon machine (I rent from softlayer). So I am using the private network ip to communicate between the 2 servers. However the queries are lagging out a lot, and causing my game servers to freeze up. I never had this problem when it ran locally, both servers are 100mbit and the mysql machine has a 15K RPM hardrive. The queries seem to not go through as fast as they did when it ran locally (of course), but not as much as I thought it would. Is there anything I can do to make it operate faster? Also, both servers are Windows 2003 Server.
I feel like I'm making this much harder than it is. I have one server with multiple IPs. I list my ns1 as (example) 1.2.3.4 and ns2 as 1.2.3.5. All of that's squared away, and it all resolves properly.
Except that tinydns only listens on 1.2.3.4, and I can't for the life of me figure out how to make it listen on the second IP too. Consequently, queries to ns2 fail.
What I ended up doing was just starting a second session with /etc/tinydns2 (and /service/tinydns2)... This is surely not the right solution, but it's made even worse because my "cp -R /service/tinydns /service/tinydns2" command doesn't do anything.
to move data (a lot) from one server to another. The thing is that the old server's host will not allow SSH access, not even just for a few hours. The new server is a dedicated, so I will of course have SSH there, but how should I handle this situation?
The data in question is massive..Much too much to download to the PC via FTP and upload to the new server. I'm not too familiar with FTP on linux. Could I use SSH on the new machine to FTP into the old machine and recursively grab everything (IIRC, the FTP protocol doesn't allow recursive gets...although it's been awhile since I've used CLI FTP)
I have a small linux box that I use as a router (CentOS 4.4 on OpenVZ).
I have quite a few clients connecting to it and using it as a gateway.
I would like to monitor their bandwidth usage if possible, I have iptables installed and am using iptables -L -v -n, which shows me the data transferred on specific ports that I am forwarding to them.
So, is there an bit of software out there that will monitor each IP for all UDP and TCP traffic, and wont be lost if I restart iptables.
I have looked at Cacti, but have never managed to get it to work...?
For some reason, every time I setup the SSL cert on my Windows box, I receive a error 400 when I try to visit the site. When I take the SSL cert out then the site loads again. Anyone know what the problem is here?
Currently running a server with a little less than 100 accounts, running 32 bit CentOS 4.6 on a single chip running a current version of Cpanel and we're going to upgrade to a two quad core chips, and I figured I'd take this time to upgrade to a 64 bit version of CentOS 5.1. Here's my question: We're running an old version of Apache, and PHP
and while all we have on the server now are wordpress blogs and a few forums, I have to confess a level of uninformed-ness when it comes to upgrading Apache and then transferring all the accounts.
These will be different machines, as I hope to make the backup/restore fairly seemless.
Should I upgrade Apache to the Apache2.* before I make the transfer, or does it matter? Same with PHP to PHP5.* There shouldn't be any conflicts, but I'm posting this because I haven't had to deal with it as yet and thought someone knew of an large issue I could be on the lookout.
Also, the RPM/Perl modules's I've installed over time. Uh, is there a "differential" list or am I going to have to create a list of the RPMs I've got now, and then check after I make the build on the new machine? There's probably a dozen or so ones specific to some applications on the server that aren't required for the core operation, and damned if I can remember which they were...
I have more worries, but most of it involves hand holding issues I think I'll work on privately, but if I do have any more questions, I'll add it to this thread.
I am currently researching the options open to me for Virtualisation, the two main ones I have seen are Xen or KVM.
I mainly use CentOS (RHEL), but have read that the version of Xen with it is very old, broken and unstable. KVM isn't included in the kernel that ships with CentOS, as it is too old, apparently it was first featured in Kernel v2.6.20. There isn't likely to be an update till RHEL6, which is due for release first quarter of 2010. I can't wait over a year, so need to find another Distro for use as the Host OS/Hypervisor.
I have built a pretty powerful server, it has an Intel Xeon 3230 which has VT - so I might be better off using KVM over Xen. I am going to collocate this server, so realistically I can make this decision only once - as it would be a PITA to re-install a host Linux distro remotely.
I did a search on distrowatch for distros with the latest version of the kernel, and Slackware came up as being just one minor version behind the most current (v2.6.27.7).
Now this distro is very mature, so should be a fairly safe bet, but it is a 32bit version and can't host 64bit VMs. I have 8GB of ram so want to be able to use it all, and offer the choice of 32/64bit VMs. So that's that out of the window.
I have used Arch Linux on and off for a couple of years as a workstation OS, but because it is so bleeding edge, when pacman updates it can break itself. But I suppose if I just use it as the Host OS, and never let it update/reboot, then it won't break. It should be fairly lightweight and stable, as I will be installing the bare minimum packages. I have a management card, so if the server fails to boot, then I can still remote in to fix it.
If I do want to update the kernel, is it possible to update without rebooting? I think it is somehow... unless I can just reboot during an unused time at 3am or something.
As you can tell I am leaning towards KVM on Arch Linux (x86_64). Is this a good plan?
Basically I have installed asp.net and mssql on my machine at home and I was wondering what steps I need to do so that I can have my machine act as a web server?
I've set up a website using a no ip-account, nothing fancy or business-oriented, and I have it working this way:
no-ip(port 80 redirect to port 6500)->router(redirect port 6500 to pc2 with ip 192.168.1.3)->pc2(vmware server redirects port 6500 to virtual machine with ip 192.168.60.100, which is running an asp .net server serving on port 6500)
The router does not forward any port other than that one, but I would like to know if this has any risk for the other machines on the lan.
I was wondering if anyone can help me out or point me to a guide that will allow me to to create a Windows 2003 x64 Virtual MAchine using VMWare server on a CentOS or Redhat installation.
I am planning to get a Juniper firewall, but due to SSG140 has a maximum of 48,000 concurrent sessions per second, so it triggers me how do I measure the concurrent session of a linux server of the total throughput instead of just port 80?