I currently have a website that is on a dedicated, managed server and on this I have a few websites.
I'm looking at buying a video upload script, similar to youtube. So that people can upload DIY related videos.
My dedicated managed server is quite full and so I have just leased another dedicated server.
Originally I intended having the video upload site on a seperate domain name, but then I wondered if I created a directory on my current site, would it be possible to have the script on my main site in a directory but use the new server to run the script and hold the video etc?
So on my existing site I would have a directory like this /video
But instead of this using my resources on my existing server use the new server that I am leasing!
I am not sure if this is possible or not and so I thought I would ask the experts.
There is a directory say, "Master" and inside, "Master" there is sub-directory, "Slave". A user who has access to, "Master" should be able to access, "Slave" automatically. However, a user who has access to, "Slave" should not have access to, "Master". Inside cPanel this type of protection is not possible.
I want to move the entire contents of a directory tree to another directory.
So for example we may have a directory with 15 directories inside, each directory contains files itself. I want to copy all the files from the directory tree into another directory located somewhere else one the file system. I want only the "files" to end up in the other directory and not the file structure too.
I was just wondering is it possible to move a directory to another server through SSH?
I bought a website a few weeks ago and I have all the files on the server except 30gb of content, now I've tried to put the folder into a tar archive but every time I try it brings up this error:
So I need a way to drag that folder from the old server to my new server through SSH because it will take me ages downloading and uploading all the content.
I notice there is a "mail" directory in the home directories of most of my cpanel accounts. It's in the main home directory, along with public_html, www, etc...
Can I just transfer this entire directory over to a new cpanel server, and everything should just work? (that is, after I change DNS, etc...).
Will the old account be completely restored on the new server?
I am fairly sure this is impossible, but I figured I'd ask anyway.....
Right now I have a forum setup at http://support.myhostingcompanydomain.com/forum/ and I would like to move the forum to another server on another network, so I'm wondering if its possible to have
http://support.myhostingcompanydomain.com hosted on one server and
http://support.myhostingcompanydomain.com/forum/ hosted on another?
Which protocol I should use for performing a backup of important server files and folders. I do make a snapshot but i found that's not enough.
My Centos 6.6 has ProFTPD configuration setup only for webhosts so no root access. I came across SCP, but don't know how to secure that like a Pro. I have also heard about SSHFS for linux.
I have recently switched from Ubuntu to Mint because I would like to enjoy out of the box functionality. I installed Apache, PHP and Mysql. When I go to localhost, the page displays correctly indicating that the server is functioning. In /var/www/html I created devPages. In devPages I created a symlink to a directory in my home directory where I keep all of my html/php files.
When I point my browser to http://localhost/devPages/1/ I get "You don't have permission to access /devPages/1/ on this server". This appears to be a permissions/ownership issue. I don't want to change the permissions for the directory or files in the home directory because they are set to work properly on my production server. Is there a way to get apache to accept files in my home directory?
I have set owner and group for devPages and the symlink to www-data. Set permissions to 755 for devPages.
By the way, under Ubuntu, I have always used a symlink and it always worked without any issues.
If we want to Create output of "Tar" a directory ; to different server, how can we do that ?
I tried this command, but it does not works
tar -cvf oldserverzip.tar /directory | ssh newserver@newserver.com Basically, what we want is to shift a site "dump" to another server, without creating a copy at source server (as there is no space available there to create a "dump")
how I go about unlocking my bin directory and making it writable again. I tried to do installworld with the latest freebsd release and as soon as it gets to the bin directory I get a permissions denied error. I looked at the permissions and ownership compared to another freebsd box and they all look the same. I tried to create a file by using touch just to test it and I received permission denied error.
I want to have a directory that i want to set up so you can only a script can see files in a directory. They will be getting the file names from the database, i have turned off directory browsing. I have also set up a Robots.txt and stopped robots from scanning the directory. Is there anything else i can do?
Also is there and software that allows you to download an entire site. Even if it is gigs of content. If so, how do i set it so it will stop that from happening. I am on linux. Centos running Cpanel.
I know that /www is an alias (symlink) for public_html, but I came across an account with virtual_html. Is there a difference between virtual_html and public_html? And while I assume secure_html is for secure files, how does it tie in?
I have 10 GB partition and we have mount /var direcotry and now it has been full 97% what can i do and how to manage it and i have no other option because no other space is available to create new partition and mount there.
the caching directory I created for eaccelerator deleted by itself. There are 2 copies of eaccelerator, one for apache and one for litespeed. The one for Apache got deleted today. Could it be due to the fact that the directory hasn't been modified or accessed in nearly 2 weeks? It's been that long since Apache was used.
I need to be able to access resources such as PDFs, videos from my website once a user has logged in but I need to block users being able to access the resources from a url?
i updated my old pages on the ftp directory wwwroot.. i deleted the old files and inserted new pages with new names.. now i am not able to see the new website if i type the exact address..the old page is still being viewed. wheras if i omit the www part from the addres i am able to see the new page..
I would like to make some kind of script (probably .sh?) that automatically takes a directory, makes a copy of it then makes a gzip tar and then shoots it over to an FTP server. I would like this to happen twice a day (ie. every 12 hours).
<Directory "/usr/local/apache/cgi-bin"> AllowOverride None Options None Order allow,deny Allow from all
</Directory>
And the error logs say..
[Sun Jan 20 18:09:56 2008] [error] [client xx.xx.xx.xx] File does not exist: /home/goewowc/public_html/404.shtml [Sun Jan 20 18:09:56 2008] [error] [client xx.xx.xx.xx] script not found or unable to stat: /usr/local/apache/cgi-bin/first.txt
The CGI-bin directory is chmodded correctly, the files are also chmodded and belong to the correct group