I'm using WAMP and IIS for the first time so bear with me if I overlooked something obvious.
I have a simple script that creates a few files and then attempts to move them to a subdirectory. I am getting a permission denied error when attempting to move the scripts in WAMP. I am uncertain how to set directory permissions in WAMP. I am also unsure how this will affect the script when I migrate it to an IIS server.
I just got 2 dedicateds, and while creating software raid 1, upon initial sync up I'm getting around 7 megabytes per second (6700 kb/s) in write speed I assume. This is a quad core, sata2 setup...
config.php to not have write permissions for everyone? I am running cpanel 10x with whm/extras Here is full report when i try and load fantastico scripts. You must secure this program. Insecure permissions on config.php While installing CSLH you might of needed to change the permissions of config.php so that it is writable by the web server. config.php no longer needs to be written to so please chmod config.php to not have write permissions for everyone. you can do this by UNCHECKING the box that reads write permissions for the file:
I'm trying to create a directory, "hello", so I log in via SSH and CD to /proc, as this is where I want the directory to be. I type, "mkdir hello" and I get the following;
mkdir: cannot create directory `hello': No such file or directory
I've searched around for the error, and I've found two mentioned problems;
Persmissions Parent directory doesn't exist.
However, the parent directory; /proc does exist and it has the permissions 7777.
I have an upload feature in my site.. On upload i check the file extention of the uploaded file.. But recently a person uploaded a file like xx.php.jar and as the extention is jar it got uploaded.. But when we go to the file through url it gets executed as php... So any solution how can i stop execution of php files in my upload directory
I have an ecommerce store. dns at godaddy. A record www points to the 3dr party hosted shopping cart solution on ip a.a.a.a. I also have a blog. A record blog subdomain points to another server b on ip b.b.b.b hosted away from the shopping cart server a completely.
So www goes one place and blog goes to another.The problem is I need to somehow make the blog appear within the main site as URL...blog instead of as a subdomain blog.domain.com.
I cannot access or modify the hosted shopping cart server in any way. I cannot load wordpress there, that is why it is on a completely separate server.I've been told this is perhaps possible with a reverse proxy configuraton on server b where ip b.b.b.b is.So that even though blog. points to server b it can appear that the url is www.mydomain.com/blog and the blog pages off that directory as well.Can this be done via reverse proxy?
I have put up a WAMP server that is based on Apache2.4.4. I am trying to utilize the rotatelogs.exe to rotate my access log every night. I have this working on IBM HTTP installations. IBM's HTTP is based on Apache 2.2. why the CustomLog command is not working? The command under my Virtual Hosts that I am using is:
I am a beginner in Apache Web server, whether it is possible to make some rewrite function ( or something similar ) to add some text on url from access ?
For example URL....
I need some function in httpd.conf where in SetHandler server-status ( URL... ) can see users who are connected like /testurl username=someusername:somepassword ?
I did link http://someusername:somepassword@domen.com/testurl?username=someusername:somepassword but some users delete from ?username=someusername:somepassword from link and I have a problem to monitoring that lines !
I need to be able to access resources such as PDFs, videos from my website once a user has logged in but I need to block users being able to access the resources from a url?
Ok, here's the deal. I am about to start a site to allow users to log into a password protected area and then, based on each individual user, they will be given access to a .pdf file which contains pricing and rate information that is specific to the particular user.
What is the best way for me to secure the location so that a person could not go to - for instance - www.site.com/docs/xxx.pdf and view someone else's file?
I am using a Win Server 2003/IIS 6 environment. The site will be written in php, and will have a SQL Server 2000 backend to store the userid/pwd combinations, and the name of the particular user's file. I know to make it where the contents of the directory can't be listed, but I need to make sure a person could not quess the filename of another users .pdf and be able to view it.
ive got a flash music player that gets its tracks from a dedicated directory on my server. there's about 10GB of music in there (we own it) and i want to stop people getting at the files (they can see the path in the source of the page that has the flash player).
i tried an htaccess directive that stops listing the directory contents but that obviously wont work. what is the best & most secure strategy to achieve this, blocking all ip addresses apart from my server's?
I'm trying to control download access in a particular directory. I don't want to hide the directory behind a password.
What I am looking to do is to permit access to the files to people who have clicked on my download link only (referrer). Direct links, and hot links to be denied.
I can disable hot linking easily enough via .htaccess. But I'm not savy enough to pull off referral only access.
We currently have a site hosted on a dedicated server, and we use Ensim to manage the site.
Although we can use Ensim to FTP, we mainly use Dreamweaver or other FTP clients to FTP.
What we are wanting to do, is set up and FTP user, so that they can only have access to certain directories. Is this possible? If so, what would I need to do?
Also, when they connect via an FTP program (assuming that the access has been set up as desired) will they only be able to see those directories specified?
I have a set of confidential files that I want to make accessible over the internet to members overseas.
Members will access the file links on a secure web application.
I want to restrict access to the files so that they only open when the user clicked them from the web application. ie. if they paste the url into a browser it should not open the file.
I managed to do this in Apache, but I need to do it in IIS - is it possible?
in wordpress is wp-includes folder, i dont want anyone just execute file from this directory so it just serve wordpress not any malicious file be executed from there,
so i added this rule into /wp-includes/.htaccess: RewriteRule ^(wp-includes)/.*$ ./ [NC,R=301,L]
I have recently switched from Ubuntu to Mint because I would like to enjoy out of the box functionality. I installed Apache, PHP and Mysql. When I go to localhost, the page displays correctly indicating that the server is functioning. In /var/www/html I created devPages. In devPages I created a symlink to a directory in my home directory where I keep all of my html/php files.
When I point my browser to http://localhost/devPages/1/ I get "You don't have permission to access /devPages/1/ on this server". This appears to be a permissions/ownership issue. I don't want to change the permissions for the directory or files in the home directory because they are set to work properly on my production server. Is there a way to get apache to accept files in my home directory?
I have set owner and group for devPages and the symlink to www-data. Set permissions to 755 for devPages.
By the way, under Ubuntu, I have always used a symlink and it always worked without any issues.
I seem to have an error with backing up to my personal FTP repository. I have only just noticed the issue, but i believe it originated when I upgraded Plesk Panel 11 to Plesk Panel 12. I am currently running version 12.0.18 on Cent OS 6.5.
Note whether i check the "Use passive mode" option makes no difference to my problem.Below is what I get from the panel.log under /usr/local/psa/admin/logs
I recently worked on an issue involving a severe performance issue between "write back" and "write through" caching on the RAID/HD.
Long story short, we purchased 12 IBM x3550 M2s came with LSI SAS1068E/SR-BR10i (gimp redheaded stepchild of the MR series. No BBU, no onboard DIMM.) RAID controller and had very bad and inconsistent write throughput with it. Sometimes it writes out 300-400MB/s (dd test, I know... don't flame. I know dd is NOT a good test.), somethings as low as 30MB/s. The server is configured with 2.5" SATA 500GB on HW RAID-1.
From dmesg log it was default to "write through" on sda. I figured out via lsiutil, you can set the drives to "write back". Once we did that, the write performance is more consistent. NOTE, this enables write back on the SATA drives, NOT the controller itself.
I loop out lspci on all our VPS servers. Found out those with LSI (SAS 8344ELP) cards have sda set to "write through w/ FUA". Those are already all RAID-10s, and I have not heard a single complaint from any customer stating poor I/O performance.
I believe the 8344ELP do have a BBU, I can double-check with DC. The DC is on UPS as well. So that rules out the shortcomings of enabling write back caching.
I want to ask those of you using Xen (3.3 & 3.4). Do you guys have better I/O performance with "write back" or "write through" caching? I'm actually looking for real world results. Where you guys actually deployed production VPS server with clients on it.
There is a directory say, "Master" and inside, "Master" there is sub-directory, "Slave". A user who has access to, "Master" should be able to access, "Slave" automatically. However, a user who has access to, "Slave" should not have access to, "Master". Inside cPanel this type of protection is not possible.