Removing Large Files Or Folders CentOS

Jul 13, 2008

Something weird happening here. I have tried every string possible...

There are a number of folders I want to remove off my server, tried the good old and simple...

rm -r /folder/

And then went and ended up with a string as long as my screen. No matter what I do, as it goes recursive in to the directory it asks me if I want to remove each file individually. No matter what string or action I take it insists on asking me as it goes to delete each file.

Could this be a configuration option in CentOS?

View 3 Replies


Plesk 12.x / Linux :: Files And Folders Generated By PHP Hidden In FTP - Centos 6.5 / FastCGI

Jun 23, 2014

I have the following problem, the files and folders generated by PHP are hidden in FTP.

I can see in the Plesk file manager that users, permissions and groups are the same for all files and folders (those that can be seen and those that cannot be seen from FTP).

All the options in the server are set by default. It is a new installation of Plesk 12.0.18 #4 in CentOS 6.5 (Final).

SO: CentOS 6.5 (Final)
Plesk: 12.0.18 #4
Apache Modules:

View 3 Replies View Related

Downloading LARGE Folders On Linux

Jan 15, 2007

On my site, there is a folder containing somewheres around 75,000 images.

I'm on shared hosting, but I do have shell access.

I want to basically copy this folder to another directory in the site. Can anyone give me any pointers?

View 1 Replies View Related

Best Way Of Removing Files And Cleaning Up When Finishing A Server Rental

Feb 14, 2008

What's the best way of removing files and cleaning up when finishing a server rental?

View 9 Replies View Related

Plesk 11.x / Windows :: Prevent Removing Files On File Manager

Jun 4, 2014

How Can I prevent users removing file web.config on FileManager?

I try set attribute read-only, but not effect.

View 4 Replies View Related

Plesk 12.x / Linux :: Delete Subdomain Without Removing Files Or Database?

Aug 30, 2014

I cannot seem to find a conclusive answer to this. I use subdomains to create client websites then once they are in their own domains the subdomains still remain. Can I delete only the subdomain and leave all the files or at least leave the database since it is still connected to the new domain. Example below. uses database1 still uses database1

Is there a way to copy the database into the new domain?

View 1 Replies View Related

Best Permission For SMF Files And Folders

Feb 22, 2008

I'd like to know how to secure an SMF site. Is the default permission good enough?

View 4 Replies View Related

Apache Treats Files As Folders

Mar 5, 2014

I have just discovered in my access log some people accessing urls like this: or even

I have opened those links in browser and they worked. Why?

My .htaccess is empty and I have no 'index.php' folder. But I do have 'index.php' file.

I think this is happening to most php sites, not just mine. This is a random example:

[URL] ....

My question is how to force apache to return 404 error when such url is accessed?

View 1 Replies View Related

Offsite Backups (via Rsync) - What Files/folders To Backup

Jun 3, 2009

I'm currently using (amongst other backup systems) rsync to an offsite space (am using BQBackup at the moment)

I'm just wondering - apart from backing up all of /home/, /var/lib/mysql/ and the important config files (httpd.conf, php.conf, etc etc) is there anything else that *needs* to be backed up?

Obviously in a worst case scenario, a new machine would be deployed with a fresh OS install (and a fresh WHM/cPanel install) so I wouldn't worry about backing up OS files or cPanel core files, although I'm wondering if there's anything apart from the /home/ directory and the MySQL databases which would be lost (and so need backing up) in the event of a crash?

View 10 Replies View Related

Simple Script To Find Suspected Files And Folders

Sep 1, 2007

Try this useful script to find all 777 permission files and folders in /home directory

also it can find all names of suspected folders and files you want

and then you can take the required action

to install this follow the steps

login as root

cd /root

Enter this code and in the 5th line from the end change to your email

# This file will help you to find suspected folders and files in /home directory
# Coded and desgined by Alrutani Web Hosting , for more informations please contact us.
echo " " > /root/perdfmbc
echo "################# Folders with 777 permission #################" >> /root/perdfmbc
echo " " >> /root/perdfmbc
find /home -type d -perm 777 |egrep -v "./cpapachebuild|./.cpan|./src" >> /root/perdfmbc
echo " " >> /root/perdfmbc
echo "################## Files with 777 permission ##################" >> /root/perdfmbc
echo " " >> /root/perdfmbc
find /home -type f -perm 777 >> /root/perdfmbc
echo " " >> /root/perdfmbc
echo "############### Folders & files must be checked ###############" >> /root/perdfmbc
echo " " >> /root/perdfmbc
find /home -name forum >> /root/perdfmbc
find /home -name upload >> /root/perdfmbc
find /home -name 4images >> /root/perdfmbc
find /home -name gallery >> /root/perdfmbc
find /home -name uploader >> /root/perdfmbc
find /home -name up >> /root/perdfmbc
find /home -name r57shell >> /root/perdfmbc
find /home -name r57shell.php >> /root/perdfmbc
find /home -name r57.php >> /root/perdfmbc
find /home -name c99shell >> /root/perdfmbc
find /home -name c99shell.php >> /root/perdfmbc
find /home -name c99.php >> /root/perdfmbc
find /home -name shell.php >> /root/perdfmbc
echo " " >> /root/perdfmbc
echo "###############################################################" >> /root/perdfmbc
echo "Developed by Alrutani Web Hosting" >> /root/perdfmbc
echo "For more informations please contact us." >> /root/perdfmbc
echo " " >> /root/perdfmbc
cat /root/perdfmbc | mail -s "Suspected files & folders in your server"
cd /root
rm -rf perdfmbc
# This file will help you to find suspected folders and files in /home directory
# Coded and desgined by Alrutani Web Hosting , for more informations please contact us.
To add more files and folders that you want the system to list


find /home -name upload >> /root/perdfmbc
after it add

find /home -name xxxxxx >> /root/perdfmbc
where xxxxx is the name of the file or the folder you want

Save file Ctrl X
select yes then click enter

chmod 755
To make the script works daily

crontab -e
At the end enter

* 3 * * * sh /root/
save and exit done !!

now to test the script

cd /root

you will receive email from the server

View 2 Replies View Related

Plesk 12.x / Windows :: How To View Attributes For Files / Folders

Aug 7, 2014

If users copy files/folders on Plesk Panel 11/12 with attribute "read-only", I don't see this attribute on File Manager, every time do next "attrib -r <path to file/folder".

Maybe you can improve functionality of "File Manager" on Plesk Panel 12?

View 1 Replies View Related

Plesk 12.x / Linux :: Write Recursive Rights On Files / Folders

Feb 20, 2015

I have a concern during the installation of PrestaShop. I created a subdomain through Plesk, put my files on my server via FTP. During installation of PrestaShop, I have a problem:

Recursive write permissions for the Apache user on the ~ / config /

My user is "NOWIS" and my group is "psacln" ....

I think the user "NOWIS" has no rights as "Apache" ....

View 3 Replies View Related

Disk Quota Issue - Locating The Folders With The Highest Number Of Files

Jun 6, 2008

we've been having issues with reaching or exceeding our disk quotas. I've checked carefully, and while I've cleared our mail queue's, I don't think that's the issue.

So what I'm really looking for is a way to figure out why we have so many files (our quota is 220000, and I'm pretty confident that we aren't intentionally doing anything to create so many files). There's lots of information on finding the folders with the largest file sizes, but locating the folders with the greatest number of files isn't quite so simple--or at least, it doesn't seem very clear to me.

Does anyone know a way via the command line to figure out this information, short of going through every single folder and figuring out how many files are in the specific folder?

View 6 Replies View Related

Mod_Security Log Files Too Large

Nov 7, 2008

Linux Fedora 6, Apache 2 with Mod Security, MySQL.

Our mod_sec logs get incredibly large very quickly. In the configuration for mod_security, we have specified logging options as
SecAuditEngine RelevantOnly
SecAuditLogRelevantStatus "^[45]"

but the mod_sec.log gets to almost 10 GB (in a matter of 5-6 days) before it is truncated to mod_sec.log.1 and a new one is created.

Is there a way we can specify that a max size of one log file is 1 GB, for example?
Or another question, how come it gets so huge so quickly? We thought that logging "RelevantOnly" will only display errors / requests that are deemed security risks.

View 2 Replies View Related

Hosting Large Flv Files

Oct 6, 2008

I have a customer who wants to sell access to videos of conferences he runs.

Each flv vid is approx 1 - 1 1/2 hors long approx 380MB each and there will be about 12 videos per conference.

approx 4 - 8 conferences per year.

My customer suggests 10 - 20 people will buy access to watch each video.

Access to watch the videos will be through a password protected webpage.

issue - the current site hosting company only allow uploads up to 150MB per file.

Can I host the flash videos elsewhere and deliver them through the password protected web page without anyone else being able to see them via server they are hosted on?

This would also reduce the bandwidth going through his current site server.

View 14 Replies View Related

Locate Large Files

Jul 7, 2008

I am trying to locate what large file are filling up the / on the server but I am having trouble using the find command to do this.

View 1 Replies View Related

Speed Up Serving Large Files

Nov 9, 2009

I'm working on a web site which will basically be a flash games portal. I have a dedicated server running Apache 2 on a 100mbit dedicated line but my download speed for large files (flash files of over 5mbs) is really slow. I am thinking this is because of Apache but I don't know much about this. I've read that I should change for a lighter http server for serving static files. The way my server is set up is I have 2 virtual machines running, one doing the PHP processing and the other serving static files, both running Apache, so if I have to change HTTP server for the static files it would be very easy. Although I am not sure if this is necessary or if I can tune Apache to push files faster than this.

View 8 Replies View Related

FTP :: Stop Uploading Large Files

Jul 17, 2008

I'm facing a very strange FTP issue with one of my shared-hosting accounts, while all of my other servers are having no problems but only this one, when I try to upload a file (whatever file) larger than 500kb from my local PCs, in most cases, the file would stop uploading during the process and hang there until it times out.

There are 2 interesting things though:
The file transmission typically hangs when approximately 248kb of the file have been transferred, pls see the attached screenshot for example.

If you look at the attached screenshot, you will notice that the uploading transmission hangs when 248kb of the file have been transferred. This is very strange and what I mean is that for example, I randomly pick up a file, and attempt to upload it onto my host for 10 times, now see, 5 times it will hang when 248kb of the total size have been transferred, 3 times it will hang at other points *near* 248kb (224kb or 280kb typically), 1 time it will hang at another random point, and 1 time it might be uploaded successfully (yes, there is still a tiny chance for the file to be uploaded successfully).

My default internet uploading speed is 80kb/s-100kb/s, lately I found that, when I limit the uploading speed on my FTP client (e.g. max. 30kb/s), everything WILL WORK without any problem! No hangs, no interrupt.. Whereas when I free up the uploading speed limitation and let it upload with my regular speed, the problem appears again.

It seems to me that the FTP hangs only when the uploading speed is higher than 60kb/s. However my host provider told me that they have customers uploading without any problem at over 400kb/s, and they said "there's no problem or limitations on the server at all".

Up until now, I have done following things to troubleshoot the issue but with no luck:

Contacted my host.
Disabled/Enabled the PASV mode on my FTP client.
Tried different FTP clients on different computers (FlashFXP and Filezilla).
Rebooted my router and reseted everything with the factory default settings.
Contacted my ISP for the issue, they "did something" but nothing were helpful.
Rebooted all my PCs.
Disabled both firewalls on my PC and on the router.

Furthermore, I have asked another friend of mine in another city with another ISP to test the FTP uploading, but unfortunately he got the exact same problem. And I've done some search on the internet for hours but no one seemed to have the same problem..

View 12 Replies View Related

VPS :: Large Files And Taking Too Much Space?

Nov 22, 2008

I just logged into my VPS and was astonished by how much space I have in use.
8.09GB... but I can't figure out what's using up so much space!?

How can I find out were large files are located ? Since it's increasing daily

I use LXAdmin with HyperVM Control Panel

View 10 Replies View Related

Multiple Large Core Files

May 20, 2007

Just noticed quite a few large Core. files within one of our websites (within a sub folder of public_html). Anyone knwo what these are and how they got there?

View 3 Replies View Related

Transfer Large Data Files

Jun 15, 2008

I have a debian box, and have archived a gallery in to a .tar file, 5.77gb.

I have a centOS box, and have used wget to bring the data file over to the new server.

However upon doing so it only detects it as 1.8gb when it starts downloading.

I have terminal access to both servers, just trying to bring my files over from one server to another.

View 4 Replies View Related

Large Media Files & Backups

Jul 16, 2008

I've been using Lypha for the past 4 years, but they've taken the last straw (gigabytes of backups went missing and they wont reply to emails as to why).

Looking for a web hosting package for under $10/month that has large enough disk-space/bandwidth to allow me to backup large audio / video files to it, as well as the normal site operation (I use it for portfolio website, as well as hosting additional domains)

View 17 Replies View Related

Large Video Files Need To Be Streamed

Mar 30, 2007

I am developing a web application for a private investigative firm. They do surveillance work and therefore have surveillance videos. I would like the capabilities of uploading the videos online and allowing the client to login and view their surveillance video online.

Currently, we get the video from the PI, put it on a DVD and then mail it to the client.

This takes too long. We want the client to be able to view the video online.

Some of these videos can be up to 2 hours long.

First, is this even possible?

- how much bandwidth would a website like this take?
- Is there a host that can hold hundreds of GB of video?

I want to convert it to flash to save file size and also so I can stream it.

View 3 Replies View Related

How To Move Large Files To New Host

Mar 21, 2007

I have some 100's of MB's to move and I'm definitely not doing it by transferring it via my PC / FTP.

I seen all the tutorials on how to move your MySQL databases, but what about full folders etc, how do I move those (putty?)?

View 1 Replies View Related

Optimizing Lighttpd For Large Files (180mb Avg)

May 9, 2008

I have four servers with a quad xeon, 4gb ram, and 2x300GB SAS 15K RAID0 harddrives, pushing a total of 1.6gbits. It serves a lot of zip files with an average flesize of 180mb. My question is, how can I optimize lighttpd 1.4.19 to push its max with very low IO-wait. I've looked up some stuff and only found options that deal with lighttpd 1.5 and use Linux-AIO for the backend network. Currently I use writev with 16 workers and an read/write idle timeout of 10s. Logging is off, too.

View 14 Replies View Related

Daily Backup + Rsync Ssh + Large Number Of Files

Oct 29, 2006

i just wana know is it safe to do remote daily backup for about 70,000 files?

file sizes is about 200kb and every day i have about 1000 new file, so rsync first should check old files becouse i am deleting about 30-50 of them daily and them backup new 1000 files ,
so how much it will take every time to compare that 70,000 files?

i have 2 option now:

1-using second hdd and raid 1
2-using rsync and backuping to my second server , so i can save about $70 each month.

View 9 Replies View Related

Writing Large Files - Risk Of Damaging Filesystem

Jan 19, 2008

Does writing large files (ie, 10GB backups in one archive) cause any risk of damaging a linux filesystem?

View 1 Replies View Related

How Do You Host Large Video Files And Support The Bandwidth

Feb 8, 2007

I've got a client who wants to host audio files... Here are the sizes:

50 x 75MBs
300 x 10MBs
400 x 5MBs

That totals 8750MBs or 8.75GBs... If he gets hundreds of visitors, it could end up being 1000's of GBBs or bandwidth.

I don't know what to look for to support so much bandwidth... Do you buy bandwidth? Are their special companies out their that host it for you?

View 6 Replies View Related

Plesk 12.x / Linux :: Content-length Limit When Uploading Large Files

Jun 18, 2015

Domain has PHP Settings in Plesk set to 2G and I get this error when uploading a 48MB file using Wordpress. I assume I need ot modify this manually in conf file somewhere to allow uploading large files?

Requested content-length of 48443338 is larger than the configured limit of 10240000..

mod_fcgid: error reading data, FastCGI server closed connection...

View 1 Replies View Related

Copyrights 2005-15, All rights reserved