how would i search for any files modified between such and such date using SSH
some one seems to have downloaded some malicious codes on to some on one of my sites and the files are infected
however i have 1000's of files and some are infected wheras some hasnt so iw ant to check what files have been modified between a date range or on a particular day
if Web servers might prioritize servicing Regular GET or a Conditional GET.
I have seen busy servers take 9 seconds to respond with a (304 Not modified). On fast networks (LANs), the file size is no issue and, it seems that a (200 OK) with the object is downloaded faster than the (304 Not modified) response alone.
I have a problem with time at one RHEL 4 server, it is a plain box, and from time to time the time (hours) seems to be auto modified, no one is touching the box and then from once it got modified to a few hours less than what we specified, to change time and date we always use:
Code: date --set "2007-10-24 13:35" hwclock --set --date="2007-10-24 13:35"
The only thing that I noted is that ntpd daemon was running, while in the rest of our plain box it is not running, maybe this is the reason the system is auto chaning the time?
I temporarily enabled and activated Plesk firewall module (which I wish I didn't the first time) and for some reason it seems to have overwritten the default iptables configuration that was set, leaving my ftp unable to be logged into. I tried to disable the firewall module and reboot the server. It didn't work.
I also noticed that it somehow seem to have changed my hostname to my previous server hostname as well
Is there any way to completely revert back to original iptables settings before enabling the Firewall module?
We have many webservers in our environment, Few webservers serves static contents. In one application, we often change the excel file. Every day or 2 days once we will modified the excel sheet content. File name will be the same, only the content will be changed. We will modify the content of the sheet and upload through FTP to the docroot. After we done this if we access the application URL, its displaying the old content. Its takes time to reflect. sometime with in 3 hours sometimes after a day only reflecting. We are not sure what is the issue. we cleared the cache in browser and tried then also its showing the old content. We are using DNS,network load balancer, proxy between the browser and the webserver.
I tired accessing through FQDN, it showed old content, then i accessed through the LB IP it showed the modifed file. For testing i changed the content again and then accessed. This time even for IP it showed old content. Same i tried with the instance 1 IP and Instance 2 IP. On first time it showing properly but after i cahnged the file content and accessed it showing the old file I tried accesing from a different PC where we havent accessed before, there also it showed old content, SO i feel browser cache is not an issue.
We using Source subnet mask IP persistence in load balancer. I am not sure where the old file is cached actually. Will it be cached in Load balancer or proxy or somewhere in webserver. Just we are placing the file in docroot and accessing it in the URL
I couldn't keep my mouth shut (technically fingers). A customer wanted to upgrade servers and he needed a way to move the data across. Since I don't allow hard drives to be swapped, they have to do it manually all by themselves. I generally allow up-to 4 days for them to transfer data and make DNS changes, etc. But this time, I offered help! I agreed to move the data (darn me) and it just came out of me, involuntarily.
God knows what just happened... but in a positive way, customer is extremely happy!
So...
Both servers are on cPanel - with root access (duh)
200 odd files which total to 25 GB
1 database about 100 MB in size (no biggie)
I was planning on using one of my Windows 2003 servers (via remote desktop) to download the 25 GB and upload the 25 GB, but that sounds like a waste of resources and time.
When I upload a file in FTP I can not see the file, When I upload the file again asked me to file any replaced it exists, When using Cpanel can upload any file, And working well.
When watching the size of folders I find little in the FTP but sized Cpanel see him very different.
I found that recently a lot of nobody files appear in my /tmp.
I delete and delete.. by still same. I don't know how to trace where they from. I suspect is from my hosting users, but I don't know how to check and trace. Anybody can give me some guide?
I've just uploaded my website files onto the server, the website is working fine, but I'm not sure my files are 'arranged' right. I uploaded my images in a folder, but all the rest of the files are 'loose' along with the index file. If I go to either of my domain names, the website appears, which is ok for now, but I only really intended it to under one domain name, so there may be problems if I want another site up there. I think my files should be in a folder. Is this right? Should the index file be in the same folder, or should it be outside the folder with the links changed accordingly. I've been emailing my hosting service, and they 've been trying to help, but I feel a bit thick because I don't really understand what they're saying. Could someone tell me in very plain english how the files should be arranged?
i just bought php file management script, its running smoothly, the only problem was i cant upload big file ( > 1 Gigs ) using that script, then i ask the developer and he said it shouldnt be a problem since he sold that software before people and he never got a problem like that.
OS : Slamd64 apache 2.2.10 php 5.2.8
i tried to changed post_max_size = 1900M upload_max_filesize = 1500M
There is a directory with many files and subdirectories. I want to copy ALL the filesnames beginning with post-1111 to another folder. What is the command to do this?
I run a fairly popular file hosting service and in the process of moving servers.
At the moment i have a system that automatic create a new folder each day and store new uploaded files into this folder. The name of the folder match the current date etc.
Example: /2009/06/15/...
I find this system a bit "outdated" and belive there is a smarter way to do this.
My new fileservers contain ~10TB of available diskspace (each). Average filesize is around 10MB.
I was told not to put all the files into one folder, still the case?
im on a vps with linux and directadmin installed, how do i activate the function to show all files in ftp including hidden files and also how to prevent hidden files from creating?
I need to perform this functions on multiple accounts under my vps, how do i do it?
I have a problem with a .dat file on my webserver. It seems to download correctly (I have added the mime type), but my application refuses to use/read the file correctly.
This is moving from an Apache server, where it worked perfectly, to IIS6. Apache reported the mime as text/plain, whereas I've tried both this and application/octet-stream on IIS.
The dat-file is actually a ZIP password protected file containing an XML-file.
Could there perhaps be some issues with dato format? I'm most greatful for any insigth into the matter.
im in the midst of doing my school final year project and i done some codings of asp pages but when i tried to upload to a linux server, i can only view the webpage with codes instead of the gui interface....i tried to upload my files to free asp host but same thing i cant view my webpages...
to upload this files online and present it to my teachers for grading...
the files are here [url]I hope i can find a free host dun need domain name with free ftp to upload these files and work properly...