to set up a serious of scheduled tasks. For example at 9:35:12 PM I would need a php script to run. And its very important that it run at the 12 second mark for the 35th minute. This seems like an application for a cronjob. Anyway, it looks like you can only have cronjobs run every minute, not seconds.
Problem: the machine crashes exactly every 10 minutes. The crash occurs with no entry on the logs and with 0.00 load. It is as if someone take out the current every 10 minutes.
Here are the specs:
- 2 CPU Intel Xeon 2.0 - 8 Gb RAM ECC - 2 x 250 Gb HDs
This machine needs plenty of current. I wonder if I am not going over the rack power quota. May be there is a system to allow overages for then 10 minutes, then it cut back the current to the rack quota.
i have a small cluster (one web and one db) setup and i host a rather popular group of 4-5 sites that allow users to dynamicly create their own mobile chat communitys automaticly. each site gets its own mysql db created and populated automaticly.
this is all fine,
but in the last 24hours weird things have begun happening, previously i had the sql max_connections set to 500 and this was perfectly adquate for the demand but now even when i set the connection to 4000+ they are all maxxed out within 5-10 minutes, and mysql processlist shows thousands of unauthenticated user connections sitting at login status,
i have gone through the sites and all their mysql configs are fine so i cant see what the issue is.
server specs below
db server: dual amd opteron 246 8GB ram 120gb hd(64gb free) 33gb swap (rarly used but their for emergencys) centos 5 64bit. direct 100mbit lan to web serv only mysql,ssh and webmin running, no other apps installed
I have an apache server on windows which I wanted to speed up using caching systems.
I tried installing squid, and I got it working with a basic example config. It seemed to work well, however under more heavy load I experienced some weird behaviour where network output is 0 for several seconds at the time, and all clients just hang in the wind and wait for a respons.
Remote to the server is working fine, so it's definitely a squid issue.. with only apache running the server is constantly sending out data, no halts there.
My site will wait for 30s almost everytime before loading any of the page itself.Specs of my install:
- DigitalOcean Droplet (VPS) with Ubuntu Server 12.10: 512 Ram and 20GB SSD (not even coming close to needing more RAM, still have 240MB free according to top) - Wordpress 3.6.1 - 5 plugins: W3 Total Cache, Wordpress SEO by yoast, WP Better Security, WP Smush.it, and Redirection (problem occured before adding the last 2, I can't remember about the others) - No traffic to speak of. I get maybe 10 uniques/day. - Apache 2.2.22 - MySQL 5.5.32
I've optimized my site itself the best I can, minifying and combine js and css files, using the WP Smush. It plugin to compress images, serving jQuery from a CDN, but none of that worked the 30 second wait (though it did shave about 10 seconds off the load time after the wait for response).
I was using cloudflare and had to fiddle with the nameservers of my domain, but cloudflare didn't work at all and I switched the nameservers back to normal pointing DNS directly at my site to eliminate the obvious causes. I'm comfortable with Linux and the command line. This is the link to my site: [URL] ....
I have installed NGINX + PHP-FPM + PHP5Memcached + Memcached + APC on my VServer. My Problem is Page loading time of around 40-60 seconds. Installed is 12.0.18 Update #49
[URL] .....
Connection Setup: TIME Stalled: 1.860 ms Request/Response TIME Request sent: 0.214 ms Waiting (TTFB): 47.08 s Content Download: 41.757 ms Explanation 47.12
I have reconfigured the Domains but nothing changed and Bootstrap Repair
Code: /usr/local/psa/admin/bin/httpdmng --reconfigure-all /usr/local/psa/bootstrapper/pp12.0.18-bootstrapper/bootstrapper.sh repair If I disable Nginx and use "Apache Module" for PHP than the TTFB time is shorter but is not the prefered way. If I use "CGI" or "FASTCGI" without Nginx than the Page loading time is a little bit shorter but with Erros on the Website.
I tried to reinstall Nginx from Plesk but this gave me Error after enable it
I want to copy a sql file from my secondary hdd to my NAS daily at 5AM and overwrite the older file with the same filename without prompting for confirmation.
how to save the work after you are done editing in SSH.
Currently I want to edit my cronjob in ssh so i type "crontab -e" and it shows the screen which displays cronjobs of the server so after all this how you save it after your done editing it in ssh?
So I am unsure where it is getting the "300 seconds" from.
Even though the error states 300 seconds, this problem happens way before 300 seconds every time and has happened during the transfer of files (when the connection has been active and in use).
I've tried 3 different FTP clients and used the "Keep Alive" option in each and it has absolutely no effect.
I am unsure if APF is causing the problem, I can't see any problems in any of the server logs in relation to ftp.
while i am restoring db (110MB) via SSH following error occur
Code: ERROR 1064 (42000) at line 145689: You have an error in your SQL syntax; check the manual that corresponds to your MySQL server version for the right syntax to use near ' Fatal error: Maximum execution time of 30 seconds exceeded in ' at line 1
I have a reseller account with Reseller Zoom. My problem started 72 hours ago when I noticed my WHMCS was not working on my website. I was also recieving Cron Daemon emails stating:
Site error: the file <b>/home/tmass10/public_html/clients/admin/cron.php</b> requires the ionCube PHP Loader ioncube_loader_lin_5.2.so to be installed by the site administrator.
at this time i was using the standard WHMCS cron: php -q /home/tmass10/public_html/clients/admin/cron.php
RZ told me they made a custom php.ini file and placed it in my directory. Then said to change my Cron job to
I have a cronjob that is still entered in my server, but due to my own lack of knowledge i deleted the domain from my server without removing the cronjob and now i keep getting the emails from my server
Code: /bin/sh: line 1: /home/userID_Removed/public_html/cgi-bin/script-removed/cron.cgi: No such file or directory I just need to know how to remove the cronjob via SSH, i logged in as Root to my SSH and then did crontab -l and I see the two cronjobs I want to remove, but can't find the code to do so anywhere via Google.
There is 3 total that I want to remove but there is about 10 cronjobs setup there so i want to remove those three and make sure the other 7 stay running as they should.
It's a php script with some mysql queries. For testing purposes in the end of script I added command to send me mail when cronjob is done. I never get mail.
If I click on "Run now" in Webmin interface, everything works fine - cronjob is done and I get mail. So, I am sure that there is nothing wrong with script, but for some reason my cronjob just doesn't run.
i'm install whmcs in My plesk 12 for Windows now i set Cron Job but system Send email with subject "Parallels Panel Scheduler notification" and attached Files and Write>> Could not open input file: php ...
It seems that since my latest update of 12.0.18. Now everytime I login to the Admin interface of plesk. I am prompted to relogin several times when browsing using file manager or clicking on domains or sitting idle for like 15 seconds.
It can do this up to 5 times in a 5 minute period or while not even being idle.
I am trying to setup a cronjob but cPanel, (Linux Centos 5), tells me that user/bin/crontab permissions are incorrect and need to be changed to 4755 - I try to change permissions in WinSCP [Properties] but it doesn't 'take'.
how to force a change of permissions with SSH (Putty)
Seems if I want to back up a domain's emails to a remote server (3x a week) I might be able to accomplish that with a Scheduled Task / Cronjob? I have a weekly full domain backup from the Plesk scheduled backup. But I can only set one backup per client. So I'd like to backup email separately, several times a week.