Large File Sizes
Jul 29, 2008
My client's website needs to hold files that are around 60 or 70 MB. The host only allows files up to 10 MB. Is that typical?
Right now I'm linking to a file storage but would rather make the files available from my site without going to a 3rd Party Site. He doesn't want to zip his files either - just to be a straight download.
View 7 Replies
ADVERTISEMENT
Mar 28, 2008
I have a situation here.
I wanted to transfer a file (about 10GB) from one server to another server. Both server are on the same LAN.
Inside server1, I zipped a folder as file.tar.gz
From server2, I uses wget http://server1.com/file.tar.gz and it says "404 Not Found"
I suspect it won't allow me to download because the file is too big.
View 7 Replies
View Related
Apr 14, 2009
So I've recently ordered a Supermicro 4U server with 24x1TB HDs, 64GB RAM and put it in RAID 10. I'm running Debian 5.0 and have installed lighttpd. All the content I serve are video files (AVi, MP4, MKV, OGM) and each file is about 100-500mb in size. I'm wondering how can I optimize lighttpd to get the best performance out of it. I look forward to your replies.
View 14 Replies
View Related
Mar 26, 2008
Can anyone tell me the format of SCP command for transfering an 18gig TAR.GZ file from one srever to another
say im logged in to the old host via SSH and the backup is located at /home/public_html/cpbackup/blahblah.tar.gz
what command would i put down to transfer to the new server
View 3 Replies
View Related
Apr 6, 2015
I have been trying quite unsuccessfully to import a large sql db file via phpMyAdmin for one of my clients. Since the db file is about 250mb I get a server timeout error.how I can do this via SSH...I have a CentOS server 6.5, 64 bit that runs Plesk v 12.0.18
View 4 Replies
View Related
Aug 21, 2007
on good hosting setups for getting large amounts of disk space.
I would like to be able to offer up to 2Gb storage space for 100s, maybe up to a few 1000 users - any solution should scale well. The files would be static files that might be up to 400Mb in size.
It would be nice to be able to give users FTP access to their disk space, although it's not a core requirement.
View 4 Replies
View Related
Jan 18, 2008
Using Cpanel:
-I went to backup option and downloaded the database backup. 951kb
-I went into phpadmin and checked addrop table. 1.5m
Using SSH:
command mysqldump -u myusername_usernam -pmypassword myusername_name > /home/name/backupz/name.sql = 949k
3 different sizes, I would figure maybe adding droptables would amount to higher size but the first way and third way i use, how come there is a size difference of 2kb?
Which is the right way to backup mysql database safely?
View 6 Replies
View Related
Aug 10, 2007
I have a dell 2950 server with raid 5 + a hotswap.
There are 4 73G 15K drives with TWO dual quad process.
4G mem
I was reading that the swap space s/b 2X the mem or 8Gigs.
The swap drive should be at the outer edge of the drive for speed.
Or drive C, but with these fast drives is it really going to matter
My layout idea so far is as follows
8 swap
20 OS and programs
20 SQL databases and websites
10 log files
10 downloads and upgrade files.
View 13 Replies
View Related
Jun 10, 2008
Knowing Qoodaa was quite by chance.
A few days ago, my friends studying in America recommended me a new popular transfer tool—Qoodaa. And he told me that it was a quite good software to download files and movies. At first,I was skeptical, but after using it, I found it’s a good choice to choose Qoodaa. And I have summarized the some features of Qoodaa:
1.Its speed is faster than any other softwares I used before to upload movies.
2.It can download files quickly through downloading links, in a word, it is time-saver and with high efficiency.
3.No limit of space.No matter where you are, it can download fast.
4. Qoodaa is a green software with high security and easy use.
It really can give you unexpected surprise.
I am a person who would like to share with others, and if you have sth good pls share with me
View 0 Replies
View Related
Jul 24, 2007
I'm currently running on a VPS. My site allows for large file uploads and downloads, with files over 600mb in size.
The server has issues when the site gets three or more requests for large file downloads. I'm trying to grow this site to thousands of users and it is hard to do when the site can't handle even three.
I've been told by my host that I need to upgrade to dedicated. My VPS only has 512mb RAM and one large file download is eating up that RAM. This is causing the issue.
I'm a newbie and while I knew I was risking a bit by going with VPS I do find it a bit annoying that these guys advertise 1TB of bandwidth per month but I can't even support downloading 1GB at the same time....maybe it's just me...
Anyway, I am now looking into moving the large files and the upload/download over to Amazon S3. If I do this I am expecting my RAM usage on the VPS to greatly decrease. Is this correct? If my PHP code is running on the VPS, but the actual file download via HTTP is coming from S3, that should not be a heavy load on my box, correct?
any opinions on S3?
View 2 Replies
View Related
Feb 10, 2015
I'm build Plesk Panel for Linux and Presence Builder, I don't want my user can upload their website to hosting via File Manager. How can I do it...
View 2 Replies
View Related
May 26, 2008
Say I have 2 websites and they all use file.php which is located on mainserver.com/file.php.
I want to use the file like this:
website1.com/file.php
website2.com/file.php
View 2 Replies
View Related
Mar 6, 2008
Rapidly growing error logs showing the same message
$ug-non-zts-20020429/ffmpeg.so' - /usr/local/lib/php/extensions/no-debug-non-zts-20020429//usr/local/lib/php/extensions/no-debug-non-zts-20020429/ffmpeg.so: cannot open shared object file: No such file or directory in Unknown on line 0
root@server [~]# ls /usr/local/lib/php/extensions/no-debug-non-zts-20020429
./ ../ eaccelerator.so*
root@server [~]# ls /usr/local/lib/php/extensions/no-debug-non-zts-20020429
./ ../ eaccelerator.so*
Using cpanel 11 / centos 4
View 1 Replies
View Related
Jun 16, 2008
i have a server with centos,
i need to edit the hidden file .htaccess from the file management tool of cpanel,
but the hidden files not shown,
ow can i modify the setting and let the files shown in the file management tool of cpanel?
View 6 Replies
View Related
Sep 17, 2014
How can we stop Plesk resetting the file permissions on a dll file that is found in
C:Program Files (x86)
??
Specifically,
we have a file, jmail.dll,
here
C:Program Files (x86)Dimacw3JMail
By default Plesk permissions are set to DENY for PSACLN.
But the JMail plugin cannot work with these permissions !
We change this to be ALLOW for READ & EXECUTE and DENY for WRITE.
And everything works fine.
But every time Plesk does an update it reverts it back !
This means that a number of our customers contact forms stop working !
View 12 Replies
View Related
Nov 22, 2008
I'm trying to do this
/usr/bin/gzip -p /home/mysite/public_html/shop/feeds/myfile.xml.zip > /home/mysite/public_html/shop/feeds/myfile.xml
But it just tells me
/usr/bin/gzip: invalid option -- p
X-Powered-By: PHP/5.2.5
Content-type: text/html
How do I find the correct option to unzip first file to second file?
View 8 Replies
View Related
Oct 10, 2014
i manage linux apache webserver with a few wordpress blogs and from time to time i see someone inject a malicious .php file into wp-content/uploads/2014/10/ directory.
i think its some bad plugin or theme, but these is more blogs, i ugrade, update, WP, but
how can i setup some monitor to tell me which php file (or even line in php file) injected that malicious .php ? I have linux root access so i can setup anythingÂ
View 3 Replies
View Related
Apr 5, 2007
I'm having a lengthy issue where my databases are to large to import in phpmyadmin using plesk. Unfortunately I dont have direct access to phpmyadmin and can only access it by DB user through plesk.
I have tried to edit php.ini in the following locations:
upload_max_filesize = changed this to 64M
post_max_size = changed this to 32M
maximum_execution_time = changed this to 300
maximum_input_time = changed this to 300
Why am I still not able to import my DB's which are about 8MB each?
View 4 Replies
View Related
Oct 2, 2009
what type of hosting do use big sites like Deviantart, or any kind of site with a lot of Gb or Tb. and how much it costs?
View 7 Replies
View Related
May 20, 2008
I have a website which has about 20K users, and now I am using VPS plan at LunarPages.
However, I have encountered a trouble of out-of-memory. Although I have configured my Apache and MySQL carefully, the 512M memory is not enough. Therefore, the users' expirence is not good these days because my site is very unstable.
I contacted Lunarpages, asking them whether I can upgrade my VPS to bigger RAM, but they said the ONLY way to get a RAM bigger than 512M is to upgrade to dedicated hosting plan.
The following are some stats of my website:
Total Members: 20k
Online at the same time: max 600, average 300
The Lunarpages VPS plan:
www[dot]lunarpages[dot]com/virtual-private-server/
disk space: 20G
RAM: 512M
price: $42 / mo
Now I am not sure whether to migrate to didicated hosting plan, because currently, the main problem is just the size of RAM. Other resources e.g (CPU, network etc. ) are not my bottleneck. So I think it seems not worthwhile for me to migrate to the dedicated hosting plan with a doubled price (even more, almost 3x if I need 1G RAM), just for a larger size of RAM.
Can you guys give some suggestions to choose a VPS provider for my site?
The factors taken into my consideration include:
* RAM size: at least (1G for peak, 768M garantee). The bigger, the better. Nice if can choose larger size when needed.
* price
* bandwidth: 1T/mon?
* easy to upgrade to dedicated host: just in case that one day I will have to use dedicated.
* whether there are coupons for a lower price.
View 9 Replies
View Related
May 22, 2007
I've been with zone.net for a couple months now, and I have a guaranteed 512MB of memory, which I seem to constantly be hitting, which seems to result in processes being killed and http access vanishing. Growing quite annoying.
I'm looking into moving onto a new provider that can provide more guaranteed RAM for about the same price.
Space isn't a huge deal, I'd do fine with a meager 5GB. Bandwidth I need at least 200GB, but wouldn't mind more.
I'd like to stay managed if possible, as I'm not as well versed in server workings as I should be. Also am in need of cPanel, which I know is a spendy sucker.
My budget is something around $70 a month, and I don't really want to go much higher than that. Still a poor college boy :/
Can anyone suggest such a provider? I've browsed around a lot of the VPS hosts but can't seem to find one that has as much RAM as I need for a decent price. All the ones that seem to have 512MB+ are pretty expensive, and offer a lot more other stuff (space/bandwidth) than I need.
As a final note, the line speed isn't that big of a deal. I'm currently on a 3mbit and am surviving, but going back to a higher speed line would be great
View 11 Replies
View Related
Jan 17, 2007
Just had a quick question about backing up a large MySQL DB. I have a database that is 50gb with about half a billion entries in it. One table itself is about 40gb, the other 10gb consists of smaller tables.
The problem is, I want to back the database up and be able to keep it LIVE at the same time (as it will fall behind quickly if it's pulled for more than a few hours, as there are somewhere in the area of a million entries an hour, plus other deletions and queries).
View 3 Replies
View Related
May 30, 2007
I'm currently using iptables to ban IP addresses from the servers, like:
Code:
iptables -A INPUT -s xxx.xxx.xxx.xxx -j DROP
I ran a "spam trap" for the last few months and now I have over 11000 IP addresses who were trying to spam on my website (guestbooks, phpBB and forms) and I want to ban them all (pretty sure bots run from them).
My question - is iptables the way to do it? I mean does banning such a large number of addresses have any significant performance or other issues I should be aware of (except of the fact I may be banning some legitimate traffic)? Is the -A INPUT the way to ban them all or is there a more appropriate way of baning such a number of addresses?
I'm on CentOS 4.5 i686, Apache/1.3.37, Pentium D 930, 2GB RAM.
View 11 Replies
View Related
Jul 9, 2007
Today I found some cstomer on the servers make a link for named it file.txt and link it to other customer php file.
so that customer have the ability to show the other custoer file content when visiting the url because it is a text wile originally it is a php file.
the php file was a config file, so now he know the database password , and because he is in the same server he can use that databse.
the question , how to avoide this prolem in the future?
notes , the SuExec is rnning and the open_basedir protection is enabled, but the problem still exists.
View 12 Replies
View Related
Feb 2, 2008
I wasn't sure where to post this so here goes, I need to migrate a MySQL DB, in the past I have just created an SQL file and used that method (sometimes having to split the SQL file up) but now the DB is about 50 meg and 733,233 records.
Is there an easier way to migrate the Database from one server to another?
View 8 Replies
View Related
Apr 5, 2009
I'm trying to find a server that can offer about 7-8 TB with 6-8 GB of ram. Does anyone know a good provider?
View 14 Replies
View Related
Mar 28, 2009
any recommendation for a good e-mail service (IMAP) for large accounts? between 5 and 10GB each?
View 13 Replies
View Related
Mar 9, 2009
I'm selling downloads of music files. The zip files are quite large. I've had several people complain that they get a message that the server resets their connection before the download finishes.
What can I do to allow these large downloads?
I'm on IIS.
View 21 Replies
View Related