web site sAVE

justinups

New Member
Is It Posibale To Save An Entire Web Site And I Mean Links And All.

SO WHEN YOU ARE OFF LINE YOU CAN VIEW THOUSE LINKS WITH OUT ANY PROLBEMS
 
Last edited:
Yes there a program that grabs the whole web site and saves it (including all the medias,scripts and stuff) talking about that would be AGAINST the forum RULE
 
i think a Gold member like you understand why.
if you really want the program name google for it
as you wont find it here
 
no, i am not interested, just strickes me a odd being aganced the rules. (have to re look at them) :P

edit: not to be rude or anything (just would like to know) what rule it is braking. all i could see it under is maybe illegal but i dont understan how. could you please explane. thanks
 
Last edited:
no, i am not interested, just strickes me a odd being aganced the rules. (have to re look at them) :P

edit: not to be rude or anything (just would like to know) what rule it is braking. all i could see it under is maybe illegal but i dont understan how. could you please explane. thanks

Downloading peoples content and work, without the consent of the author

EDIT : would you like it you made a website and someones just comes at the click of a mouse and download your codes etc..and make them their own.
i guess not !!
 
ok, just though it was public just beacause you can easly use the view sorce butten. o well. thanks patrickv.
 
Almost all web browsers have the option to save a webpage as a HTML file right on your desktop.

If it is the whole content of the page you want you may look into some Linux/Unix commands.

the curl command comes into mind.
 
Almost all web browsers have the option to save a webpage as a HTML file right on your desktop.

If it is the whole content of the page you want you may look into some Linux/Unix commands.

the curl command comes into mind.
You can save an individual page like that, but a program such as I suggested lets you download a set number of links from the home page. I'm sure Linux would work, but it's not as practical for most users.
 
just to give an example of how powerful *nix terminal usage can be:

this is what I get when I use curl on www.google.com

Code:
tlarkins-power-mac-g4:~ tlarkin$ curl www.google.com
<html><head><meta http-equiv="content-type" content="text/html; charset=ISO-8859-1"><title>Google</title><style><!--
body,td,a,p,.h{font-family:arial,sans-serif}
.h{font-size:20px}
.h{color:#3366cc}
.q{color:#00c}
.ts td{padding:0}.ts{border-collapse:collapse}#gbar{float:left;font-weight:bold;height:22px;padding-left:2px}#gbh{border-top:1px solid #c9d7f1;font-size:0;height:0;position:absolute;right:0;top:24px;width:200%}#gbi{background:#fff;border:1px solid;border-color:#c9d7f1 #36c #36c #a2bae7;font-size:13px;top:24px;z-index:1000}#guser{padding-bottom:7px !important}#gbar,#guser{font-size:13px;padding-top:1px !important}.gb1,.gb3{display:inline;height:22px;margin-right:1em;vertical-align:top}#gbi,.gb2{display:none;position:absolute;width:7em}.gb2{z-index:1001}#gbar a,#gbar a:active,#gbar a:visited{color:#00c;font-weight:normal}.gb2 a,.gb3 a{text-decoration:none}.gb2 a{display:block;padding:.2em .5em}#gbar .gb2 a:hover{background:#36c;color:#fff}--></style>
<script>
<!--
window.google={kEI:"wlWiRqzhF4O4jgHvtsiFDA",kEXPI:"17259,17497",kHL:"en"};function sf(){document.f.q.focus();}
window.gbar={};(function(){;var h=window.gbar,a,g,f;function m(b,e,d){b.display=b.display=="block"?"none":"block";b.left=e+"px";b.top=d+"px"}h.tg=function(b){var e=0,d,c,i,j=0,k=window.navExtra;!g&&(g=document.getElementById("gbar"));!f&&(f=g.getElementsByTagName("div"));(b||window.event).cancelBubble=true;if(!a){a=document.createElement(Array.every||window.createPopup?"iframe":"div");a.frameBorder="0";a.id="gbi";a.scrolling="no";a.src="#";document.body.appendChild(a);if(k)for(var n in k){var l=document.createElement("div");l.appendChild(k[n]);l.className="gb2";g.appendChild(l)}document.onclick=h.close}for(;f[j];j++){c=f[j];i=c.className;if(i=="gb3"){d=c.offsetLeft;while(c=c.offsetParent)d+=c.offsetLeft;m(a.style,d,24)}else if(i=="gb2"){m(c.style,d+1,25+e);e+=20}}a.style.height=e+"px"};h.close=function(b){a&&a.style.display=="block"&&h.tg(b)};})();// -->
</script>
</head><body bgcolor=#ffffff text=#000000 link=#0000cc vlink=#551a8b alink=#ff0000 onload="sf();if(document.images){new Image().src='/images/nav_logo3.png'}" topmargin=3 marginheight=3><div id=gbar><nobr><div class=gb1>Web</a></div><div class=gb1><a href=http://images.google.com/imghp?tab=wi>Images</a></div><div class=gb1><a href=http://video.google.com/?tab=wv>Video</a></div><div class=gb1><a href=http://news.google.com/nwshp?tab=wn>News</a></div><div class=gb1><a href=http://maps.google.com/maps?tab=wl>Maps</a></div><div class=gb1><a href=http://mail.google.com/mail?tab=wm>Gmail</a></div><div class=gb3><a href=http://www.google.com/intl/en/options/ onclick="this.blur();gbar.tg(event);return false"><u>more</u> <span style=font-size:11px>▼</span></a></div><div class=gb2><a href=http://blogsearch.google.com/?tab=wb>Blog Search</a></div><div class=gb2><a href=http://www.blogger.com/?tab=wj>Blogger</a></div><div class=gb2><a href=http://books.google.com/bkshp?tab=wp>Books</a></div><div class=gb2><a href=http://www.google.com/calendar?tab=wc>Calendar</a></div><div class=gb2><a href=http://docs.google.com/?tab=wo>Documents</a></div><div class=gb2><a href=http://finance.google.com/finance?tab=we>Finance</a></div><div class=gb2><a href=http://groups.google.com/grphp?tab=wg>Groups</a></div><div class=gb2><a href=http://labs.google.com/>Labs</a></div><div class=gb2><a href=http://www.orkut.com/?tab=w0>Orkut</a></div><div class=gb2><a href=http://www.google.com/ptshp?tab=wt>Patents</a></div><div class=gb2><a href=http://picasaweb.google.com/home?tab=wq>Photos</a></div><div class=gb2><a href=http://www.google.com/prdhp?tab=wf>Products</a></div><div class=gb2><a href=http://www.google.com/reader?tab=wy>Reader</a></div><div class=gb2><a href=http://scholar.google.com/schhp?tab=ws>Scholar</a></div></nobr></div><div id=gbh></div><div align=right id=guser style="font-size:84%;padding:0 0 4px" width=100%><nobr><a href="/url?sa=p&pref=ig&pval=3&q=http://www.google.com/ig%3Fhl%3Den&usg=AFQjCNEj49wK5T88bDwGcaZsW52jiYwJwg">iGoogle</a> | <a href="https://www.google.com/accounts/Login?continue=http://www.google.com/&hl=en">Sign in</a></nobr></div><center><br id=lgpd><img alt="Google" height=110 src="/intl/en_ALL/images/logo.gif" width=276><br><br><form action="/search" name=f><table cellpadding=0 cellspacing=0><tr valign=top><td width=25%>&nbsp;</td><td align=center nowrap><input name=hl type=hidden value=en><input type=hidden name=ie value="ISO-8859-1"><input maxlength=2048 name=q size=55 title="Google Search" value=""><br><input name=btnG type=submit value="Google Search"><input name=btnI type=submit value="I'm Feeling Lucky"></td><td nowrap width=25%><font size=-2>&nbsp;&nbsp;<a href=/advanced_search?hl=en>Advanced Search</a><br>&nbsp;&nbsp;<a href=/preferences?hl=en>Preferences</a><br>&nbsp;&nbsp;<a href=/language_tools?hl=en>Language Tools</a></font></td></tr></table></form><br><br><font size=-1><a href="/intl/en/ads/">Advertising&nbsp;Programs</a> - <a href="/services/">Business Solutions</a> - <a href="/intl/en/about.html">About Google</a></font><p><font size=-2>&copy;2007 Google</font></p></center></body></html>tlarkins-power-mac-g4:~ tlarkin$

here goe curl's options in BASH
Code:
tlarkins-power-mac-g4:~ tlarkin$ curl --help
Usage: curl [options...] <url>
Options: (H) means HTTP/HTTPS only, (F) means FTP only
 -a/--append        Append to target file when uploading (F)
 -A/--user-agent <string> User-Agent to send to server (H)
    --anyauth       Tell curl to choose authentication method (H)
 -b/--cookie <name=string/file> Cookie string or file to read cookies from (H)
    --basic         Enable HTTP Basic Authentication (H)
 -B/--use-ascii     Use ASCII/text transfer
 -c/--cookie-jar <file> Write cookies to this file after operation (H)
 -C/--continue-at <offset> Resumed transfer offset
 -d/--data <data>   HTTP POST data (H)
    --data-ascii <data>   HTTP POST ASCII data (H)
    --data-binary <data>  HTTP POST binary data (H)
    --negotiate     Enable HTTP Negotiate Authentication (H)
    --digest        Enable HTTP Digest Authentication (H)
    --disable-eprt  Prevent curl from using EPRT or LPRT (F)
    --disable-epsv  Prevent curl from using EPSV (F)
 -D/--dump-header <file> Write the headers to this file
    --egd-file <file> EGD socket path for random data (SSL)
    --tcp-nodelay   Set the TCP_NODELAY option
 -e/--referer       Referer URL (H)
 -E/--cert <cert[:passwd]> Client certificate file and password (SSL)
    --cert-type <type> Certificate file type (DER/PEM/ENG) (SSL)
    --key <key>     Private key file name (SSL)
    --key-type <type> Private key file type (DER/PEM/ENG) (SSL)
    --pass  <pass>  Pass phrase for the private key (SSL)
    --engine <eng>  Crypto engine to use (SSL). "--engine list" for list
    --cacert <file> CA certificate to verify peer against (SSL)
    --capath <directory> CA directory (made using c_rehash) to verify
                    peer against (SSL)
    --ciphers <list> SSL ciphers to use (SSL)
    --compressed    Request compressed response (using deflate or gzip)
    --connect-timeout <seconds> Maximum time allowed for connection
    --create-dirs   Create necessary local directory hierarchy
    --crlf          Convert LF to CRLF in upload
 -f/--fail          Fail silently (no output at all) on errors (H)
    --ftp-create-dirs Create the remote dirs if not present (F)
    --ftp-pasv      Use PASV instead of PORT (F)
    --ftp-ssl       Enable SSL/TLS for the ftp transfer (F)
 -F/--form <name=content> Specify HTTP multipart POST data (H)
 -g/--globoff       Disable URL sequences and ranges using {} and []
 -G/--get           Send the -d data with a HTTP GET (H)
 -h/--help          This help text
 -H/--header <line> Custom header to pass to server (H)
 -i/--include       Include protocol headers in the output (H/F)
 -I/--head          Show document info only
 -j/--junk-session-cookies Ignore session cookies read from file (H)
    --interface <interface> Specify network interface to use
    --krb4 <level>  Enable krb4 with specified security level (F)
 -k/--insecure      Allow curl to connect to SSL sites without certs (H)
 -K/--config        Specify which config file to read
 -l/--list-only     List only names of an FTP directory (F)
    --limit-rate <rate> Limit transfer speed to this rate
 -L/--location      Follow Location: hints (H)
    --location-trusted Follow Location: and send authentication even 
                    to other hostnames (H)
 -m/--max-time <seconds> Maximum time allowed for the transfer
    --max-redirs <num> Maximum number of redirects allowed (H)
    --max-filesize <bytes> Maximum file size to download (H/F)
 -M/--manual        Display the full manual
 -n/--netrc         Must read .netrc for user name and password
    --netrc-optional Use either .netrc or URL; overrides -n
    --ntlm          Enable HTTP NTLM authentication (H)
 -N/--no-buffer     Disable buffering of the output stream
 -o/--output <file> Write output to <file> instead of stdout
 -O/--remote-name   Write output to a file named as the remote file
 -p/--proxytunnel   Operate through a HTTP proxy tunnel (using CONNECT)
    --proxy-basic   Enable Basic authentication on the proxy (H)
    --proxy-digest  Enable Digest authentication on the proxy (H)
    --proxy-ntlm    Enable NTLM authentication on the proxy (H)
 -P/--ftp-port <address> Use PORT with address instead of PASV (F)
 -q                 If used as the first parameter disables .curlrc
 -Q/--quote <cmd>   Send command(s) to server before file transfer (F)
 -r/--range <range> Retrieve a byte range from a HTTP/1.1 or FTP server
    --random-file <file> File for reading random data from (SSL)
 -R/--remote-time   Set the remote file's time on the local output
    --retry <num>   Retry request <num> times if transient problems occur
    --retry-delay <seconds> When retrying, wait this many seconds between each
    --retry-max-time <seconds> Retry only within this period
 -s/--silent        Silent mode. Don't output anything
 -S/--show-error    Show error. With -s, make curl show errors when they occur
    --socks <host[:port]> Use SOCKS5 proxy on given host + port
    --stderr <file> Where to redirect stderr. - means stdout
 -t/--telnet-option <OPT=val> Set telnet option
    --trace <file>  Write a debug trace to the given file
    --trace-ascii <file> Like --trace but without the hex output
 -T/--upload-file <file> Transfer <file> to remote site
    --url <URL>     Spet URL to work with
 -u/--user <user[:password]> Set server user and password
 -U/--proxy-user <user[:password]> Set proxy user and password
 -v/--verbose       Make the operation more talkative
 -V/--version       Show version number and quit
 -w/--write-out [format] What to output after completion
 -x/--proxy <host[:port]> Use HTTP proxy on given port
 -X/--request <command> Specify request command to use
 -y/--speed-time    Time needed to trig speed-limit abort. Defaults to 30
 -Y/--speed-limit   Stop transfer if below speed-limit for 'speed-time' secs
 -z/--time-cond <time> Transfer based on a time condition
 -0/--http1.0       Use HTTP 1.0 (H)
 -1/--tlsv1         Use TLSv1 (SSL)
 -2/--sslv2         Use SSLv2 (SSL)
 -3/--sslv3         Use SSLv3 (SSL)
    --3p-quote      like -Q for the source URL for 3rd party transfer (F)
    --3p-url        source URL to activate 3rd party transfer (F)
    --3p-user       user and password for source 3rd party transfer (F)
 -4/--ipv4          Resolve name to IPv4 address
 -6/--ipv6          Resolve name to IPv6 address
 -#/--progress-bar  Display transfer progress as a progress bar

Another great terminal command is wget which allows to you to directly download files from the terminal.

I am not a web developer but those who are can plainly see how powerful this tool is, and also how it could be put to bad use if someone was so inclined. Its greatest asset is if you are doing work on someone's webpage for them, you can use this command to simply download their code, modify whatever you need and send it back to the person.

There are easy ways to run Linux with in windows also since VPC is now a free download from MS. You can simply run it, then download any flavor of linux you want (the ISO file) and then just run that ISO as a virtual machine, do your terminal work, output to a text file and then copy it back over to windows. Not very hard at all.

I did this on my powermac G4 running OS X 10.4.10 just an FYI.
 
Back
Top