$ wget \ --recursive \ --no-clobber \ --page-requisites \ --html-extension \ --convert-links \ --restrict-file-names=windows \ --domains website.org \ --no-parent \ www.website.org/tutorials/html/
source: Linux Journal
$ wget \ --recursive \ --no-clobber \ --page-requisites \ --html-extension \ --convert-links \ --restrict-file-names=windows \ --domains website.org \ --no-parent \ www.website.org/tutorials/html/
source: Linux Journal
Sometimes you have to remove directories from your VCS tool.
find . -name .svn -exec rm -rf {} \;
find . -name .svn -exec echo {} \;
--head
– performs HEAD http request (not GET).
curl --head https://www.wp.pl/
--get
– performs GET request
-i
– include headers
curl --get -i https://www.wp.pl/
Simple thing if you want to have min
and max
in SQL query.
ROUND((RAND() * (max-min))+min)
So, you have to copy one directory into another on your server and you don’t want to kill the machine?
Instead of
cp -r master master_copy
use
rsync --bwlimit=30000 -av master/ master_copy/
The –bwlimit=KBPS switch is for limiting I/O bandwidth.
Small thing but extremely useful. Getting context with grep
.
-C
– Print num lines of leading and trailing context surrounding each match. The default is 2 and is equivalent to -A 2 -B 2. Note: no whitespace may be given between the option and its argument.
grep -C 5 file.txt
Title: Agile w pracy agencyjnej – z dwóch perspektyw – programisty i project managera
Venue: LoGeek Night #2, 23.01.2014
Slides: Slideshare
If you have to remove large number of files, like hundreds of thousands, and don’t want to kill your production server, instead of rm -fr DIR
you can use this little fella.
while [ true ]; do ionice -c 3 find DIRECTORY_TO_BE_DELETED -type f -print | head -50000 | xargs rm -f; sleep 50; done
Simple, yet useful. -e
option does the trick.
zip -r -9 -e name.zip folder/
Title: Architektura to podstawa. Efektywny rozwój rozbudowanych projektów przy użyciu Symfony 2
Venue: DevCon GigaCon Wrocław, 23.06.2013
Slides: Slideshare