Under construction!
I'm migrating this blog from an ancient BlogEngine (.NET, XML) to Hugo (Go, MD)...

Parsing

2013
Recently I needed to acquire some database. Unfortunately it was published only as a website that presented 50 records per single page. Whole database had more than 150 thousand records. What to do in such situation? Click through 3000 pages, manually collecting data in a text file? One week and it’s done! ;) Better to write a program (so called scraper) which will do the work for you. The program has to do three things:
2007
Sorry, this post is not yet available in English. See Polish version.