summaryrefslogtreecommitdiff
path: root/posts/2009-06-06-modules-i-like-web-scraper.md
diff options
context:
space:
mode:
Diffstat (limited to '')
-rw-r--r--posts/2009-06-06-modules-i-like-web-scraper.md (renamed from content/post/2009-06-06-modules-i-like-web-scraper.md)6
1 files changed, 0 insertions, 6 deletions
diff --git a/content/post/2009-06-06-modules-i-like-web-scraper.md b/posts/2009-06-06-modules-i-like-web-scraper.md
index ba383d1..519fc10 100644
--- a/content/post/2009-06-06-modules-i-like-web-scraper.md
+++ b/posts/2009-06-06-modules-i-like-web-scraper.md
@@ -1,9 +1,3 @@
----
-date: 2009-06-06T00:00:00Z
-summary: In which I talk about Web::Scraper.
-title: Modules I like Web::Scraper
----
-
For [$work](http://rtgi.fr) I need to write scrapers. It used to be boring and painful. But thanks to [miyagawa](http://search.cpan.org/~miyagawa/), this is not true anymore. [Web::Scraper](http://search.cpan.org/perldoc?Web::Scraper) offer a nice API: you can write your rules using XPath, you can chaine rules, a nice and simple syntax, etc.
I wanted to export my data from my last.fm account but there is no API for this, so I would need to scrap them. All the data are available [as a web page](http://www.last.fm/user/franckcuny/tracks) that list your music. So the scraper need to find how many pages, and find the content on each page to extract a list of your listening.