The 'adult' content policy messages could have been a PR move because they really had no idea what was going on. No company ever tells the public that it's dealing with a massive crisis that hasn't surfaced yet--there's no need if it's not public. If that was the case, it was still a bad move.
Actually, ironically, just last week I got an email reply from Amazon's digital music department that not only answered my question but gave me a phone number to call to walk me through my number (which I called, and which walked me through my problem).
Well, sure. links is a real program that really does go out and fetch pages from the Web.
But if you run links as:
links -dump
You get a formatted version of the page; you don't get the raw HTML, which is what you need if you want to extract the links.
I mean, it's a side point, because it's obviously possible to write a script that'll extract every product ID on Amazon, but the code's undeniably buggy as posted.
Elinks 0.10.6 is what is installed by default on ubuntu 6.02.2 LTS when I type "apt-get install links". It does dump links at the end of the file, which are trivially grepped for.
Please type links --version and find out what version of links you are using. Then install elinks 0.10.6 and retry the code.
Also, it doesn't account for all of the books that were taken down. Amongst other things, straight erotica also went down, and that wouldn't go down with a "I'm targeting GLTB books" code.
If someone can come up with evidence that it worked prior to today I'll recant, because turning off a Web page is exactly how I'd turn off that capability without having to do a code push. But I wanna see some evidence first.
Comments 161
Form a common sense standpoint, his story would not explain the two Amazon reps' response template about adult content.
Thank you.
Reply
Reply
(The comment has been removed)
Reply
Reply
merk@locke:~ [31/83]$ links --version
ELinks 0.10.6 (built on Sep 25 2007 18:50:54)
Features:
Standard, Fastmem, IPv6, gzip, bzip2, Cascading Style Sheets, Protocol (File, FTP, HTTP, NNTP, SMB, URI rewrite, User protocols),
SSL (GnuTLS), MIME (Option system, Mailcap, Mimetypes files), LED indicators,
Bookmarks, Cookies, Form History, Global History, Scripting (Lua, Perl)
Reply
But if you run links as:
links -dump
You get a formatted version of the page; you don't get the raw HTML, which is what you need if you want to extract the links.
I mean, it's a side point, because it's obviously possible to write a script that'll extract every product ID on Amazon, but the code's undeniably buggy as posted.
Reply
http://www.youtube.com/v/hYk-F5FXIb0
Elinks 0.10.6 is what is installed by default on ubuntu 6.02.2 LTS when I type "apt-get install links". It does dump links at the end of the file, which are trivially grepped for.
Please type links --version and find out what version of links you are using. Then install elinks 0.10.6 and retry the code.
I'd appreciate a retraction, thanks.
Reply
$ head -n 1 amazon
104. http://www.amazon.com/Lessons-Kim-Pritekel/dp/1933720085/ref=sr_1_1/182-9014588-9333611?ie=UTF8&s=books&qid=1239630457&sr=1-1
$ head -n 1 amazon|sed s/.*dp\\/// |sed s/\\/ref.*//
1933720085
Reply
Reply
That the code doesn't work at all, though......
Reply
He does insert the product IDs after http://www.amazon.com/ri/product-listing/ like so:
http://www.amazon.com/ri/product-listing/0830823794
But still results in 404.
Reply
If someone can come up with evidence that it worked prior to today I'll recant, because turning off a Web page is exactly how I'd turn off that capability without having to do a code push. But I wanna see some evidence first.
Reply
Leave a comment