e*****r 发帖数: 379 | 1 Hi, guys,
How to fetch PDF files from a journal website automatically?
Most journals today tell you the link of PDF to the paper, and when you
click the link in web explorer, and then right click again for saving to
PDFs.
This is very time consuming! I wonder there must be a way to do it
automatically.
I have experience on vb et al, but not on Web programming
Thanks! | u**d 发帖数: 211 | 2 批量下载,人家立马就把你的 ip 封了
【在 e*****r 的大作中提到】 : Hi, guys, : How to fetch PDF files from a journal website automatically? : Most journals today tell you the link of PDF to the paper, and when you : click the link in web explorer, and then right click again for saving to : PDFs. : This is very time consuming! I wonder there must be a way to do it : automatically. : I have experience on vb et al, but not on Web programming : Thanks!
| n****g 发帖数: 150 | | s****h 发帖数: 15 | 4 有些可以下载的,如ACM Digital Library.但是你要遵守它们的规则。
看看www.acm.org/robots.txt这个文件。
好像是间隔是80s, 看看googlebot,给的是20s :) | K****n 发帖数: 5970 | | p*******d 发帖数: 359 | 6 没那么严重把,都publish了还怕人看。写一个parser读前一叶的link,大不了让
thread等2分钟在下下一篇。
【在 K****n 的大作中提到】 : 这你要研究他们的命名规则,相当与骇客
|
|