I need a Perl script to craw 10 subtitles websites and download all their subtitles.
http://www.opensubtitles.org/
http://subscene.com/
http://www.divxsubtitles.net/
http://www.subtitlesource.org/
http://www.podnapisi.net/
http://www.mysubtitles.com/
http://www.allsubs.org/
http://www.subtitleonline.com/
http://www.tvsubtitles.net/
http://www.sub-titles.net/
In some websites the subtitles are .SRT, .SUB, .TXT, .SSA, .SMI, or .MPL while in other websites they are inside .ZIP files. Your script should accommodate for that.
The downloads should be in 10 different folders (one folder per each web site, named accordingly).
Your scripts should be able to avoid downloading unnecessary files (no need to download JPGs if all I need is to crawl the HTML in order to identify the SRT files). So please make a list of downlaodable / parsable files (that I can modify later on).
The script should be commented so that I can modify it later on.
接包方 | 国家/地区 | |
---|---|---|
![]() |
3
Ks_unni
|
|
![]() |
3
Tonychan1
|
|
![]() |
3
Mpla
|
|
![]() |
3
Xercestechnologies
(中标)
|
|
![]() |
2
Peternlewis
|
|
![]() |
2
Dynamixsoftware
|