112 lines
3.9 KiB
Java
112 lines
3.9 KiB
Java
package com.rarchives.ripme.ripper.rippers;
|
|
|
|
import java.io.IOException;
|
|
import java.net.MalformedURLException;
|
|
import java.net.URL;
|
|
import java.util.ArrayList;
|
|
import java.util.List;
|
|
import java.util.regex.Matcher;
|
|
import java.util.regex.Pattern;
|
|
|
|
import org.jsoup.nodes.Document;
|
|
import org.jsoup.nodes.Element;
|
|
|
|
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
|
import com.rarchives.ripme.utils.Http;
|
|
|
|
public class Hentai2readRipper extends AbstractHTMLRipper {
|
|
String lastPage;
|
|
|
|
public Hentai2readRipper(URL url) throws IOException {
|
|
super(url);
|
|
}
|
|
|
|
@Override
|
|
public String getHost() {
|
|
return "hentai2read";
|
|
}
|
|
|
|
@Override
|
|
public String getDomain() {
|
|
return "hentai2read.com";
|
|
}
|
|
|
|
@Override
|
|
public String getGID(URL url) throws MalformedURLException {
|
|
Pattern p = Pattern.compile("https?://hentai2read\\.com/([a-zA-Z0-9_-]*)/(\\d)?/?");
|
|
Matcher m = p.matcher(url.toExternalForm());
|
|
if (m.matches()) {
|
|
return m.group(1);
|
|
}
|
|
throw new MalformedURLException("Expected hentai2read.com URL format: " +
|
|
"hbrowse.com/COMICID - got " + url + " instead");
|
|
}
|
|
|
|
@Override
|
|
public Document getFirstPage() throws IOException {
|
|
Document tempDoc;
|
|
// get the first page of the comic
|
|
if (url.toExternalForm().substring(url.toExternalForm().length() - 1).equals("/")) {
|
|
tempDoc = Http.url(url + "1").get();
|
|
} else {
|
|
tempDoc = Http.url(url + "/1").get();
|
|
}
|
|
for (Element el : tempDoc.select("ul.nav > li > a")) {
|
|
if (el.attr("href").startsWith("https://hentai2read.com/thumbnails/")) {
|
|
// Get the page with the thumbnails
|
|
return Http.url(el.attr("href")).get();
|
|
}
|
|
}
|
|
throw new IOException("Unable to get first page");
|
|
}
|
|
|
|
@Override
|
|
public String getAlbumTitle(URL url) throws MalformedURLException {
|
|
try {
|
|
return getHost() + "_" + getGID(url);
|
|
} catch (Exception e) {
|
|
// Fall back to default album naming convention
|
|
logger.warn("Failed to get album title from " + url, e);
|
|
}
|
|
return super.getAlbumTitle(url);
|
|
}
|
|
|
|
@Override
|
|
public List<String> getURLsFromPage(Document doc) {
|
|
List<String> result = new ArrayList<String>();
|
|
for (Element el : doc.select("div.block-content > div > div.img-container > a > img.img-responsive")) {
|
|
String imageURL = "https:" + el.attr("src");
|
|
imageURL = imageURL.replace("hentaicdn.com", "static.hentaicdn.com");
|
|
imageURL = imageURL.replace("thumbnails/", "");
|
|
imageURL = imageURL.replace("tmb", "");
|
|
result.add(imageURL);
|
|
}
|
|
return result;
|
|
}
|
|
|
|
@Override
|
|
public Document getNextPage(Document doc) throws IOException {
|
|
// Find next page
|
|
String nextUrl = "";
|
|
Element elem = doc.select("div.bg-white > ul.pagination > li > a").last();
|
|
if (elem == null) {
|
|
throw new IOException("No more pages");
|
|
}
|
|
nextUrl = elem.attr("href");
|
|
// We use the global lastPage to check if we've already ripped this page
|
|
// and is so we quit as there are no more pages
|
|
if (nextUrl.equals(lastPage)) {
|
|
throw new IOException("No more pages");
|
|
}
|
|
lastPage = nextUrl;
|
|
// Sleep for half a sec to avoid getting IP banned
|
|
sleep(500);
|
|
return Http.url(nextUrl).get();
|
|
}
|
|
|
|
@Override
|
|
public void downloadURL(URL url, int index) {
|
|
addURLToDownload(url, getPrefix(index));
|
|
}
|
|
}
|