Fixed RedditRipper (Removed GoneWildRipper).
RedditRipper retries download on timeout (should be a global setting)
This commit is contained in:
parent
b662b46a44
commit
e49cab1254
@ -1,112 +0,0 @@
|
|||||||
package com.rarchives.ripme.ripper.rippers;
|
|
||||||
|
|
||||||
import java.io.IOException;
|
|
||||||
import java.net.MalformedURLException;
|
|
||||||
import java.net.URL;
|
|
||||||
import java.util.regex.Matcher;
|
|
||||||
import java.util.regex.Pattern;
|
|
||||||
|
|
||||||
import org.apache.log4j.Logger;
|
|
||||||
import org.json.JSONArray;
|
|
||||||
import org.json.JSONObject;
|
|
||||||
import org.jsoup.Jsoup;
|
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractRipper;
|
|
||||||
import com.rarchives.ripme.utils.Utils;
|
|
||||||
|
|
||||||
public class GonewildRipper extends AbstractRipper {
|
|
||||||
|
|
||||||
private static final String HOST = "gonewild";
|
|
||||||
private static final Logger logger = Logger.getLogger(GonewildRipper.class);
|
|
||||||
private static final int SLEEP_TIME = 1000;
|
|
||||||
|
|
||||||
private static String API_DOMAIN;
|
|
||||||
private String username;
|
|
||||||
|
|
||||||
public GonewildRipper(URL url) throws IOException {
|
|
||||||
super(url);
|
|
||||||
API_DOMAIN = Utils.getConfigString("gw.api", "gonewild");
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public boolean canRip(URL url) {
|
|
||||||
return getUsernameMatcher(url).matches();
|
|
||||||
}
|
|
||||||
|
|
||||||
private Matcher getUsernameMatcher(URL url) {
|
|
||||||
Pattern p = Pattern.compile("^https?://[a-z]{0,3}\\.?reddit\\.com/(u|user)/([a-zA-Z0-9\\-]{3,})/?.*$");
|
|
||||||
return p.matcher(url.toExternalForm());
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public URL sanitizeURL(URL url) throws MalformedURLException {
|
|
||||||
return url;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void rip() throws IOException {
|
|
||||||
int start = 0,
|
|
||||||
count = 50;
|
|
||||||
String baseGwURL = "http://" + API_DOMAIN + ".rarchives.com/api.cgi"
|
|
||||||
+ "?method=get_user"
|
|
||||||
+ "&user=" + username
|
|
||||||
+ "&count=" + count;
|
|
||||||
String gwURL, jsonString, imagePath;
|
|
||||||
JSONArray posts, images;
|
|
||||||
JSONObject json, post, image;
|
|
||||||
while (true) {
|
|
||||||
logger.info(" Retrieving posts by " + username);
|
|
||||||
gwURL = baseGwURL
|
|
||||||
+ "&start=" + start;
|
|
||||||
start += count;
|
|
||||||
jsonString = Jsoup.connect(gwURL)
|
|
||||||
.ignoreContentType(true)
|
|
||||||
.execute()
|
|
||||||
.body();
|
|
||||||
json = new JSONObject(jsonString);
|
|
||||||
if (json.has("error")) {
|
|
||||||
logger.error("Error while retrieving user posts:" + json.getString("error"));
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
posts = json.getJSONArray("posts");
|
|
||||||
if (posts.length() == 0) {
|
|
||||||
break; // No more posts to get
|
|
||||||
}
|
|
||||||
for (int i = 0; i < posts.length(); i++) {
|
|
||||||
post = (JSONObject) posts.get(i);
|
|
||||||
images = post.getJSONArray("images");
|
|
||||||
for (int j = 0; j < images.length(); j++) {
|
|
||||||
image = (JSONObject) images.get(j);
|
|
||||||
imagePath = image.getString("path");
|
|
||||||
if (imagePath.startsWith("..")) {
|
|
||||||
imagePath = imagePath.substring(2);
|
|
||||||
}
|
|
||||||
imagePath = "http://" + API_DOMAIN + ".rarchives.com" + imagePath;
|
|
||||||
logger.info(" Found file: " + imagePath);
|
|
||||||
addURLToDownload(new URL(imagePath));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
try {
|
|
||||||
Thread.sleep(SLEEP_TIME);
|
|
||||||
} catch (InterruptedException e) {
|
|
||||||
logger.error("[!] Interrupted while waiting to load more posts", e);
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
waitForThreads();
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public String getHost() {
|
|
||||||
return HOST;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public String getGID(URL url) throws MalformedURLException {
|
|
||||||
Matcher m = getUsernameMatcher(url);
|
|
||||||
if (m.matches()) {
|
|
||||||
this.username = m.group(m.groupCount());
|
|
||||||
}
|
|
||||||
return username;
|
|
||||||
}
|
|
||||||
}
|
|
@ -16,6 +16,8 @@ import org.jsoup.nodes.Document;
|
|||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractRipper;
|
import com.rarchives.ripme.ripper.AbstractRipper;
|
||||||
import com.rarchives.ripme.utils.RipUtils;
|
import com.rarchives.ripme.utils.RipUtils;
|
||||||
|
import com.rarchives.ripme.utils.Utils;
|
||||||
|
import java.net.SocketTimeoutException;
|
||||||
|
|
||||||
public class RedditRipper extends AbstractRipper {
|
public class RedditRipper extends AbstractRipper {
|
||||||
|
|
||||||
@ -26,7 +28,7 @@ public class RedditRipper extends AbstractRipper {
|
|||||||
private static final String HOST = "reddit";
|
private static final String HOST = "reddit";
|
||||||
private static final String DOMAIN = "reddit.com";
|
private static final String DOMAIN = "reddit.com";
|
||||||
|
|
||||||
private static final Logger logger = Logger.getLogger(GonewildRipper.class);
|
private static final Logger logger = Logger.getLogger(RedditRipper.class);
|
||||||
private static final int SLEEP_TIME = 2000;
|
private static final int SLEEP_TIME = 2000;
|
||||||
|
|
||||||
//private static final String USER_AGENT = "ripme by /u/4_pr0n github.com/4pr0n/ripme";
|
//private static final String USER_AGENT = "ripme by /u/4_pr0n github.com/4pr0n/ripme";
|
||||||
@ -67,6 +69,8 @@ public class RedditRipper extends AbstractRipper {
|
|||||||
waitForThreads();
|
waitForThreads();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
private URL getAndParseAndReturnNext(URL url) throws IOException {
|
private URL getAndParseAndReturnNext(URL url) throws IOException {
|
||||||
JSONArray jsonArray = getJsonArrayFromURL(url), children;
|
JSONArray jsonArray = getJsonArrayFromURL(url), children;
|
||||||
JSONObject json, data;
|
JSONObject json, data;
|
||||||
@ -85,7 +89,7 @@ public class RedditRipper extends AbstractRipper {
|
|||||||
parseJsonChild(children.getJSONObject(j));
|
parseJsonChild(children.getJSONObject(j));
|
||||||
}
|
}
|
||||||
if (data.has("after") && !data.isNull("after")) {
|
if (data.has("after") && !data.isNull("after")) {
|
||||||
String nextURLString = url.toExternalForm();
|
String nextURLString = Utils.stripURLParameter(url.toExternalForm(), "after");
|
||||||
if (nextURLString.contains("?")) {
|
if (nextURLString.contains("?")) {
|
||||||
nextURLString = nextURLString.concat("&after=" + data.getString("after"));
|
nextURLString = nextURLString.concat("&after=" + data.getString("after"));
|
||||||
}
|
}
|
||||||
@ -111,11 +115,21 @@ public class RedditRipper extends AbstractRipper {
|
|||||||
}
|
}
|
||||||
lastRequestTime = System.currentTimeMillis();
|
lastRequestTime = System.currentTimeMillis();
|
||||||
|
|
||||||
|
int attempts = 0;
|
||||||
|
Document doc = null;
|
||||||
logger.info(" Retrieving " + url);
|
logger.info(" Retrieving " + url);
|
||||||
Document doc= Jsoup.connect(url.toExternalForm())
|
while(doc == null && attempts++ < 3) {
|
||||||
.ignoreContentType(true)
|
try {
|
||||||
.userAgent(USER_AGENT)
|
doc= Jsoup.connect(url.toExternalForm())
|
||||||
.get();
|
.ignoreContentType(true)
|
||||||
|
.userAgent(USER_AGENT)
|
||||||
|
.get();
|
||||||
|
} catch(SocketTimeoutException ex) {
|
||||||
|
if(attempts >= 3) throw ex;
|
||||||
|
logger.warn(String.format("[!] Connection timed out (attempt %d)", attempts));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
String jsonString = doc.body().html().replaceAll(""", "\"");
|
String jsonString = doc.body().html().replaceAll(""", "\"");
|
||||||
|
|
||||||
Object jsonObj = new JSONTokener(jsonString).nextValue();
|
Object jsonObj = new JSONTokener(jsonString).nextValue();
|
||||||
|
@ -84,6 +84,28 @@ public class Utils {
|
|||||||
return prettySaveAs;
|
return prettySaveAs;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public static String stripURLParameter(String url, String parameter) {
|
||||||
|
int paramIndex = url.indexOf("?" + parameter);
|
||||||
|
boolean wasFirstParam = true;
|
||||||
|
if(paramIndex < 0) {
|
||||||
|
wasFirstParam = false;
|
||||||
|
paramIndex = url.indexOf("&" + parameter);
|
||||||
|
}
|
||||||
|
|
||||||
|
if(paramIndex > 0) {
|
||||||
|
int nextParam = url.indexOf("&", paramIndex+1);
|
||||||
|
if(nextParam != -1) {
|
||||||
|
String c = "&";
|
||||||
|
if(wasFirstParam) c = "?";
|
||||||
|
url = url.substring(0, paramIndex) + c + url.substring(nextParam+1, url.length());
|
||||||
|
} else {
|
||||||
|
url = url.substring(0, paramIndex);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return url;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Removes the current working directory from a given filename
|
* Removes the current working directory from a given filename
|
||||||
* @param file
|
* @param file
|
||||||
|
@ -1,31 +0,0 @@
|
|||||||
package com.rarchives.ripme.tst.ripper.rippers;
|
|
||||||
|
|
||||||
import java.io.IOException;
|
|
||||||
import java.net.URL;
|
|
||||||
import java.util.ArrayList;
|
|
||||||
import java.util.List;
|
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.rippers.GonewildRipper;
|
|
||||||
|
|
||||||
public class GonewildRipperTest extends RippersTest {
|
|
||||||
|
|
||||||
public void testInstagramAlbums() throws IOException {
|
|
||||||
if (!DOWNLOAD_CONTENT) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
List<URL> contentURLs = new ArrayList<URL>();
|
|
||||||
contentURLs.add(new URL("http://reddit.com/u/amle69"));
|
|
||||||
for (URL url : contentURLs) {
|
|
||||||
try {
|
|
||||||
GonewildRipper ripper = new GonewildRipper(url);
|
|
||||||
ripper.rip();
|
|
||||||
assert(ripper.getWorkingDir().listFiles().length > 1);
|
|
||||||
deleteDir(ripper.getWorkingDir());
|
|
||||||
} catch (Exception e) {
|
|
||||||
e.printStackTrace();
|
|
||||||
fail("Error while ripping URL " + url + ": " + e.getMessage());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
Loading…
Reference in New Issue
Block a user