Merge branch 'formatting': Various code cleanup and formatting.
This commit is contained in:
commit
dc2c100d9d
@ -1,4 +1,3 @@
|
|||||||
|
|
||||||
package com.rarchives.ripme.ripper.rippers;
|
package com.rarchives.ripme.ripper.rippers;
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
||||||
@ -8,24 +7,16 @@ import com.rarchives.ripme.utils.Utils;
|
|||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.io.UnsupportedEncodingException;
|
import java.io.UnsupportedEncodingException;
|
||||||
import java.net.MalformedURLException;
|
import java.net.MalformedURLException;
|
||||||
import java.net.URI;
|
|
||||||
import java.net.URISyntaxException;
|
|
||||||
import java.net.URL;
|
import java.net.URL;
|
||||||
import java.net.URLDecoder;
|
import java.net.URLDecoder;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.logging.Level;
|
import java.util.logging.Level;
|
||||||
import java.util.logging.Logger;
|
import java.util.logging.Logger;
|
||||||
import java.util.regex.Matcher;
|
|
||||||
import java.util.regex.Pattern;
|
|
||||||
import org.jsoup.nodes.Document;
|
import org.jsoup.nodes.Document;
|
||||||
import org.jsoup.nodes.Element;
|
import org.jsoup.nodes.Element;
|
||||||
import org.jsoup.select.Elements;
|
import org.jsoup.select.Elements;
|
||||||
|
|
||||||
/**
|
|
||||||
*
|
|
||||||
* @author
|
|
||||||
*/
|
|
||||||
public class E621Ripper extends AbstractHTMLRipper {
|
public class E621Ripper extends AbstractHTMLRipper {
|
||||||
public static final int POOL_IMAGES_PER_PAGE = 24;
|
public static final int POOL_IMAGES_PER_PAGE = 24;
|
||||||
|
|
||||||
@ -177,5 +168,4 @@ public class E621Ripper extends AbstractHTMLRipper {
|
|||||||
|
|
||||||
return Utils.filesystemSafe(prefix + getTerm(url));
|
return Utils.filesystemSafe(prefix + getTerm(url));
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
@ -9,7 +9,6 @@ import java.io.IOException;
|
|||||||
import java.net.MalformedURLException;
|
import java.net.MalformedURLException;
|
||||||
import java.net.URL;
|
import java.net.URL;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Arrays;
|
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.regex.Matcher;
|
import java.util.regex.Matcher;
|
||||||
import java.util.regex.Pattern;
|
import java.util.regex.Pattern;
|
||||||
@ -18,10 +17,8 @@ import org.jsoup.Connection.Response;
|
|||||||
import org.jsoup.nodes.Document;
|
import org.jsoup.nodes.Document;
|
||||||
import org.jsoup.nodes.Element;
|
import org.jsoup.nodes.Element;
|
||||||
import org.jsoup.select.Elements;
|
import org.jsoup.select.Elements;
|
||||||
import org.jsoup.Connection.Method;
|
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
||||||
import com.rarchives.ripme.ui.RipStatusMessage.STATUS;
|
|
||||||
import com.rarchives.ripme.utils.Http;
|
import com.rarchives.ripme.utils.Http;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -11,11 +11,8 @@ import java.util.regex.Pattern;
|
|||||||
import org.json.JSONArray;
|
import org.json.JSONArray;
|
||||||
import org.json.JSONException;
|
import org.json.JSONException;
|
||||||
import org.json.JSONObject;
|
import org.json.JSONObject;
|
||||||
import org.jsoup.nodes.Document;
|
|
||||||
import org.jsoup.nodes.Element;
|
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractJSONRipper;
|
import com.rarchives.ripme.ripper.AbstractJSONRipper;
|
||||||
import com.rarchives.ripme.ui.RipStatusMessage.STATUS;
|
|
||||||
import com.rarchives.ripme.utils.Http;
|
import com.rarchives.ripme.utils.Http;
|
||||||
|
|
||||||
public class InstagramRipper extends AbstractJSONRipper {
|
public class InstagramRipper extends AbstractJSONRipper {
|
||||||
|
@ -2,24 +2,15 @@ package com.rarchives.ripme.ripper.rippers;
|
|||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
||||||
import com.rarchives.ripme.utils.Http;
|
import com.rarchives.ripme.utils.Http;
|
||||||
import com.rarchives.ripme.utils.Utils;
|
|
||||||
import java.io.File;
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.net.MalformedURLException;
|
import java.net.MalformedURLException;
|
||||||
import java.net.URI;
|
|
||||||
import java.net.URISyntaxException;
|
|
||||||
import java.net.URL;
|
import java.net.URL;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.HashMap;
|
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
|
||||||
import java.util.logging.Level;
|
|
||||||
import java.util.logging.Logger;
|
|
||||||
import java.util.regex.Matcher;
|
import java.util.regex.Matcher;
|
||||||
import java.util.regex.Pattern;
|
import java.util.regex.Pattern;
|
||||||
import org.jsoup.nodes.Document;
|
import org.jsoup.nodes.Document;
|
||||||
import org.jsoup.nodes.Element;
|
import org.jsoup.nodes.Element;
|
||||||
import org.jsoup.select.Elements;
|
|
||||||
|
|
||||||
public class MyhentaicomicsRipper extends AbstractHTMLRipper {
|
public class MyhentaicomicsRipper extends AbstractHTMLRipper {
|
||||||
public static boolean isTag;
|
public static boolean isTag;
|
||||||
|
@ -4,7 +4,6 @@ import java.io.IOException;
|
|||||||
import java.net.MalformedURLException;
|
import java.net.MalformedURLException;
|
||||||
import java.net.URL;
|
import java.net.URL;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Arrays;
|
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.regex.Matcher;
|
import java.util.regex.Matcher;
|
||||||
import java.util.regex.Pattern;
|
import java.util.regex.Pattern;
|
||||||
@ -13,7 +12,6 @@ import org.jsoup.nodes.Document;
|
|||||||
import org.jsoup.nodes.Element;
|
import org.jsoup.nodes.Element;
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
||||||
import com.rarchives.ripme.ripper.rippers.ripperhelpers.ChanSite;
|
|
||||||
import com.rarchives.ripme.utils.Http;
|
import com.rarchives.ripme.utils.Http;
|
||||||
|
|
||||||
public class NatalieMuRipper extends AbstractHTMLRipper {
|
public class NatalieMuRipper extends AbstractHTMLRipper {
|
||||||
|
@ -21,10 +21,6 @@ import org.jsoup.nodes.Document;
|
|||||||
import org.jsoup.nodes.Element;
|
import org.jsoup.nodes.Element;
|
||||||
import org.jsoup.select.Elements;
|
import org.jsoup.select.Elements;
|
||||||
|
|
||||||
/**
|
|
||||||
*
|
|
||||||
* @author
|
|
||||||
*/
|
|
||||||
public class PahealRipper extends AbstractHTMLRipper {
|
public class PahealRipper extends AbstractHTMLRipper {
|
||||||
private static Map<String, String> cookies = null;
|
private static Map<String, String> cookies = null;
|
||||||
private static Pattern gidPattern = null;
|
private static Pattern gidPattern = null;
|
||||||
@ -59,9 +55,10 @@ public class PahealRipper extends AbstractHTMLRipper{
|
|||||||
@Override
|
@Override
|
||||||
public Document getNextPage(Document page) throws IOException {
|
public Document getNextPage(Document page) throws IOException {
|
||||||
for (Element e : page.select("#paginator a")) {
|
for (Element e : page.select("#paginator a")) {
|
||||||
if(e.text().toLowerCase().equals("next"))
|
if (e.text().toLowerCase().equals("next")) {
|
||||||
return Http.url(e.absUrl("href")).cookies(getCookies()).get();
|
return Http.url(e.absUrl("href")).cookies(getCookies()).get();
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
@ -71,8 +68,9 @@ public class PahealRipper extends AbstractHTMLRipper{
|
|||||||
Elements elements = page.select(".shm-thumb.thumb>a").not(".shm-thumb-link");
|
Elements elements = page.select(".shm-thumb.thumb>a").not(".shm-thumb-link");
|
||||||
List<String> res = new ArrayList<String>(elements.size());
|
List<String> res = new ArrayList<String>(elements.size());
|
||||||
|
|
||||||
for(Element e:elements)
|
for (Element e : elements) {
|
||||||
res.add(e.absUrl("href"));
|
res.add(e.absUrl("href"));
|
||||||
|
}
|
||||||
|
|
||||||
return res;
|
return res;
|
||||||
}
|
}
|
||||||
@ -89,7 +87,11 @@ public class PahealRipper extends AbstractHTMLRipper{
|
|||||||
name = name.substring(0, name.length() - ext.length());
|
name = name.substring(0, name.length() - ext.length());
|
||||||
}
|
}
|
||||||
|
|
||||||
addURLToDownload(url,new File(workingDir.getCanonicalPath()+File.separator+Utils.filesystemSafe(new URI(name).getPath())+ext));
|
File outFile = new File(workingDir.getCanonicalPath()
|
||||||
|
+ File.separator
|
||||||
|
+ Utils.filesystemSafe(new URI(name).getPath())
|
||||||
|
+ ext);
|
||||||
|
addURLToDownload(url, outFile);
|
||||||
} catch (IOException ex) {
|
} catch (IOException ex) {
|
||||||
Logger.getLogger(PahealRipper.class.getName()).log(Level.SEVERE, null, ex);
|
Logger.getLogger(PahealRipper.class.getName()).log(Level.SEVERE, null, ex);
|
||||||
} catch (URISyntaxException ex) {
|
} catch (URISyntaxException ex) {
|
||||||
@ -98,12 +100,14 @@ public class PahealRipper extends AbstractHTMLRipper{
|
|||||||
}
|
}
|
||||||
|
|
||||||
private String getTerm(URL url) throws MalformedURLException {
|
private String getTerm(URL url) throws MalformedURLException {
|
||||||
if(gidPattern==null)
|
if (gidPattern == null) {
|
||||||
gidPattern = Pattern.compile("^https?://(www\\.)?rule34\\.paheal\\.net/post/list/([a-zA-Z0-9$_.+!*'(),%-]+)(/.*)?(#.*)?$");
|
gidPattern = Pattern.compile("^https?://(www\\.)?rule34\\.paheal\\.net/post/list/([a-zA-Z0-9$_.+!*'(),%-]+)(/.*)?(#.*)?$");
|
||||||
|
}
|
||||||
|
|
||||||
Matcher m = gidPattern.matcher(url.toExternalForm());
|
Matcher m = gidPattern.matcher(url.toExternalForm());
|
||||||
if(m.matches())
|
if (m.matches()) {
|
||||||
return m.group(2);
|
return m.group(2);
|
||||||
|
}
|
||||||
|
|
||||||
throw new MalformedURLException("Expected paheal.net URL format: rule34.paheal.net/post/list/searchterm - got " + url + " instead");
|
throw new MalformedURLException("Expected paheal.net URL format: rule34.paheal.net/post/list/searchterm - got " + url + " instead");
|
||||||
}
|
}
|
||||||
@ -118,5 +122,4 @@ public class PahealRipper extends AbstractHTMLRipper{
|
|||||||
|
|
||||||
throw new MalformedURLException("Expected paheal.net URL format: rule34.paheal.net/post/list/searchterm - got " + url + " instead");
|
throw new MalformedURLException("Expected paheal.net URL format: rule34.paheal.net/post/list/searchterm - got " + url + " instead");
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -11,19 +11,18 @@ import java.util.regex.Pattern;
|
|||||||
|
|
||||||
import org.jsoup.nodes.Document;
|
import org.jsoup.nodes.Document;
|
||||||
import org.jsoup.nodes.Element;
|
import org.jsoup.nodes.Element;
|
||||||
import org.jsoup.select.Elements;
|
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
||||||
import com.rarchives.ripme.utils.Http;
|
import com.rarchives.ripme.utils.Http;
|
||||||
|
|
||||||
public class WordpressComicRipper extends AbstractHTMLRipper {
|
public class WordpressComicRipper extends AbstractHTMLRipper {
|
||||||
|
|
||||||
public WordpressComicRipper(URL url) throws IOException {
|
public WordpressComicRipper(URL url) throws IOException {
|
||||||
super(url);
|
super(url);
|
||||||
}
|
}
|
||||||
|
|
||||||
public static List<String> explicit_domains = Arrays.asList("www.totempole666.com",
|
public static List<String> explicit_domains = Arrays.asList("www.totempole666.com",
|
||||||
"buttsmithy.com", "themonsterunderthebed.net", "prismblush.com");
|
"buttsmithy.com", "themonsterunderthebed.net", "prismblush.com");
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public String getHost() {
|
public String getHost() {
|
||||||
String host = url.toExternalForm().split("/")[2];
|
String host = url.toExternalForm().split("/")[2];
|
||||||
@ -63,8 +62,8 @@ public class WordpressComicRipper extends AbstractHTMLRipper {
|
|||||||
if (prismblushMat.matches()) {
|
if (prismblushMat.matches()) {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -97,7 +96,6 @@ public class WordpressComicRipper extends AbstractHTMLRipper {
|
|||||||
return super.getAlbumTitle(url);
|
return super.getAlbumTitle(url);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public String getGID(URL url) throws MalformedURLException {
|
public String getGID(URL url) throws MalformedURLException {
|
||||||
String url_name = url.toExternalForm();
|
String url_name = url.toExternalForm();
|
||||||
@ -129,10 +127,10 @@ public class WordpressComicRipper extends AbstractHTMLRipper {
|
|||||||
}
|
}
|
||||||
nextPage = elem.attr("href");
|
nextPage = elem.attr("href");
|
||||||
}
|
}
|
||||||
|
|
||||||
if (nextPage == "") {
|
if (nextPage == "") {
|
||||||
throw new IOException("No more pages");
|
throw new IOException("No more pages");
|
||||||
}
|
} else {
|
||||||
else {
|
|
||||||
return Http.url(nextPage).get();
|
return Http.url(nextPage).get();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -160,6 +158,4 @@ public class WordpressComicRipper extends AbstractHTMLRipper {
|
|||||||
public void downloadURL(URL url, int index) {
|
public void downloadURL(URL url, int index) {
|
||||||
addURLToDownload(url, getPrefix(index));
|
addURLToDownload(url, getPrefix(index));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -1,4 +1,3 @@
|
|||||||
|
|
||||||
package com.rarchives.ripme.ripper.rippers;
|
package com.rarchives.ripme.ripper.rippers;
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
||||||
@ -18,10 +17,6 @@ import java.util.regex.Pattern;
|
|||||||
import org.jsoup.nodes.Document;
|
import org.jsoup.nodes.Document;
|
||||||
import org.jsoup.nodes.Element;
|
import org.jsoup.nodes.Element;
|
||||||
|
|
||||||
/**
|
|
||||||
*
|
|
||||||
* @author
|
|
||||||
*/
|
|
||||||
public class XbooruRipper extends AbstractHTMLRipper {
|
public class XbooruRipper extends AbstractHTMLRipper {
|
||||||
private static Pattern gidPattern = null;
|
private static Pattern gidPattern = null;
|
||||||
|
|
||||||
@ -53,8 +48,9 @@ public class XbooruRipper extends AbstractHTMLRipper{
|
|||||||
int offset = Integer.parseInt(doc.getElementsByTag("posts").first().attr("offset"));
|
int offset = Integer.parseInt(doc.getElementsByTag("posts").first().attr("offset"));
|
||||||
int num = Integer.parseInt(doc.getElementsByTag("posts").first().attr("count"));
|
int num = Integer.parseInt(doc.getElementsByTag("posts").first().attr("count"));
|
||||||
|
|
||||||
if(offset+100>num)
|
if (offset + 100 > num) {
|
||||||
return null;
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
return Http.url(getPage(offset / 100 + 1)).get();
|
return Http.url(getPage(offset / 100 + 1)).get();
|
||||||
}
|
}
|
||||||
@ -62,8 +58,9 @@ public class XbooruRipper extends AbstractHTMLRipper{
|
|||||||
@Override
|
@Override
|
||||||
public List<String> getURLsFromPage(Document page) {
|
public List<String> getURLsFromPage(Document page) {
|
||||||
List<String> res = new ArrayList<String>(100);
|
List<String> res = new ArrayList<String>(100);
|
||||||
for(Element e:page.getElementsByTag("post"))
|
for (Element e : page.getElementsByTag("post")) {
|
||||||
res.add(e.absUrl("file_url") + "#" + e.attr("id"));
|
res.add(e.absUrl("file_url") + "#" + e.attr("id"));
|
||||||
|
}
|
||||||
return res;
|
return res;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -73,12 +70,14 @@ public class XbooruRipper extends AbstractHTMLRipper{
|
|||||||
}
|
}
|
||||||
|
|
||||||
private String getTerm(URL url) throws MalformedURLException {
|
private String getTerm(URL url) throws MalformedURLException {
|
||||||
if(gidPattern==null)
|
if (gidPattern == null) {
|
||||||
gidPattern = Pattern.compile("^https?://(www\\.)?xbooru\\.com/(index.php)?.*([?&]tags=([a-zA-Z0-9$_.+!*'(),%-]+))(\\&|(#.*)?$)");
|
gidPattern = Pattern.compile("^https?://(www\\.)?xbooru\\.com/(index.php)?.*([?&]tags=([a-zA-Z0-9$_.+!*'(),%-]+))(\\&|(#.*)?$)");
|
||||||
|
}
|
||||||
|
|
||||||
Matcher m = gidPattern.matcher(url.toExternalForm());
|
Matcher m = gidPattern.matcher(url.toExternalForm());
|
||||||
if(m.matches())
|
if (m.matches()) {
|
||||||
return m.group(4);
|
return m.group(4);
|
||||||
|
}
|
||||||
|
|
||||||
throw new MalformedURLException("Expected xbooru.com URL format: xbooru.com/index.php?tags=searchterm - got " + url + " instead");
|
throw new MalformedURLException("Expected xbooru.com URL format: xbooru.com/index.php?tags=searchterm - got " + url + " instead");
|
||||||
}
|
}
|
||||||
@ -93,5 +92,4 @@ public class XbooruRipper extends AbstractHTMLRipper{
|
|||||||
|
|
||||||
throw new MalformedURLException("Expected xbooru.com URL format: xbooru.com/index.php?tags=searchterm - got " + url + " instead");
|
throw new MalformedURLException("Expected xbooru.com URL format: xbooru.com/index.php?tags=searchterm - got " + url + " instead");
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@ -4,7 +4,6 @@ import java.io.IOException;
|
|||||||
import java.net.MalformedURLException;
|
import java.net.MalformedURLException;
|
||||||
import java.net.URL;
|
import java.net.URL;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Collections;
|
|
||||||
import java.util.HashMap;
|
import java.util.HashMap;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
@ -14,10 +13,8 @@ import java.util.regex.Pattern;
|
|||||||
import org.jsoup.Connection.Response;
|
import org.jsoup.Connection.Response;
|
||||||
import org.jsoup.nodes.Document;
|
import org.jsoup.nodes.Document;
|
||||||
import org.jsoup.nodes.Element;
|
import org.jsoup.nodes.Element;
|
||||||
import org.jsoup.select.Elements;
|
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
||||||
import com.rarchives.ripme.ui.RipStatusMessage.STATUS;
|
|
||||||
import com.rarchives.ripme.utils.Http;
|
import com.rarchives.ripme.utils.Http;
|
||||||
|
|
||||||
public class ZizkiRipper extends AbstractHTMLRipper {
|
public class ZizkiRipper extends AbstractHTMLRipper {
|
||||||
|
@ -9,16 +9,20 @@ public class ChanSite {
|
|||||||
public List<String> cdnDomains;
|
public List<String> cdnDomains;
|
||||||
|
|
||||||
public ChanSite(List<String> Domains, List<String> CdnDomains) {
|
public ChanSite(List<String> Domains, List<String> CdnDomains) {
|
||||||
if(Domains.isEmpty())
|
if (Domains.isEmpty()) {
|
||||||
throw new IllegalArgumentException("Domains");
|
throw new IllegalArgumentException("Domains");
|
||||||
if(CdnDomains.isEmpty())
|
}
|
||||||
|
if (CdnDomains.isEmpty()) {
|
||||||
throw new IllegalArgumentException("CdnDomains");
|
throw new IllegalArgumentException("CdnDomains");
|
||||||
|
}
|
||||||
domains = Domains;
|
domains = Domains;
|
||||||
cdnDomains = CdnDomains;
|
cdnDomains = CdnDomains;
|
||||||
}
|
}
|
||||||
|
|
||||||
public ChanSite(List<String> Domains) {
|
public ChanSite(List<String> Domains) {
|
||||||
if(Domains.isEmpty())
|
if (Domains.isEmpty()) {
|
||||||
throw new IllegalArgumentException("Domains");
|
throw new IllegalArgumentException("Domains");
|
||||||
|
}
|
||||||
domains = Domains;
|
domains = Domains;
|
||||||
cdnDomains = Domains;
|
cdnDomains = Domains;
|
||||||
}
|
}
|
||||||
|
@ -3,7 +3,6 @@ package com.rarchives.ripme.utils;
|
|||||||
/**
|
/**
|
||||||
* Base64 encoder/decoder
|
* Base64 encoder/decoder
|
||||||
* From http://stackoverflow.com/a/4265472
|
* From http://stackoverflow.com/a/4265472
|
||||||
*
|
|
||||||
*/
|
*/
|
||||||
public class Base64 {
|
public class Base64 {
|
||||||
private final static char[] ALPHABET = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/".toCharArray();
|
private final static char[] ALPHABET = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/".toCharArray();
|
||||||
|
@ -23,7 +23,6 @@ import com.rarchives.ripme.ripper.rippers.VidbleRipper;
|
|||||||
import com.rarchives.ripme.ripper.rippers.video.GfycatRipper;
|
import com.rarchives.ripme.ripper.rippers.video.GfycatRipper;
|
||||||
import com.rarchives.ripme.ripper.rippers.EroShareRipper;
|
import com.rarchives.ripme.ripper.rippers.EroShareRipper;
|
||||||
|
|
||||||
|
|
||||||
public class RipUtils {
|
public class RipUtils {
|
||||||
private static final Logger logger = Logger.getLogger(RipUtils.class);
|
private static final Logger logger = Logger.getLogger(RipUtils.class);
|
||||||
|
|
||||||
@ -165,18 +164,6 @@ public class RipUtils {
|
|||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "vinebox", "http://finebox.co/u/", "");
|
if (url == null) url = urlFromSiteDirectoryName(dir, "vinebox", "http://finebox.co/u/", "");
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "imgbox", "http://imgbox.com/g/", "");
|
if (url == null) url = urlFromSiteDirectoryName(dir, "imgbox", "http://imgbox.com/g/", "");
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "modelmayhem", "http://www.modelmayhem.com/", "");
|
if (url == null) url = urlFromSiteDirectoryName(dir, "modelmayhem", "http://www.modelmayhem.com/", "");
|
||||||
/*
|
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "", "", "");
|
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "", "", "");
|
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "", "", "");
|
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "", "", "");
|
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "", "", "");
|
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "", "", "");
|
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "", "", "");
|
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "", "", "");
|
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "", "", "");
|
|
||||||
if (url == null) url = urlFromSiteDirectoryName(dir, "", "", "");
|
|
||||||
*/
|
|
||||||
//if (url == null) url = urlFromSiteDirectoryName(dir, "8muses", "http://www.8muses.com/index/category/", "");
|
//if (url == null) url = urlFromSiteDirectoryName(dir, "8muses", "http://www.8muses.com/index/category/", "");
|
||||||
return url;
|
return url;
|
||||||
}
|
}
|
||||||
@ -249,8 +236,7 @@ public class RipUtils {
|
|||||||
String album = fields.get(1);
|
String album = fields.get(1);
|
||||||
String url = "http://";
|
String url = "http://";
|
||||||
if ((fields.contains("top") || fields.contains("new"))
|
if ((fields.contains("top") || fields.contains("new"))
|
||||||
&& (fields.contains("year") || fields.contains("month") || fields.contains("week") || fields.contains("all"))
|
&& (fields.contains("year") || fields.contains("month") || fields.contains("week") || fields.contains("all"))) {
|
||||||
) {
|
|
||||||
// Subreddit
|
// Subreddit
|
||||||
fields.remove(0); // "imgur"
|
fields.remove(0); // "imgur"
|
||||||
String sub = "";
|
String sub = "";
|
||||||
|
@ -34,7 +34,6 @@ import com.rarchives.ripme.ripper.AbstractRipper;
|
|||||||
* Common utility functions used in various places throughout the project.
|
* Common utility functions used in various places throughout the project.
|
||||||
*/
|
*/
|
||||||
public class Utils {
|
public class Utils {
|
||||||
|
|
||||||
public static final String RIP_DIRECTORY = "rips";
|
public static final String RIP_DIRECTORY = "rips";
|
||||||
private static final String configFile = "rip.properties";
|
private static final String configFile = "rip.properties";
|
||||||
private static final Logger logger = Logger.getLogger(Utils.class);
|
private static final Logger logger = Logger.getLogger(Utils.class);
|
||||||
@ -180,7 +179,9 @@ public class Utils {
|
|||||||
int nextParam = url.indexOf("&", paramIndex+1);
|
int nextParam = url.indexOf("&", paramIndex+1);
|
||||||
if (nextParam != -1) {
|
if (nextParam != -1) {
|
||||||
String c = "&";
|
String c = "&";
|
||||||
if(wasFirstParam) c = "?";
|
if (wasFirstParam) {
|
||||||
|
c = "?";
|
||||||
|
}
|
||||||
url = url.substring(0, paramIndex) + c + url.substring(nextParam+1, url.length());
|
url = url.substring(0, paramIndex) + c + url.substring(nextParam+1, url.length());
|
||||||
} else {
|
} else {
|
||||||
url = url.substring(0, paramIndex);
|
url = url.substring(0, paramIndex);
|
||||||
|
Loading…
x
Reference in New Issue
Block a user