Merge pull request #530 from cyian-1756/code_cleanup
Some minor code clean up
This commit is contained in:
commit
5147562f90
@ -55,14 +55,8 @@ public class BlackbrickroadofozRipper extends AbstractHTMLRipper {
|
|||||||
throw new IOException("No more pages");
|
throw new IOException("No more pages");
|
||||||
}
|
}
|
||||||
String nextPage = elem.attr("href");
|
String nextPage = elem.attr("href");
|
||||||
// Some times this returns a empty string
|
return Http.url(nextPage).get();
|
||||||
// This for stops that
|
|
||||||
if (nextPage == "") {
|
|
||||||
throw new IOException("No more pages");
|
|
||||||
}
|
|
||||||
else {
|
|
||||||
return Http.url(nextPage).get();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -59,7 +59,7 @@ public class CfakeRipper extends AbstractHTMLRipper {
|
|||||||
String nextPage = elem.attr("href");
|
String nextPage = elem.attr("href");
|
||||||
// Some times this returns a empty string
|
// Some times this returns a empty string
|
||||||
// This for stops that
|
// This for stops that
|
||||||
if (nextPage == "") {
|
if (nextPage.equals("")) {
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
|
@ -4,7 +4,6 @@ import java.io.IOException;
|
|||||||
import java.net.MalformedURLException;
|
import java.net.MalformedURLException;
|
||||||
import java.net.URL;
|
import java.net.URL;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Collections;
|
|
||||||
import java.util.HashMap;
|
import java.util.HashMap;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
|
@ -86,7 +86,7 @@ public class EroShareRipper extends AbstractHTMLRipper {
|
|||||||
throw new IOException("No more pages");
|
throw new IOException("No more pages");
|
||||||
}
|
}
|
||||||
nextUrl = elem.attr("href");
|
nextUrl = elem.attr("href");
|
||||||
if (nextUrl == "") {
|
if (nextUrl.equals("")) {
|
||||||
throw new IOException("No more pages");
|
throw new IOException("No more pages");
|
||||||
}
|
}
|
||||||
return Http.url("eroshae.com" + nextUrl).get();
|
return Http.url("eroshae.com" + nextUrl).get();
|
||||||
|
@ -13,7 +13,6 @@ import java.util.regex.Matcher;
|
|||||||
import java.util.regex.Pattern;
|
import java.util.regex.Pattern;
|
||||||
|
|
||||||
import com.rarchives.ripme.utils.Utils;
|
import com.rarchives.ripme.utils.Utils;
|
||||||
import org.jsoup.Connection.Method;
|
|
||||||
import org.jsoup.Connection.Response;
|
import org.jsoup.Connection.Response;
|
||||||
import org.jsoup.Jsoup;
|
import org.jsoup.Jsoup;
|
||||||
import org.jsoup.nodes.Document;
|
import org.jsoup.nodes.Document;
|
||||||
@ -23,7 +22,6 @@ import org.jsoup.select.Elements;
|
|||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
||||||
import com.rarchives.ripme.ripper.DownloadThreadPool;
|
import com.rarchives.ripme.ripper.DownloadThreadPool;
|
||||||
import com.rarchives.ripme.utils.Base64;
|
|
||||||
import com.rarchives.ripme.utils.Http;
|
import com.rarchives.ripme.utils.Http;
|
||||||
|
|
||||||
public class FuraffinityRipper extends AbstractHTMLRipper {
|
public class FuraffinityRipper extends AbstractHTMLRipper {
|
||||||
@ -162,10 +160,6 @@ public class FuraffinityRipper extends AbstractHTMLRipper {
|
|||||||
if (!subdirectory.equals("")) {
|
if (!subdirectory.equals("")) {
|
||||||
subdirectory = File.separator + subdirectory;
|
subdirectory = File.separator + subdirectory;
|
||||||
}
|
}
|
||||||
int o = url.toString().lastIndexOf('/')-1;
|
|
||||||
String test = url.toString().substring(url.toString().lastIndexOf('/',o)+1);
|
|
||||||
test = test.replace("/",""); // This is probably not the best way to do this.
|
|
||||||
test = test.replace("\\",""); // CLOSE ENOUGH!
|
|
||||||
saveFileAs = new File(
|
saveFileAs = new File(
|
||||||
workingDir.getCanonicalPath()
|
workingDir.getCanonicalPath()
|
||||||
+ subdirectory
|
+ subdirectory
|
||||||
|
@ -50,7 +50,7 @@ public class LusciousRipper extends AbstractHTMLRipper {
|
|||||||
// This is here for pages with mp4s instead of images
|
// This is here for pages with mp4s instead of images
|
||||||
String video_image = "";
|
String video_image = "";
|
||||||
video_image = page.select("div > video > source").attr("src");
|
video_image = page.select("div > video > source").attr("src");
|
||||||
if (video_image != "") {
|
if (!video_image.equals("")) {
|
||||||
urls.add(video_image);
|
urls.add(video_image);
|
||||||
}
|
}
|
||||||
return urls;
|
return urls;
|
||||||
|
@ -73,7 +73,7 @@ public class MyhentaicomicsRipper extends AbstractHTMLRipper {
|
|||||||
if (m.matches()) {
|
if (m.matches()) {
|
||||||
nextUrl = "http://myhentaicomics.com" + m.group(0);
|
nextUrl = "http://myhentaicomics.com" + m.group(0);
|
||||||
}
|
}
|
||||||
if (nextUrl == "") {
|
if (nextUrl.equals("")) {
|
||||||
throw new IOException("No more pages");
|
throw new IOException("No more pages");
|
||||||
}
|
}
|
||||||
// Sleep for half a sec to avoid getting IP banned
|
// Sleep for half a sec to avoid getting IP banned
|
||||||
@ -100,7 +100,7 @@ public class MyhentaicomicsRipper extends AbstractHTMLRipper {
|
|||||||
Element elem = nextAlbumPage.select("a.ui-icon-right").first();
|
Element elem = nextAlbumPage.select("a.ui-icon-right").first();
|
||||||
String nextPage = elem.attr("href");
|
String nextPage = elem.attr("href");
|
||||||
pageNumber = pageNumber + 1;
|
pageNumber = pageNumber + 1;
|
||||||
if (nextPage == "") {
|
if (nextPage.equals("")) {
|
||||||
logger.info("Got " + pageNumber + " pages");
|
logger.info("Got " + pageNumber + " pages");
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
@ -220,7 +220,7 @@ public class MyhentaicomicsRipper extends AbstractHTMLRipper {
|
|||||||
// If true the page is a page of albums
|
// If true the page is a page of albums
|
||||||
if (doc.toString().contains("class=\"g-item g-album\"")) {
|
if (doc.toString().contains("class=\"g-item g-album\"")) {
|
||||||
// This if checks that there is more than 1 page
|
// This if checks that there is more than 1 page
|
||||||
if (doc.select("a.ui-icon-right").last().attr("href") != "") {
|
if (!doc.select("a.ui-icon-right").last().attr("href").equals("")) {
|
||||||
// There is more than one page so we call getListOfPages
|
// There is more than one page so we call getListOfPages
|
||||||
List<String> pagesToRip = getListOfPages(doc);
|
List<String> pagesToRip = getListOfPages(doc);
|
||||||
logger.debug("Pages to rip = " + pagesToRip);
|
logger.debug("Pages to rip = " + pagesToRip);
|
||||||
|
@ -57,7 +57,7 @@ public class SinfestRipper extends AbstractHTMLRipper {
|
|||||||
String nextPage = elem.parent().attr("href");
|
String nextPage = elem.parent().attr("href");
|
||||||
// Some times this returns a empty string
|
// Some times this returns a empty string
|
||||||
// This for stops that
|
// This for stops that
|
||||||
if (nextPage == "") {
|
if (nextPage.equals("")) {
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
|
@ -4,14 +4,12 @@ import java.io.IOException;
|
|||||||
import java.net.MalformedURLException;
|
import java.net.MalformedURLException;
|
||||||
import java.net.URL;
|
import java.net.URL;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Arrays;
|
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.regex.Matcher;
|
import java.util.regex.Matcher;
|
||||||
import java.util.regex.Pattern;
|
import java.util.regex.Pattern;
|
||||||
|
|
||||||
import org.jsoup.nodes.Document;
|
import org.jsoup.nodes.Document;
|
||||||
import org.jsoup.nodes.Element;
|
import org.jsoup.nodes.Element;
|
||||||
import org.jsoup.select.Elements;
|
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
import com.rarchives.ripme.ripper.AbstractHTMLRipper;
|
||||||
import com.rarchives.ripme.utils.Http;
|
import com.rarchives.ripme.utils.Http;
|
||||||
|
@ -247,7 +247,7 @@ public class WordpressComicRipper extends AbstractHTMLRipper {
|
|||||||
nextPage = elem.attr("href");
|
nextPage = elem.attr("href");
|
||||||
}
|
}
|
||||||
|
|
||||||
if (nextPage == "") {
|
if (nextPage.equals("")) {
|
||||||
throw new IOException("No more pages");
|
throw new IOException("No more pages");
|
||||||
} else {
|
} else {
|
||||||
return Http.url(nextPage).get();
|
return Http.url(nextPage).get();
|
||||||
|
@ -72,7 +72,7 @@ public class XhamsterRipper extends AbstractHTMLRipper {
|
|||||||
|
|
||||||
@Override
|
@Override
|
||||||
public Document getNextPage(Document doc) throws IOException {
|
public Document getNextPage(Document doc) throws IOException {
|
||||||
if (doc.select("a.next").first().attr("href") != "") {
|
if (!doc.select("a.next").first().attr("href").equals("")) {
|
||||||
return Http.url(doc.select("a.next").first().attr("href")).get();
|
return Http.url(doc.select("a.next").first().attr("href")).get();
|
||||||
} else {
|
} else {
|
||||||
throw new IOException("No more pages");
|
throw new IOException("No more pages");
|
||||||
|
@ -3,7 +3,6 @@ package com.rarchives.ripme.ripper.rippers.video;
|
|||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.net.MalformedURLException;
|
import java.net.MalformedURLException;
|
||||||
import java.net.URL;
|
import java.net.URL;
|
||||||
import java.net.URLDecoder;
|
|
||||||
import java.util.regex.Matcher;
|
import java.util.regex.Matcher;
|
||||||
import java.util.regex.Pattern;
|
import java.util.regex.Pattern;
|
||||||
|
|
||||||
|
@ -3,7 +3,7 @@ package com.rarchives.ripme.tst.ripper.rippers;
|
|||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.net.URL;
|
import java.net.URL;
|
||||||
|
|
||||||
import com.rarchives.ripme.ripper.rippers.AerisdiesRipper;;
|
import com.rarchives.ripme.ripper.rippers.AerisdiesRipper;
|
||||||
|
|
||||||
public class AerisdiesRipperTest extends RippersTest {
|
public class AerisdiesRipperTest extends RippersTest {
|
||||||
public void testAlbum() throws IOException {
|
public void testAlbum() throws IOException {
|
||||||
|
Loading…
Reference in New Issue
Block a user