More logging
Signed-off-by: akaessens <24660231+akaessens@users.noreply.github.com>
This commit is contained in:
parent
866889db27
commit
32da94c275
|
@ -1,5 +1,7 @@
|
||||||
package com.akdev.nofbeventscraper;
|
package com.akdev.nofbeventscraper;
|
||||||
|
|
||||||
|
import android.util.Log;
|
||||||
|
|
||||||
import org.jsoup.Connection;
|
import org.jsoup.Connection;
|
||||||
import org.jsoup.Jsoup;
|
import org.jsoup.Jsoup;
|
||||||
import org.jsoup.nodes.Element;
|
import org.jsoup.nodes.Element;
|
||||||
|
@ -17,12 +19,16 @@ public class DocumentReceiver {
|
||||||
// use default android user agent
|
// use default android user agent
|
||||||
String user_agent = "Mozilla/5.0 (X11; Linux x86_64)";
|
String user_agent = "Mozilla/5.0 (X11; Linux x86_64)";
|
||||||
|
|
||||||
|
Log.d("scraperLog", "DocumentReceiver: "+url);
|
||||||
|
|
||||||
Connection connection = Jsoup.connect(url).userAgent(user_agent).followRedirects(true);
|
Connection connection = Jsoup.connect(url).userAgent(user_agent).followRedirects(true);
|
||||||
|
|
||||||
Connection.Response response = connection.execute();
|
Connection.Response response = connection.execute();
|
||||||
|
|
||||||
document = response.parse();
|
document = response.parse();
|
||||||
|
|
||||||
|
Log.d("scraperLog", "Document title: "+document.title());
|
||||||
|
|
||||||
try {
|
try {
|
||||||
// accept cookies needed?
|
// accept cookies needed?
|
||||||
Element form = document.select("form[method=post]").first();
|
Element form = document.select("form[method=post]").first();
|
||||||
|
@ -46,7 +52,8 @@ public class DocumentReceiver {
|
||||||
} catch (Exception ignore) {
|
} catch (Exception ignore) {
|
||||||
}
|
}
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
return null;
|
e.printStackTrace();
|
||||||
|
return null;
|
||||||
}
|
}
|
||||||
return document;
|
return document;
|
||||||
}
|
}
|
||||||
|
|
|
@ -153,6 +153,8 @@ public class FbScraper {
|
||||||
*/
|
*/
|
||||||
void scrapeEvent(String event_url) {
|
void scrapeEvent(String event_url) {
|
||||||
FbEventScraper scraper = new FbEventScraper(this, event_url);
|
FbEventScraper scraper = new FbEventScraper(this, event_url);
|
||||||
|
|
||||||
|
Log.d("scraperLog", "scrapeEvent: "+event_url);
|
||||||
tasks.add(scraper);
|
tasks.add(scraper);
|
||||||
scraper.executeOnExecutor(AsyncTask.THREAD_POOL_EXECUTOR);
|
scraper.executeOnExecutor(AsyncTask.THREAD_POOL_EXECUTOR);
|
||||||
}
|
}
|
||||||
|
@ -182,6 +184,8 @@ public class FbScraper {
|
||||||
void scrapePage(String page_url) {
|
void scrapePage(String page_url) {
|
||||||
FbPageScraper scraper = new FbPageScraper(this, page_url);
|
FbPageScraper scraper = new FbPageScraper(this, page_url);
|
||||||
|
|
||||||
|
Log.d("scraperLog", "scrapePage: "+page_url);
|
||||||
|
|
||||||
tasks.add(scraper);
|
tasks.add(scraper);
|
||||||
scraper.executeOnExecutor(AsyncTask.THREAD_POOL_EXECUTOR);
|
scraper.executeOnExecutor(AsyncTask.THREAD_POOL_EXECUTOR);
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue