com.jaeksoft.searchlib.crawler.web.database.UrlCrawlQueue Maven / Gradle / Ivy
Show all versions of opensearchserver Show documentation
/**
* License Agreement for OpenSearchServer
*
* Copyright (C) 2008-2013 Emmanuel Keller / Jaeksoft
*
* http://www.open-search-server.com
*
* This file is part of OpenSearchServer.
*
* OpenSearchServer is free software: you can redistribute it and/or
* modify it under the terms of the GNU General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* OpenSearchServer is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with OpenSearchServer.
* If not, see .
**/
package com.jaeksoft.searchlib.crawler.web.database;
import com.jaeksoft.searchlib.SearchLibException;
import com.jaeksoft.searchlib.config.Config;
import com.jaeksoft.searchlib.crawler.common.process.CrawlQueueAbstract;
import com.jaeksoft.searchlib.crawler.common.process.CrawlStatistics;
import com.jaeksoft.searchlib.crawler.web.spider.Crawl;
import com.jaeksoft.searchlib.util.ReadWriteLock;
import org.apache.http.HttpException;
import java.io.IOException;
import java.net.URISyntaxException;
import java.security.NoSuchAlgorithmException;
import java.util.ArrayList;
import java.util.List;
public class UrlCrawlQueue extends CrawlQueueAbstract {
final private ReadWriteLock rwl = new ReadWriteLock();
private List updateCrawlList;
private List insertUrlList;
private List deleteUrlList;
private List workingUpdateCrawlList;
private List workingInsertUrlList;
private List workingDeleteUrlList;
public UrlCrawlQueue(Config config) {
super(config);
updateCrawlList = new ArrayList(0);
insertUrlList = new ArrayList(0);
deleteUrlList = new ArrayList(0);
workingUpdateCrawlList = null;
workingInsertUrlList = null;
workingDeleteUrlList = null;
}
public void add(CrawlStatistics currentStats, Crawl crawl)
throws NoSuchAlgorithmException, IOException, SearchLibException, URISyntaxException {
rwl.r.lock();
try {
updateCrawlList.add(crawl);
currentStats.incPendingUpdateCount();
final List discoverLinks = crawl.getDiscoverLinks();
final UrlManager urlManager = getConfig().getUrlManager();
if (discoverLinks != null) {
for (LinkItem link : discoverLinks)
insertUrlList.add(urlManager.getNewUrlItem(link));
currentStats.addPendingNewUrlCount(discoverLinks.size());
}
} finally {
rwl.r.unlock();
}
}
public void delete(CrawlStatistics currentStats, String url) {
rwl.r.lock();
try {
if (url == null)
return;
deleteUrlList.add(url);
currentStats.incPendingDeleteCount();
} finally {
rwl.r.unlock();
}
}
@Override
protected boolean shouldWePersist() {
rwl.r.lock();
try {
if (updateCrawlList.size() > getMaxBufferSize())
return true;
if (deleteUrlList.size() > getMaxBufferSize() * 10)
return true;
if (insertUrlList.size() > getMaxBufferSize() * 10)
return true;
return false;
} finally {
rwl.r.unlock();
}
}
@Override
protected boolean workingInProgress() {
rwl.r.lock();
try {
if (workingUpdateCrawlList != null)
return true;
if (workingInsertUrlList != null)
return true;
if (workingDeleteUrlList != null)
return true;
return false;
} finally {
rwl.r.unlock();
}
}
@Override
protected void initWorking() {
rwl.w.lock();
try {
workingUpdateCrawlList = updateCrawlList;
workingInsertUrlList = insertUrlList;
workingDeleteUrlList = deleteUrlList;
updateCrawlList = new ArrayList(0);
insertUrlList = new ArrayList(0);
deleteUrlList = new ArrayList(0);
if (getSessionStats() != null)
getSessionStats().resetPending();
} finally {
rwl.w.unlock();
}
}
@Override
protected void resetWork() {
rwl.w.lock();
try {
workingUpdateCrawlList = null;
workingInsertUrlList = null;
workingDeleteUrlList = null;
} finally {
rwl.w.unlock();
}
}
@Override
protected void indexWork()
throws SearchLibException, IOException, URISyntaxException, InstantiationException, IllegalAccessException,
ClassNotFoundException, HttpException {
CrawlStatistics sessionStats = getSessionStats();
deleteCollection(workingDeleteUrlList, sessionStats);
updateCrawls(workingUpdateCrawlList, sessionStats);
insertCollection(workingInsertUrlList, sessionStats);
}
private boolean deleteCollection(List workDeleteUrlList, CrawlStatistics sessionStats)
throws SearchLibException {
if (workDeleteUrlList.size() == 0)
return false;
UrlManager urlManager = getConfig().getUrlManager();
urlManager.deleteUrls(workDeleteUrlList);
if (sessionStats != null)
sessionStats.addDeletedCount(workDeleteUrlList.size());
return true;
}
private boolean updateCrawls(List workUpdateCrawlList, CrawlStatistics sessionStats)
throws IOException, SearchLibException {
if (workUpdateCrawlList.size() == 0)
return false;
UrlManager urlManager = getConfig().getUrlManager();
urlManager.updateCrawlTarget(workUpdateCrawlList,
getConfig().getWebPropertyManager().getPropagateDeletion().getValue());
urlManager.updateCrawlUrlDb(workUpdateCrawlList);
if (sessionStats != null)
sessionStats.addUpdatedCount(workUpdateCrawlList.size());
return true;
}
private boolean insertCollection(List workInsertUrlList, CrawlStatistics sessionStats)
throws SearchLibException {
if (workInsertUrlList.size() == 0)
return false;
UrlManager urlManager = getConfig().getUrlManager();
urlManager.updateUrlItems(workInsertUrlList);
if (sessionStats != null)
sessionStats.addNewUrlCount(workInsertUrlList.size());
return true;
}
}