Add start/chunk/sleep logging to crawler
This commit is contained in:
parent
dceebc1c8d
commit
b1274125c9
|
@ -106,7 +106,10 @@ public class AccountDatabaseCrawler implements Managed, Runnable {
|
||||||
}
|
}
|
||||||
final long endTimeMs = System.currentTimeMillis();
|
final long endTimeMs = System.currentTimeMillis();
|
||||||
final long sleepIntervalMs = chunkIntervalMs - (endTimeMs - startTimeMs);
|
final long sleepIntervalMs = chunkIntervalMs - (endTimeMs - startTimeMs);
|
||||||
if (sleepIntervalMs > 0) sleepWhileRunning(sleepIntervalMs);
|
if (sleepIntervalMs > 0) {
|
||||||
|
logger.info("Sleeping {}ms", sleepIntervalMs);
|
||||||
|
sleepWhileRunning(sleepIntervalMs);
|
||||||
|
}
|
||||||
} finally {
|
} finally {
|
||||||
cache.releaseActiveWork(workerId);
|
cache.releaseActiveWork(workerId);
|
||||||
}
|
}
|
||||||
|
@ -122,6 +125,7 @@ public class AccountDatabaseCrawler implements Managed, Runnable {
|
||||||
final Optional<UUID> fromUuid = getLastUuid(useDynamo);
|
final Optional<UUID> fromUuid = getLastUuid(useDynamo);
|
||||||
|
|
||||||
if (fromUuid.isEmpty()) {
|
if (fromUuid.isEmpty()) {
|
||||||
|
logger.info("Started crawl");
|
||||||
listeners.forEach(AccountDatabaseCrawlerListener::onCrawlStart);
|
listeners.forEach(AccountDatabaseCrawlerListener::onCrawlStart);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -133,6 +137,7 @@ public class AccountDatabaseCrawler implements Managed, Runnable {
|
||||||
cacheLastUuid(Optional.empty(), useDynamo);
|
cacheLastUuid(Optional.empty(), useDynamo);
|
||||||
cache.setAccelerated(false);
|
cache.setAccelerated(false);
|
||||||
} else {
|
} else {
|
||||||
|
logger.info("Processing chunk");
|
||||||
try {
|
try {
|
||||||
for (AccountDatabaseCrawlerListener listener : listeners) {
|
for (AccountDatabaseCrawlerListener listener : listeners) {
|
||||||
listener.timeAndProcessCrawlChunk(fromUuid, chunkAccounts.getAccounts());
|
listener.timeAndProcessCrawlChunk(fromUuid, chunkAccounts.getAccounts());
|
||||||
|
|
Loading…
Reference in New Issue