Add start/chunk/sleep logging to crawler

This commit is contained in:
Chris Eager 2021-09-03 14:18:39 -07:00 committed by Chris Eager
parent dceebc1c8d
commit b1274125c9
1 changed files with 6 additions and 1 deletions

View File

@ -106,7 +106,10 @@ public class AccountDatabaseCrawler implements Managed, Runnable {
} }
final long endTimeMs = System.currentTimeMillis(); final long endTimeMs = System.currentTimeMillis();
final long sleepIntervalMs = chunkIntervalMs - (endTimeMs - startTimeMs); final long sleepIntervalMs = chunkIntervalMs - (endTimeMs - startTimeMs);
if (sleepIntervalMs > 0) sleepWhileRunning(sleepIntervalMs); if (sleepIntervalMs > 0) {
logger.info("Sleeping {}ms", sleepIntervalMs);
sleepWhileRunning(sleepIntervalMs);
}
} finally { } finally {
cache.releaseActiveWork(workerId); cache.releaseActiveWork(workerId);
} }
@ -122,6 +125,7 @@ public class AccountDatabaseCrawler implements Managed, Runnable {
final Optional<UUID> fromUuid = getLastUuid(useDynamo); final Optional<UUID> fromUuid = getLastUuid(useDynamo);
if (fromUuid.isEmpty()) { if (fromUuid.isEmpty()) {
logger.info("Started crawl");
listeners.forEach(AccountDatabaseCrawlerListener::onCrawlStart); listeners.forEach(AccountDatabaseCrawlerListener::onCrawlStart);
} }
@ -133,6 +137,7 @@ public class AccountDatabaseCrawler implements Managed, Runnable {
cacheLastUuid(Optional.empty(), useDynamo); cacheLastUuid(Optional.empty(), useDynamo);
cache.setAccelerated(false); cache.setAccelerated(false);
} else { } else {
logger.info("Processing chunk");
try { try {
for (AccountDatabaseCrawlerListener listener : listeners) { for (AccountDatabaseCrawlerListener listener : listeners) {
listener.timeAndProcessCrawlChunk(fromUuid, chunkAccounts.getAccounts()); listener.timeAndProcessCrawlChunk(fromUuid, chunkAccounts.getAccounts());