mirror of
https://github.com/overcuriousity/autopsy-flatpak.git
synced 2025-07-14 17:06:16 +00:00
Merge pull request #725 from rcordovano/ingest_scheduler_job_cancel_enhancement
Ingest scheduler job cancel enhancement
This commit is contained in:
commit
35a2ffa896
@ -65,13 +65,12 @@ final class IngestJob {
|
||||
long jobId = nextIngestJobId.incrementAndGet();
|
||||
IngestJob job = new IngestJob(jobId, dataSource, ingestModuleTemplates, processUnallocatedSpace);
|
||||
ingestJobsById.put(jobId, job);
|
||||
IngestManager.getInstance().fireIngestJobStarted(jobId);
|
||||
List<IngestModuleError> errors = job.start();
|
||||
if (errors.isEmpty()) {
|
||||
taskScheduler.scheduleTasksForIngestJob(job, dataSource);
|
||||
IngestManager.getInstance().fireIngestJobStarted(jobId);
|
||||
taskScheduler.addTasksForIngestJob(job, dataSource);
|
||||
} else {
|
||||
ingestJobsById.remove(jobId);
|
||||
IngestManager.getInstance().fireIngestJobCancelled(jobId);
|
||||
}
|
||||
return errors;
|
||||
}
|
||||
@ -106,7 +105,7 @@ final class IngestJob {
|
||||
return processUnallocatedSpace;
|
||||
}
|
||||
|
||||
List<IngestModuleError> start() throws InterruptedException {
|
||||
private List<IngestModuleError> start() throws InterruptedException {
|
||||
List<IngestModuleError> errors = startUpIngestPipelines();
|
||||
if (errors.isEmpty()) {
|
||||
startFileIngestProgressBar();
|
||||
@ -117,7 +116,7 @@ final class IngestJob {
|
||||
|
||||
private List<IngestModuleError> startUpIngestPipelines() throws InterruptedException {
|
||||
IngestJobContext context = new IngestJobContext(this);
|
||||
|
||||
|
||||
dataSourceIngestPipeline = new DataSourceIngestPipeline(context, ingestModuleTemplates);
|
||||
List<IngestModuleError> errors = new ArrayList<>();
|
||||
errors.addAll(dataSourceIngestPipeline.startUp());
|
||||
@ -180,24 +179,28 @@ final class IngestJob {
|
||||
}
|
||||
|
||||
void process(DataSourceIngestTask task) throws InterruptedException {
|
||||
// If the job is not cancelled, complete the task, otherwise just flush
|
||||
// it.
|
||||
if (!isCancelled()) {
|
||||
List<IngestModuleError> errors = new ArrayList<>();
|
||||
errors.addAll(dataSourceIngestPipeline.process(task.getDataSource(), dataSourceTasksProgress));
|
||||
if (!errors.isEmpty()) {
|
||||
logIngestModuleErrors(errors);
|
||||
}
|
||||
dataSourceTasksProgress.finish();
|
||||
} else {
|
||||
taskScheduler.removeTasksForIngestJob(id);
|
||||
}
|
||||
|
||||
// Because there is only one data source task per job, it is o.k. to
|
||||
// call ProgressHandle.finish() now that the data source ingest modules
|
||||
// are through using the progress bar via the DataSourceIngestModuleProgress wrapper.
|
||||
// Calling ProgressHandle.finish() again in finish() will be harmless.
|
||||
dataSourceTasksProgress.finish();
|
||||
|
||||
if (taskScheduler.isLastTaskForIngestJob(task)) {
|
||||
finish();
|
||||
}
|
||||
}
|
||||
|
||||
void process(FileIngestTask task) throws InterruptedException {
|
||||
// If the job is not cancelled, complete the task, otherwise just flush
|
||||
// it.
|
||||
if (!isCancelled()) {
|
||||
AbstractFile file = task.getFile();
|
||||
synchronized (this) {
|
||||
@ -215,7 +218,10 @@ final class IngestJob {
|
||||
if (!errors.isEmpty()) {
|
||||
logIngestModuleErrors(errors);
|
||||
}
|
||||
} else {
|
||||
taskScheduler.removeTasksForIngestJob(id);
|
||||
}
|
||||
|
||||
if (taskScheduler.isLastTaskForIngestJob(task)) {
|
||||
finish();
|
||||
}
|
||||
@ -227,13 +233,13 @@ final class IngestJob {
|
||||
FileIngestPipeline pipeline = fileIngestPipelines.poll();
|
||||
errors.addAll(pipeline.shutDown());
|
||||
}
|
||||
fileTasksProgress.finish();
|
||||
if (!errors.isEmpty()) {
|
||||
logIngestModuleErrors(errors);
|
||||
}
|
||||
|
||||
dataSourceTasksProgress.finish();
|
||||
fileTasksProgress.finish();
|
||||
ingestJobsById.remove(id);
|
||||
if (!cancelled) {
|
||||
if (!isCancelled()) {
|
||||
IngestManager.getInstance().fireIngestJobCompleted(id);
|
||||
}
|
||||
}
|
||||
@ -248,10 +254,8 @@ final class IngestJob {
|
||||
return cancelled;
|
||||
}
|
||||
|
||||
void cancel() {
|
||||
private void cancel() {
|
||||
cancelled = true;
|
||||
fileTasksProgress.finish();
|
||||
dataSourceTasksProgress.finish();
|
||||
IngestManager.getInstance().fireIngestJobCancelled(id);
|
||||
}
|
||||
}
|
||||
|
@ -21,6 +21,7 @@ package org.sleuthkit.autopsy.ingest;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collection;
|
||||
import java.util.Comparator;
|
||||
import java.util.Iterator;
|
||||
import java.util.List;
|
||||
import java.util.TreeSet;
|
||||
import java.util.concurrent.LinkedBlockingQueue;
|
||||
@ -55,22 +56,15 @@ final class IngestScheduler {
|
||||
private IngestScheduler() {
|
||||
}
|
||||
|
||||
synchronized void scheduleTasksForIngestJob(IngestJob job, Content dataSource) throws InterruptedException {
|
||||
// Enqueue a data source ingest task for the data source.
|
||||
// If the thread executing this code is interrupted, it is because the
|
||||
// the number of ingest threads has been decreased while ingest jobs are
|
||||
// running. The calling thread will exit in an orderly fashion, but the
|
||||
// task still needs to be enqueued rather than lost, hence the loop.
|
||||
synchronized void addTasksForIngestJob(IngestJob job, Content dataSource) throws InterruptedException {
|
||||
// Enqueue a data source ingest task for the data source.
|
||||
DataSourceIngestTask task = new DataSourceIngestTask(job, dataSource);
|
||||
while (true) {
|
||||
try {
|
||||
dataSourceTasks.put(task);
|
||||
break;
|
||||
} catch (InterruptedException ex) {
|
||||
// Reset the interrupted status of the thread so the orderly
|
||||
// exit can occur in the intended place.
|
||||
Thread.currentThread().interrupt();
|
||||
}
|
||||
try {
|
||||
dataSourceTasks.put(task);
|
||||
} catch (InterruptedException ex) {
|
||||
Thread.currentThread().interrupt();
|
||||
logger.log(Level.FINE, "Task scheduling for ingest job interrupted", ex); //NON-NLS
|
||||
return;
|
||||
}
|
||||
|
||||
// Get the top level files of the data source.
|
||||
@ -117,7 +111,40 @@ final class IngestScheduler {
|
||||
void addFileTaskToIngestJob(IngestJob job, AbstractFile file) {
|
||||
FileIngestTask task = new FileIngestTask(job, file);
|
||||
if (shouldEnqueueFileTask(task)) {
|
||||
addTaskToFileQueue(task);
|
||||
try {
|
||||
fileTasks.put(task);
|
||||
} catch (InterruptedException ex) {
|
||||
Thread.currentThread().interrupt();
|
||||
logger.log(Level.FINE, "Task scheduling for ingest job interrupted", ex); //NON-NLS
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
synchronized void removeTasksForIngestJob(long ingestJobId) {
|
||||
// Remove all tasks for this ingest job that are not in progress.
|
||||
Iterator<FileIngestTask> fileTasksIterator = fileTasks.iterator();
|
||||
while (fileTasksIterator.hasNext()) {
|
||||
if (fileTasksIterator.next().getIngestJob().getId() == ingestJobId) {
|
||||
fileTasksIterator.remove();
|
||||
}
|
||||
}
|
||||
Iterator<FileIngestTask> directoryTasksIterator = directoryTasks.iterator();
|
||||
while (directoryTasksIterator.hasNext()) {
|
||||
if (directoryTasksIterator.next().getIngestJob().getId() == ingestJobId) {
|
||||
directoryTasksIterator.remove();
|
||||
}
|
||||
}
|
||||
Iterator<FileIngestTask> rootDirectoryTasksIterator = rootDirectoryTasks.iterator();
|
||||
while (rootDirectoryTasksIterator.hasNext()) {
|
||||
if (rootDirectoryTasksIterator.next().getIngestJob().getId() == ingestJobId) {
|
||||
rootDirectoryTasksIterator.remove();
|
||||
}
|
||||
}
|
||||
Iterator<DataSourceIngestTask> dataSourceTasksIterator = dataSourceTasks.iterator();
|
||||
while (dataSourceTasksIterator.hasNext()) {
|
||||
if (dataSourceTasksIterator.next().getIngestJob().getId() == ingestJobId) {
|
||||
dataSourceTasksIterator.remove();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@ -148,7 +175,13 @@ final class IngestScheduler {
|
||||
final AbstractFile parentFile = parentTask.getFile();
|
||||
// add itself to the file list
|
||||
if (shouldEnqueueFileTask(parentTask)) {
|
||||
addTaskToFileQueue(parentTask);
|
||||
try {
|
||||
fileTasks.put(parentTask);
|
||||
} catch (InterruptedException ex) {
|
||||
Thread.currentThread().interrupt();
|
||||
logger.log(Level.FINE, "Task scheduling for ingest job interrupted", ex); //NON-NLS
|
||||
return;
|
||||
}
|
||||
}
|
||||
// add its children to the file and directory lists
|
||||
try {
|
||||
@ -160,7 +193,13 @@ final class IngestScheduler {
|
||||
if (childFile.hasChildren()) {
|
||||
directoryTasks.add(childTask);
|
||||
} else if (shouldEnqueueFileTask(childTask)) {
|
||||
addTaskToFileQueue(childTask);
|
||||
try {
|
||||
fileTasks.put(childTask);
|
||||
} catch (InterruptedException ex) {
|
||||
Thread.currentThread().interrupt();
|
||||
logger.log(Level.FINE, "Task scheduling for ingest job interrupted", ex); //NON-NLS
|
||||
return;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -170,23 +209,6 @@ final class IngestScheduler {
|
||||
}
|
||||
}
|
||||
|
||||
private void addTaskToFileQueue(FileIngestTask task) {
|
||||
// If the thread executing this code is interrupted, it is because the
|
||||
// the number of ingest threads has been decreased while ingest jobs are
|
||||
// running. The calling thread will exit in an orderly fashion, but the
|
||||
// task still needs to be enqueued rather than lost.
|
||||
while (true) {
|
||||
try {
|
||||
fileTasks.put(task);
|
||||
break;
|
||||
} catch (InterruptedException ex) {
|
||||
// Reset the interrupted status of the thread so the orderly
|
||||
// exit can occur in the intended place.
|
||||
Thread.currentThread().interrupt();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private static boolean shouldEnqueueFileTask(final FileIngestTask processTask) {
|
||||
final AbstractFile aFile = processTask.getFile();
|
||||
//if it's unalloc file, skip if so scheduled
|
||||
|
Loading…
x
Reference in New Issue
Block a user