mirror of
https://github.com/sp-tarkov/loot-dump-processor.git
synced 2025-02-13 08:10:45 -05:00
245 lines
8.6 KiB
C#
245 lines
8.6 KiB
C#
using System.Collections.Concurrent;
|
|
using System.Diagnostics;
|
|
using LootDumpProcessor.Model.Input;
|
|
using LootDumpProcessor.Process.Collector;
|
|
using LootDumpProcessor.Process.Processor.DumpProcessor;
|
|
using LootDumpProcessor.Process.Processor.FileProcessor;
|
|
using LootDumpProcessor.Process.Reader.Filters;
|
|
using LootDumpProcessor.Process.Reader.Intake;
|
|
using LootDumpProcessor.Process.Reader.PreProcess;
|
|
using LootDumpProcessor.Process.Writer;
|
|
using LootDumpProcessor.Serializers.Json;
|
|
using LootDumpProcessor.Storage;
|
|
using LootDumpProcessor.Utils;
|
|
using Microsoft.Extensions.Logging;
|
|
|
|
namespace LootDumpProcessor.Process;
|
|
|
|
public class QueuePipeline(
|
|
IFileProcessor fileProcessor, IDumpProcessor dumpProcessor, ILogger<QueuePipeline> logger,
|
|
IPreProcessReader preProcessReader, IFileFilter fileFilter, IIntakeReader intakeReader
|
|
)
|
|
: IPipeline
|
|
{
|
|
private readonly IFileProcessor _fileProcessor =
|
|
fileProcessor ?? throw new ArgumentNullException(nameof(fileProcessor));
|
|
|
|
private readonly IDumpProcessor _dumpProcessor =
|
|
dumpProcessor ?? throw new ArgumentNullException(nameof(dumpProcessor));
|
|
|
|
private readonly ILogger<QueuePipeline> _logger = logger ?? throw new ArgumentNullException(nameof(logger));
|
|
|
|
private readonly IPreProcessReader _preProcessReader =
|
|
preProcessReader ?? throw new ArgumentNullException(nameof(preProcessReader));
|
|
|
|
private readonly IFileFilter _fileFilter = fileFilter ?? throw new ArgumentNullException(nameof(fileFilter));
|
|
|
|
private readonly IIntakeReader
|
|
_intakeReader = intakeReader ?? throw new ArgumentNullException(nameof(intakeReader));
|
|
|
|
private readonly List<string> _filesToRename = new();
|
|
private readonly BlockingCollection<string> _filesToProcess = new();
|
|
|
|
private readonly List<string> _mapNames = LootDumpProcessorContext.GetConfig().MapsToProcess;
|
|
|
|
|
|
public async Task Execute()
|
|
{
|
|
var stopwatch = Stopwatch.StartNew();
|
|
// Single collector instance to collect results
|
|
var collector = CollectorFactory.GetInstance();
|
|
collector.Setup();
|
|
|
|
_logger.LogInformation("Gathering files to begin processing");
|
|
|
|
try
|
|
{
|
|
await FixFilesFromDumps();
|
|
foreach (var mapName in _mapNames)
|
|
{
|
|
ProcessFilesFromDumpsPerMap(collector, mapName);
|
|
}
|
|
}
|
|
finally
|
|
{
|
|
_preProcessReader.Dispose();
|
|
stopwatch.Stop();
|
|
_logger.LogInformation("Dumps processed in {@Time}", stopwatch.Elapsed);
|
|
}
|
|
}
|
|
|
|
private List<string> GatherFiles()
|
|
{
|
|
// Read locations
|
|
var inputPath = LootDumpProcessorContext.GetConfig().ReaderConfig.DumpFilesLocation;
|
|
|
|
if (inputPath == null || inputPath.Count == 0)
|
|
{
|
|
throw new Exception("Reader dumpFilesLocations must be set to a valid value");
|
|
}
|
|
|
|
// We gather up all files into a queue
|
|
var queuedFilesToProcess = GetFileQueue(inputPath);
|
|
// Then we preprocess the files in the queue and get them ready for processing
|
|
return PreProcessQueuedFiles(queuedFilesToProcess);
|
|
}
|
|
|
|
private List<string> PreProcessQueuedFiles(Queue<string> queuedFilesToProcess)
|
|
{
|
|
var gatheredFiles = new List<string>();
|
|
|
|
if (queuedFilesToProcess.Count == 0)
|
|
{
|
|
throw new Exception("No files matched accepted extension types in configs");
|
|
}
|
|
|
|
while (queuedFilesToProcess.TryDequeue(out var file))
|
|
{
|
|
var extensionFull = Path.GetExtension(file);
|
|
if (extensionFull.Length > 1)
|
|
{
|
|
// if the preprocessor found something new to process or generated new files, add them to the queue
|
|
if (extensionFull == "7z" &&
|
|
_preProcessReader.TryPreProcess(file, out var outputFiles, out var outputDirectories))
|
|
{
|
|
// all new directories need to be processed as well
|
|
GetFileQueue(outputDirectories).ToList().ForEach(queuedFilesToProcess.Enqueue);
|
|
// all output files need to be queued as well
|
|
outputFiles.ForEach(queuedFilesToProcess.Enqueue);
|
|
}
|
|
|
|
else
|
|
{
|
|
// if there is no preprocessor for the file, means its ready to filter or accept
|
|
|
|
if (_fileFilter.Accept(file))
|
|
{
|
|
gatheredFiles.Add(file);
|
|
}
|
|
|
|
else
|
|
{
|
|
gatheredFiles.Add(file);
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
// Handle invalid extension
|
|
_logger.LogWarning("File '{File}' has an invalid extension.", file);
|
|
}
|
|
}
|
|
|
|
return gatheredFiles;
|
|
}
|
|
|
|
private Queue<string> GetFileQueue(List<string> inputPath)
|
|
{
|
|
var queuedPathsToProcess = new Queue<string>();
|
|
var queuedFilesToProcess = new Queue<string>();
|
|
|
|
inputPath.ForEach(p => queuedPathsToProcess.Enqueue(p));
|
|
|
|
while (queuedPathsToProcess.TryDequeue(out var path))
|
|
{
|
|
// Check the input file to be sure its going to have data on it.
|
|
if (!Directory.Exists(path))
|
|
{
|
|
throw new Exception($"Input directory \"{path}\" could not be found");
|
|
}
|
|
|
|
// If we should process subfolder, queue them up as well
|
|
if (LootDumpProcessorContext.GetConfig().ReaderConfig.ProcessSubFolders)
|
|
{
|
|
foreach (var directory in Directory.GetDirectories(path))
|
|
{
|
|
queuedPathsToProcess.Enqueue(directory);
|
|
}
|
|
}
|
|
|
|
var files = Directory.GetFiles(path);
|
|
|
|
foreach (var file in files)
|
|
{
|
|
queuedFilesToProcess.Enqueue(file);
|
|
}
|
|
}
|
|
|
|
return queuedFilesToProcess;
|
|
}
|
|
|
|
private void ProcessFilesFromDumpsPerMap(ICollector collector, string mapName)
|
|
{
|
|
// Gather all files, sort them by date descending and then add them into the processing queue
|
|
GatherFiles().FindAll(f => f.ToLower().Contains($"{mapName}--")).OrderByDescending(f =>
|
|
{
|
|
FileDateParser.TryParseFileDate(f, out var date);
|
|
return date;
|
|
}
|
|
).ToList().ForEach(f => _filesToProcess.Add(f));
|
|
|
|
_logger.LogInformation("Files sorted and ready to begin pre-processing");
|
|
|
|
Parallel.ForEach(_filesToProcess, file =>
|
|
{
|
|
try
|
|
{
|
|
if (_intakeReader.Read(file, out var basicInfo))
|
|
{
|
|
collector.Hold(_fileProcessor.Process(basicInfo));
|
|
}
|
|
}
|
|
catch (Exception e)
|
|
{
|
|
_logger.LogError(e, "Error occurred while processing file {File}", file);
|
|
}
|
|
});
|
|
|
|
_logger.LogInformation("Pre-processing finished");
|
|
|
|
// Single writer instance to collect results
|
|
var writer = WriterFactory.GetInstance();
|
|
// Single collector instance to collect results
|
|
writer.WriteAll(_dumpProcessor.ProcessDumps(collector.Retrieve()));
|
|
|
|
// clear collector and datastorage as we process per map now
|
|
collector.Clear();
|
|
DataStorageFactory.GetInstance().Clear();
|
|
}
|
|
|
|
/// <summary>
|
|
/// Adds map name to file if they don't have it already.
|
|
/// </summary>
|
|
private async Task FixFilesFromDumps()
|
|
{
|
|
var inputPath = LootDumpProcessorContext.GetConfig().ReaderConfig.DumpFilesLocation;
|
|
|
|
if (inputPath == null || inputPath.Count == 0)
|
|
{
|
|
throw new Exception("Reader dumpFilesLocations must be set to a valid value");
|
|
}
|
|
|
|
GetFileQueue(inputPath).ToList().ForEach(f => _filesToRename.Add(f));
|
|
|
|
var jsonUtil = JsonSerializerFactory.GetInstance(JsonSerializerTypes.DotNet);
|
|
|
|
await Parallel.ForEachAsync(_filesToRename, async (file, cancellationToken) =>
|
|
{
|
|
if (_mapNames.Any(file.Contains)) return;
|
|
|
|
try
|
|
{
|
|
var data = await File.ReadAllTextAsync(file, cancellationToken);
|
|
var fileData = jsonUtil.Deserialize<RootData>(data);
|
|
var newPath = file.Replace("resp", $"{fileData.Data.LocationLoot.Id.ToLower()}--resp");
|
|
File.Move(file, newPath);
|
|
}
|
|
catch (Exception e)
|
|
{
|
|
_logger.LogError(e, "Error occurred while processing file {File}", file);
|
|
}
|
|
});
|
|
|
|
_logger.LogInformation("File-processing finished");
|
|
}
|
|
} |