BotGenerator/Generator/BotParser.cs
DrakiaXYZ 9ddbc91977 Drastically reduce memory footprint at the cost of removing threading
- Drastically improve performance, where non-threaded perf is now better than previous threaded perf
- Process all data in-line, as opposed to building huge lists of bot data that is then processed in groups later
- Memory footprint should now be relatively static no matter the amount of dumps
2024-11-03 23:29:50 -08:00

252 lines
8.9 KiB
C#

using System.Collections.Concurrent;
using Common.Models.Input;
using Newtonsoft.Json.Linq;
using System.Collections.Generic;
using System.Diagnostics;
using System.IO;
using System.Linq;
using System.Text.Json;
using System.Threading.Tasks;
using Common.Models.Output;
using Common.Models;
using Generator;
using Generator.Helpers.Gear;
namespace Common.Bots;
public static class BotParser
{
private static readonly JsonSerializerOptions serialiserOptions = new() { };
public static List<Bot> Parse(string dumpPath, HashSet<string> botTypes)
{
var stopwatch = Stopwatch.StartNew();
// Build the list of base bot data
var baseBots = new HashSet<Bot>();
foreach (var botType in botTypes)
{
var typeToAdd = (BotType)Enum.Parse(typeof(BotType), botType);
baseBots.Add(new Bot(typeToAdd));
}
DiskHelpers.CreateDirIfDoesntExist(dumpPath);
var botFiles = Directory.GetFiles(dumpPath, "*.json", SearchOption.TopDirectoryOnly);
LoggingHelpers.LogToConsole($"{botFiles.Length} bot dump files found");
// Store a list of parsed bots so we don't parse the same bot twice
int totalDupeCount = 0;
var parsedBotIds = new HashSet<string>();
int i = 0;
foreach (var filePath in botFiles)
{
i++;
if (i % 100 == 0) Console.WriteLine($"Processing file {i}");
ProcessBotFileSync(baseBots, filePath, parsedBotIds, totalDupeCount);
}
// Handle things we can only do once all data has been processed, or only needs to run once per bot type
foreach (var bot in baseBots)
{
// Skip any bots we didn't handle
if (bot.botCount == 0) continue;
BaseBotGenerator.AddDifficulties(bot);
GearChanceHelpers.CalculateModChances(bot);
GearChanceHelpers.CalculateEquipmentModChances(bot);
GearChanceHelpers.CalculateEquipmentChances(bot);
GearChanceHelpers.ApplyModChanceOverrides(bot);
GearChanceHelpers.ApplyEquipmentChanceOverrides(bot);
GearHelpers.ReduceAmmoWeightValues(bot);
GearHelpers.ReduceEquipmentWeightValues(bot.inventory.equipment);
GearHelpers.ReduceWeightValues(bot.appearance.voice);
GearHelpers.ReduceWeightValues(bot.appearance.feet);
GearHelpers.ReduceWeightValues(bot.appearance.body);
GearHelpers.ReduceWeightValues(bot.appearance.head);
GearHelpers.ReduceWeightValues(bot.appearance.hands);
}
stopwatch.Stop();
LoggingHelpers.LogToConsole($"{totalDupeCount} dupes were ignored. Took {LoggingHelpers.LogTimeTaken(stopwatch.Elapsed.TotalSeconds)} seconds");
return baseBots.ToList();
}
private static void ProcessBotFileSync(
HashSet<Bot> baseBots,
string filePath,
HashSet<string> parsedBotIds,
int totalDupeCount)
{
var splitFilePath = filePath.Split("\\");
int dupeCount = 0;
List<Datum> bots = [];
try
{
// Parse the bots inside the json file
using (var reader = new StreamReader(filePath))
{
var deSerialisedObject = JsonSerializer.Deserialize<Root>(reader.ReadToEnd(), serialiserOptions);
foreach (var botData in deSerialisedObject.data)
{
// Bot fucks up something, never allow it in
if (botData._id == "6483938c53cc9087c70eae86")
{
Console.WriteLine("oh no");
continue;
}
var baseBot = baseBots.SingleOrDefault(bot => bot.botType.ToString().Equals(botData.Info.Settings.Role, StringComparison.OrdinalIgnoreCase));
if (baseBot == null)
{
//Console.WriteLine($"Skipping {botData._id} due to unknown role {botData.Info.Settings.Role}");
continue;
}
// Add bot if not already added
if (!parsedBotIds.Add(botData._id))
{
dupeCount++;
}
baseBot.botCount += 1;
BaseBotGenerator.UpdateBaseDetails(baseBot, botData);
BotGearGenerator.AddGear(baseBot, botData);
BotLootGenerator.AddLoot(baseBot, botData);
BotChancesGenerator.AddChances(baseBot, botData);
}
}
}
catch (Exception ex)
{
Console.WriteLine($"File parse fucked up: {filePath}");
throw;
}
totalDupeCount += dupeCount;
}
public static async Task<List<Datum>> ParseAsync(string dumpPath, HashSet<string> botTypes)
{
var stopwatch = Stopwatch.StartNew();
DiskHelpers.CreateDirIfDoesntExist(dumpPath);
var botFiles = Directory.GetFiles(dumpPath, "*.json", SearchOption.TopDirectoryOnly);
LoggingHelpers.LogToConsole($"{botFiles.Length} bot dump files found");
// key = bot type
// Store bots keyed against their ID so we never get duplicates
var parsedBotsDict = new ConcurrentDictionary<string, Datum>();
int totalDupeCount = 0;
var tasks = new List<Task>();
foreach (var filePath in botFiles)
{
tasks.Add(ProcessBotFile(botTypes, filePath, parsedBotsDict, totalDupeCount));
}
await Task.WhenAll(tasks.ToArray());
stopwatch.Stop();
LoggingHelpers.LogToConsole($"Cleaned and Parsed: {parsedBotsDict.Count} bots. {totalDupeCount} dupes were ignored. Took {LoggingHelpers.LogTimeTaken(stopwatch.Elapsed.TotalSeconds)} seconds");
return [.. parsedBotsDict.Values];
}
private static async Task<int> ProcessBotFile(
HashSet<string> botTypes,
string filePath,
ConcurrentDictionary<string, Datum> parsedBotsDict,
int totalDupeCount)
{
var splitFilePath = filePath.Split("\\");
int dupeCount = 0;
List<Datum> bots = [];
try
{
// Parse the bots inside the json file
using (var reader = new StreamReader(filePath))
{
var deSerialisedObject = JsonSerializer.Deserialize<Root>(reader.ReadToEnd(), serialiserOptions);
bots.AddRange(deSerialisedObject.data.Where(botData => botTypes.Contains(botData.Info.Settings.Role.ToLower())));
}
}
catch (Exception ex)
{
Console.WriteLine($"File parse fucked up: {filePath}");
throw;
}
if (bots == null || bots.Count == 0)
{
Console.WriteLine($"Skipping file: {splitFilePath.Last()}. no bots found, ");
return totalDupeCount;
}
//Console.WriteLine($"parsing: {bots.Count} bots in file {splitFilePath.Last()}");
foreach (var bot in bots)
{
// Bot fucks up something, never allow it in
if (bot._id == "6483938c53cc9087c70eae86")
{
Console.WriteLine("oh no");
continue;
}
// null out unnecessary data to save ram
bot.Stats = null;
bot.Encyclopedia = null;
bot.Hideout = null;
bot.TaskConditionCounters = null;
bot.Bonuses = null;
bot.InsuredItems = null;
// Add bot if not already added
if (!parsedBotsDict.TryAdd(bot._id, bot))
{
dupeCount++;
}
}
totalDupeCount += dupeCount;
//Console.WriteLine($"Parsed file: {filePath}");
return totalDupeCount;
}
private static IEnumerable<Datum> ParseJson(string json)
{
var deSerialisedObject = JsonSerializer.Deserialize<Root>(json, serialiserOptions);
return deSerialisedObject.data;
}
private static string PruneMalformedBsgJson(string json, string fileName)
{
// Bsg send json where an item has a location of 1 but it should be an object with x/y/z coords
var o = JObject.Parse(json);
var jItemsToReplace = o.SelectTokens("$.data[*].Inventory.items[?(@.location == 1)].location");
//var jItemsToReplace = o.SelectTokens("$.data[*].Inventory.items[?(@.location == 1 && @.slotId == 'cartridges')].location");
if (jItemsToReplace != null && jItemsToReplace.Any())
{
LoggingHelpers.LogToConsole($"file {fileName} has {jItemsToReplace.Count()} json issues, cleaning up.", ConsoleColor.Yellow);
foreach (var item in jItemsToReplace)
{
var obj = new { x = 1, y = 0, r = 0 };
item.Replace(JToken.FromObject(obj));
}
}
var returnString = o.ToString();
o = null;
jItemsToReplace = null;
return returnString;
}
}