using System; using System.Collections.Generic; using System.IO; using System.IO.Compression; using System.Linq; using System.Threading.Tasks; using Wabbajack.Common; using Wabbajack.Lib.Downloaders; using Wabbajack.VirtualFileSystem; using Path = Alphaleonis.Win32.Filesystem.Path; namespace Wabbajack.Lib { public abstract class AInstaller : ABatchProcessor { public bool IgnoreMissingFiles { get; internal set; } = false; public AbsolutePath OutputFolder { get; private set; } public AbsolutePath DownloadFolder { get; private set; } public abstract ModManager ModManager { get; } public AbsolutePath ModListArchive { get; private set; } public ModList ModList { get; private set; } public Dictionary HashedArchives { get; } = new Dictionary(); public GameMetaData Game { get; } public SystemParameters? SystemParameters { get; set; } public bool UseCompression { get; set; } public TempFolder? ExtractedModlistFolder { get; set; } = null; public AInstaller(AbsolutePath archive, ModList modList, AbsolutePath outputFolder, AbsolutePath downloadFolder, SystemParameters? parameters, int steps, Game game) : base(steps) { ModList = modList; ModListArchive = archive; OutputFolder = outputFolder; DownloadFolder = downloadFolder; SystemParameters = parameters; Game = game.MetaData(); } public async Task ExtractModlist() { ExtractedModlistFolder = await TempFolder.Create(); await FileExtractor2.ExtractAll(Queue, ModListArchive, ExtractedModlistFolder.Dir); } public void Info(string msg) { Utils.Log(msg); } public void Status(string msg) { Queue.Report(msg, Percent.Zero); } public void Error(string msg) { Utils.Log(msg); throw new Exception(msg); } public async Task LoadBytesFromPath(RelativePath path) { var fullPath = ExtractedModlistFolder!.Dir.Combine(path); if (!fullPath.IsFile) throw new Exception($"Cannot load inlined data {path} file does not exist"); return await fullPath.ReadAllBytesAsync(); } public static ModList LoadFromFile(AbsolutePath path) { using var fs = new FileStream((string)path, FileMode.Open, FileAccess.Read, FileShare.Read); using var ar = new ZipArchive(fs, ZipArchiveMode.Read); var entry = ar.GetEntry("modlist"); if (entry == null) { entry = ar.GetEntry("modlist.json"); if (entry == null) throw new Exception("Invalid Wabbajack Installer"); using var e = entry.Open(); return e.FromJson(); } using (var e = entry.Open()) return e.FromJson(); } /// /// We don't want to make the installer index all the archives, that's just a waste of time, so instead /// we'll pass just enough information to VFS to let it know about the files we have. /// protected async Task PrimeVFS() { VFS.AddKnown(ModList.Directives.OfType().Select(d => d.ArchiveHashPath), HashedArchives); await VFS.BackfillMissing(); } public void BuildFolderStructure() { Info("Building Folder Structure"); ModList.Directives .Select(d => OutputFolder.Combine(d.To.Parent)) .Distinct() .Do(f => f.CreateDirectory()); } public async Task InstallArchives() { var grouped = ModList.Directives .OfType() .Select(a => new {VF = VFS.Index.FileForArchiveHashPath(a.ArchiveHashPath), Directive = a}) .GroupBy(a => a.VF) .ToDictionary(a => a.Key); if (grouped.Count == 0) return; await VFS.Extract(Queue, grouped.Keys.ToHashSet(), async (vf, sf) => { foreach (var directive in grouped[vf]) { var file = directive.Directive; switch (file) { case PatchedFromArchive pfa: { await using var s = await sf.GetStream(); s.Position = 0; var patchData = await LoadBytesFromPath(pfa.PatchID); var toFile = file.To.RelativeTo(OutputFolder); { await using var os = await toFile.Create(); Utils.ApplyPatch(s, () => new MemoryStream(patchData), os); } if (await VirusScanner.ShouldScan(toFile) && await ClientAPI.GetVirusScanResult(toFile) == VirusScanner.Result.Malware) { await toFile.DeleteAsync(); Utils.ErrorThrow(new Exception($"Virus scan of patched executable reported possible malware: {toFile.ToString()} ({(long)await toFile.FileHashCachedAsync()})")); } } break; case FromArchive _: if (grouped[vf].Count() == 1) { await sf.Move(directive.Directive.To.RelativeTo(OutputFolder)); } else { await using var s = await sf.GetStream(); await directive.Directive.To.RelativeTo(OutputFolder).WriteAllAsync(s, false); } break; default: throw new Exception($"No handler for {directive}"); } if (file is PatchedFromArchive) { await file.To.RelativeTo(OutputFolder).FileHashAsync(); } else { file.To.RelativeTo(OutputFolder).FileHashWriteCache(file.Hash); } if (UseCompression) { Utils.Status($"Compacting {file.To}"); await file.To.RelativeTo(OutputFolder).Compact(FileCompaction.Algorithm.XPRESS16K); } } }, tempFolder: OutputFolder, updateTracker: UpdateTracker); } public async Task DownloadArchives() { var missing = ModList.Archives.Where(a => !HashedArchives.ContainsKey(a.Hash)).ToList(); Info($"Missing {missing.Count} archives"); Info("Getting Nexus API Key, if a browser appears, please accept"); var dispatchers = missing.Select(m => m.State.GetDownloader()).Distinct(); await Task.WhenAll(dispatchers.Select(d => d.Prepare())); var nexusDownloader = dispatchers.OfType().FirstOrDefault(); if (nexusDownloader != null && !await nexusDownloader.HaveEnoughAPICalls(missing)) { throw new Exception($"Not enough Nexus API calls to download this list, please try again after midnight GMT when your API limits reset"); } await DownloadMissingArchives(missing); } public async Task DownloadMissingArchives(List missing, bool download = true) { if (download) { var result = SendDownloadMetrics(missing); foreach (var a in missing.Where(a => a.State.GetType() == typeof(ManualDownloader.State))) { var outputPath = DownloadFolder.Combine(a.Name); await a.State.Download(a, outputPath); } } DesiredThreads.OnNext(DownloadThreads); await missing.Where(a => a.State.GetType() != typeof(ManualDownloader.State)) .PMap(Queue, UpdateTracker, async archive => { Info($"Downloading {archive.Name}"); var outputPath = DownloadFolder.Combine(archive.Name); if (download) { if (outputPath.Exists) { var origName = Path.GetFileNameWithoutExtension(archive.Name); var ext = Path.GetExtension(archive.Name); var uniqueKey = archive.State.PrimaryKeyString.StringSha256Hex(); outputPath = DownloadFolder.Combine(origName + "_" + uniqueKey + "_" + ext); await outputPath.DeleteAsync(); } } return await DownloadArchive(archive, download, outputPath); }); DesiredThreads.OnNext(DiskThreads); } private async Task SendDownloadMetrics(List missing) { var grouped = missing.GroupBy(m => m.State.GetType()); foreach (var group in grouped) { await Metrics.Send($"downloading_{group.Key.FullName!.Split(".").Last().Split("+").First()}", group.Sum(g => g.Size).ToString()); } } public async Task DownloadArchive(Archive archive, bool download, AbsolutePath? destination = null) { try { destination ??= DownloadFolder.Combine(archive.Name); var result = await DownloadDispatcher.DownloadWithPossibleUpgrade(archive, destination.Value); if (result == DownloadDispatcher.DownloadResult.Update) { await destination.Value.MoveToAsync(destination.Value.Parent.Combine(archive.Hash.ToHex())); } } catch (Exception ex) { Utils.Log($"Download error for file {archive.Name}"); Utils.Log(ex.ToString()); return false; } return false; } public async Task HashArchives() { Utils.Log("Looking for files to hash"); // Enumerate all files. var allFiles = DownloadFolder.EnumerateFiles() .Concat(Game.GameLocation().EnumerateFiles()) .ToList(); // Create a Dictionary with filesize as the key. var hashDict = new Dictionary>(); // Populate the Dictionary with all files we found, using a list for the paths to allow for duplicated filesizes. allFiles.ForEach(o => { if (hashDict.ContainsKey(o.Size)) { hashDict[o.Size].Add(o); } else { hashDict.Add(o.Size, new List { o }); } }); // Create a list for the files we filter. var toHash = new List(); // Iterate through required archives, adding any files we have that match filesize. ModList.Archives.ForEach(o => { if (hashDict.ContainsKey(o.Size)) { hashDict[o.Size].ForEach(o => toHash.Add(o)); } }); Utils.Log($"Found {allFiles.Count} total files, {toHash.Count} matching filesize"); // Hash all the files we added. var hashResults = await toHash .PMap(Queue, UpdateTracker,async e => (await e.FileHashCachedAsync(), e)); HashedArchives.SetTo(hashResults .OrderByDescending(e => e.Item2.LastModified) .GroupBy(e => e.Item1) .Select(e => e.First()) .Select(e => new KeyValuePair(e.Item1, e.Item2))); } /// /// Disabled /// public void ValidateFreeSpace() { return; // Disabled, caused more problems than it was worth. /* DiskSpaceInfo DriveInfo(string path) { return Volume.GetDiskFreeSpace(Volume.GetUniqueVolumeNameForPath(path)); } var paths = new[] {(OutputFolder, ModList.InstallSize), (DownloadFolder, ModList.DownloadSize), (Directory.GetCurrentDirectory(), ModList.ScratchSpaceSize)}; paths.GroupBy(f => DriveInfo(f.Item1).DriveName) .Do(g => { var required = g.Sum(i => i.Item2); var contains = g.Sum(folder => Directory.EnumerateFiles(folder.Item1, "*", DirectoryEnumerationOptions.Recursive) .Sum(file => new FileInfo(file).Length)); var available = DriveInfo(g.Key).FreeBytesAvailable; if (required - contains > available) throw new NotEnoughDiskSpaceException( $"This ModList requires {required.ToFileSizeString()} on {g.Key} but only {available.ToFileSizeString()} is available."); }); */ } /// /// The user may already have some files in the OutputFolder. If so we can go through these and /// figure out which need to be updated, deleted, or left alone /// public async Task OptimizeModlist() { Utils.Log("Optimizing ModList directives"); // Clone the ModList so our changes don't modify the original data ModList = ModList.Clone(); var indexed = ModList.Directives.ToDictionary(d => d.To); var profileFolder = OutputFolder.Combine("profiles"); var savePath = (RelativePath)"saves"; UpdateTracker.NextStep("Looking for files to delete"); await OutputFolder.EnumerateFiles() .PMap(Queue, UpdateTracker, async f => { var relativeTo = f.RelativeTo(OutputFolder); if (indexed.ContainsKey(relativeTo) || f.InFolder(DownloadFolder)) return; if (f.InFolder(profileFolder) && f.Parent.FileName == savePath) return; Utils.Log($"Deleting {relativeTo} it's not part of this ModList"); await f.DeleteAsync(); }); Utils.Log("Cleaning empty folders"); var expectedFolders = indexed.Keys .Select(f => f.RelativeTo(OutputFolder)) // We ignore the last part of the path, so we need a dummy file name .Append(DownloadFolder.Combine("_")) .Where(f => f.InFolder(OutputFolder)) .SelectMany(path => { // Get all the folders and all the folder parents // so for foo\bar\baz\qux.txt this emits ["foo", "foo\\bar", "foo\\bar\\baz"] var split = ((string)path.RelativeTo(OutputFolder)).Split('\\'); return Enumerable.Range(1, split.Length - 1).Select(t => string.Join("\\", split.Take(t))); }) .Distinct() .Select(p => OutputFolder.Combine(p)) .ToHashSet(); try { var toDelete = OutputFolder.EnumerateDirectories(true) .Where(p => !expectedFolders.Contains(p)) .OrderByDescending(p => ((string)p).Length) .ToList(); foreach (var dir in toDelete) { await dir.DeleteDirectory(dontDeleteIfNotEmpty:true); } } catch (Exception) { // ignored because it's not worth throwing a fit over Utils.Log("Error when trying to clean empty folders. This doesn't really matter."); } var existingfiles = OutputFolder.EnumerateFiles().ToHashSet(); UpdateTracker.NextStep("Looking for unmodified files"); (await indexed.Values.PMap(Queue, UpdateTracker, async d => { // Bit backwards, but we want to return null for // all files we *want* installed. We return the files // to remove from the install list. var path = OutputFolder.Combine(d.To); if (!existingfiles.Contains(path)) return null; return await path.FileHashCachedAsync() == d.Hash ? d : null; })) .Do(d => { if (d != null) { indexed.Remove(d.To); } }); UpdateTracker.NextStep("Updating ModList"); Utils.Log($"Optimized {ModList.Directives.Count} directives to {indexed.Count} required"); var requiredArchives = indexed.Values.OfType() .GroupBy(d => d.ArchiveHashPath.BaseHash) .Select(d => d.Key) .ToHashSet(); ModList.Archives = ModList.Archives.Where(a => requiredArchives.Contains(a.Hash)).ToList(); ModList.Directives = indexed.Values.ToList(); } } public class NotEnoughDiskSpaceException : Exception { public NotEnoughDiskSpaceException(string s) : base(s) { } } }