mirror of
https://github.com/Thealexbarney/LibHac.git
synced 2024-11-14 10:49:41 +01:00
Merge branch 'master' of https://github.com/Thealexbarney/LibHac
This commit is contained in:
commit
f56500e257
37 changed files with 896 additions and 532 deletions
|
@ -105,11 +105,6 @@ namespace LibHac
|
|||
Counter[8] = (byte)((Counter[8] & 0xF0) | (int)(off & 0x0F));
|
||||
}
|
||||
|
||||
public override void Flush()
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
}
|
||||
|
||||
public override long Seek(long offset, SeekOrigin origin)
|
||||
{
|
||||
switch (origin)
|
||||
|
|
|
@ -21,6 +21,14 @@ namespace LibHac
|
|||
return comp;
|
||||
}
|
||||
|
||||
public static byte[] ComputeSha256(byte[] data, int offset, int count)
|
||||
{
|
||||
using (SHA256 sha = SHA256.Create())
|
||||
{
|
||||
return sha.ComputeHash(data, offset, count);
|
||||
}
|
||||
}
|
||||
|
||||
public static void DecryptEcb(byte[] key, byte[] src, int srcIndex, byte[] dest, int destIndex, int length)
|
||||
{
|
||||
using (Aes aes = Aes.Create())
|
||||
|
|
|
@ -39,31 +39,33 @@ namespace LibHac
|
|||
}
|
||||
|
||||
/// <summary>
|
||||
/// Checks the hashes of any unchecked blocks and returns the <see cref="Validity"/> of the hash level.
|
||||
/// Checks the hashes of any unchecked blocks and returns the <see cref="Validity"/> of the data.
|
||||
/// </summary>
|
||||
/// <param name="level">The level of hierarchical hashes to check.</param>
|
||||
/// <param name="returnOnError">If <see langword="true"/>, return as soon as an invalid block is found.</param>
|
||||
/// <param name="logger">An optional <see cref="IProgressReport"/> for reporting progress.</param>
|
||||
/// <returns>The <see cref="Validity"/> of the data of the specified hash level.</returns>
|
||||
public Validity ValidateLevel(int level, bool returnOnError, IProgressReport logger = null)
|
||||
public Validity Validate(bool returnOnError, IProgressReport logger = null)
|
||||
{
|
||||
Validity[] validities = LevelValidities[level];
|
||||
IntegrityVerificationStream levelStream = IntegrityStreams[level];
|
||||
Validity[] validities = LevelValidities[LevelValidities.Length - 1];
|
||||
IntegrityVerificationStream stream = IntegrityStreams[IntegrityStreams.Length - 1];
|
||||
|
||||
// The original position of the stream must be restored when we're done validating
|
||||
long initialPosition = levelStream.Position;
|
||||
// Restore the original position of the stream when we're done validating
|
||||
long initialPosition = stream.Position;
|
||||
|
||||
var buffer = new byte[levelStream.SectorSize];
|
||||
long blockSize = stream.SectorSize;
|
||||
int blockCount = (int)Util.DivideByRoundUp(Length, blockSize);
|
||||
|
||||
var buffer = new byte[blockSize];
|
||||
var result = Validity.Valid;
|
||||
|
||||
logger?.SetTotal(levelStream.SectorCount);
|
||||
logger?.SetTotal(blockCount);
|
||||
|
||||
for (int i = 0; i < levelStream.SectorCount; i++)
|
||||
for (int i = 0; i < blockCount; i++)
|
||||
{
|
||||
if (validities[i] == Validity.Unchecked)
|
||||
{
|
||||
levelStream.Position = (long)levelStream.SectorSize * i;
|
||||
levelStream.Read(buffer, 0, buffer.Length, IntegrityCheckLevel.IgnoreOnInvalid);
|
||||
stream.Position = blockSize * i;
|
||||
stream.Read(buffer, 0, buffer.Length, IntegrityCheckLevel.IgnoreOnInvalid);
|
||||
}
|
||||
|
||||
if (validities[i] == Validity.Invalid)
|
||||
|
@ -76,13 +78,13 @@ namespace LibHac
|
|||
}
|
||||
|
||||
logger?.SetTotal(0);
|
||||
levelStream.Position = initialPosition;
|
||||
stream.Position = initialPosition;
|
||||
return result;
|
||||
}
|
||||
|
||||
public override void Flush()
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
DataLevel.Flush();
|
||||
}
|
||||
|
||||
public override long Seek(long offset, SeekOrigin origin)
|
||||
|
@ -115,7 +117,7 @@ namespace LibHac
|
|||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
DataLevel.Write(buffer, offset, count);
|
||||
}
|
||||
|
||||
public override bool CanRead => DataLevel.CanRead;
|
||||
|
@ -128,4 +130,32 @@ namespace LibHac
|
|||
set => DataLevel.Position = value;
|
||||
}
|
||||
}
|
||||
|
||||
public static class HierarchicalIntegrityVerificationStreamExtensions
|
||||
{
|
||||
internal static void SetLevelValidities(this HierarchicalIntegrityVerificationStream stream, IvfcHeader header)
|
||||
{
|
||||
for (int i = 0; i < stream.Levels.Length - 1; i++)
|
||||
{
|
||||
Validity[] level = stream.LevelValidities[i];
|
||||
var levelValidity = Validity.Valid;
|
||||
|
||||
foreach (Validity block in level)
|
||||
{
|
||||
if (block == Validity.Invalid)
|
||||
{
|
||||
levelValidity = Validity.Invalid;
|
||||
break;
|
||||
}
|
||||
|
||||
if (block == Validity.Unchecked && levelValidity != Validity.Invalid)
|
||||
{
|
||||
levelValidity = Validity.Unchecked;
|
||||
}
|
||||
}
|
||||
|
||||
header.LevelHeaders[i].HashValidity = levelValidity;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -30,11 +30,6 @@ namespace LibHac
|
|||
BlockValidities = new Validity[SectorCount];
|
||||
}
|
||||
|
||||
public override void Flush()
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
}
|
||||
|
||||
public override long Seek(long offset, SeekOrigin origin)
|
||||
{
|
||||
switch (origin)
|
||||
|
@ -76,6 +71,7 @@ namespace LibHac
|
|||
if (Type == IntegrityStreamType.Save && _hashBuffer.IsEmpty())
|
||||
{
|
||||
Array.Clear(buffer, offset, SectorSize);
|
||||
BlockValidities[blockNum] = Validity.Valid;
|
||||
return bytesRead;
|
||||
}
|
||||
|
||||
|
@ -90,7 +86,7 @@ namespace LibHac
|
|||
bytesToHash = bytesRead;
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
if (BlockValidities[blockNum] == Validity.Invalid && integrityCheckLevel == IntegrityCheckLevel.ErrorOnInvalid)
|
||||
{
|
||||
throw new InvalidDataException("Hash error!");
|
||||
|
@ -100,23 +96,7 @@ namespace LibHac
|
|||
|
||||
if (BlockValidities[blockNum] != Validity.Unchecked) return bytesRead;
|
||||
|
||||
_hash.Initialize();
|
||||
|
||||
if (Type == IntegrityStreamType.Save)
|
||||
{
|
||||
_hash.TransformBlock(Salt, 0, Salt.Length, null, 0);
|
||||
}
|
||||
|
||||
_hash.TransformBlock(buffer, offset, bytesToHash, null, 0);
|
||||
_hash.TransformFinalBlock(buffer, 0, 0);
|
||||
|
||||
byte[] hash = _hash.Hash;
|
||||
|
||||
if (Type == IntegrityStreamType.Save)
|
||||
{
|
||||
// This bit is set on all save hashes
|
||||
hash[0x1F] |= 0x80;
|
||||
}
|
||||
byte[] hash = DoHash(buffer, offset, bytesToHash);
|
||||
|
||||
Validity validity = Util.ArraysEqual(_hashBuffer, hash) ? Validity.Valid : Validity.Invalid;
|
||||
BlockValidities[blockNum] = validity;
|
||||
|
@ -131,7 +111,48 @@ namespace LibHac
|
|||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
long blockNum = CurrentSector;
|
||||
int toWrite = (int)Math.Min(count, Length - Position);
|
||||
byte[] hash = DoHash(buffer, offset, toWrite);
|
||||
|
||||
if (Type == IntegrityStreamType.Save && buffer.IsEmpty())
|
||||
{
|
||||
Array.Clear(hash, 0, DigestSize);
|
||||
}
|
||||
|
||||
base.Write(buffer, offset, count);
|
||||
|
||||
HashStream.Position = blockNum * DigestSize;
|
||||
HashStream.Write(hash, 0, DigestSize);
|
||||
}
|
||||
|
||||
private byte[] DoHash(byte[] buffer, int offset, int count)
|
||||
{
|
||||
_hash.Initialize();
|
||||
|
||||
if (Type == IntegrityStreamType.Save)
|
||||
{
|
||||
_hash.TransformBlock(Salt, 0, Salt.Length, null, 0);
|
||||
}
|
||||
|
||||
_hash.TransformBlock(buffer, offset, count, null, 0);
|
||||
_hash.TransformFinalBlock(buffer, 0, 0);
|
||||
|
||||
byte[] hash = _hash.Hash;
|
||||
|
||||
if (Type == IntegrityStreamType.Save)
|
||||
{
|
||||
// This bit is set on all save hashes
|
||||
hash[0x1F] |= 0x80;
|
||||
}
|
||||
|
||||
return hash;
|
||||
}
|
||||
|
||||
public override void Flush()
|
||||
{
|
||||
HashStream.Flush();
|
||||
base.Flush();
|
||||
}
|
||||
|
||||
public override bool CanRead => true;
|
||||
|
|
|
@ -544,6 +544,7 @@ namespace LibHac
|
|||
|
||||
public enum KeyType
|
||||
{
|
||||
None,
|
||||
Common,
|
||||
Unique,
|
||||
Title
|
||||
|
|
|
@ -16,7 +16,7 @@
|
|||
<RepositoryType>git</RepositoryType>
|
||||
<RepositoryUrl>https://github.com/Thealexbarney/LibHac</RepositoryUrl>
|
||||
|
||||
<VersionPrefix>0.1.2</VersionPrefix>
|
||||
<VersionPrefix>0.1.3</VersionPrefix>
|
||||
<PathMap>$(MSBuildProjectDirectory)=C:/LibHac/</PathMap>
|
||||
<IncludeSymbols>true</IncludeSymbols>
|
||||
<IncludeSource>true</IncludeSource>
|
||||
|
|
|
@ -75,5 +75,25 @@ namespace LibHac
|
|||
base.GetObjectData(info, context);
|
||||
info.AddValue(nameof(Name), Name);
|
||||
}
|
||||
|
||||
public override string Message
|
||||
{
|
||||
get
|
||||
{
|
||||
string s = base.Message;
|
||||
|
||||
if (Type != KeyType.None)
|
||||
{
|
||||
s += $"{Environment.NewLine}Key Type: {Type}";
|
||||
}
|
||||
|
||||
if (Name != null)
|
||||
{
|
||||
s += $"{Environment.NewLine}Key Name: {Name}";
|
||||
}
|
||||
|
||||
return s;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -381,7 +381,6 @@ namespace LibHac
|
|||
stream.Read(hashTable, 0, hashTable.Length);
|
||||
|
||||
sect.MasterHashValidity = Crypto.CheckMemoryHashTable(hashTable, expected, 0, hashTable.Length);
|
||||
if (sect.Header.HashType == NcaHashType.Ivfc) sect.Header.IvfcInfo.LevelHeaders[0].HashValidity = sect.MasterHashValidity;
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
|
@ -400,7 +399,21 @@ namespace LibHac
|
|||
public int SectionNum { get; set; }
|
||||
public long Offset { get; set; }
|
||||
public long Size { get; set; }
|
||||
public Validity MasterHashValidity { get; set; }
|
||||
|
||||
public Validity MasterHashValidity
|
||||
{
|
||||
get
|
||||
{
|
||||
if (Header.HashType == NcaHashType.Ivfc) return Header.IvfcInfo.LevelHeaders[0].HashValidity;
|
||||
if (Header.HashType == NcaHashType.Sha256) return Header.Sha256Info.MasterHashValidity;
|
||||
return Validity.Unchecked;
|
||||
}
|
||||
set
|
||||
{
|
||||
if (Header.HashType == NcaHashType.Ivfc) Header.IvfcInfo.LevelHeaders[0].HashValidity = value;
|
||||
if (Header.HashType == NcaHashType.Sha256) Header.Sha256Info.MasterHashValidity = value;
|
||||
}
|
||||
}
|
||||
|
||||
public byte[] GetMasterHash()
|
||||
{
|
||||
|
@ -489,25 +502,18 @@ namespace LibHac
|
|||
if (stream == null) return Validity.Unchecked;
|
||||
|
||||
if (!quiet) logger?.LogMessage($"Verifying section {index}...");
|
||||
Validity validity = stream.Validate(true, logger);
|
||||
|
||||
for (int i = 0; i < stream.Levels.Length - 1; i++)
|
||||
if (hashType == NcaHashType.Ivfc)
|
||||
{
|
||||
if (!quiet) logger?.LogMessage($" Verifying Hash Level {i}...");
|
||||
Validity levelValidity = stream.ValidateLevel(i, true, logger);
|
||||
|
||||
if (hashType == NcaHashType.Ivfc)
|
||||
{
|
||||
sect.Header.IvfcInfo.LevelHeaders[i].HashValidity = levelValidity;
|
||||
}
|
||||
else if (hashType == NcaHashType.Sha256 && i == stream.Levels.Length - 2)
|
||||
{
|
||||
sect.Header.Sha256Info.HashValidity = levelValidity;
|
||||
}
|
||||
|
||||
if (levelValidity == Validity.Invalid) return Validity.Invalid;
|
||||
stream.SetLevelValidities(sect.Header.IvfcInfo);
|
||||
}
|
||||
else if (hashType == NcaHashType.Sha256)
|
||||
{
|
||||
sect.Header.Sha256Info.HashValidity = validity;
|
||||
}
|
||||
|
||||
return Validity.Valid;
|
||||
return validity;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -158,8 +158,8 @@ namespace LibHac
|
|||
{
|
||||
public string Magic;
|
||||
public int Version;
|
||||
public uint MasterHashSize;
|
||||
public uint NumLevels;
|
||||
public int MasterHashSize;
|
||||
public int NumLevels;
|
||||
public IvfcLevelHeader[] LevelHeaders = new IvfcLevelHeader[6];
|
||||
public byte[] SaltSource;
|
||||
public byte[] MasterHash;
|
||||
|
@ -167,10 +167,10 @@ namespace LibHac
|
|||
public IvfcHeader(BinaryReader reader)
|
||||
{
|
||||
Magic = reader.ReadAscii(4);
|
||||
Version = reader.ReadInt16();
|
||||
reader.BaseStream.Position += 2;
|
||||
MasterHashSize = reader.ReadUInt32();
|
||||
NumLevels = reader.ReadUInt32();
|
||||
Version = reader.ReadInt16();
|
||||
MasterHashSize = reader.ReadInt32();
|
||||
NumLevels = reader.ReadInt32();
|
||||
|
||||
for (int i = 0; i < LevelHeaders.Length; i++)
|
||||
{
|
||||
|
@ -210,8 +210,9 @@ namespace LibHac
|
|||
public long DataOffset;
|
||||
public long DataSize;
|
||||
|
||||
public Validity MasterHashValidity = Validity.Unchecked;
|
||||
public Validity HashValidity = Validity.Unchecked;
|
||||
|
||||
|
||||
public Sha256Info(BinaryReader reader)
|
||||
{
|
||||
MasterHash = reader.ReadBytes(0x20);
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
using System.IO;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class AllocationTable
|
||||
{
|
|
@ -1,6 +1,6 @@
|
|||
using System;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class AllocationTableIterator
|
||||
{
|
|
@ -1,7 +1,7 @@
|
|||
using System;
|
||||
using System.IO;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class AllocationTableStream : Stream
|
||||
{
|
||||
|
@ -22,7 +22,7 @@ namespace LibHac.Savefile
|
|||
|
||||
public override void Flush()
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
Data.Flush();
|
||||
}
|
||||
|
||||
public override int Read(byte[] buffer, int offset, int count)
|
||||
|
@ -53,7 +53,20 @@ namespace LibHac.Savefile
|
|||
|
||||
public override long Seek(long offset, SeekOrigin origin)
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
switch (origin)
|
||||
{
|
||||
case SeekOrigin.Begin:
|
||||
Position = offset;
|
||||
break;
|
||||
case SeekOrigin.Current:
|
||||
Position += offset;
|
||||
break;
|
||||
case SeekOrigin.End:
|
||||
Position = Length - offset;
|
||||
break;
|
||||
}
|
||||
|
||||
return Position;
|
||||
}
|
||||
|
||||
public override void SetLength(long value)
|
||||
|
@ -63,12 +76,29 @@ namespace LibHac.Savefile
|
|||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
int remaining = count;
|
||||
int outOffset = offset;
|
||||
|
||||
while (remaining > 0)
|
||||
{
|
||||
int remainingInSegment = Iterator.CurrentSegmentSize * BlockSize - SegmentPos;
|
||||
int bytesToWrite = Math.Min(remaining, remainingInSegment);
|
||||
Data.Write(buffer, outOffset, bytesToWrite);
|
||||
|
||||
outOffset += bytesToWrite;
|
||||
remaining -= bytesToWrite;
|
||||
|
||||
if (SegmentPos >= Iterator.CurrentSegmentSize * BlockSize)
|
||||
{
|
||||
if (!Iterator.MoveNext()) return;
|
||||
Data.Position = Iterator.PhysicalBlock * BlockSize;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
public override bool CanRead => true;
|
||||
public override bool CanSeek => true;
|
||||
public override bool CanWrite => false;
|
||||
public override bool CanWrite => true;
|
||||
public override long Length { get; }
|
||||
|
||||
public override long Position
|
||||
|
@ -95,5 +125,11 @@ namespace LibHac.Savefile
|
|||
Data.Position = Iterator.PhysicalBlock * BlockSize + segmentPos;
|
||||
}
|
||||
}
|
||||
|
||||
protected override void Dispose(bool disposing)
|
||||
{
|
||||
Flush();
|
||||
base.Dispose(disposing);
|
||||
}
|
||||
}
|
||||
}
|
|
@ -2,7 +2,7 @@
|
|||
using System.Collections;
|
||||
using System.IO;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class DuplexBitmap
|
||||
{
|
|
@ -1,7 +1,7 @@
|
|||
using System;
|
||||
using System.IO;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class DuplexFs : Stream
|
||||
{
|
||||
|
@ -28,7 +28,9 @@ namespace LibHac.Savefile
|
|||
|
||||
public override void Flush()
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
BitmapStream?.Flush();
|
||||
DataA?.Flush();
|
||||
DataB?.Flush();
|
||||
}
|
||||
|
||||
public override int Read(byte[] buffer, int offset, int count)
|
||||
|
@ -57,6 +59,30 @@ namespace LibHac.Savefile
|
|||
return totalBytesRead;
|
||||
}
|
||||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
long remaining = Math.Min(count, Length - Position);
|
||||
if (remaining <= 0) return;
|
||||
|
||||
int inOffset = offset;
|
||||
|
||||
while (remaining > 0)
|
||||
{
|
||||
int blockNum = (int)(Position / BlockSize);
|
||||
int blockPos = (int)(Position % BlockSize);
|
||||
int bytesToWrite = (int)Math.Min(remaining, BlockSize - blockPos);
|
||||
|
||||
Stream data = Bitmap.Bitmap[blockNum] ? DataB : DataA;
|
||||
data.Position = blockNum * BlockSize + blockPos;
|
||||
|
||||
data.Write(buffer, inOffset, bytesToWrite);
|
||||
|
||||
inOffset += bytesToWrite;
|
||||
remaining -= bytesToWrite;
|
||||
Position += bytesToWrite;
|
||||
}
|
||||
}
|
||||
|
||||
public override long Seek(long offset, SeekOrigin origin)
|
||||
{
|
||||
switch (origin)
|
||||
|
@ -80,11 +106,6 @@ namespace LibHac.Savefile
|
|||
throw new NotImplementedException();
|
||||
}
|
||||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
}
|
||||
|
||||
public override bool CanRead => true;
|
||||
public override bool CanSeek => true;
|
||||
public override bool CanWrite => false;
|
|
@ -3,7 +3,7 @@ using System.Diagnostics;
|
|||
using System.IO;
|
||||
using System.Text;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
namespace LibHac.Save
|
||||
{
|
||||
[DebuggerDisplay("{" + nameof(FullPath) + "}")]
|
||||
public abstract class FsEntry
|
|
@ -1,7 +1,8 @@
|
|||
using System;
|
||||
using System.IO;
|
||||
using LibHac.Streams;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class Header
|
||||
{
|
||||
|
@ -24,13 +25,17 @@ namespace LibHac.Savefile
|
|||
public byte[] DuplexMasterA { get; }
|
||||
public byte[] DuplexMasterB { get; }
|
||||
|
||||
public Stream MasterHash { get; }
|
||||
|
||||
public Validity SignatureValidity { get; }
|
||||
public Validity HeaderHashValidity { get; }
|
||||
|
||||
public byte[] Data { get; }
|
||||
|
||||
public Header(Keyset keyset, BinaryReader reader)
|
||||
public Header(Keyset keyset, SharedStreamSource streamSource)
|
||||
{
|
||||
var reader = new BinaryReader(streamSource.CreateStream());
|
||||
|
||||
reader.BaseStream.Position = 0;
|
||||
Data = reader.ReadBytes(0x4000);
|
||||
reader.BaseStream.Position = 0;
|
||||
|
@ -65,6 +70,8 @@ namespace LibHac.Savefile
|
|||
reader.BaseStream.Position = Layout.IvfcMasterHashOffsetB;
|
||||
MasterHashB = reader.ReadBytes((int)Layout.IvfcMasterHashSize);
|
||||
|
||||
MasterHash = streamSource.CreateStream(Layout.IvfcMasterHashOffsetA, Layout.IvfcMasterHashSize);
|
||||
|
||||
reader.BaseStream.Position = Layout.DuplexMasterOffsetA;
|
||||
DuplexMasterA = reader.ReadBytes((int)Layout.DuplexMasterSize);
|
||||
reader.BaseStream.Position = Layout.DuplexMasterOffsetB;
|
||||
|
@ -192,11 +199,11 @@ namespace LibHac.Savefile
|
|||
|
||||
public class RemapHeader
|
||||
{
|
||||
public string Magic { get; set; }
|
||||
public uint MagicNum { get; set; }
|
||||
public int MapEntryCount { get; set; }
|
||||
public int MapSegmentCount { get; set; }
|
||||
public int Field10 { get; set; }
|
||||
public string Magic { get; }
|
||||
public uint MagicNum { get; }
|
||||
public int MapEntryCount { get; }
|
||||
public int MapSegmentCount { get; }
|
||||
public int SegmentBits { get; }
|
||||
|
||||
public RemapHeader(BinaryReader reader)
|
||||
{
|
||||
|
@ -204,7 +211,7 @@ namespace LibHac.Savefile
|
|||
MagicNum = reader.ReadUInt32();
|
||||
MapEntryCount = reader.ReadInt32();
|
||||
MapSegmentCount = reader.ReadInt32();
|
||||
Field10 = reader.ReadInt32();
|
||||
SegmentBits = reader.ReadInt32();
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -320,7 +327,8 @@ namespace LibHac.Savefile
|
|||
public long PhysicalOffset { get; }
|
||||
public long Size { get; }
|
||||
public int Alignment { get; }
|
||||
public int StorageType { get; }
|
||||
public int Field1C { get; }
|
||||
|
||||
public long VirtualOffsetEnd => VirtualOffset + Size;
|
||||
public long PhysicalOffsetEnd => PhysicalOffset + Size;
|
||||
internal RemapSegment Segment { get; set; }
|
||||
|
@ -332,7 +340,7 @@ namespace LibHac.Savefile
|
|||
PhysicalOffset = reader.ReadInt64();
|
||||
Size = reader.ReadInt64();
|
||||
Alignment = reader.ReadInt32();
|
||||
StorageType = reader.ReadInt32();
|
||||
Field1C = reader.ReadInt32();
|
||||
}
|
||||
}
|
||||
|
|
@ -1,7 +1,7 @@
|
|||
using System;
|
||||
using System.IO;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class JournalStream : Stream
|
||||
{
|
||||
|
@ -45,6 +45,25 @@ namespace LibHac.Savefile
|
|||
return count;
|
||||
}
|
||||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
long remaining = Math.Min(Length - Position, count);
|
||||
if (remaining <= 0) return;
|
||||
|
||||
int inPos = offset;
|
||||
|
||||
while (remaining > 0)
|
||||
{
|
||||
long remainInEntry = BlockSize - Position % BlockSize;
|
||||
int toRead = (int)Math.Min(remaining, remainInEntry);
|
||||
BaseStream.Write(buffer, inPos, toRead);
|
||||
|
||||
inPos += toRead;
|
||||
remaining -= toRead;
|
||||
Position += toRead;
|
||||
}
|
||||
}
|
||||
|
||||
public override long Seek(long offset, SeekOrigin origin)
|
||||
{
|
||||
switch (origin)
|
||||
|
@ -64,11 +83,15 @@ namespace LibHac.Savefile
|
|||
}
|
||||
|
||||
public override void SetLength(long value) => throw new NotSupportedException();
|
||||
public override void Write(byte[] buffer, int offset, int count) => throw new NotSupportedException();
|
||||
public override void Flush() => throw new NotSupportedException();
|
||||
|
||||
public override void Flush()
|
||||
{
|
||||
BaseStream.Flush();
|
||||
}
|
||||
|
||||
public override bool CanRead => true;
|
||||
public override bool CanSeek => true;
|
||||
public override bool CanWrite => false;
|
||||
public override bool CanWrite => true;
|
||||
public override long Length { get; }
|
||||
public override long Position
|
||||
{
|
|
@ -1,6 +1,6 @@
|
|||
using System.IO;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class LayeredDuplexFs : Stream
|
||||
{
|
||||
|
@ -32,7 +32,7 @@ namespace LibHac.Savefile
|
|||
|
||||
public override void Flush()
|
||||
{
|
||||
throw new System.NotImplementedException();
|
||||
DataLayer.Flush();
|
||||
}
|
||||
|
||||
public override int Read(byte[] buffer, int offset, int count)
|
||||
|
@ -52,7 +52,7 @@ namespace LibHac.Savefile
|
|||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
throw new System.NotImplementedException();
|
||||
DataLayer.Write(buffer, offset, count);
|
||||
}
|
||||
|
||||
public override bool CanRead => DataLayer.CanRead;
|
116
LibHac/Save/RemapStorage.cs
Normal file
116
LibHac/Save/RemapStorage.cs
Normal file
|
@ -0,0 +1,116 @@
|
|||
using System;
|
||||
using System.IO;
|
||||
using LibHac.Streams;
|
||||
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class RemapStorage
|
||||
{
|
||||
private SharedStreamSource StreamSource { get; }
|
||||
private RemapHeader Header { get; }
|
||||
public MapEntry[] MapEntries { get; set; }
|
||||
public RemapSegment[] Segments { get; set; }
|
||||
|
||||
/// <summary>
|
||||
/// Creates a new <see cref="RemapStorage"/>
|
||||
/// </summary>
|
||||
/// <param name="data">A <see cref="Stream"/> of the main data of the RemapStream.
|
||||
/// The <see cref="RemapStorage"/> object takes complete ownership of the Stream.</param>
|
||||
/// <param name="header">The header for this RemapStorage.</param>
|
||||
/// <param name="mapEntries">The remapping entries for this RemapStorage.</param>
|
||||
public RemapStorage(Stream data, RemapHeader header, MapEntry[] mapEntries)
|
||||
{
|
||||
StreamSource = new SharedStreamSource(data);
|
||||
Header = header;
|
||||
MapEntries = mapEntries;
|
||||
|
||||
Segments = InitSegments(Header, MapEntries);
|
||||
}
|
||||
|
||||
public Stream OpenStream(long offset, long size)
|
||||
{
|
||||
int segmentIdx = GetSegmentFromVirtualOffset(offset);
|
||||
long segmentOffset = GetOffsetFromVirtualOffset(offset);
|
||||
|
||||
if (segmentIdx > Segments.Length)
|
||||
{
|
||||
throw new ArgumentOutOfRangeException(nameof(offset));
|
||||
}
|
||||
|
||||
RemapSegment segment = Segments[GetSegmentFromVirtualOffset(offset)];
|
||||
|
||||
if (segmentOffset > segment.Length)
|
||||
{
|
||||
throw new ArgumentOutOfRangeException(nameof(offset));
|
||||
}
|
||||
|
||||
Stream stream = new RemapStream(StreamSource.CreateStream(), segment);
|
||||
|
||||
return new SubStream(stream, offset, size);
|
||||
}
|
||||
|
||||
public Stream OpenSegmentStream(int segment)
|
||||
{
|
||||
long offset = ToVirtualOffset(segment, 0);
|
||||
long size = Segments[segment].Length;
|
||||
|
||||
return OpenStream(offset, size);
|
||||
}
|
||||
|
||||
private static RemapSegment[] InitSegments(RemapHeader header, MapEntry[] mapEntries)
|
||||
{
|
||||
var segments = new RemapSegment[header.MapSegmentCount];
|
||||
int entryIdx = 0;
|
||||
|
||||
for (int i = 0; i < header.MapSegmentCount; i++)
|
||||
{
|
||||
var seg = new RemapSegment();
|
||||
seg.Entries.Add(mapEntries[entryIdx]);
|
||||
seg.Offset = mapEntries[entryIdx].VirtualOffset;
|
||||
mapEntries[entryIdx].Segment = seg;
|
||||
entryIdx++;
|
||||
|
||||
while (entryIdx < mapEntries.Length &&
|
||||
mapEntries[entryIdx - 1].VirtualOffsetEnd == mapEntries[entryIdx].VirtualOffset)
|
||||
{
|
||||
mapEntries[entryIdx].Segment = seg;
|
||||
mapEntries[entryIdx - 1].Next = mapEntries[entryIdx];
|
||||
seg.Entries.Add(mapEntries[entryIdx]);
|
||||
entryIdx++;
|
||||
}
|
||||
|
||||
seg.Length = seg.Entries[seg.Entries.Count - 1].VirtualOffsetEnd - seg.Entries[0].VirtualOffset;
|
||||
segments[i] = seg;
|
||||
}
|
||||
|
||||
return segments;
|
||||
}
|
||||
|
||||
private int GetSegmentFromVirtualOffset(long virtualOffset)
|
||||
{
|
||||
return (int)((ulong)virtualOffset >> (64 - Header.SegmentBits));
|
||||
}
|
||||
|
||||
private long GetOffsetFromVirtualOffset(long virtualOffset)
|
||||
{
|
||||
return virtualOffset & GetOffsetMask();
|
||||
}
|
||||
|
||||
private long ToVirtualOffset(int segment, long offset)
|
||||
{
|
||||
long seg = (segment << (64 - Header.SegmentBits)) & GetSegmentMask();
|
||||
long off = offset & GetOffsetMask();
|
||||
return seg | off;
|
||||
}
|
||||
|
||||
private long GetOffsetMask()
|
||||
{
|
||||
return (1 << (64 - Header.SegmentBits)) - 1;
|
||||
}
|
||||
|
||||
private long GetSegmentMask()
|
||||
{
|
||||
return ~GetOffsetMask();
|
||||
}
|
||||
}
|
||||
}
|
|
@ -3,46 +3,21 @@ using System.Collections.Generic;
|
|||
using System.IO;
|
||||
using System.Linq;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class RemapStream : Stream
|
||||
{
|
||||
private long _position;
|
||||
private Stream BaseStream { get; }
|
||||
public MapEntry[] MapEntries { get; set; }
|
||||
public MapEntry CurrentEntry { get; set; }
|
||||
public RemapSegment[] Segments { get; set; }
|
||||
private RemapSegment Segment { get; }
|
||||
private MapEntry CurrentEntry { get; set; }
|
||||
|
||||
public RemapStream(Stream baseStream, MapEntry[] entries, int segmentCount)
|
||||
public RemapStream(Stream baseStream, RemapSegment segment)
|
||||
{
|
||||
BaseStream = baseStream;
|
||||
MapEntries = entries;
|
||||
Segments = new RemapSegment[segmentCount];
|
||||
|
||||
int entryIdx = 0;
|
||||
for (int i = 0; i < segmentCount; i++)
|
||||
{
|
||||
var seg = new RemapSegment();
|
||||
seg.Entries.Add(MapEntries[entryIdx]);
|
||||
seg.Offset = MapEntries[entryIdx].VirtualOffset;
|
||||
MapEntries[entryIdx].Segment = seg;
|
||||
entryIdx++;
|
||||
|
||||
while (entryIdx < MapEntries.Length &&
|
||||
MapEntries[entryIdx - 1].VirtualOffsetEnd == MapEntries[entryIdx].VirtualOffset)
|
||||
{
|
||||
MapEntries[entryIdx].Segment = seg;
|
||||
MapEntries[entryIdx - 1].Next = MapEntries[entryIdx];
|
||||
seg.Entries.Add(MapEntries[entryIdx]);
|
||||
entryIdx++;
|
||||
}
|
||||
|
||||
seg.Length = seg.Entries[seg.Entries.Count - 1].VirtualOffsetEnd - seg.Entries[0].VirtualOffset;
|
||||
Segments[i] = seg;
|
||||
}
|
||||
|
||||
CurrentEntry = GetMapEntry(0);
|
||||
UpdateBaseStreamPosition();
|
||||
Segment = segment;
|
||||
CurrentEntry = segment.Entries[0];
|
||||
Length = segment.Length;
|
||||
}
|
||||
|
||||
public override int Read(byte[] buffer, int offset, int count)
|
||||
|
@ -68,6 +43,27 @@ namespace LibHac.Savefile
|
|||
return count;
|
||||
}
|
||||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
if (CurrentEntry == null) throw new EndOfStreamException();
|
||||
|
||||
long remaining = Math.Min(CurrentEntry.VirtualOffsetEnd - Position, count);
|
||||
if (remaining <= 0) return;
|
||||
|
||||
int inPos = offset;
|
||||
|
||||
while (remaining > 0)
|
||||
{
|
||||
long remainInEntry = CurrentEntry.VirtualOffsetEnd - Position;
|
||||
int toWrite = (int)Math.Min(remaining, remainInEntry);
|
||||
BaseStream.Write(buffer, inPos, toWrite);
|
||||
|
||||
inPos += toWrite;
|
||||
remaining -= toWrite;
|
||||
Position += toWrite;
|
||||
}
|
||||
}
|
||||
|
||||
public override long Seek(long offset, SeekOrigin origin)
|
||||
{
|
||||
switch (origin)
|
||||
|
@ -91,20 +87,14 @@ namespace LibHac.Savefile
|
|||
throw new NotSupportedException();
|
||||
}
|
||||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
throw new NotSupportedException();
|
||||
}
|
||||
|
||||
public override void Flush()
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
BaseStream.Flush();
|
||||
}
|
||||
|
||||
private MapEntry GetMapEntry(long offset)
|
||||
{
|
||||
// todo: is O(n) search a possible performance issue?
|
||||
MapEntry entry = MapEntries.FirstOrDefault(x => offset >= x.VirtualOffset && offset < x.VirtualOffsetEnd);
|
||||
MapEntry entry = Segment.Entries.FirstOrDefault(x => offset >= x.VirtualOffset && offset < x.VirtualOffsetEnd);
|
||||
if (entry == null) throw new ArgumentOutOfRangeException(nameof(offset));
|
||||
return entry;
|
||||
}
|
||||
|
@ -120,7 +110,7 @@ namespace LibHac.Savefile
|
|||
public override bool CanRead => true;
|
||||
public override bool CanSeek => true;
|
||||
public override bool CanWrite => false;
|
||||
public override long Length { get; } = -1;
|
||||
public override long Length { get; }
|
||||
|
||||
public override long Position
|
||||
{
|
147
LibHac/Save/SaveFs.cs
Normal file
147
LibHac/Save/SaveFs.cs
Normal file
|
@ -0,0 +1,147 @@
|
|||
using System.Collections.Generic;
|
||||
using System.IO;
|
||||
using LibHac.Streams;
|
||||
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class SaveFs
|
||||
{
|
||||
private SharedStreamSource StreamSource { get; }
|
||||
private AllocationTable AllocationTable { get; }
|
||||
private SaveHeader Header { get; }
|
||||
|
||||
public DirectoryEntry RootDirectory { get; private set; }
|
||||
public FileEntry[] Files { get; private set; }
|
||||
public DirectoryEntry[] Directories { get; private set; }
|
||||
public Dictionary<string, FileEntry> FileDictionary { get; }
|
||||
|
||||
public SaveFs(Stream storage, Stream allocationTable, SaveHeader header)
|
||||
{
|
||||
StreamSource = new SharedStreamSource(storage);
|
||||
AllocationTable = new AllocationTable(allocationTable);
|
||||
Header = header;
|
||||
|
||||
ReadFileInfo();
|
||||
var dictionary = new Dictionary<string, FileEntry>();
|
||||
foreach (FileEntry entry in Files)
|
||||
{
|
||||
dictionary[entry.FullPath] = entry;
|
||||
}
|
||||
|
||||
FileDictionary = dictionary;
|
||||
}
|
||||
|
||||
public Stream OpenFile(string filename)
|
||||
{
|
||||
if (!FileDictionary.TryGetValue(filename, out FileEntry file))
|
||||
{
|
||||
throw new FileNotFoundException();
|
||||
}
|
||||
|
||||
return OpenFile(file);
|
||||
}
|
||||
|
||||
public Stream OpenFile(FileEntry file)
|
||||
{
|
||||
if (file.BlockIndex < 0)
|
||||
{
|
||||
return Stream.Null;
|
||||
}
|
||||
|
||||
return OpenFatBlock(file.BlockIndex, file.FileSize);
|
||||
}
|
||||
|
||||
public bool FileExists(string filename) => FileDictionary.ContainsKey(filename);
|
||||
|
||||
public Stream OpenRawSaveFs() => StreamSource.CreateStream();
|
||||
|
||||
private void ReadFileInfo()
|
||||
{
|
||||
// todo: Query the FAT for the file size when none is given
|
||||
AllocationTableStream dirTableStream = OpenFatBlock(Header.DirectoryTableBlock, 1000000);
|
||||
AllocationTableStream fileTableStream = OpenFatBlock(Header.FileTableBlock, 1000000);
|
||||
|
||||
DirectoryEntry[] dirEntries = ReadDirEntries(dirTableStream);
|
||||
FileEntry[] fileEntries = ReadFileEntries(fileTableStream);
|
||||
|
||||
foreach (DirectoryEntry dir in dirEntries)
|
||||
{
|
||||
if (dir.NextSiblingIndex != 0) dir.NextSibling = dirEntries[dir.NextSiblingIndex];
|
||||
if (dir.FirstChildIndex != 0) dir.FirstChild = dirEntries[dir.FirstChildIndex];
|
||||
if (dir.FirstFileIndex != 0) dir.FirstFile = fileEntries[dir.FirstFileIndex];
|
||||
if (dir.NextInChainIndex != 0) dir.NextInChain = dirEntries[dir.NextInChainIndex];
|
||||
if (dir.ParentDirIndex != 0 && dir.ParentDirIndex < dirEntries.Length)
|
||||
dir.ParentDir = dirEntries[dir.ParentDirIndex];
|
||||
}
|
||||
|
||||
foreach (FileEntry file in fileEntries)
|
||||
{
|
||||
if (file.NextSiblingIndex != 0) file.NextSibling = fileEntries[file.NextSiblingIndex];
|
||||
if (file.NextInChainIndex != 0) file.NextInChain = fileEntries[file.NextInChainIndex];
|
||||
if (file.ParentDirIndex != 0 && file.ParentDirIndex < dirEntries.Length)
|
||||
file.ParentDir = dirEntries[file.ParentDirIndex];
|
||||
}
|
||||
|
||||
RootDirectory = dirEntries[2];
|
||||
|
||||
FileEntry fileChain = fileEntries[1].NextInChain;
|
||||
var files = new List<FileEntry>();
|
||||
while (fileChain != null)
|
||||
{
|
||||
files.Add(fileChain);
|
||||
fileChain = fileChain.NextInChain;
|
||||
}
|
||||
|
||||
DirectoryEntry dirChain = dirEntries[1].NextInChain;
|
||||
var dirs = new List<DirectoryEntry>();
|
||||
while (dirChain != null)
|
||||
{
|
||||
dirs.Add(dirChain);
|
||||
dirChain = dirChain.NextInChain;
|
||||
}
|
||||
|
||||
Files = files.ToArray();
|
||||
Directories = dirs.ToArray();
|
||||
|
||||
FsEntry.ResolveFilenames(Files);
|
||||
FsEntry.ResolveFilenames(Directories);
|
||||
}
|
||||
|
||||
private FileEntry[] ReadFileEntries(Stream stream)
|
||||
{
|
||||
var reader = new BinaryReader(stream);
|
||||
int count = reader.ReadInt32();
|
||||
|
||||
reader.BaseStream.Position -= 4;
|
||||
|
||||
var entries = new FileEntry[count];
|
||||
for (int i = 0; i < count; i++)
|
||||
{
|
||||
entries[i] = new FileEntry(reader);
|
||||
}
|
||||
|
||||
return entries;
|
||||
}
|
||||
|
||||
private DirectoryEntry[] ReadDirEntries(Stream stream)
|
||||
{
|
||||
var reader = new BinaryReader(stream);
|
||||
int count = reader.ReadInt32();
|
||||
|
||||
reader.BaseStream.Position -= 4;
|
||||
|
||||
var entries = new DirectoryEntry[count];
|
||||
for (int i = 0; i < count; i++)
|
||||
{
|
||||
entries[i] = new DirectoryEntry(reader);
|
||||
}
|
||||
|
||||
return entries;
|
||||
}
|
||||
|
||||
private AllocationTableStream OpenFatBlock(int blockIndex, long size)
|
||||
{
|
||||
return new AllocationTableStream(StreamSource.CreateStream(), AllocationTable, (int)Header.BlockSize, blockIndex, size);
|
||||
}
|
||||
}
|
||||
}
|
200
LibHac/Save/Savefile.cs
Normal file
200
LibHac/Save/Savefile.cs
Normal file
|
@ -0,0 +1,200 @@
|
|||
using System.IO;
|
||||
using System.Security.Cryptography;
|
||||
using System.Text;
|
||||
using LibHac.Streams;
|
||||
|
||||
namespace LibHac.Save
|
||||
{
|
||||
public class Savefile
|
||||
{
|
||||
public Header Header { get; }
|
||||
public SharedStreamSource SavefileSource { get; }
|
||||
|
||||
public SharedStreamSource JournalStreamSource { get; }
|
||||
private HierarchicalIntegrityVerificationStream IvfcStream { get; }
|
||||
public SharedStreamSource IvfcStreamSource { get; }
|
||||
public SaveFs SaveFs { get; }
|
||||
|
||||
public RemapStorage DataRemapStorage { get; }
|
||||
public RemapStorage MetaRemapStorage { get; }
|
||||
|
||||
public LayeredDuplexFs DuplexData { get; }
|
||||
|
||||
public DirectoryEntry RootDirectory => SaveFs.RootDirectory;
|
||||
public FileEntry[] Files => SaveFs.Files;
|
||||
public DirectoryEntry[] Directories => SaveFs.Directories;
|
||||
|
||||
public Savefile(Keyset keyset, Stream file, IntegrityCheckLevel integrityCheckLevel)
|
||||
{
|
||||
SavefileSource = new SharedStreamSource(file);
|
||||
|
||||
Header = new Header(keyset, SavefileSource);
|
||||
FsLayout layout = Header.Layout;
|
||||
|
||||
DataRemapStorage = new RemapStorage(SavefileSource.CreateStream(layout.FileMapDataOffset, layout.FileMapDataSize),
|
||||
Header.FileRemap, Header.FileMapEntries);
|
||||
|
||||
DuplexData = InitDuplexStream(DataRemapStorage, Header);
|
||||
|
||||
MetaRemapStorage = new RemapStorage(DuplexData, Header.MetaRemap, Header.MetaMapEntries);
|
||||
|
||||
Stream journalTable = MetaRemapStorage.OpenStream(layout.JournalTableOffset, layout.JournalTableSize);
|
||||
|
||||
MappingEntry[] journalMap = JournalStream.ReadMappingEntries(journalTable, Header.Journal.MainDataBlockCount);
|
||||
|
||||
Stream journalData = DataRemapStorage.OpenStream(layout.JournalDataOffset,
|
||||
layout.JournalDataSizeB + layout.SizeReservedArea);
|
||||
var journalStream = new JournalStream(journalData, journalMap, (int)Header.Journal.BlockSize);
|
||||
JournalStreamSource = new SharedStreamSource(journalStream);
|
||||
|
||||
IvfcStream = InitIvfcStream(integrityCheckLevel);
|
||||
|
||||
SaveFs = new SaveFs(IvfcStream, MetaRemapStorage.OpenStream(layout.FatOffset, layout.FatSize), Header.Save);
|
||||
|
||||
IvfcStreamSource = new SharedStreamSource(IvfcStream);
|
||||
}
|
||||
|
||||
private static LayeredDuplexFs InitDuplexStream(RemapStorage baseStorage, Header header)
|
||||
{
|
||||
FsLayout layout = header.Layout;
|
||||
var duplexLayers = new DuplexFsLayerInfo[3];
|
||||
|
||||
duplexLayers[0] = new DuplexFsLayerInfo
|
||||
{
|
||||
DataA = new MemoryStream(header.DuplexMasterA),
|
||||
DataB = new MemoryStream(header.DuplexMasterB),
|
||||
Info = header.Duplex.Layers[0]
|
||||
};
|
||||
|
||||
duplexLayers[1] = new DuplexFsLayerInfo
|
||||
{
|
||||
DataA = baseStorage.OpenStream(layout.DuplexL1OffsetA, layout.DuplexL1Size),
|
||||
DataB = baseStorage.OpenStream(layout.DuplexL1OffsetB, layout.DuplexL1Size),
|
||||
Info = header.Duplex.Layers[1]
|
||||
};
|
||||
|
||||
duplexLayers[2] = new DuplexFsLayerInfo
|
||||
{
|
||||
DataA = baseStorage.OpenStream(layout.DuplexDataOffsetA, layout.DuplexDataSize),
|
||||
DataB = baseStorage.OpenStream(layout.DuplexDataOffsetB, layout.DuplexDataSize),
|
||||
Info = header.Duplex.Layers[2]
|
||||
};
|
||||
|
||||
return new LayeredDuplexFs(duplexLayers, layout.DuplexIndex == 1);
|
||||
}
|
||||
|
||||
private HierarchicalIntegrityVerificationStream InitIvfcStream(IntegrityCheckLevel integrityCheckLevel)
|
||||
{
|
||||
IvfcHeader ivfc = Header.Ivfc;
|
||||
|
||||
const int ivfcLevels = 5;
|
||||
var initInfo = new IntegrityVerificationInfo[ivfcLevels];
|
||||
|
||||
initInfo[0] = new IntegrityVerificationInfo
|
||||
{
|
||||
Data = Header.MasterHash,
|
||||
BlockSize = 0,
|
||||
Type = IntegrityStreamType.Save
|
||||
};
|
||||
|
||||
for (int i = 1; i < ivfcLevels; i++)
|
||||
{
|
||||
IvfcLevelHeader level = ivfc.LevelHeaders[i - 1];
|
||||
|
||||
Stream data = i == ivfcLevels - 1
|
||||
? JournalStreamSource.CreateStream()
|
||||
: MetaRemapStorage.OpenStream(level.LogicalOffset, level.HashDataSize);
|
||||
|
||||
initInfo[i] = new IntegrityVerificationInfo
|
||||
{
|
||||
Data = data,
|
||||
BlockSize = 1 << level.BlockSizePower,
|
||||
Salt = new HMACSHA256(Encoding.ASCII.GetBytes(SaltSources[i - 1])).ComputeHash(ivfc.SaltSource),
|
||||
Type = IntegrityStreamType.Save
|
||||
};
|
||||
}
|
||||
|
||||
return new HierarchicalIntegrityVerificationStream(initInfo, integrityCheckLevel);
|
||||
}
|
||||
|
||||
public Stream OpenFile(string filename)
|
||||
{
|
||||
return SaveFs.OpenFile(filename);
|
||||
}
|
||||
|
||||
public Stream OpenFile(FileEntry file)
|
||||
{
|
||||
return SaveFs.OpenFile(file);
|
||||
}
|
||||
|
||||
public bool FileExists(string filename) => SaveFs.FileExists(filename);
|
||||
|
||||
public bool CommitHeader(Keyset keyset)
|
||||
{
|
||||
SharedStream headerStream = SavefileSource.CreateStream();
|
||||
|
||||
var hashData = new byte[0x3d00];
|
||||
|
||||
headerStream.Position = 0x300;
|
||||
headerStream.Read(hashData, 0, hashData.Length);
|
||||
|
||||
byte[] hash = Crypto.ComputeSha256(hashData, 0, hashData.Length);
|
||||
headerStream.Position = 0x108;
|
||||
headerStream.Write(hash, 0, hash.Length);
|
||||
|
||||
if (keyset.SaveMacKey.IsEmpty()) return false;
|
||||
|
||||
var cmacData = new byte[0x200];
|
||||
var cmac = new byte[0x10];
|
||||
|
||||
headerStream.Position = 0x100;
|
||||
headerStream.Read(cmacData, 0, 0x200);
|
||||
|
||||
Crypto.CalculateAesCmac(keyset.SaveMacKey, cmacData, 0, cmac, 0, 0x200);
|
||||
|
||||
headerStream.Position = 0;
|
||||
headerStream.Write(cmac, 0, 0x10);
|
||||
headerStream.Flush();
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
public Validity Verify(IProgressReport logger = null)
|
||||
{
|
||||
Validity validity = IvfcStream.Validate(true, logger);
|
||||
IvfcStream.SetLevelValidities(Header.Ivfc);
|
||||
|
||||
return validity;
|
||||
}
|
||||
|
||||
private string[] SaltSources =
|
||||
{
|
||||
"HierarchicalIntegrityVerificationStorage::Master",
|
||||
"HierarchicalIntegrityVerificationStorage::L1",
|
||||
"HierarchicalIntegrityVerificationStorage::L2",
|
||||
"HierarchicalIntegrityVerificationStorage::L3",
|
||||
"HierarchicalIntegrityVerificationStorage::L4",
|
||||
"HierarchicalIntegrityVerificationStorage::L5"
|
||||
};
|
||||
}
|
||||
|
||||
public static class SavefileExtensions
|
||||
{
|
||||
public static void Extract(this Savefile save, string outDir, IProgressReport logger = null)
|
||||
{
|
||||
foreach (FileEntry file in save.Files)
|
||||
{
|
||||
Stream stream = save.OpenFile(file);
|
||||
string outName = outDir + file.FullPath;
|
||||
string dir = Path.GetDirectoryName(outName);
|
||||
if (!string.IsNullOrWhiteSpace(dir)) Directory.CreateDirectory(dir);
|
||||
|
||||
using (var outFile = new FileStream(outName, FileMode.Create, FileAccess.ReadWrite))
|
||||
{
|
||||
logger?.LogMessage(file.FullPath);
|
||||
stream.CopyStream(outFile, stream.Length, logger);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
|
@ -1,317 +0,0 @@
|
|||
using System.Collections.Generic;
|
||||
using System.IO;
|
||||
using System.Security.Cryptography;
|
||||
using System.Text;
|
||||
using LibHac.Streams;
|
||||
|
||||
namespace LibHac.Savefile
|
||||
{
|
||||
public class Savefile
|
||||
{
|
||||
public Header Header { get; }
|
||||
private RemapStream FileRemap { get; }
|
||||
public SharedStreamSource SavefileSource { get; }
|
||||
public SharedStreamSource FileRemapSource { get; }
|
||||
private RemapStream MetaRemap { get; }
|
||||
public SharedStreamSource MetaRemapSource { get; }
|
||||
private JournalStream JournalStream { get; }
|
||||
public SharedStreamSource JournalStreamSource { get; }
|
||||
private HierarchicalIntegrityVerificationStream IvfcStream { get; }
|
||||
public SharedStreamSource IvfcStreamSource { get; }
|
||||
private AllocationTable AllocationTable { get; }
|
||||
|
||||
public Stream DuplexL1A { get; }
|
||||
public Stream DuplexL1B { get; }
|
||||
public Stream DuplexDataA { get; }
|
||||
public Stream DuplexDataB { get; }
|
||||
public LayeredDuplexFs DuplexData { get; }
|
||||
public Stream JournalData { get; }
|
||||
|
||||
public Stream JournalTable { get; }
|
||||
public Stream JournalBitmapUpdatedPhysical { get; }
|
||||
public Stream JournalBitmapUpdatedVirtual { get; }
|
||||
public Stream JournalBitmapUnassigned { get; }
|
||||
public Stream JournalLayer1Hash { get; }
|
||||
public Stream JournalLayer2Hash { get; }
|
||||
public Stream JournalLayer3Hash { get; }
|
||||
public Stream JournalFat { get; }
|
||||
|
||||
public DirectoryEntry RootDirectory { get; private set; }
|
||||
public FileEntry[] Files { get; private set; }
|
||||
public DirectoryEntry[] Directories { get; private set; }
|
||||
private Dictionary<string, FileEntry> FileDict { get; }
|
||||
|
||||
public Savefile(Keyset keyset, Stream file, IntegrityCheckLevel integrityCheckLevel)
|
||||
{
|
||||
SavefileSource = new SharedStreamSource(file);
|
||||
|
||||
using (var reader = new BinaryReader(SavefileSource.CreateStream(), Encoding.Default, true))
|
||||
{
|
||||
Header = new Header(keyset, reader);
|
||||
FsLayout layout = Header.Layout;
|
||||
|
||||
FileRemap = new RemapStream(
|
||||
SavefileSource.CreateStream(layout.FileMapDataOffset, layout.FileMapDataSize),
|
||||
Header.FileMapEntries, Header.FileRemap.MapSegmentCount);
|
||||
|
||||
FileRemapSource = new SharedStreamSource(FileRemap);
|
||||
|
||||
var duplexLayers = new DuplexFsLayerInfo[3];
|
||||
|
||||
duplexLayers[0] = new DuplexFsLayerInfo
|
||||
{
|
||||
DataA = new MemoryStream(Header.DuplexMasterA),
|
||||
DataB = new MemoryStream(Header.DuplexMasterB),
|
||||
Info = Header.Duplex.Layers[0]
|
||||
};
|
||||
|
||||
duplexLayers[1] = new DuplexFsLayerInfo
|
||||
{
|
||||
DataA = FileRemapSource.CreateStream(layout.DuplexL1OffsetA, layout.DuplexL1Size),
|
||||
DataB = FileRemapSource.CreateStream(layout.DuplexL1OffsetB, layout.DuplexL1Size),
|
||||
Info = Header.Duplex.Layers[1]
|
||||
};
|
||||
|
||||
duplexLayers[2] = new DuplexFsLayerInfo
|
||||
{
|
||||
DataA = FileRemapSource.CreateStream(layout.DuplexDataOffsetA, layout.DuplexDataSize),
|
||||
DataB = FileRemapSource.CreateStream(layout.DuplexDataOffsetB, layout.DuplexDataSize),
|
||||
Info = Header.Duplex.Layers[2]
|
||||
};
|
||||
|
||||
DuplexL1A = FileRemapSource.CreateStream(layout.DuplexL1OffsetA, layout.DuplexL1Size);
|
||||
DuplexL1B = FileRemapSource.CreateStream(layout.DuplexL1OffsetB, layout.DuplexL1Size);
|
||||
DuplexDataA = FileRemapSource.CreateStream(layout.DuplexDataOffsetA, layout.DuplexDataSize);
|
||||
DuplexDataB = FileRemapSource.CreateStream(layout.DuplexDataOffsetB, layout.DuplexDataSize);
|
||||
JournalData = FileRemapSource.CreateStream(layout.JournalDataOffset, layout.JournalDataSizeB + layout.SizeReservedArea);
|
||||
|
||||
DuplexData = new LayeredDuplexFs(duplexLayers, Header.Layout.DuplexIndex == 1);
|
||||
MetaRemap = new RemapStream(DuplexData, Header.MetaMapEntries, Header.MetaRemap.MapSegmentCount);
|
||||
MetaRemapSource = new SharedStreamSource(MetaRemap);
|
||||
|
||||
JournalTable = MetaRemapSource.CreateStream(layout.JournalTableOffset, layout.JournalTableSize);
|
||||
JournalBitmapUpdatedPhysical = MetaRemapSource.CreateStream(layout.JournalBitmapUpdatedPhysicalOffset, layout.JournalBitmapUpdatedPhysicalSize);
|
||||
JournalBitmapUpdatedVirtual = MetaRemapSource.CreateStream(layout.JournalBitmapUpdatedVirtualOffset, layout.JournalBitmapUpdatedVirtualSize);
|
||||
JournalBitmapUnassigned = MetaRemapSource.CreateStream(layout.JournalBitmapUnassignedOffset, layout.JournalBitmapUnassignedSize);
|
||||
JournalLayer1Hash = MetaRemapSource.CreateStream(layout.IvfcL1Offset, layout.IvfcL1Size);
|
||||
JournalLayer2Hash = MetaRemapSource.CreateStream(layout.IvfcL2Offset, layout.IvfcL2Size);
|
||||
JournalLayer3Hash = MetaRemapSource.CreateStream(layout.IvfcL3Offset, layout.IvfcL3Size);
|
||||
JournalFat = MetaRemapSource.CreateStream(layout.FatOffset, layout.FatSize);
|
||||
AllocationTable = new AllocationTable(JournalFat);
|
||||
|
||||
MappingEntry[] journalMap = JournalStream.ReadMappingEntries(JournalTable, Header.Journal.MainDataBlockCount);
|
||||
|
||||
SharedStream journalData = FileRemapSource.CreateStream(layout.JournalDataOffset,
|
||||
layout.JournalDataSizeB + layout.SizeReservedArea);
|
||||
JournalStream = new JournalStream(journalData, journalMap, (int)Header.Journal.BlockSize);
|
||||
JournalStreamSource = new SharedStreamSource(JournalStream);
|
||||
|
||||
IvfcStream = InitIvfcStream(integrityCheckLevel);
|
||||
IvfcStreamSource = new SharedStreamSource(IvfcStream);
|
||||
|
||||
ReadFileInfo();
|
||||
var dictionary = new Dictionary<string, FileEntry>();
|
||||
foreach (FileEntry entry in Files)
|
||||
{
|
||||
dictionary[entry.FullPath] = entry;
|
||||
}
|
||||
|
||||
FileDict = dictionary;
|
||||
}
|
||||
}
|
||||
|
||||
private HierarchicalIntegrityVerificationStream InitIvfcStream(IntegrityCheckLevel integrityCheckLevel)
|
||||
{
|
||||
IvfcHeader ivfc = Header.Ivfc;
|
||||
|
||||
const int ivfcLevels = 5;
|
||||
var initInfo = new IntegrityVerificationInfo[ivfcLevels];
|
||||
|
||||
initInfo[0] = new IntegrityVerificationInfo
|
||||
{
|
||||
Data = new MemoryStream(Header.MasterHashA),
|
||||
BlockSize = 0,
|
||||
Type = IntegrityStreamType.Save
|
||||
};
|
||||
|
||||
for (int i = 1; i < ivfcLevels; i++)
|
||||
{
|
||||
IvfcLevelHeader level = ivfc.LevelHeaders[i - 1];
|
||||
|
||||
Stream data = i == ivfcLevels - 1
|
||||
? (Stream)JournalStream
|
||||
: MetaRemapSource.CreateStream(level.LogicalOffset, level.HashDataSize);
|
||||
|
||||
initInfo[i] = new IntegrityVerificationInfo
|
||||
{
|
||||
Data = data,
|
||||
BlockSize = 1 << level.BlockSizePower,
|
||||
Salt = new HMACSHA256(Encoding.ASCII.GetBytes(SaltSources[i - 1])).ComputeHash(ivfc.SaltSource),
|
||||
Type = IntegrityStreamType.Save
|
||||
};
|
||||
}
|
||||
|
||||
return new HierarchicalIntegrityVerificationStream(initInfo, integrityCheckLevel);
|
||||
}
|
||||
|
||||
public Stream OpenFile(string filename)
|
||||
{
|
||||
if (!FileDict.TryGetValue(filename, out FileEntry file))
|
||||
{
|
||||
throw new FileNotFoundException();
|
||||
}
|
||||
|
||||
return OpenFile(file);
|
||||
}
|
||||
|
||||
public Stream OpenFile(FileEntry file)
|
||||
{
|
||||
if (file.BlockIndex < 0)
|
||||
{
|
||||
return Stream.Null;
|
||||
}
|
||||
|
||||
return OpenFatBlock(file.BlockIndex, file.FileSize);
|
||||
}
|
||||
|
||||
private AllocationTableStream OpenFatBlock(int blockIndex, long size)
|
||||
{
|
||||
return new AllocationTableStream(IvfcStreamSource.CreateStream(), AllocationTable, (int)Header.Save.BlockSize, blockIndex, size);
|
||||
}
|
||||
|
||||
public bool FileExists(string filename) => FileDict.ContainsKey(filename);
|
||||
|
||||
private void ReadFileInfo()
|
||||
{
|
||||
// todo: Query the FAT for the file size when none is given
|
||||
AllocationTableStream dirTableStream = OpenFatBlock(Header.Save.DirectoryTableBlock, 1000000);
|
||||
AllocationTableStream fileTableStream = OpenFatBlock(Header.Save.FileTableBlock, 1000000);
|
||||
|
||||
DirectoryEntry[] dirEntries = ReadDirEntries(dirTableStream);
|
||||
FileEntry[] fileEntries = ReadFileEntries(fileTableStream);
|
||||
|
||||
foreach (DirectoryEntry dir in dirEntries)
|
||||
{
|
||||
if (dir.NextSiblingIndex != 0) dir.NextSibling = dirEntries[dir.NextSiblingIndex];
|
||||
if (dir.FirstChildIndex != 0) dir.FirstChild = dirEntries[dir.FirstChildIndex];
|
||||
if (dir.FirstFileIndex != 0) dir.FirstFile = fileEntries[dir.FirstFileIndex];
|
||||
if (dir.NextInChainIndex != 0) dir.NextInChain = dirEntries[dir.NextInChainIndex];
|
||||
if (dir.ParentDirIndex != 0 && dir.ParentDirIndex < dirEntries.Length)
|
||||
dir.ParentDir = dirEntries[dir.ParentDirIndex];
|
||||
}
|
||||
|
||||
foreach (FileEntry file in fileEntries)
|
||||
{
|
||||
if (file.NextSiblingIndex != 0) file.NextSibling = fileEntries[file.NextSiblingIndex];
|
||||
if (file.NextInChainIndex != 0) file.NextInChain = fileEntries[file.NextInChainIndex];
|
||||
if (file.ParentDirIndex != 0 && file.ParentDirIndex < dirEntries.Length)
|
||||
file.ParentDir = dirEntries[file.ParentDirIndex];
|
||||
}
|
||||
|
||||
RootDirectory = dirEntries[2];
|
||||
|
||||
FileEntry fileChain = fileEntries[1].NextInChain;
|
||||
var files = new List<FileEntry>();
|
||||
while (fileChain != null)
|
||||
{
|
||||
files.Add(fileChain);
|
||||
fileChain = fileChain.NextInChain;
|
||||
}
|
||||
|
||||
DirectoryEntry dirChain = dirEntries[1].NextInChain;
|
||||
var dirs = new List<DirectoryEntry>();
|
||||
while (dirChain != null)
|
||||
{
|
||||
dirs.Add(dirChain);
|
||||
dirChain = dirChain.NextInChain;
|
||||
}
|
||||
|
||||
Files = files.ToArray();
|
||||
Directories = dirs.ToArray();
|
||||
|
||||
FsEntry.ResolveFilenames(Files);
|
||||
FsEntry.ResolveFilenames(Directories);
|
||||
}
|
||||
|
||||
private FileEntry[] ReadFileEntries(Stream stream)
|
||||
{
|
||||
var reader = new BinaryReader(stream);
|
||||
int count = reader.ReadInt32();
|
||||
|
||||
reader.BaseStream.Position -= 4;
|
||||
|
||||
var entries = new FileEntry[count];
|
||||
for (int i = 0; i < count; i++)
|
||||
{
|
||||
entries[i] = new FileEntry(reader);
|
||||
}
|
||||
|
||||
return entries;
|
||||
}
|
||||
|
||||
private DirectoryEntry[] ReadDirEntries(Stream stream)
|
||||
{
|
||||
var reader = new BinaryReader(stream);
|
||||
int count = reader.ReadInt32();
|
||||
|
||||
reader.BaseStream.Position -= 4;
|
||||
|
||||
var entries = new DirectoryEntry[count];
|
||||
for (int i = 0; i < count; i++)
|
||||
{
|
||||
entries[i] = new DirectoryEntry(reader);
|
||||
}
|
||||
|
||||
return entries;
|
||||
}
|
||||
|
||||
public bool SignHeader(Keyset keyset)
|
||||
{
|
||||
if (keyset.SaveMacKey.IsEmpty()) return false;
|
||||
|
||||
var data = new byte[0x200];
|
||||
var cmac = new byte[0x10];
|
||||
|
||||
SharedStream headerStream = SavefileSource.CreateStream();
|
||||
headerStream.Position = 0x100;
|
||||
headerStream.Read(data, 0, 0x200);
|
||||
|
||||
Crypto.CalculateAesCmac(keyset.SaveMacKey, data, 0, cmac, 0, 0x200);
|
||||
|
||||
headerStream.Position = 0;
|
||||
headerStream.Write(cmac, 0, 0x10);
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
private string[] SaltSources =
|
||||
{
|
||||
"HierarchicalIntegrityVerificationStorage::Master",
|
||||
"HierarchicalIntegrityVerificationStorage::L1",
|
||||
"HierarchicalIntegrityVerificationStorage::L2",
|
||||
"HierarchicalIntegrityVerificationStorage::L3",
|
||||
"HierarchicalIntegrityVerificationStorage::L4",
|
||||
"HierarchicalIntegrityVerificationStorage::L5"
|
||||
};
|
||||
}
|
||||
|
||||
public static class SavefileExtensions
|
||||
{
|
||||
public static void Extract(this Savefile save, string outDir, IProgressReport logger = null)
|
||||
{
|
||||
foreach (FileEntry file in save.Files)
|
||||
{
|
||||
Stream stream = save.OpenFile(file);
|
||||
string outName = outDir + file.FullPath;
|
||||
string dir = Path.GetDirectoryName(outName);
|
||||
if (!string.IsNullOrWhiteSpace(dir)) Directory.CreateDirectory(dir);
|
||||
|
||||
using (var outFile = new FileStream(outName, FileMode.Create, FileAccess.ReadWrite))
|
||||
{
|
||||
logger?.LogMessage(file.FullPath);
|
||||
stream.CopyStream(outFile, stream.Length, logger);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
|
@ -44,8 +44,9 @@ namespace LibHac.Streams
|
|||
if (_readBytes == 0 || !_bufferDirty) return;
|
||||
|
||||
_baseStream.Position = _currentSector * _bufferSize;
|
||||
_baseStream.Write(_buffer, 0, _readBytes);
|
||||
_baseStream.Write(_buffer, 0, _bufferSize);
|
||||
|
||||
_bufferPos = 0;
|
||||
_readBytes = 0;
|
||||
_bufferDirty = false;
|
||||
}
|
||||
|
@ -53,6 +54,7 @@ namespace LibHac.Streams
|
|||
public override void Flush()
|
||||
{
|
||||
WriteSectorIfDirty();
|
||||
_baseStream.Flush();
|
||||
}
|
||||
|
||||
public override int Read(byte[] buffer, int offset, int count)
|
||||
|
|
|
@ -112,7 +112,9 @@ namespace LibHac.Streams
|
|||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
ValidateSize(count);
|
||||
_baseStream.Write(buffer, offset, count);
|
||||
int toWrite = (int)Math.Min(count, Length - Position);
|
||||
|
||||
_baseStream.Write(buffer, offset, toWrite);
|
||||
CurrentSector += count / SectorSize;
|
||||
}
|
||||
|
||||
|
|
|
@ -32,7 +32,13 @@ namespace LibHac.Streams
|
|||
return new SharedStream(this, offset, length);
|
||||
}
|
||||
|
||||
public void Flush() => BaseStream.Flush();
|
||||
public void Flush()
|
||||
{
|
||||
lock (Locker)
|
||||
{
|
||||
BaseStream.Flush();
|
||||
}
|
||||
}
|
||||
|
||||
public int Read(long readOffset, byte[] buffer, int bufferOffset, int count)
|
||||
{
|
||||
|
|
|
@ -21,7 +21,7 @@ namespace LibHac.Streams
|
|||
baseStream.Seek(offset, SeekOrigin.Begin);
|
||||
}
|
||||
|
||||
public SubStream(Stream baseStream, long offset)
|
||||
public SubStream(Stream baseStream, long offset)
|
||||
: this(baseStream, offset, baseStream.Length - offset) { }
|
||||
|
||||
public override int Read(byte[] buffer, int offset, int count)
|
||||
|
@ -32,6 +32,14 @@ namespace LibHac.Streams
|
|||
return BaseStream.Read(buffer, offset, count);
|
||||
}
|
||||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
long remaining = Math.Min(Length - Position, count);
|
||||
if (remaining <= 0) return;
|
||||
|
||||
BaseStream.Write(buffer, offset, (int)remaining);
|
||||
}
|
||||
|
||||
public override long Length { get; }
|
||||
public override bool CanRead => BaseStream.CanRead;
|
||||
public override bool CanWrite => BaseStream.CanWrite;
|
||||
|
@ -70,10 +78,5 @@ namespace LibHac.Streams
|
|||
public override void SetLength(long value) => throw new NotSupportedException();
|
||||
|
||||
public override void Flush() => BaseStream.Flush();
|
||||
|
||||
public override void Write(byte[] buffer, int offset, int count)
|
||||
{
|
||||
throw new NotImplementedException();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -4,6 +4,7 @@ using System.Diagnostics;
|
|||
using System.IO;
|
||||
using System.Linq;
|
||||
using System.Text;
|
||||
using LibHac.Save;
|
||||
using LibHac.Streams;
|
||||
|
||||
namespace LibHac
|
||||
|
@ -16,7 +17,7 @@ namespace LibHac
|
|||
public string SaveDir { get; }
|
||||
|
||||
public Dictionary<string, Nca> Ncas { get; } = new Dictionary<string, Nca>(StringComparer.OrdinalIgnoreCase);
|
||||
public Dictionary<string, Savefile.Savefile> Saves { get; } = new Dictionary<string, Savefile.Savefile>(StringComparer.OrdinalIgnoreCase);
|
||||
public Dictionary<string, Savefile> Saves { get; } = new Dictionary<string, Savefile>(StringComparer.OrdinalIgnoreCase);
|
||||
public Dictionary<ulong, Title> Titles { get; } = new Dictionary<ulong, Title>();
|
||||
public Dictionary<ulong, Application> Applications { get; } = new Dictionary<ulong, Application>();
|
||||
|
||||
|
@ -117,7 +118,7 @@ namespace LibHac
|
|||
|
||||
foreach (string file in files)
|
||||
{
|
||||
Savefile.Savefile save = null;
|
||||
Savefile save = null;
|
||||
string saveName = Path.GetFileNameWithoutExtension(file);
|
||||
|
||||
try
|
||||
|
@ -126,7 +127,7 @@ namespace LibHac
|
|||
|
||||
string sdPath = "/" + Util.GetRelativePath(file, SaveDir).Replace('\\', '/');
|
||||
var nax0 = new Nax0(Keyset, stream, sdPath, false);
|
||||
save = new Savefile.Savefile(Keyset, nax0.Stream, IntegrityCheckLevel.None);
|
||||
save = new Savefile(Keyset, nax0.Stream, IntegrityCheckLevel.None);
|
||||
}
|
||||
catch (Exception ex)
|
||||
{
|
||||
|
|
|
@ -4,7 +4,7 @@ using System.Collections.Generic;
|
|||
using System.IO;
|
||||
using LibHac;
|
||||
using LibHac.Nand;
|
||||
using LibHac.Savefile;
|
||||
using LibHac.Save;
|
||||
|
||||
namespace NandReader
|
||||
{
|
||||
|
|
|
@ -7,7 +7,7 @@ using GalaSoft.MvvmLight;
|
|||
using GalaSoft.MvvmLight.Command;
|
||||
using LibHac;
|
||||
using LibHac.Nand;
|
||||
using LibHac.Savefile;
|
||||
using LibHac.Save;
|
||||
using LibHac.Streams;
|
||||
|
||||
namespace NandReaderGui.ViewModel
|
||||
|
|
|
@ -44,6 +44,7 @@ namespace hactoolnet
|
|||
new CliOption("listapps", 0, (o, a) => o.ListApps = true),
|
||||
new CliOption("listtitles", 0, (o, a) => o.ListTitles = true),
|
||||
new CliOption("listromfs", 0, (o, a) => o.ListRomFs = true),
|
||||
new CliOption("listfiles", 0, (o, a) => o.ListFiles = true),
|
||||
new CliOption("sign", 0, (o, a) => o.SignSave = true),
|
||||
new CliOption("title", 1, (o, a) => o.TitleId = ParseTitleId(a[0])),
|
||||
};
|
||||
|
@ -203,9 +204,10 @@ namespace hactoolnet
|
|||
sb.AppendLine(" --outdir <dir> Specify directory path to save contents to.");
|
||||
sb.AppendLine(" --debugoutdir <dir> Specify directory path to save intermediate data to for debugging.");
|
||||
sb.AppendLine(" --sign Sign the save file. (Requires device_key in key file)");
|
||||
sb.AppendLine(" --listfiles List files in save file.");
|
||||
sb.AppendLine("Keygen options:");
|
||||
sb.AppendLine(" --outdir <dir> Specify directory path to save key files to.");
|
||||
|
||||
|
||||
return sb.ToString();
|
||||
}
|
||||
|
||||
|
|
|
@ -34,11 +34,19 @@ namespace hactoolnet
|
|||
public bool ListApps;
|
||||
public bool ListTitles;
|
||||
public bool ListRomFs;
|
||||
public bool ListFiles;
|
||||
public bool SignSave;
|
||||
public ulong TitleId;
|
||||
|
||||
public IntegrityCheckLevel IntegrityLevel =>
|
||||
EnableHash ? IntegrityCheckLevel.ErrorOnInvalid : IntegrityCheckLevel.None;
|
||||
public IntegrityCheckLevel IntegrityLevel
|
||||
{
|
||||
get
|
||||
{
|
||||
if (Validate) return IntegrityCheckLevel.IgnoreOnInvalid;
|
||||
if (EnableHash) return IntegrityCheckLevel.ErrorOnInvalid;
|
||||
return IntegrityCheckLevel.None;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
internal enum FileType
|
||||
|
|
|
@ -1,4 +1,5 @@
|
|||
using System.Text;
|
||||
using System;
|
||||
using System.Text;
|
||||
using LibHac;
|
||||
|
||||
namespace hactoolnet
|
||||
|
@ -26,5 +27,42 @@ namespace hactoolnet
|
|||
default: return string.Empty;
|
||||
}
|
||||
}
|
||||
|
||||
public static void PrintIvfcHash(StringBuilder sb, int colLen, int indentSize, IvfcHeader ivfcInfo, IntegrityStreamType type)
|
||||
{
|
||||
string prefix = new string(' ', indentSize);
|
||||
string prefix2 = new string(' ', indentSize + 4);
|
||||
|
||||
if (type == IntegrityStreamType.RomFs)
|
||||
PrintItem(sb, colLen, $"{prefix}Master Hash{ivfcInfo.LevelHeaders[0].HashValidity.GetValidityString()}:", ivfcInfo.MasterHash);
|
||||
|
||||
PrintItem(sb, colLen, $"{prefix}Magic:", ivfcInfo.Magic);
|
||||
PrintItem(sb, colLen, $"{prefix}Version:", ivfcInfo.Version);
|
||||
|
||||
if (type == IntegrityStreamType.Save)
|
||||
PrintItem(sb, colLen, $"{prefix}Salt Seed:", ivfcInfo.SaltSource);
|
||||
|
||||
int levelCount = Math.Max(ivfcInfo.NumLevels - 1, 0);
|
||||
if (type == IntegrityStreamType.Save) levelCount = 4;
|
||||
|
||||
int offsetLen = type == IntegrityStreamType.Save ? 16 : 12;
|
||||
|
||||
for (int i = 0; i < levelCount; i++)
|
||||
{
|
||||
IvfcLevelHeader level = ivfcInfo.LevelHeaders[i];
|
||||
long hashOffset = 0;
|
||||
|
||||
if (i != 0)
|
||||
{
|
||||
hashOffset = ivfcInfo.LevelHeaders[i - 1].LogicalOffset;
|
||||
}
|
||||
|
||||
sb.AppendLine($"{prefix}Level {i}{level.HashValidity.GetValidityString()}:");
|
||||
PrintItem(sb, colLen, $"{prefix2}Data Offset:", $"0x{level.LogicalOffset.ToString($"x{offsetLen}")}");
|
||||
PrintItem(sb, colLen, $"{prefix2}Data Size:", $"0x{level.HashDataSize.ToString($"x{offsetLen}")}");
|
||||
PrintItem(sb, colLen, $"{prefix2}Hash Offset:", $"0x{hashOffset.ToString($"x{offsetLen}")}");
|
||||
PrintItem(sb, colLen, $"{prefix2}Hash BlockSize:", $"0x{1 << level.BlockSizePower:x8}");
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -174,7 +174,7 @@ namespace hactoolnet
|
|||
PrintSha256Hash(sect);
|
||||
break;
|
||||
case NcaHashType.Ivfc:
|
||||
PrintIvfcHash(sect);
|
||||
PrintIvfcHash(sb, colLen, 8, sect.Header.IvfcInfo, IntegrityStreamType.RomFs);
|
||||
break;
|
||||
default:
|
||||
sb.AppendLine(" Unknown/invalid superblock!");
|
||||
|
@ -196,32 +196,6 @@ namespace hactoolnet
|
|||
PrintItem(sb, colLen, " PFS0 Offset:", $"0x{hashInfo.DataOffset:x12}");
|
||||
PrintItem(sb, colLen, " PFS0 Size:", $"0x{hashInfo.DataSize:x12}");
|
||||
}
|
||||
|
||||
void PrintIvfcHash(NcaSection sect)
|
||||
{
|
||||
IvfcHeader ivfcInfo = sect.Header.IvfcInfo;
|
||||
|
||||
PrintItem(sb, colLen, $" Master Hash{sect.MasterHashValidity.GetValidityString()}:", ivfcInfo.MasterHash);
|
||||
PrintItem(sb, colLen, " Magic:", ivfcInfo.Magic);
|
||||
PrintItem(sb, colLen, " Version:", $"{ivfcInfo.Version:x8}");
|
||||
|
||||
for (int i = 0; i < Romfs.IvfcMaxLevel; i++)
|
||||
{
|
||||
IvfcLevelHeader level = ivfcInfo.LevelHeaders[i];
|
||||
long hashOffset = 0;
|
||||
|
||||
if (i != 0)
|
||||
{
|
||||
hashOffset = ivfcInfo.LevelHeaders[i - 1].LogicalOffset;
|
||||
}
|
||||
|
||||
sb.AppendLine($" Level {i}{level.HashValidity.GetValidityString()}:");
|
||||
PrintItem(sb, colLen, " Data Offset:", $"0x{level.LogicalOffset:x12}");
|
||||
PrintItem(sb, colLen, " Data Size:", $"0x{level.HashDataSize:x12}");
|
||||
PrintItem(sb, colLen, " Hash Offset:", $"0x{hashOffset:x12}");
|
||||
PrintItem(sb, colLen, " Hash BlockSize:", $"0x{1 << level.BlockSizePower:x8}");
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,9 +1,8 @@
|
|||
using System;
|
||||
using System.IO;
|
||||
using System.Linq;
|
||||
using System.Text;
|
||||
using LibHac;
|
||||
using LibHac.Savefile;
|
||||
using LibHac.Save;
|
||||
using static hactoolnet.Print;
|
||||
|
||||
namespace hactoolnet
|
||||
|
@ -16,6 +15,11 @@ namespace hactoolnet
|
|||
{
|
||||
var save = new Savefile(ctx.Keyset, file, ctx.Options.IntegrityLevel);
|
||||
|
||||
if (ctx.Options.Validate)
|
||||
{
|
||||
save.Verify(ctx.Logger);
|
||||
}
|
||||
|
||||
if (ctx.Options.OutDir != null)
|
||||
{
|
||||
save.Extract(ctx.Options.OutDir, ctx.Logger);
|
||||
|
@ -26,32 +30,50 @@ namespace hactoolnet
|
|||
string dir = ctx.Options.DebugOutDir;
|
||||
Directory.CreateDirectory(dir);
|
||||
|
||||
FsLayout layout = save.Header.Layout;
|
||||
|
||||
File.WriteAllBytes(Path.Combine(dir, "L0_0_MasterHashA"), save.Header.MasterHashA);
|
||||
File.WriteAllBytes(Path.Combine(dir, "L0_1_MasterHashB"), save.Header.MasterHashB);
|
||||
File.WriteAllBytes(Path.Combine(dir, "L0_2_DuplexMasterA"), save.Header.DuplexMasterA);
|
||||
File.WriteAllBytes(Path.Combine(dir, "L0_3_DuplexMasterB"), save.Header.DuplexMasterB);
|
||||
save.DuplexL1A.WriteAllBytes(Path.Combine(dir, "L0_4_DuplexL1A"), ctx.Logger);
|
||||
save.DuplexL1B.WriteAllBytes(Path.Combine(dir, "L0_5_DuplexL1B"), ctx.Logger);
|
||||
save.DuplexDataA.WriteAllBytes(Path.Combine(dir, "L0_6_DuplexDataA"), ctx.Logger);
|
||||
save.DuplexDataB.WriteAllBytes(Path.Combine(dir, "L0_7_DuplexDataB"), ctx.Logger);
|
||||
save.JournalData.WriteAllBytes(Path.Combine(dir, "L0_9_JournalData"), ctx.Logger);
|
||||
|
||||
Stream duplexL1A = save.DataRemapStorage.OpenStream(layout.DuplexL1OffsetA, layout.DuplexL1Size);
|
||||
Stream duplexL1B = save.DataRemapStorage.OpenStream(layout.DuplexL1OffsetB, layout.DuplexL1Size);
|
||||
Stream duplexDataA = save.DataRemapStorage.OpenStream(layout.DuplexDataOffsetA, layout.DuplexDataSize);
|
||||
Stream duplexDataB = save.DataRemapStorage.OpenStream(layout.DuplexDataOffsetB, layout.DuplexDataSize);
|
||||
Stream journalData = save.DataRemapStorage.OpenStream(layout.JournalDataOffset, layout.JournalDataSizeB + layout.SizeReservedArea);
|
||||
|
||||
duplexL1A.WriteAllBytes(Path.Combine(dir, "L0_4_DuplexL1A"), ctx.Logger);
|
||||
duplexL1B.WriteAllBytes(Path.Combine(dir, "L0_5_DuplexL1B"), ctx.Logger);
|
||||
duplexDataA.WriteAllBytes(Path.Combine(dir, "L0_6_DuplexDataA"), ctx.Logger);
|
||||
duplexDataB.WriteAllBytes(Path.Combine(dir, "L0_7_DuplexDataB"), ctx.Logger);
|
||||
journalData.WriteAllBytes(Path.Combine(dir, "L0_9_JournalData"), ctx.Logger);
|
||||
save.DuplexData.WriteAllBytes(Path.Combine(dir, "L1_0_DuplexData"), ctx.Logger);
|
||||
save.JournalTable.WriteAllBytes(Path.Combine(dir, "L2_0_JournalTable"), ctx.Logger);
|
||||
save.JournalBitmapUpdatedPhysical.WriteAllBytes(Path.Combine(dir, "L2_1_JournalBitmapUpdatedPhysical"), ctx.Logger);
|
||||
save.JournalBitmapUpdatedVirtual.WriteAllBytes(Path.Combine(dir, "L2_2_JournalBitmapUpdatedVirtual"), ctx.Logger);
|
||||
save.JournalBitmapUnassigned.WriteAllBytes(Path.Combine(dir, "L2_3_JournalBitmapUnassigned"), ctx.Logger);
|
||||
save.JournalLayer1Hash.WriteAllBytes(Path.Combine(dir, "L2_4_Layer1Hash"), ctx.Logger);
|
||||
save.JournalLayer2Hash.WriteAllBytes(Path.Combine(dir, "L2_5_Layer2Hash"), ctx.Logger);
|
||||
save.JournalLayer3Hash.WriteAllBytes(Path.Combine(dir, "L2_6_Layer3Hash"), ctx.Logger);
|
||||
save.JournalFat.WriteAllBytes(Path.Combine(dir, "L2_7_FAT"), ctx.Logger);
|
||||
|
||||
Stream journalTable = save.MetaRemapStorage.OpenStream(layout.JournalTableOffset, layout.JournalTableSize);
|
||||
Stream journalBitmapUpdatedPhysical = save.MetaRemapStorage.OpenStream(layout.JournalBitmapUpdatedPhysicalOffset, layout.JournalBitmapUpdatedPhysicalSize);
|
||||
Stream journalBitmapUpdatedVirtual = save.MetaRemapStorage.OpenStream(layout.JournalBitmapUpdatedVirtualOffset, layout.JournalBitmapUpdatedVirtualSize);
|
||||
Stream journalBitmapUnassigned = save.MetaRemapStorage.OpenStream(layout.JournalBitmapUnassignedOffset, layout.JournalBitmapUnassignedSize);
|
||||
Stream journalLayer1Hash = save.MetaRemapStorage.OpenStream(layout.IvfcL1Offset, layout.IvfcL1Size);
|
||||
Stream journalLayer2Hash = save.MetaRemapStorage.OpenStream(layout.IvfcL2Offset, layout.IvfcL2Size);
|
||||
Stream journalLayer3Hash = save.MetaRemapStorage.OpenStream(layout.IvfcL3Offset, layout.IvfcL3Size);
|
||||
Stream journalFat = save.MetaRemapStorage.OpenStream(layout.FatOffset, layout.FatSize);
|
||||
|
||||
journalTable.WriteAllBytes(Path.Combine(dir, "L2_0_JournalTable"), ctx.Logger);
|
||||
journalBitmapUpdatedPhysical.WriteAllBytes(Path.Combine(dir, "L2_1_JournalBitmapUpdatedPhysical"), ctx.Logger);
|
||||
journalBitmapUpdatedVirtual.WriteAllBytes(Path.Combine(dir, "L2_2_JournalBitmapUpdatedVirtual"), ctx.Logger);
|
||||
journalBitmapUnassigned.WriteAllBytes(Path.Combine(dir, "L2_3_JournalBitmapUnassigned"), ctx.Logger);
|
||||
journalLayer1Hash.WriteAllBytes(Path.Combine(dir, "L2_4_Layer1Hash"), ctx.Logger);
|
||||
journalLayer2Hash.WriteAllBytes(Path.Combine(dir, "L2_5_Layer2Hash"), ctx.Logger);
|
||||
journalLayer3Hash.WriteAllBytes(Path.Combine(dir, "L2_6_Layer3Hash"), ctx.Logger);
|
||||
journalFat.WriteAllBytes(Path.Combine(dir, "L2_7_FAT"), ctx.Logger);
|
||||
|
||||
save.IvfcStreamSource.CreateStream().WriteAllBytes(Path.Combine(dir, "L3_0_SaveData"), ctx.Logger);
|
||||
}
|
||||
|
||||
if (ctx.Options.SignSave)
|
||||
{
|
||||
if (save.SignHeader(ctx.Keyset))
|
||||
if (save.CommitHeader(ctx.Keyset))
|
||||
{
|
||||
ctx.Logger.LogMessage("Successfully signed save file");
|
||||
}
|
||||
|
@ -61,6 +83,14 @@ namespace hactoolnet
|
|||
}
|
||||
}
|
||||
|
||||
if (ctx.Options.ListFiles)
|
||||
{
|
||||
foreach (FileEntry fileEntry in save.Files)
|
||||
{
|
||||
ctx.Logger.LogMessage(fileEntry.FullPath);
|
||||
}
|
||||
}
|
||||
|
||||
ctx.Logger.LogMessage(save.Print());
|
||||
}
|
||||
}
|
||||
|
@ -72,7 +102,7 @@ namespace hactoolnet
|
|||
sb.AppendLine();
|
||||
|
||||
sb.AppendLine("Savefile:");
|
||||
PrintItem(sb, colLen, "CMAC Signature:", save.Header.Cmac);
|
||||
PrintItem(sb, colLen, $"CMAC Signature{save.Header.SignatureValidity.GetValidityString()}:", save.Header.Cmac);
|
||||
PrintItem(sb, colLen, "Title ID:", $"{save.Header.ExtraData.TitleId:x16}");
|
||||
PrintItem(sb, colLen, "User ID:", save.Header.ExtraData.UserId);
|
||||
PrintItem(sb, colLen, "Save ID:", $"{save.Header.ExtraData.SaveId:x16}");
|
||||
|
@ -82,17 +112,9 @@ namespace hactoolnet
|
|||
PrintItem(sb, colLen, "Save Data Size:", $"0x{save.Header.ExtraData.DataSize:x16} ({Util.GetBytesReadable(save.Header.ExtraData.DataSize)})");
|
||||
PrintItem(sb, colLen, "Journal Size:", $"0x{save.Header.ExtraData.JournalSize:x16} ({Util.GetBytesReadable(save.Header.ExtraData.JournalSize)})");
|
||||
PrintItem(sb, colLen, $"Header Hash{save.Header.HeaderHashValidity.GetValidityString()}:", save.Header.Layout.Hash);
|
||||
PrintItem(sb, colLen, "IVFC Salt Seed:", save.Header.Ivfc.SaltSource);
|
||||
PrintItem(sb, colLen, "Number of Files:", save.Files.Length);
|
||||
|
||||
if (save.Files.Length > 0 && save.Files.Length < 100)
|
||||
{
|
||||
sb.AppendLine("Files:");
|
||||
foreach (FileEntry file in save.Files.OrderBy(x => x.FullPath))
|
||||
{
|
||||
sb.AppendLine(file.FullPath);
|
||||
}
|
||||
}
|
||||
PrintIvfcHash(sb, colLen, 4, save.Header.Ivfc, IntegrityStreamType.Save);
|
||||
|
||||
return sb.ToString();
|
||||
}
|
||||
|
|
|
@ -4,7 +4,7 @@ using System.IO;
|
|||
using System.Linq;
|
||||
using System.Text;
|
||||
using LibHac;
|
||||
using LibHac.Savefile;
|
||||
using LibHac.Save;
|
||||
|
||||
namespace hactoolnet
|
||||
{
|
||||
|
|
|
@ -7,7 +7,7 @@
|
|||
</PropertyGroup>
|
||||
|
||||
<PropertyGroup>
|
||||
<VersionPrefix>0.1.2</VersionPrefix>
|
||||
<VersionPrefix>0.1.3</VersionPrefix>
|
||||
<PathMap>$(MSBuildProjectDirectory)=C:/hactoolnet/</PathMap>
|
||||
</PropertyGroup>
|
||||
|
||||
|
|
Loading…
Reference in a new issue