mirror of
https://github.com/Kaisaan/lostkefin.git
synced 2025-06-18 08:35:38 -04:00
Add files
This commit is contained in:
parent
bc97822ef3
commit
b2dd9a4f3a
29
README.md
29
README.md
@ -1,2 +1,27 @@
|
||||
# lostkefin
|
||||
English translation for Ys V: Lost Kefin - Kingdom of Sand (PS2)
|
||||
# Ys V: Lost Kefin - Kingdom of Sand (PS2) English Translation
|
||||
Currently using Sam Farron's translation from his [Translation Series](https://www.youtube.com/watch?v=LfZZPwIdhzg&list=PLoD4gkRCJkUcgfpU5puBqYy5DX-RJK--b) with permission
|
||||
|
||||
# Building
|
||||
- Copy the contents of the game's .iso into the `extracted` and `translated` folders
|
||||
- Run `build.bat` or manually run the commands
|
||||
|
||||
# Hacking Notes
|
||||
- The game uses SHIFT-JIS encoding for the most part
|
||||
- The game's base pointer is $FFF80
|
||||
- Extracted .bin files with `_anm` in the filename are animation files with indexed 8BPP graphics
|
||||
- Extracted .HGB files are texture files with 32BPP RGBA graphics
|
||||
- Music files are `.hd` (header), `.bd` (header), and `.sq` (sequence) files
|
||||
- In `SLPM_663.60` the font is located at $1A3E90 as 4BPP graphics, its palette is stored at $25E4C0, and the fontmap is at $1A31F0
|
||||
|
||||
# Extracting the DATA.BIN Files
|
||||
`extract.py` extracts all the files from DATA.BIN and its folders into a `DATA` folder but does not extract the files into their correct folders yet, a `logfile.txt` is also created for fixing issues with the script
|
||||
|
||||
# To do
|
||||
- Create a proper table file (since the game does not use SHIFT-JIS encoding fully)
|
||||
- Updated extractiong script to extract `DATA0.BIN`, `DATA1.BIN`, and `SLPM_663.60`
|
||||
- Add more hacking notes (my notes.txt file is a mess so I haven't added it here)
|
||||
- Continue inserting the English script
|
||||
|
||||
# Credits
|
||||
- [Hilltop](https://x.com/HilltopWorks) - Providing valuable videos such as [hacking with Ghidra](https://youtu.be/qCEZC3cPc1s) and [PS1/PS2 graphics](https://youtu.be/lePKUCYakqM)
|
||||
- [Life Bottle Productions](https://www.lifebottle.org/#/) - Providing me with their [isotool.py script](https://github.com/lifebottle/PythonLib/blob/main/isotool.py) and their tutorial for [finding the base pointer](https://youtu.be/q5aEj-aSw50)
|
BIN
armips.exe
Normal file
BIN
armips.exe
Normal file
Binary file not shown.
5
build.bat
Normal file
5
build.bat
Normal file
@ -0,0 +1,5 @@
|
||||
armips patch.asm
|
||||
copy DATA.BIN translated/DATA.BIN
|
||||
copy "SLPM_663.60" "translated/SLPM_663.60"
|
||||
isotool.py -m insert --iso english.iso -o english.iso --filelist filelist.txt --files translated
|
||||
@pause
|
83
extract.py
Normal file
83
extract.py
Normal file
@ -0,0 +1,83 @@
|
||||
import os
|
||||
|
||||
slpm = open("extracted/SLPM_663.60", "rb")
|
||||
datafile = open('extracted/DATA.BIN', 'rb')
|
||||
|
||||
fileptrstart = 0x001829D0
|
||||
fileptrend = 0x001875E0
|
||||
|
||||
fileptr = fileptrstart
|
||||
|
||||
baseptr = 0xFFF80
|
||||
sector = 0x800
|
||||
|
||||
fileorder = 0
|
||||
filenamesize = 32
|
||||
terminator = "\x00"
|
||||
|
||||
logfile = open("logfile.txt", "w", encoding="shift-jis")
|
||||
logfile.write("\nDATA.BIN Files\n\n")
|
||||
|
||||
while fileptr < fileptrend:
|
||||
|
||||
slpm.seek(fileptr)
|
||||
filenameptr = slpm.read(4)
|
||||
filesize = slpm.read(4)
|
||||
filesectorstart = slpm.read(4)
|
||||
filesectorsize = slpm.read(4)
|
||||
|
||||
filenameptr = int.from_bytes(filenameptr, "little")
|
||||
filesize = int.from_bytes(filesize, "little")
|
||||
filesectorstart = int.from_bytes(filesectorstart, "little")
|
||||
filesectorsize = int.from_bytes(filesectorsize, "little")
|
||||
|
||||
filenameloc = filenameptr - baseptr
|
||||
slpm.seek(filenameloc)
|
||||
filename = slpm.read(filenamesize)
|
||||
filename = filename.decode(encoding="shift-jis", errors="backslashreplace")
|
||||
filename = filename[:filename.find(terminator)]
|
||||
|
||||
filesectorstart = filesectorstart * sector
|
||||
filesectorsize = filesectorstart * sector
|
||||
|
||||
datafile.seek(filesectorstart)
|
||||
filedata = datafile.read(filesize)
|
||||
|
||||
newfile = open(f"DATA/{filename}", "wb")
|
||||
newfile.write(filedata)
|
||||
newfile.close
|
||||
|
||||
logfile.write(f"Current Pointer: {fileptr:X}\tFile name: {fileorder:X} {filename}\tFilename location: {filenameloc:X}\tFilename pointer: {filenameptr:X}\tFile size: {filesize:X}\tFile data start in DATA.BIN: {filesectorstart:X}\tFile data sector size: {filesectorsize:X}\n")
|
||||
|
||||
fileptr += 16
|
||||
fileorder += 1
|
||||
|
||||
logfile.write("\nDATA.BIN Folders\n\n")
|
||||
|
||||
folderptrstart = 0x001875E0
|
||||
folderptrend = 0x00187A18
|
||||
|
||||
folderptr = folderptrstart
|
||||
|
||||
while folderptr < folderptrend:
|
||||
|
||||
slpm.seek(folderptr)
|
||||
foldernameptr = slpm.read(4)
|
||||
folderorderstart = slpm.read(4)
|
||||
foldercount = slpm.read(4)
|
||||
|
||||
foldernameptr = int.from_bytes(foldernameptr, "little")
|
||||
folderorderstart = int.from_bytes(folderorderstart, "little")
|
||||
foldercount = int.from_bytes(foldercount, "little")
|
||||
|
||||
foldernameloc = foldernameptr - baseptr
|
||||
slpm.seek(foldernameloc)
|
||||
foldername = slpm.read(filenamesize)
|
||||
foldername = foldername.decode(encoding="shift-jis", errors="backslashreplace")
|
||||
foldername = foldername[:foldername.find(terminator)]
|
||||
|
||||
os.makedirs(f"C:/Users/kaisaan/Downloads/ps2 stuff/lostkefin/DATA/{foldername}", exist_ok=True)
|
||||
|
||||
logfile.write(f"Current Pointer: {folderptr:X}\tFolder name: {foldername}\tFoldername location: {foldernameloc:X}\tFilename pointer: {foldernameptr:X}\tFolder Order start {folderorderstart}\tFolder file count: {foldercount}\n")
|
||||
|
||||
folderptr += 12
|
556
isotool.py
Normal file
556
isotool.py
Normal file
@ -0,0 +1,556 @@
|
||||
import io
|
||||
import struct
|
||||
import argparse
|
||||
from pathlib import Path
|
||||
from dataclasses import dataclass, field
|
||||
from typing import BinaryIO
|
||||
|
||||
SCRIPT_VERSION = "1.7"
|
||||
SECTOR_SIZE = 0x800
|
||||
READ_CHUNK = 0x50_0000 # 5MiB
|
||||
SYSTEM_AREA_SIZE = 0x10 * SECTOR_SIZE
|
||||
LAYER0_PVD_LOCATION = SYSTEM_AREA_SIZE
|
||||
MAX_VOLUME_SECTOR_COUNT = 0x1FD060
|
||||
MAX_VOLUME_SIZE = MAX_VOLUME_SECTOR_COUNT * SECTOR_SIZE
|
||||
VOLUME_ALIGN = 0x8000
|
||||
PAD_NONE = 0
|
||||
PAD_20MB = 1
|
||||
PAD_VOLUME = 2
|
||||
|
||||
def align(x: int, alg: int) -> int:
|
||||
return (x + (alg-1)) & ~(alg-1)
|
||||
|
||||
|
||||
@dataclass
|
||||
class FileListData:
|
||||
path: Path
|
||||
inode: int
|
||||
lba: int = 0
|
||||
size: int = 0
|
||||
|
||||
|
||||
@dataclass
|
||||
class FileListInfo:
|
||||
files: list[FileListData]
|
||||
total_inodes: int
|
||||
|
||||
@dataclass
|
||||
class IsoLayer:
|
||||
header: bytes = b""
|
||||
footer: bytes = b""
|
||||
offset: int = 0
|
||||
meta: FileListInfo = field(default_factory=lambda: FileListInfo([], 0))
|
||||
|
||||
@dataclass
|
||||
class Iso:
|
||||
has_second_layer: bool = False
|
||||
layers: list[IsoLayer] = field(default_factory=lambda: [IsoLayer(), IsoLayer()])
|
||||
|
||||
|
||||
def main():
|
||||
print(f"pyPS2 ISO Rebuilder v{SCRIPT_VERSION}")
|
||||
print("Original by Raynê Games")
|
||||
print()
|
||||
|
||||
args = get_arguments()
|
||||
|
||||
if args.mode == "extract":
|
||||
dump_iso(args.iso, args.filelist, args.files, args.dry)
|
||||
print("dumping finished")
|
||||
else:
|
||||
if args.with_padding:
|
||||
pad_mode = PAD_20MB
|
||||
elif args.max_size:
|
||||
pad_mode = PAD_VOLUME
|
||||
else:
|
||||
pad_mode = PAD_NONE
|
||||
|
||||
rebuild_iso(args.iso, args.filelist, args.files, args.output, pad_mode)
|
||||
print("rebuild finished")
|
||||
|
||||
|
||||
def get_arguments(argv=None):
|
||||
# Init argument parser
|
||||
parser = argparse.ArgumentParser()
|
||||
|
||||
parser.add_argument(
|
||||
"-m",
|
||||
"--mode",
|
||||
choices=["extract", "insert"],
|
||||
required=True,
|
||||
metavar="operation",
|
||||
help="Options: extract, insert",
|
||||
)
|
||||
|
||||
parser.add_argument(
|
||||
"--iso",
|
||||
required=True,
|
||||
type=Path,
|
||||
metavar="original_iso",
|
||||
help="input game iso file path",
|
||||
)
|
||||
|
||||
group = parser.add_mutually_exclusive_group()
|
||||
|
||||
group.add_argument(
|
||||
"--with-padding",
|
||||
required=False,
|
||||
action="store_true",
|
||||
help="flag to control outermost iso padding",
|
||||
)
|
||||
|
||||
group.add_argument(
|
||||
"--max-size",
|
||||
required=False,
|
||||
action="store_true",
|
||||
help="flag to make the biggest iso possible",
|
||||
)
|
||||
|
||||
parser.add_argument(
|
||||
"--dry",
|
||||
required=False,
|
||||
action="store_false",
|
||||
help="dry run, parses the iso without saving the files",
|
||||
)
|
||||
|
||||
parser.add_argument(
|
||||
"-o",
|
||||
"--output",
|
||||
required=False,
|
||||
type=Path,
|
||||
metavar="output_iso",
|
||||
help="resulting iso file name",
|
||||
)
|
||||
|
||||
parser.add_argument(
|
||||
"--filelist",
|
||||
required=False,
|
||||
type=Path,
|
||||
metavar="filelist_path",
|
||||
help="filelist.txt file path",
|
||||
)
|
||||
|
||||
parser.add_argument(
|
||||
"--files",
|
||||
required=False,
|
||||
type=Path,
|
||||
metavar="files_folder",
|
||||
help="path to folder with extracted iso files",
|
||||
)
|
||||
|
||||
args = parser.parse_args()
|
||||
curr_dir = Path("./").resolve()
|
||||
|
||||
args.iso = args.iso.resolve()
|
||||
if hasattr(args, "filelist") and not args.filelist:
|
||||
args.filelist = curr_dir / f"{args.iso.name.upper()}-FILELIST-LSN.TXT"
|
||||
|
||||
if hasattr(args, "files") and not args.files:
|
||||
args.files = curr_dir / f"@{args.iso.name.upper()}"
|
||||
|
||||
if hasattr(args, "output") and not args.output:
|
||||
args.output = curr_dir / f"NEW_{args.iso.name}"
|
||||
|
||||
return args
|
||||
|
||||
|
||||
def check_pvd(fp: BinaryIO, pvd_loc: int) -> bool:
|
||||
fp.seek(pvd_loc)
|
||||
vd_type, vd_id = struct.unpack("<B5s", fp.read(6))
|
||||
if vd_type == 1 and vd_id == b"CD001":
|
||||
return True
|
||||
else:
|
||||
return False
|
||||
|
||||
|
||||
def dump_dir_records(iso: BinaryIO, pvd_loc: int, pvd_off: int) -> FileListInfo:
|
||||
path_parts = []
|
||||
record_ends = []
|
||||
record_pos = []
|
||||
file_info = FileListInfo([], 0)
|
||||
|
||||
# get the root directory record off the PVD
|
||||
iso.seek(pvd_loc + 0x9E)
|
||||
dr_data_pos, dr_data_len = struct.unpack("<I4xI", iso.read(12))
|
||||
dr_data_pos *= SECTOR_SIZE
|
||||
dr_data_pos += pvd_off
|
||||
record_ends.append(dr_data_pos + dr_data_len)
|
||||
record_pos.append(0)
|
||||
|
||||
iso.seek(dr_data_pos)
|
||||
|
||||
# Traverse directory records recursively
|
||||
# Did I mention that I won't do function recursion?
|
||||
while True:
|
||||
# Have we reached the end of current dir record?
|
||||
if iso.tell() >= record_ends[-1]:
|
||||
if len(record_ends) == 1:
|
||||
# If it's the last one, we finished
|
||||
break
|
||||
else:
|
||||
# Otherwise keep reading the previous one
|
||||
record_ends.pop()
|
||||
path_parts.pop()
|
||||
iso.seek(record_pos.pop())
|
||||
continue
|
||||
|
||||
# Parse the record
|
||||
inode = iso.tell()
|
||||
|
||||
dr_len = struct.unpack("<B", iso.read(1))[0]
|
||||
if dr_len == 0:
|
||||
# Found a 0-sized entry but record is not over
|
||||
# align to boundary or bail out otherwise
|
||||
if inode < record_ends[-1]:
|
||||
iso.seek(align(iso.tell(), SECTOR_SIZE))
|
||||
continue
|
||||
else:
|
||||
print("ERROR: Invalid directory record found")
|
||||
exit(1)
|
||||
|
||||
dr_blob = iso.read(dr_len - 1)
|
||||
|
||||
(
|
||||
dr_ea_len,
|
||||
dr_data_pos,
|
||||
dr_data_len,
|
||||
dr_flags,
|
||||
dr_inter,
|
||||
dr_volume,
|
||||
dr_name_len,
|
||||
) = struct.unpack_from("<BI4xI4x7xBHH2xB", dr_blob)
|
||||
|
||||
assert dr_ea_len == 0, "ISOs with extra attributes are not supported!"
|
||||
assert dr_inter == 0, "Interleaved ISOs are not supported!"
|
||||
assert dr_volume == 1, "multi-volume ISOs are not supported!"
|
||||
assert (dr_flags & 0b1000000) == 0, "4GiB+ files are not supported!"
|
||||
|
||||
# the dir records always en on even addresses
|
||||
if (iso.tell() % 2) != 0:
|
||||
iso.read(1)
|
||||
|
||||
dr_data_pos *= 0x800
|
||||
dr_data_pos += pvd_off
|
||||
|
||||
dr_name = dr_blob[32 : 32 + dr_name_len]
|
||||
|
||||
# record with these names are the '.' and '..'
|
||||
# directories respectively, so skip them
|
||||
if dr_name == b"\x00" or dr_name == b"\x01":
|
||||
continue
|
||||
|
||||
dr_name = dr_name.decode()
|
||||
if dr_name.endswith(";1"):
|
||||
dr_name = dr_name[:-2]
|
||||
path_parts.append(dr_name)
|
||||
|
||||
file_info.total_inodes += 1
|
||||
|
||||
# is it a directory?
|
||||
if (dr_flags & 0b10) != 0:
|
||||
# Go to its directory record
|
||||
record_pos.append(iso.tell())
|
||||
record_ends.append(dr_data_pos + dr_data_len)
|
||||
iso.seek(dr_data_pos)
|
||||
continue
|
||||
else:
|
||||
# Otherwise dump the file
|
||||
fp = "/".join(path_parts)
|
||||
|
||||
file_info.files.append(
|
||||
FileListData(Path(fp), inode - pvd_off, dr_data_pos, dr_data_len)
|
||||
)
|
||||
path_parts.pop()
|
||||
|
||||
return file_info
|
||||
|
||||
|
||||
def save_iso_files(
|
||||
iso: BinaryIO, file_info: FileListInfo, base_folder: Path
|
||||
) -> None:
|
||||
for file in file_info.files:
|
||||
print(f"SAVING {file.path.as_posix()}")
|
||||
|
||||
final_path = base_folder / file.path
|
||||
final_path.parent.mkdir(exist_ok=True, parents=True)
|
||||
iso.seek(file.lba)
|
||||
|
||||
with open(final_path, "wb+") as f:
|
||||
for _ in range(file.size // READ_CHUNK):
|
||||
f.write(iso.read(READ_CHUNK))
|
||||
|
||||
if (file.size % READ_CHUNK) != 0:
|
||||
f.write(iso.read(file.size % READ_CHUNK))
|
||||
|
||||
|
||||
|
||||
def check_iso(iso: BinaryIO) -> tuple[bool, int, int]:
|
||||
# Sanity check
|
||||
assert check_pvd(iso, LAYER0_PVD_LOCATION), "No valid PVD found in Layer0!"
|
||||
|
||||
# Test dual-layer-dness
|
||||
has_second_layer = False
|
||||
|
||||
iso.seek(LAYER0_PVD_LOCATION + 0x50)
|
||||
pvd0_sector_count = struct.unpack("<I", iso.read(4))[0]
|
||||
iso.seek(0, io.SEEK_END)
|
||||
iso_sector_count = iso.tell() // SECTOR_SIZE
|
||||
pvd1_pos = pvd0_sector_count * SECTOR_SIZE
|
||||
|
||||
if iso_sector_count != pvd0_sector_count:
|
||||
# sector count of the PVD disagree with the file
|
||||
# check for another PVD at volume end
|
||||
has_second_layer = check_pvd(iso, pvd1_pos)
|
||||
|
||||
if has_second_layer:
|
||||
print("< Dual layer ISO Detected >")
|
||||
print()
|
||||
|
||||
else:
|
||||
print("WARNING: Iso data suggest this is a double layer image")
|
||||
print(
|
||||
" but no valid PVD was found for Layer1, iso might be corrupt"
|
||||
)
|
||||
print()
|
||||
|
||||
return has_second_layer, pvd1_pos, iso_sector_count * SECTOR_SIZE
|
||||
|
||||
|
||||
def dump_iso(iso_path: Path, filelist: Path, iso_files: Path, save_files: bool) -> None:
|
||||
if iso_path.exists() is False:
|
||||
print(f"Could not to find '{iso_path.name}'!")
|
||||
return
|
||||
|
||||
with open(iso_path, "rb") as iso:
|
||||
has_second_layer, pvd1_pos, _ = check_iso(iso)
|
||||
|
||||
layer0_data = dump_dir_records(iso, LAYER0_PVD_LOCATION, 0)
|
||||
layer0_data.files.sort(
|
||||
key=lambda x: x.lba
|
||||
) # The files are ordered based on their disc position
|
||||
|
||||
if has_second_layer:
|
||||
layer1_data = dump_dir_records(iso, pvd1_pos, pvd1_pos - SYSTEM_AREA_SIZE)
|
||||
layer1_data.files.sort(
|
||||
key=lambda x: x.lba
|
||||
) # The files are ordered based on their disc position
|
||||
else:
|
||||
layer1_data = FileListInfo([], 0)
|
||||
|
||||
if save_files:
|
||||
# save files (if requested)
|
||||
save_iso_files(iso, layer0_data, iso_files)
|
||||
|
||||
if has_second_layer:
|
||||
print("\n< SECOND LAYER >\n")
|
||||
|
||||
save_iso_files(iso, layer1_data, iso_files)
|
||||
|
||||
# Save filelist
|
||||
with open(filelist, "w", encoding="utf8") as f:
|
||||
if has_second_layer:
|
||||
f.write(f"//{len(layer0_data.files)}\n")
|
||||
|
||||
for d in layer0_data.files:
|
||||
f.write(f"|{d.inode}||{iso_files.name}/{d.path.as_posix()}|\n")
|
||||
f.write(f"//{layer0_data.total_inodes}")
|
||||
|
||||
if has_second_layer:
|
||||
f.write("\n")
|
||||
for d in layer1_data.files:
|
||||
f.write(f"|{d.inode}||{iso_files.name}/{d.path.as_posix()}|\n")
|
||||
f.write(f"//{layer1_data.total_inodes}")
|
||||
else:
|
||||
# if not then show found data
|
||||
for file in layer0_data.files:
|
||||
print(
|
||||
f"FOUND {file.path.as_posix()} at 0x{file.lba:08X} with size {file.size} bytes"
|
||||
)
|
||||
if has_second_layer:
|
||||
print("\n< SECOND LAYER >\n")
|
||||
for file in layer1_data.files:
|
||||
print(
|
||||
f"FOUND {file.path.as_posix()} at 0x{file.lba:08X} with size {file.size} bytes"
|
||||
)
|
||||
|
||||
|
||||
def parse_filelist(file_info: FileListInfo, lines: list[str]) -> None:
|
||||
for line in lines[:-1]:
|
||||
data = [x for x in line.split("|") if x]
|
||||
p = Path(data[1])
|
||||
file_info.files.append(FileListData(Path(*p.parts[1:]), int(data[0])))
|
||||
|
||||
if lines[-1].startswith("//") is False:
|
||||
print("Could not to find inode total!")
|
||||
return
|
||||
|
||||
file_info.total_inodes = int(lines[-1][2:])
|
||||
|
||||
|
||||
def consume_iso_header(iso: BinaryIO, pvd_off: int, inodes: int) -> int:
|
||||
iso.seek(pvd_off)
|
||||
i = 0
|
||||
data_start = -1
|
||||
for lba in range(7862):
|
||||
udf_check = struct.unpack("<269x18s1761x", iso.read(SECTOR_SIZE))[0]
|
||||
if udf_check == b"*UDF DVD CGMS Info":
|
||||
i += 1
|
||||
|
||||
if i == inodes + 1:
|
||||
data_start = (lba + 1) * SECTOR_SIZE
|
||||
break
|
||||
else:
|
||||
print(
|
||||
"ERROR: Couldn't get all the UDF file chunk, original tool would've looped here"
|
||||
)
|
||||
print("Closing instead...")
|
||||
exit(1)
|
||||
|
||||
return data_start
|
||||
|
||||
|
||||
def validate_rebuild(filelist: Path, iso_files: Path) -> bool:
|
||||
if filelist.exists() is False:
|
||||
print(f"Could not to find the '{filelist.name}' files log!")
|
||||
return False
|
||||
|
||||
if iso_files.exists() is False:
|
||||
print(f"Could not to find the '{iso_files.name}' files directory!")
|
||||
return False
|
||||
|
||||
if iso_files.is_dir() is False:
|
||||
print(f"'{iso_files.name}' is not a directory!")
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
|
||||
def write_new_pvd(iso: BinaryIO, iso_files: Path, add_padding: int, layer_info: IsoLayer, pvd_loc: int) -> int:
|
||||
iso.write(layer_info.header)
|
||||
|
||||
for inode in layer_info.meta.files:
|
||||
fp = iso_files / inode.path
|
||||
start_pos = iso.tell()
|
||||
if fp.exists() is False:
|
||||
print(f"File '{inode.path.as_posix()}' not found!")
|
||||
exit(1)
|
||||
|
||||
print(f"Inserting {str(inode.path)}...")
|
||||
|
||||
with open(fp, "rb") as g:
|
||||
while data := g.read(0x80000):
|
||||
iso.write(data)
|
||||
|
||||
end_pos = iso.tell()
|
||||
|
||||
# Align to next LBA
|
||||
al_end = align(end_pos, SECTOR_SIZE)
|
||||
iso.write(b"\x00" * (al_end - end_pos))
|
||||
|
||||
end_save = iso.tell()
|
||||
|
||||
new_lba = (start_pos - pvd_loc + SYSTEM_AREA_SIZE) // 0x800
|
||||
new_size = end_pos - start_pos
|
||||
iso.seek(inode.inode + pvd_loc - SYSTEM_AREA_SIZE + 2)
|
||||
|
||||
iso.write(struct.pack("<I", new_lba))
|
||||
iso.write(struct.pack(">I", new_lba))
|
||||
iso.write(struct.pack("<I", new_size))
|
||||
iso.write(struct.pack(">I", new_size))
|
||||
|
||||
iso.seek(end_save)
|
||||
|
||||
# Align to 0x8000
|
||||
end_pos = iso.tell()
|
||||
if (end_pos % VOLUME_ALIGN) == 0:
|
||||
al_end = align(end_pos + SECTOR_SIZE, VOLUME_ALIGN)
|
||||
else:
|
||||
al_end = align(end_pos, VOLUME_ALIGN)
|
||||
|
||||
iso.write(b"\x00" * (al_end - end_pos - SECTOR_SIZE))
|
||||
|
||||
# Sony's cdvdgen tool starting with v2.00 by default adds
|
||||
# a 20MiB padding to the end of the PVD, add it here if requested
|
||||
if add_padding == PAD_20MB:
|
||||
iso.write(b"\x00" * 0x140_0000)
|
||||
elif add_padding == PAD_VOLUME:
|
||||
empty_sector = b"\x00" * SECTOR_SIZE
|
||||
while (iso.tell() - pvd_loc + SYSTEM_AREA_SIZE) < (MAX_VOLUME_SIZE - SECTOR_SIZE):
|
||||
iso.write(empty_sector)
|
||||
|
||||
|
||||
# Last LBA includes the anchor
|
||||
last_pvd_lba = ((iso.tell() - pvd_loc + SYSTEM_AREA_SIZE) // 0x800) + 1
|
||||
|
||||
# Check if we didn't go over the maximum
|
||||
assert last_pvd_lba <= MAX_VOLUME_SECTOR_COUNT, "Iso image would go over the maximum allowed size!"
|
||||
|
||||
iso.write(layer_info.footer)
|
||||
iso.seek(pvd_loc + 0x50)
|
||||
iso.write(struct.pack("<I", last_pvd_lba))
|
||||
iso.write(struct.pack(">I", last_pvd_lba))
|
||||
iso.seek(-0x7F4, io.SEEK_END)
|
||||
iso.write(struct.pack("<I", last_pvd_lba-1))
|
||||
|
||||
iso.seek(0, io.SEEK_END)
|
||||
return iso.tell()
|
||||
|
||||
|
||||
def rebuild_iso(
|
||||
iso: Path, filelist: Path, iso_files: Path, output: Path, add_padding: int
|
||||
) -> None:
|
||||
# Validate args
|
||||
if not validate_rebuild(filelist, iso_files):
|
||||
return
|
||||
|
||||
|
||||
# Parse filelist file
|
||||
with open(filelist, "r") as f:
|
||||
lines = f.readlines()
|
||||
|
||||
iso_info = Iso()
|
||||
|
||||
# is a dual layer filelist?
|
||||
|
||||
if lines[0].startswith("//"):
|
||||
iso_info.has_second_layer = True
|
||||
|
||||
l0_files = int(lines.pop(0)[2:]) + 1
|
||||
|
||||
parse_filelist(iso_info.layers[0].meta, lines[:l0_files])
|
||||
parse_filelist(iso_info.layers[1].meta, lines[l0_files:])
|
||||
else:
|
||||
iso_info.has_second_layer = False
|
||||
|
||||
parse_filelist(iso_info.layers[0].meta, lines[:])
|
||||
|
||||
|
||||
with open(iso, "rb") as f:
|
||||
second_layer, pvd1_pos, _ = check_iso(f)
|
||||
|
||||
assert iso_info.has_second_layer == second_layer, "Filelist type and ISO type disagree!"
|
||||
l0_start = consume_iso_header(f, 0, iso_info.layers[0].meta.total_inodes)
|
||||
f.seek(0)
|
||||
iso_info.layers[0].header = f.read(l0_start)
|
||||
f.seek(pvd1_pos - SECTOR_SIZE)
|
||||
iso_info.layers[0].footer = f.read(SECTOR_SIZE)
|
||||
|
||||
if iso_info.has_second_layer:
|
||||
l1_start = consume_iso_header(f, pvd1_pos, iso_info.layers[1].meta.total_inodes)
|
||||
f.seek(pvd1_pos)
|
||||
iso_info.layers[1].header = f.read(l1_start)
|
||||
f.seek(-SECTOR_SIZE, io.SEEK_END)
|
||||
iso_info.layers[1].footer = f.read(SECTOR_SIZE)
|
||||
iso_info.layers[1].offset = pvd1_pos + SYSTEM_AREA_SIZE
|
||||
|
||||
with open(output, "wb+") as f:
|
||||
pvd1_start = write_new_pvd(f, iso_files, add_padding, iso_info.layers[0], SYSTEM_AREA_SIZE)
|
||||
|
||||
# PVD1
|
||||
if iso_info.has_second_layer:
|
||||
print("\n< SECOND LAYER >\n")
|
||||
write_new_pvd(f, iso_files, add_padding, iso_info.layers[1], pvd1_start)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
42
patch.asm
Normal file
42
patch.asm
Normal file
@ -0,0 +1,42 @@
|
||||
.erroronwarning on
|
||||
|
||||
.ps2
|
||||
|
||||
.open "extracted/DATA.BIN", "DATA.BIN", 0x0
|
||||
|
||||
.orga 0x094BC858
|
||||
|
||||
.db 0x3A, 0x80
|
||||
.ascii "<K3>"
|
||||
.ascii "Hey"
|
||||
.db 0x81, 0x43
|
||||
.ascii "where are you planning to go on that ship?"
|
||||
.db 0xFF, 0xFD, 0x19
|
||||
.ascii "<K0>"
|
||||
.db 0x07, 0x32, 0x2F, 0x01, 0x00, 0x0A, 0x50, 0x00, 0xA5, 0x00, 0x3F, 0x80 ; originally 0x53, 0x80
|
||||
.ascii "<K3>Oh really? The Safar desert on the Afroca continent?"
|
||||
.db 0xFF, 0xFD, 0x19
|
||||
.ascii "<K0>"
|
||||
.db 0x2F, 0x02, 0x00, 0x0A, 0x50, 0x00, 0x96, 0x00, 0x8B, 0x80 ; originally 0x90, 0x80
|
||||
.ascii "<K3>Does that mean you're searching for \"that\"?"
|
||||
.db 0x0A
|
||||
.ascii "The City of Sand and Mirages, Kefin?"
|
||||
.db 0x0A
|
||||
.ascii "Apparently it used to be a magnificent kingdom."
|
||||
.db 0xFF, 0xFD, 0x19
|
||||
.ascii "<K0>"
|
||||
.db 0x2F, 0x03, 0x00, 0x0A, 0x50, 0x00, 0x96, 0x00, 0x89, 0x80
|
||||
|
||||
|
||||
|
||||
.close
|
||||
|
||||
.open "extracted/SLPM_663.60", "SLPM_663.60", 0x0
|
||||
|
||||
.orga 0x0025FD45
|
||||
|
||||
.sjisn "Kaisaan"
|
||||
.db 0x81, 0x43
|
||||
|
||||
|
||||
.close
|
Loading…
Reference in New Issue
Block a user