2021-11-04 19:40:42 +01:00
|
|
|
#!/usr/bin/env python3
|
2021-11-16 13:01:17 +01:00
|
|
|
from math import ceil
|
|
|
|
from struct import unpack, pack
|
|
|
|
from pathlib import Path
|
|
|
|
from os import listdir
|
2021-11-15 20:18:43 +01:00
|
|
|
__version__ = "1.2"
|
2021-11-16 13:01:17 +01:00
|
|
|
__author__ = "rigodron, algoflash, GGLinnk"
|
2021-11-04 19:40:42 +01:00
|
|
|
__OriginalAutor__ = "infval"
|
2021-11-16 13:01:17 +01:00
|
|
|
__license__ = "MIT"
|
|
|
|
__status__ = "developpement"
|
2021-11-04 19:40:42 +01:00
|
|
|
|
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
BIT_COMPRESSION_FLAG = 0x40000000
|
|
|
|
FILE_LENGTH_MASK = 0x3FFFFFFF
|
|
|
|
CHUNK_SIZE = 0x800
|
2021-11-04 19:40:42 +01:00
|
|
|
|
2021-11-16 13:01:17 +01:00
|
|
|
|
2021-11-05 18:44:11 +01:00
|
|
|
def pzz_decompress(compressed_bytes: bytes):
|
|
|
|
uncompressed_bytes = bytearray()
|
|
|
|
compressed_bytes_size = len(compressed_bytes) // 2 * 2
|
2021-11-04 19:40:42 +01:00
|
|
|
|
|
|
|
cb = 0 # Control bytes
|
|
|
|
cb_bit = -1
|
|
|
|
i = 0
|
2021-11-05 18:44:11 +01:00
|
|
|
while i < compressed_bytes_size:
|
2021-11-04 19:40:42 +01:00
|
|
|
if cb_bit < 0:
|
2021-11-16 13:01:17 +01:00
|
|
|
cb = compressed_bytes[i + 1]
|
2021-11-05 18:44:11 +01:00
|
|
|
cb |= compressed_bytes[i + 0] << 8
|
2021-11-04 19:40:42 +01:00
|
|
|
cb_bit = 15
|
|
|
|
i += 2
|
|
|
|
continue
|
|
|
|
|
|
|
|
compress_flag = cb & (1 << cb_bit)
|
|
|
|
cb_bit -= 1
|
|
|
|
|
|
|
|
print(compress_flag)
|
|
|
|
if compress_flag:
|
2021-11-16 13:01:17 +01:00
|
|
|
c = compressed_bytes[i + 1]
|
2021-11-05 18:44:11 +01:00
|
|
|
c |= compressed_bytes[i + 0] << 8
|
2021-11-04 19:40:42 +01:00
|
|
|
offset = (c & 0x7FF) * 2
|
|
|
|
if offset == 0:
|
2021-11-16 13:01:17 +01:00
|
|
|
break # End of the compressed data
|
2021-11-04 19:40:42 +01:00
|
|
|
count = (c >> 11) * 2
|
|
|
|
if count == 0:
|
|
|
|
i += 2
|
2021-11-16 13:01:17 +01:00
|
|
|
c = compressed_bytes[i + 1]
|
2021-11-05 18:44:11 +01:00
|
|
|
c |= compressed_bytes[i + 0] << 8
|
2021-11-04 19:40:42 +01:00
|
|
|
count = c * 2
|
|
|
|
|
2021-11-05 18:44:11 +01:00
|
|
|
index = len(uncompressed_bytes) - offset
|
2021-11-04 19:40:42 +01:00
|
|
|
for j in range(count):
|
2021-11-05 18:44:11 +01:00
|
|
|
uncompressed_bytes.append(uncompressed_bytes[index + j])
|
2021-11-04 19:40:42 +01:00
|
|
|
else:
|
2021-11-05 18:44:11 +01:00
|
|
|
uncompressed_bytes.extend(compressed_bytes[i: i + 2])
|
2021-11-04 19:40:42 +01:00
|
|
|
i += 2
|
|
|
|
|
2021-11-05 18:44:11 +01:00
|
|
|
return uncompressed_bytes
|
|
|
|
|
2021-11-16 13:01:17 +01:00
|
|
|
|
2021-11-05 18:44:11 +01:00
|
|
|
def bytes_align(bout: bytes):
|
|
|
|
success = False
|
|
|
|
while not success:
|
|
|
|
bout.extend(b"\x00\x00")
|
|
|
|
address = len(bout)
|
|
|
|
if hex(address).endswith("00"):
|
|
|
|
break
|
2021-11-04 19:40:42 +01:00
|
|
|
|
2021-11-16 13:01:17 +01:00
|
|
|
|
2021-11-04 19:40:42 +01:00
|
|
|
def pzz_compress(b):
|
|
|
|
bout = bytearray()
|
|
|
|
size_b = len(b) // 2 * 2
|
|
|
|
|
|
|
|
cb = 0 # Control bytes
|
|
|
|
cb_bit = 15
|
|
|
|
cb_pos = 0
|
|
|
|
bout.extend(b"\x00\x00")
|
|
|
|
|
|
|
|
i = 0
|
|
|
|
while i < size_b:
|
2021-11-05 18:44:11 +01:00
|
|
|
start = max(i - 0x7FF * 2, 0)
|
2021-11-04 19:40:42 +01:00
|
|
|
count_r = 0
|
|
|
|
max_i = -1
|
|
|
|
tmp = b[i: i + 2]
|
|
|
|
init_count = len(tmp)
|
|
|
|
while True:
|
|
|
|
start = b.find(tmp, start, i + 1)
|
|
|
|
if start != -1 and start % 2 != 0:
|
|
|
|
start += 1
|
|
|
|
continue
|
|
|
|
if start != -1:
|
|
|
|
count = init_count
|
|
|
|
while i < size_b - count \
|
2021-11-05 18:44:11 +01:00
|
|
|
and count < 0xFFFF * 2 \
|
|
|
|
and b[start + count] == b[i + count] \
|
|
|
|
and b[start + count + 1] == b[i + count + 1]:
|
2021-11-04 19:40:42 +01:00
|
|
|
count += 2
|
|
|
|
if count_r < count:
|
|
|
|
count_r = count
|
|
|
|
max_i = start
|
|
|
|
start += 2
|
|
|
|
else:
|
|
|
|
break
|
|
|
|
start = max_i
|
|
|
|
|
|
|
|
compress_flag = 0
|
|
|
|
if count_r >= 4:
|
|
|
|
compress_flag = 1
|
|
|
|
offset = i - start
|
|
|
|
offset //= 2
|
|
|
|
count_r //= 2
|
|
|
|
c = offset
|
|
|
|
if count_r <= 0x1F:
|
|
|
|
c |= count_r << 11
|
|
|
|
bout.append((c >> 8))
|
|
|
|
bout.append(c & 0xFF)
|
|
|
|
else:
|
|
|
|
bout.append((c >> 8))
|
|
|
|
bout.append(c & 0xFF)
|
|
|
|
bout.append((count_r >> 8))
|
|
|
|
bout.append(count_r & 0xFF)
|
|
|
|
i += count_r * 2
|
|
|
|
else:
|
|
|
|
bout.extend(b[i: i + 2])
|
|
|
|
i += 2
|
|
|
|
cb |= (compress_flag << cb_bit)
|
|
|
|
cb_bit -= 1
|
|
|
|
if cb_bit < 0:
|
|
|
|
bout[cb_pos + 1] = cb & 0xFF
|
|
|
|
bout[cb_pos + 0] = cb >> 8
|
|
|
|
cb = 0x0000
|
|
|
|
cb_bit = 15
|
|
|
|
cb_pos = len(bout)
|
|
|
|
bout.extend(b"\x00\x00")
|
|
|
|
|
|
|
|
cb |= (1 << cb_bit)
|
|
|
|
bout[cb_pos + 1] = cb & 0xFF
|
|
|
|
bout[cb_pos + 0] = cb >> 8
|
|
|
|
bout.extend(b"\x00\x00")
|
|
|
|
|
2021-11-05 18:44:11 +01:00
|
|
|
bytes_align(bout)
|
|
|
|
|
2021-11-04 19:40:42 +01:00
|
|
|
return bout
|
|
|
|
|
2021-11-16 13:01:17 +01:00
|
|
|
|
2021-11-15 20:18:43 +01:00
|
|
|
def pzz_unpack(pzz_path, dest_folder):
|
2021-11-14 15:43:29 +01:00
|
|
|
# Script BMS pour les pzz de ps2 (GioGio's adventure) -> https://zenhax.com/viewtopic.php?f=9&t=8724&p=39437#p39437
|
2021-11-16 13:01:17 +01:00
|
|
|
|
|
|
|
if dest_folder != Path('.'):
|
2021-11-15 20:18:43 +01:00
|
|
|
unpacked_pzz_path = dest_folder
|
2021-11-16 13:01:17 +01:00
|
|
|
else:
|
2021-11-15 20:18:43 +01:00
|
|
|
unpacked_pzz_path = pzz_path.parent / pzz_path.stem
|
2021-11-15 01:25:01 +01:00
|
|
|
unpacked_pzz_path.mkdir(exist_ok=True)
|
2021-11-14 15:43:29 +01:00
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
with open(pzz_path, "rb") as pzz_file:
|
2021-11-14 15:43:29 +01:00
|
|
|
# file_count reçoit le nombre de fichiers présent dans le PZZ :
|
2021-11-16 13:01:17 +01:00
|
|
|
# On lit les 4 premiers octets (uint32 big-endian)
|
|
|
|
file_count, = unpack(">I", pzz_file.read(4))
|
2021-11-05 18:44:11 +01:00
|
|
|
|
2021-11-14 15:43:29 +01:00
|
|
|
# files_descriptors reçoit un tuple avec l'ensemble des descripteurs de fichiers (groupes d'uint32 big-endian)
|
2021-11-16 13:01:17 +01:00
|
|
|
files_descriptors = unpack(">{}I".format(
|
|
|
|
file_count), pzz_file.read(file_count * 4))
|
2021-11-04 19:40:42 +01:00
|
|
|
|
|
|
|
print("File count:", file_count)
|
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
offset = CHUNK_SIZE
|
2021-11-16 13:01:17 +01:00
|
|
|
# on parcours le tuple de descripteurs de fichiers
|
|
|
|
for index, file_descriptor in enumerate(files_descriptors):
|
2021-11-14 15:43:29 +01:00
|
|
|
|
2021-11-16 13:01:17 +01:00
|
|
|
# Le bit 30 correspond au flag de compression (bits numérotés de 0 à 31)
|
|
|
|
is_compressed = (file_descriptor & BIT_COMPRESSION_FLAG) != 0
|
|
|
|
if not is_compressed: # Si le fichier n'est pas compressé, on ajoute 'U' derrière l'index
|
2021-11-15 20:18:43 +01:00
|
|
|
compression_status = 'U'
|
|
|
|
comp_str = ""
|
2021-11-16 13:01:17 +01:00
|
|
|
else: # Si le fichier est compressé, on ajoute "_compressed" devant l'extension et 'C' derrière l'index
|
2021-11-15 20:18:43 +01:00
|
|
|
compression_status = 'C'
|
|
|
|
comp_str = "_compressed"
|
2021-11-14 15:43:29 +01:00
|
|
|
|
2021-11-16 13:01:17 +01:00
|
|
|
# file_descriptor reçoit maintenant les 30 premiers bits : (la taille / CHUNK_SIZE)
|
2021-11-15 01:25:01 +01:00
|
|
|
file_descriptor &= FILE_LENGTH_MASK
|
2021-11-14 15:34:54 +01:00
|
|
|
|
2021-11-14 15:43:29 +01:00
|
|
|
# file_len reçoit la taille du fichier
|
2021-11-15 01:25:01 +01:00
|
|
|
# la taille du fichier est un multiple de CHUNK_SIZE, on paddera avec des 0 jusqu'au fichier suivant
|
2021-11-16 13:01:17 +01:00
|
|
|
# file_len contient alors la taille du fichier en octets
|
|
|
|
file_len = file_descriptor * CHUNK_SIZE
|
|
|
|
|
2021-11-14 15:43:29 +01:00
|
|
|
# On forme le nom du nouveau fichier que l'on va extraire
|
2021-11-16 13:01:17 +01:00
|
|
|
filename = "{:03}{}_{}{}".format(
|
|
|
|
index, compression_status, pzz_path.stem, comp_str)
|
2021-11-15 20:18:43 +01:00
|
|
|
file_path = (unpacked_pzz_path / filename).with_suffix(".dat")
|
2021-11-04 19:40:42 +01:00
|
|
|
|
2021-11-15 20:18:43 +01:00
|
|
|
print("Offset: {:010} - {}".format(offset, file_path.stem))
|
|
|
|
|
|
|
|
# Si la taille est nulle, on créé un fichier vide et on passe au descripteur de fichier suivant
|
|
|
|
if file_len == 0:
|
|
|
|
file_path.touch()
|
|
|
|
continue
|
2021-11-04 19:40:42 +01:00
|
|
|
|
2021-11-14 15:43:29 +01:00
|
|
|
# On se positionne au début du fichier dans l'archive
|
2021-11-15 01:25:01 +01:00
|
|
|
pzz_file.seek(offset)
|
2021-11-14 15:43:29 +01:00
|
|
|
# On extrait notre fichier
|
2021-11-15 01:25:01 +01:00
|
|
|
file_path.write_bytes(pzz_file.read(file_len))
|
2021-11-14 15:43:29 +01:00
|
|
|
|
|
|
|
# Enfin, on ajoute la taille du fichier afin de pointer sur le fichier suivant
|
2021-11-15 01:25:01 +01:00
|
|
|
# La taille du fichier étant un multiple de CHUNK_SIZE, on aura complété les 2048 octets finaux avec des 0x00
|
2021-11-14 15:43:29 +01:00
|
|
|
offset += file_len
|
2021-11-04 19:40:42 +01:00
|
|
|
|
2021-11-16 13:01:17 +01:00
|
|
|
|
2021-11-15 20:18:43 +01:00
|
|
|
def pzz_pack(src_path, dest_file):
|
2021-11-15 01:25:01 +01:00
|
|
|
# On récupère les fichiers du dossier à compresser
|
|
|
|
src_files = listdir(src_path)
|
2021-11-07 20:31:19 +01:00
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
# On récupère le nombre total de fichiers
|
2021-11-15 20:18:43 +01:00
|
|
|
file_count = int(src_files[-1].split("_")[0][0:3]) + 1
|
|
|
|
|
2021-11-16 13:01:17 +01:00
|
|
|
if dest_file != Path('.'):
|
2021-11-15 20:18:43 +01:00
|
|
|
if dest_file.suffix != ".pzz":
|
|
|
|
raise("Invalid file format : dest must be a pzz")
|
|
|
|
pzz_path = dest_file
|
2021-11-16 13:01:17 +01:00
|
|
|
else:
|
2021-11-15 20:18:43 +01:00
|
|
|
pzz_path = src_path.with_suffix(".pzz")
|
|
|
|
|
|
|
|
print(str(file_count) + " files to pack in " + str(pzz_path))
|
2021-11-16 13:01:17 +01:00
|
|
|
|
|
|
|
with pzz_path.open("wb") as pzz_file:
|
2021-11-15 01:25:01 +01:00
|
|
|
# On écrit file_count au début de header
|
|
|
|
pzz_file.write(file_count.to_bytes(4, byteorder='big'))
|
2021-11-07 20:31:19 +01:00
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
# On écrit les file_descriptor dans le header du PZZ pour chaque fichier
|
2021-11-16 13:01:17 +01:00
|
|
|
for src_file_name in src_files:
|
2021-11-15 20:18:43 +01:00
|
|
|
index = int(src_file_name.split("_")[0][0:3])
|
2021-11-07 20:31:19 +01:00
|
|
|
|
2021-11-15 20:18:43 +01:00
|
|
|
# Compression status permet de verrifier si le fichier doit être finalement compressé ou non
|
|
|
|
compression_status = src_file_name.split("_")[0][3:4]
|
2021-11-16 13:01:17 +01:00
|
|
|
is_compressed = (len(src_file_name.split("_compressed")) > 1)
|
|
|
|
|
2021-11-15 20:18:43 +01:00
|
|
|
# Le fichier doit être compressé avant d'être pack
|
2021-11-16 13:01:17 +01:00
|
|
|
if compression_status == 'C' and is_compressed is False:
|
2021-11-15 20:18:43 +01:00
|
|
|
pass
|
|
|
|
# Le fichier doit être décompressé avant d'être pack
|
2021-11-16 13:01:17 +01:00
|
|
|
elif compression_status == 'U' and is_compressed is True:
|
2021-11-15 20:18:43 +01:00
|
|
|
pass
|
2021-11-07 20:31:19 +01:00
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
# file_descriptor = arrondi supérieur de la taille / CHUNK_SIZE
|
2021-11-16 13:01:17 +01:00
|
|
|
file_descriptor = ceil(
|
|
|
|
(src_path / src_file_name).stat().st_size / CHUNK_SIZE)
|
2021-11-07 20:31:19 +01:00
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
# On ajoute le flag de compression au file_descriptor
|
2021-11-16 13:01:17 +01:00
|
|
|
if is_compressed:
|
2021-11-15 01:25:01 +01:00
|
|
|
file_descriptor |= BIT_COMPRESSION_FLAG
|
2021-11-07 20:31:19 +01:00
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
# On ecrit le file_descriptor
|
|
|
|
pzz_file.write(file_descriptor.to_bytes(4, byteorder='big'))
|
2021-11-07 20:31:19 +01:00
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
# On se place à la fin du header PZZ
|
|
|
|
pzz_file.seek(CHUNK_SIZE)
|
2021-11-07 20:31:19 +01:00
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
# On écrit tous les fichiers à la suite du header
|
2021-11-16 13:01:17 +01:00
|
|
|
for src_file_name in src_files:
|
|
|
|
is_compressed = (len(src_file_name.split("_compressed")) > 1)
|
|
|
|
|
|
|
|
with (src_path / src_file_name).open("rb") as src_file:
|
|
|
|
pzz_file.write(src_file.read())
|
2021-11-07 20:31:19 +01:00
|
|
|
|
2021-11-15 01:25:01 +01:00
|
|
|
# Si le fichier n'est pas compressé, on ajoute le padding pour correspondre à un multiple de CHUNK_SIZE
|
|
|
|
if not is_compressed and (src_file.tell() % CHUNK_SIZE) > 0:
|
2021-11-16 13:01:17 +01:00
|
|
|
pzz_file.write(
|
|
|
|
b"\x00" * (CHUNK_SIZE - (src_file.tell() % CHUNK_SIZE)))
|
|
|
|
|
2021-11-04 19:40:42 +01:00
|
|
|
|
|
|
|
def get_argparser():
|
|
|
|
import argparse
|
2021-11-16 13:01:17 +01:00
|
|
|
parser = argparse.ArgumentParser(
|
|
|
|
description='PZZ (de)compressor & unpacker - [GameCube] Gotcha Force v' + __version__)
|
|
|
|
parser.add_argument('--version', action='version',
|
|
|
|
version='%(prog)s ' + __version__)
|
2021-11-15 20:18:43 +01:00
|
|
|
parser.add_argument('input_path', metavar='INPUT', help='')
|
2021-11-16 13:01:17 +01:00
|
|
|
parser.add_argument('output_path', metavar='OUTPUT',
|
|
|
|
help='', nargs='?', default="")
|
2021-11-04 19:40:42 +01:00
|
|
|
group = parser.add_mutually_exclusive_group(required=True)
|
2021-11-16 13:01:17 +01:00
|
|
|
group.add_argument('-p', '--pack', action='store_true',
|
|
|
|
help="-p source_folder dest_file.pzz(optionnal) : Pack source_folder in new file source_folder.pzz")
|
|
|
|
group.add_argument('-u', '--unpack', action='store_true',
|
|
|
|
help='-u source_folder.pzz dest_folder(optionnal) : Unpack the pzz in new folder source_folder')
|
|
|
|
group.add_argument('-bp', '--batch-pack', action='store_true',
|
|
|
|
help='-bp source_folder dest_folder(optionnal - if not specified it will pack in source_folder)')
|
|
|
|
group.add_argument('-bu', '--batch-unpack', action='store_true',
|
|
|
|
help='INPUT relative pattern; e.g. AFS_DATA\\*.pzz')
|
|
|
|
|
|
|
|
# group.add_argument('-a', '-aa',action='store_true', help='sha256')
|
|
|
|
# group.add_argument('-c', '--compress', action='store_true', help='')
|
|
|
|
# group.add_argument('-d', '--decompress', action='store_true', help='Unpacked files from PZZ')
|
|
|
|
# group.add_argument('-bc', '--batch-compress', action='store_true', help='INPUT relative pattern; e.g. AFS_DATA\\*.bin')
|
|
|
|
# group.add_argument('-bd', '--batch-decompress', action='store_true', help='INPUT relative pattern; e.g. AFS_DATA\\*_compressed.dat')
|
2021-11-04 19:40:42 +01:00
|
|
|
return parser
|
|
|
|
|
2021-11-16 13:01:17 +01:00
|
|
|
|
2021-11-04 19:40:42 +01:00
|
|
|
if __name__ == '__main__':
|
2021-11-16 13:01:17 +01:00
|
|
|
args = get_argparser().parse_args()
|
2021-11-04 19:40:42 +01:00
|
|
|
|
2021-11-16 13:01:17 +01:00
|
|
|
p_input = Path(args.input_path)
|
2021-11-04 19:40:42 +01:00
|
|
|
p_output = Path(args.output_path)
|
2021-11-15 20:18:43 +01:00
|
|
|
"""
|
2021-11-15 01:25:01 +01:00
|
|
|
if args.compress:
|
2021-11-04 19:40:42 +01:00
|
|
|
print("### Compress")
|
|
|
|
p_output.write_bytes(pzz_compress(p_input.read_bytes()))
|
|
|
|
elif args.decompress:
|
|
|
|
print("### Decompress")
|
|
|
|
p_output.write_bytes(pzz_decompress(p_input.read_bytes()))
|
|
|
|
elif args.batch_compress:
|
|
|
|
print("### Batch Compress")
|
|
|
|
p_output.mkdir(exist_ok=True)
|
|
|
|
|
|
|
|
p = Path('.')
|
|
|
|
for filename in p.glob(args.input_path):
|
|
|
|
print(filename)
|
|
|
|
b = filename.read_bytes()
|
|
|
|
(p_output / filename.name).with_suffix(".dat").write_bytes(pzz_compress(b))
|
|
|
|
elif args.batch_decompress:
|
|
|
|
print("### Batch Decompress")
|
|
|
|
p_output.mkdir(exist_ok=True)
|
|
|
|
|
|
|
|
p = Path('.')
|
|
|
|
for filename in p.glob(args.input_path):
|
|
|
|
print(filename)
|
|
|
|
try:
|
|
|
|
b = filename.read_bytes()
|
|
|
|
(p_output / filename.name).with_suffix(".bin").write_bytes(pzz_decompress(b))
|
|
|
|
except IndexError:
|
|
|
|
print("! Wrong PZZ file")
|
2021-11-15 20:18:43 +01:00
|
|
|
el
|
|
|
|
"""
|
|
|
|
if args.pack:
|
2021-11-07 20:31:19 +01:00
|
|
|
print("### Pack")
|
2021-11-15 20:18:43 +01:00
|
|
|
pzz_pack(p_input, p_output)
|
2021-11-04 19:40:42 +01:00
|
|
|
elif args.unpack:
|
|
|
|
print("### Unpack")
|
2021-11-15 20:18:43 +01:00
|
|
|
pzz_unpack(p_input, p_output)
|
|
|
|
elif args.batch_pack:
|
|
|
|
print("### Batch Pack")
|
|
|
|
p_output.mkdir(exist_ok=True)
|
|
|
|
|
|
|
|
for folder in listdir(p_input):
|
2021-11-16 13:01:17 +01:00
|
|
|
pzz_pack(p_input / folder, p_output /
|
|
|
|
Path(folder).with_suffix(".pzz"))
|
2021-11-04 19:40:42 +01:00
|
|
|
elif args.batch_unpack:
|
|
|
|
print("### Batch Unpack")
|
|
|
|
p_output.mkdir(exist_ok=True)
|
|
|
|
|
2021-11-15 20:18:43 +01:00
|
|
|
for filename in listdir(p_input):
|
|
|
|
pzz_unpack(p_input / filename, p_output / Path(filename).stem)
|
|
|
|
|
|
|
|
"""
|
2021-11-16 13:01:17 +01:00
|
|
|
Code pour le developement --> pzztool.py -a a
|
2021-11-15 20:18:43 +01:00
|
|
|
compare le sha256 de chaque PZZ du dossier pzz et pzz2 puis affiche le nom de fichier en cas de différence
|
|
|
|
import hashlib
|
|
|
|
for pzz_file in listdir("pzz"):
|
|
|
|
with open("pzz/"+pzz_file, "rb") as f1, open("pzz2/"+pzz_file, "rb") as f2:
|
|
|
|
if hashlib.sha256( f1.read() ).hexdigest() != hashlib.sha256( f2.read() ).hexdigest() :
|
|
|
|
print(pzz_file)
|
|
|
|
"""
|