mirror of
https://github.com/mat-1/azalea.git
synced 2024-09-20 07:02:31 +00:00
7d901e39bc
* start updating to 22w42a * work a bit more on 22w42a * player chat packet * serverbound hello packet * Update mod.rs * add more stuff to clientbound player chat packet * ClientboundPlayerInfoUpdatePacket * features enabled and container closed * serverbound chat packets * make it compile * 22w43a * ServerboundChatSessionUpdatePacket * profile_public_key isn't Option anymore * Update bitset.rs * joining a server works * fix entitydatavalue * backtraces + fix clientbound chat message * fix some warnings and add more ecomments * 22w44a * generate en_us.json * add updating guide to codegen/readme * fix some markdown * update list of generated things * metadata stuff * Replace PJS generator mod with PixLyzer (#38) * pixlizer extractor * start working on shape extraction * fix generating language * fix pixlyzer shape generation * use empty_shape * generate blocks and shapes * update pixlyzer dir * Revert "update pixlyzer dir" This reverts commitee9a0e7a49
. * fix * fix * Revert "fix" This reverts commitad12ddcb00
. * fix * detect pixlyzer fail * fix pixlyzer * 22w45a * gen entities * add async-trait dep * update codegen/readme.md * explain when rust_log should be used * remove some unused code * start fixing pixlyzer issues * fix a thing in codegen * almost fixed * more progress towards 1.19.3 * 1.19.3-pre2 * fixes * revert some hardcoded property names * Delete clientbound_player_info_packet.rs * handle 1.19.3 player info packets * handle playerinforemove * start updating to 1.19.3-rc1 * optional registries work * fix some issues with 1.19.3 chat doesn't work yet * aaaaaaaaaaaaaaaaa * oh * ignore unused shapes * uncomment generate_blocks * fix migrate * 1.19.3-rc2 * fix clippy warnings * 1.19.3-rc3 * split the azalea-buf macro into separate modules * improve Recipe in protocol * 1.19.3
144 lines
5.8 KiB
Python
Executable file
144 lines
5.8 KiB
Python
Executable file
from lib.code.packet import fix_state
|
|
from lib.utils import PacketIdentifier, group_packets
|
|
import lib.code.language
|
|
import lib.code.registry
|
|
import lib.code.version
|
|
import lib.code.blocks
|
|
import lib.code.packet
|
|
import lib.code.shapes
|
|
import lib.code.utils
|
|
import lib.download
|
|
import lib.extract
|
|
import sys
|
|
|
|
lib.download.clear_version_cache()
|
|
|
|
if len(sys.argv) == 1:
|
|
print('\033[91mYou must provide a version to migrate to.\033[m')
|
|
version_manifest = lib.download.get_version_manifest()
|
|
newest_version = version_manifest['latest']['snapshot']
|
|
print(f'Hint: newest version is \033[1m{newest_version}\033[m')
|
|
exit()
|
|
|
|
|
|
old_version_id = lib.code.version.get_version_id()
|
|
old_mappings = lib.download.get_mappings_for_version(old_version_id)
|
|
old_burger_data = lib.extract.get_burger_data_for_version(old_version_id)
|
|
old_packet_list = list(old_burger_data[0]['packets']['packet'].values())
|
|
|
|
new_version_id = sys.argv[1]
|
|
new_mappings = lib.download.get_mappings_for_version(new_version_id)
|
|
new_burger_data = lib.extract.get_burger_data_for_version(new_version_id)
|
|
new_packet_list = list(new_burger_data[0]['packets']['packet'].values())
|
|
|
|
|
|
old_packets: dict[PacketIdentifier, str] = {}
|
|
old_packets_data: dict[PacketIdentifier, dict] = {}
|
|
new_packets: dict[PacketIdentifier, str] = {}
|
|
new_packets_data: dict[PacketIdentifier, dict] = {}
|
|
|
|
for packet in old_packet_list:
|
|
assert packet['class'].endswith('.class')
|
|
packet_name = old_mappings.get_class(packet['class'][:-6])
|
|
packet_ident = PacketIdentifier(
|
|
packet['id'], packet['direction'].lower(), fix_state(packet['state']))
|
|
old_packets[packet_ident] = packet_name
|
|
old_packets_data[packet_ident] = packet
|
|
for packet in new_packet_list:
|
|
assert packet['class'].endswith('.class')
|
|
packet_name = new_mappings.get_class(packet['class'][:-6])
|
|
packet_ident = PacketIdentifier(
|
|
packet['id'], packet['direction'].lower(), fix_state(packet['state']))
|
|
new_packets[packet_ident] = packet_name
|
|
new_packets_data[packet_ident] = packet
|
|
|
|
# find removed packets
|
|
removed_packets: list[PacketIdentifier] = []
|
|
for packet, packet_name in old_packets.items():
|
|
if packet_name not in new_packets.values():
|
|
removed_packets.append(packet)
|
|
print('Removed packet:', packet, packet_name)
|
|
for (direction, state), packets in group_packets(removed_packets).items():
|
|
lib.code.packet.remove_packet_ids(packets, direction, state)
|
|
|
|
print()
|
|
|
|
# find packets that changed ids
|
|
changed_packets: dict[PacketIdentifier, int] = {}
|
|
for old_packet, old_packet_name in old_packets.items():
|
|
for new_packet, new_packet_name in new_packets.items():
|
|
if old_packet_name == new_packet_name and old_packet.direction == new_packet.direction and old_packet.state == new_packet.state and old_packet.packet_id != new_packet.packet_id:
|
|
changed_packets[old_packet] = new_packet.packet_id
|
|
print('Changed packet id:', old_packet, '->',
|
|
new_packet, f'({new_packet_name})')
|
|
break
|
|
for (direction, state), packets in group_packets(list(changed_packets.keys())).items():
|
|
id_map: dict[int, int] = {}
|
|
for old_packet_id in packets:
|
|
new_packet_id = changed_packets[PacketIdentifier(
|
|
old_packet_id, direction, state)]
|
|
id_map[old_packet_id] = new_packet_id
|
|
lib.code.packet.change_packet_ids(id_map, direction, state)
|
|
|
|
|
|
print()
|
|
|
|
# find added/changed packets
|
|
added_or_changed_packets: list[PacketIdentifier] = []
|
|
for new_packet, packet_name in new_packets.items():
|
|
old_packet = None
|
|
for old_packet_tmp, old_packet_name in old_packets.items():
|
|
if old_packet_name == packet_name:
|
|
old_packet = old_packet_tmp
|
|
break
|
|
|
|
if packet_name not in old_packets.values():
|
|
added_or_changed_packets.append(new_packet)
|
|
print('Added packet:', new_packet, packet_name)
|
|
elif old_packet and not lib.code.packet.are_packet_instructions_identical(new_packets_data[new_packet].get('instructions'), old_packets_data[old_packet].get('instructions')):
|
|
added_or_changed_packets.append(new_packet)
|
|
print('Changed packet:', new_packet, packet_name)
|
|
for packet in added_or_changed_packets:
|
|
lib.code.packet.generate_packet(
|
|
new_burger_data[0]['packets']['packet'], new_mappings, packet.packet_id, packet.direction, packet.state)
|
|
|
|
lib.code.version.set_protocol_version(
|
|
new_burger_data[0]['version']['protocol'])
|
|
|
|
print('Updated protocol!')
|
|
|
|
|
|
old_ordered_blocks = lib.extract.get_ordered_blocks_burger(old_version_id)
|
|
new_ordered_blocks = lib.extract.get_ordered_blocks_burger(new_version_id)
|
|
if old_ordered_blocks != new_ordered_blocks:
|
|
print('Blocks changed, updating...')
|
|
|
|
block_states_burger = lib.extract.get_block_states_burger(new_version_id)
|
|
block_states_report = lib.extract.get_block_states_report(new_version_id)
|
|
|
|
shape_datas = lib.extract.get_pixlyzer_data(
|
|
new_version_id, 'shapes')
|
|
pixlyzer_block_datas = lib.extract.get_pixlyzer_data(
|
|
new_version_id, 'blocks')
|
|
|
|
lib.code.blocks.generate_blocks(
|
|
block_states_burger, block_states_report, new_ordered_blocks, new_mappings)
|
|
lib.code.shapes.generate_block_shapes(
|
|
pixlyzer_block_datas, shape_datas['shapes'], shape_datas['aabbs'], block_states_report, block_states_burger, new_mappings)
|
|
|
|
print('Getting en_us.json...')
|
|
language = lib.extract.get_en_us_lang(new_version_id)
|
|
lib.code.language.write_language(language)
|
|
|
|
print('Generating registries...')
|
|
registries = lib.extract.get_registries_report(new_version_id)
|
|
lib.code.registry.generate_registries(registries)
|
|
|
|
print('Finishing touches, setting version in README and formatting code...')
|
|
lib.code.version.set_version_id(new_version_id)
|
|
|
|
|
|
lib.code.utils.fmt()
|
|
|
|
print('Done!')
|
|
print('Make sure to `cargo check` and look for the generated `TODO`s to make sure everything is correct!')
|