pull/520/head
lijiaxing 2023-11-29 11:08:40 +08:00
parent 4e4fb52898
commit 06cdcc3654
3 changed files with 33 additions and 15 deletions

View File

@ -178,7 +178,8 @@ def args_sanity_check():
else: else:
if ckpt.async_upload: if ckpt.async_upload:
assert "save_ckpt_folder" in ckpt assert "save_ckpt_folder" in ckpt
if "boto3:" not in ckpt.save_ckpt_folder: prefix_list = ["boto3:", "volc:", "oss2:"]
if not any(ckpt.save_ckpt_folder.startswith(prefix) for prefix in prefix_list):
if gpc.is_rank_for_log(): if gpc.is_rank_for_log():
logger.warning( logger.warning(
"Storing ckpt on file system does not support asynchronous storage, will use sync save!" "Storing ckpt on file system does not support asynchronous storage, will use sync save!"

View File

@ -721,8 +721,8 @@ class CheckpointManager:
self.load_ckpt_info["ckpt_type"] = CheckpointLoadMethod.convet_load_type(self.load_ckpt_info["ckpt_type"]) self.load_ckpt_info["ckpt_type"] = CheckpointLoadMethod.convet_load_type(self.load_ckpt_info["ckpt_type"])
# test storage setting is ok. # test storage setting is ok.
if self.enable_save_ckpt: # if self.enable_save_ckpt:
self.try_ping_storage() # self.try_ping_storage()
def quit_signal_handler(self, train_state) -> bool: def quit_signal_handler(self, train_state) -> bool:
""" """

View File

@ -37,6 +37,8 @@ except ImportError:
try: try:
import oss2 import oss2
from oss2 import SizedFileAdapter, determine_part_size
from oss2.models import PartInfo
except ImportError: except ImportError:
pass pass
@ -68,12 +70,6 @@ def llm_save(save_path: str, saved_obj: Any, **kwargs):
storage_manager.save(save_path, to_save_obj=saved_obj, **kwargs) storage_manager.save(save_path, to_save_obj=saved_obj, **kwargs)
def percentage(consumed_bytes: int, total_bytes: int, rw_once_bytes: int, type: DataTransferType):
if total_bytes and gpc.is_rank_for_log():
rate = int(100 * float(consumed_bytes) / float(total_bytes))
logger.info(f"rate:{rate}, consumed_bytes:{consumed_bytes},total_bytes{total_bytes}, rw_once_bytes:{rw_once_bytes}, type:{type}")
class StorageClient: class StorageClient:
""" """
StorageClient as a client for s3 storage access. StorageClient as a client for s3 storage access.
@ -548,27 +544,28 @@ class VolcClient(StorageClient):
total_size = os.path.getsize(local_nvme_path) total_size = os.path.getsize(local_nvme_path)
part_size = 5 * 1024 * 1024 part_size = 5 * 1024 * 1024
data_transfer_listener = MergeProcess(percentage, total_size, (total_size + part_size - 1) // part_size, 0)
multi_result = handler.client.create_multipart_upload(bucket_name, fp) multi_result = handler.client.create_multipart_upload(bucket_name, fp)
upload_id = multi_result.upload_id upload_id = multi_result.upload_id
parts = [] parts = []
# 上传分片数据 # 上传分片数据
logger.info('Begin multipart upload of one file')
with open(local_nvme_path, 'rb') as f: with open(local_nvme_path, 'rb') as f:
part_number = 1 part_number = 1
offset = 0 offset = 0
while offset < total_size: while offset < total_size:
num_to_upload = min(part_size, total_size - offset) num_to_upload = min(part_size, total_size - offset)
out = handler.client.upload_part(bucket_name, fp, upload_id, part_number, out = handler.client.upload_part(bucket_name, fp, upload_id, part_number,
content=SizeAdapter(f, num_to_upload, init_offset=offset), content=SizeAdapter(f, num_to_upload, init_offset=offset))
data_transfer_listener=data_transfer_listener)
parts.append(out) parts.append(out)
offset += num_to_upload offset += num_to_upload
part_number += 1 part_number += 1
# 完成分片上传任务 # 完成分片上传任务
handler.client.complete_multipart_upload(bucket_name, fp, upload_id, parts) handler.client.complete_multipart_upload(bucket_name, fp, upload_id, parts)
logger.info('Finish multipart upload of one file')
except handler.handler.exceptions.TosClientError as exc: except handler.handler.exceptions.TosClientError as exc:
raise RuntimeError( raise RuntimeError(
f"Volc Network Error: fail with client error, message:{exc.message}, cause: {exc.cause}" f"Volc Network Error: fail with client error, message:{exc.message}, cause: {exc.cause}"
@ -674,7 +671,27 @@ class AliClient(StorageClient):
@staticmethod @staticmethod
def async_upload_fileobj(handler, fp: str, local_nvme_path: str): def async_upload_fileobj(handler, fp: str, local_nvme_path: str):
try: try:
handler.client.put_object_from_file(fp, local_nvme_path) # handler.client.put_object_from_file(fp, local_nvme_path)
total_size = os.path.getsize(local_nvme_path)
part_size = determine_part_size(total_size, preferred_size=5 * 1024 * 1024)
upload_id = handler.client.init_multipart_upload(fp).upload_id
parts = []
with open(local_nvme_path, 'rb') as fileobj:
part_number = 1
offset = 0
while offset < total_size:
num_to_upload = min(part_size, total_size - offset)
# 调用SizedFileAdapter(fileobj, size)方法会生成一个新的文件对象,重新计算起始追加位置。
result = handler.client.upload_part(fp, upload_id, part_number,
SizedFileAdapter(fileobj, num_to_upload))
parts.append(PartInfo(part_number, result.etag))
offset += num_to_upload
part_number += 1
headers = dict()
handler.client.complete_multipart_upload(fp, upload_id, parts, headers=headers)
except Exception as e: except Exception as e:
raise e raise e