2121from datalad_service .config import AWS_S3_PUBLIC_BUCKET
2222
2323
24- def commit_files (store , dataset , files , name = None , email = None , cookies = None ):
24+ async def commit_files (store , dataset , files , name = None , email = None , cookies = None ):
2525 """
2626 Commit a list of files with the email and name provided.
2727
@@ -35,9 +35,9 @@ def commit_files(store, dataset, files, name=None, email=None, cookies=None):
3535 and pygit2 .Signature (name , email )
3636 or pygit2 .Signature (COMMITTER_NAME , COMMITTER_EMAIL )
3737 )
38- ref = git_commit (repo , files , author )
38+ ref = await git_commit (repo , files , author )
3939 # Run the validator but don't block on the request
40- asyncio . create_task ( validate_dataset .kiq (dataset , dataset_path , str (ref ), cookies ) )
40+ await validate_dataset .kiq (dataset , dataset_path , str (ref ), cookies )
4141 return ref
4242
4343
@@ -47,7 +47,7 @@ def get_tree(store, dataset, tree):
4747 return get_repo_files (dataset , dataset_path , tree )
4848
4949
50- def remove_files (store , dataset , paths , name = None , email = None , cookies = None ):
50+ async def remove_files (store , dataset , paths , name = None , email = None , cookies = None ):
5151 dataset_path = store .get_dataset_path (dataset )
5252 repo = pygit2 .Repository (dataset_path )
5353 if name and email :
@@ -57,7 +57,9 @@ def remove_files(store, dataset, paths, name=None, email=None, cookies=None):
5757 repo .index .remove_all (paths )
5858 repo .index .write ()
5959 repo .checkout_index ()
60- hexsha = str (git_commit_index (repo , author , message = '[OpenNeuro] Files removed' ))
60+ hexsha = str (
61+ await git_commit_index (repo , author , message = '[OpenNeuro] Files removed' )
62+ )
6163
6264
6365def parse_s3_annex_url (url , bucket_name = AWS_S3_PUBLIC_BUCKET ):
0 commit comments