@@ -200,7 +200,9 @@ def load_processed_data(
200200 filename = self .processed_file_names_dict [kind ]
201201 except NotImplementedError :
202202 filename = f"{ kind } .pt"
203- return torch .load (os .path .join (self .processed_dir , filename ))
203+ return torch .load (
204+ os .path .join (self .processed_dir , filename ), weights_only = False
205+ )
204206
205207 def dataloader (self , kind : str , ** kwargs ) -> DataLoader :
206208 """
@@ -519,7 +521,7 @@ def dataloader(self, kind: str, **kwargs) -> DataLoader:
519521 DataLoader: DataLoader object for the specified subset.
520522 """
521523 subdatasets = [
522- torch .load (os .path .join (s .processed_dir , f"{ kind } .pt" ))
524+ torch .load (os .path .join (s .processed_dir , f"{ kind } .pt" ), weights_only = False )
523525 for s in self .subsets
524526 ]
525527 dataset = [
@@ -1022,7 +1024,9 @@ def _retrieve_splits_from_csv(self) -> None:
10221024 splits_df = pd .read_csv (self .splits_file_path )
10231025
10241026 filename = self .processed_file_names_dict ["data" ]
1025- data = torch .load (os .path .join (self .processed_dir , filename ))
1027+ data = torch .load (
1028+ os .path .join (self .processed_dir , filename ), weights_only = False
1029+ )
10261030 df_data = pd .DataFrame (data )
10271031
10281032 train_ids = splits_df [splits_df ["split" ] == "train" ]["id" ]
@@ -1081,7 +1085,9 @@ def load_processed_data(
10811085
10821086 # If filename is provided
10831087 try :
1084- return torch .load (os .path .join (self .processed_dir , filename ))
1088+ return torch .load (
1089+ os .path .join (self .processed_dir , filename ), weights_only = False
1090+ )
10851091 except FileNotFoundError :
10861092 raise FileNotFoundError (f"File { filename } doesn't exist" )
10871093
0 commit comments