@@ -169,10 +169,17 @@ def dataset_to_libdeeplake(hub2_dataset: Dataset):
169
169
path : str = hub2_dataset .path
170
170
171
171
token = (
172
- hub2_dataset .client .get_token ()
173
- if (hub2_dataset .token is None or hub2_dataset ._token == "" )
174
- and hub2_dataset .client
175
- else hub2_dataset .token
172
+ hub2_dataset .token
173
+ if hasattr (hub2_dataset , "token" ) and hub2_dataset .token is not None
174
+ else (
175
+ getattr (hub2_dataset , "_token" , None )
176
+ if hasattr (hub2_dataset , "_token" ) and hub2_dataset ._token != ""
177
+ else (
178
+ hub2_dataset .client .get_token ()
179
+ if hasattr (hub2_dataset , "client" ) and hub2_dataset .client
180
+ else ""
181
+ )
182
+ )
176
183
)
177
184
178
185
if hub2_dataset .libdeeplake_dataset is not None :
@@ -239,15 +246,18 @@ def dataset_to_libdeeplake(hub2_dataset: Dataset):
239
246
hub2_dataset .libdeeplake_dataset = libdeeplake_dataset
240
247
241
248
assert libdeeplake_dataset is not None
242
- if hasattr (hub2_dataset .storage , "cache_size" ):
243
- libdeeplake_dataset ._max_cache_size = max (
244
- hub2_dataset .storage .cache_size , libdeeplake_dataset ._max_cache_size
245
- )
246
- commit_id = hub2_dataset .pending_commit_id
247
- libdeeplake_dataset .checkout (commit_id )
248
- slice_ = hub2_dataset .index .values [0 ].value
249
- if slice_ != slice (None ):
250
- if isinstance (slice_ , tuple ):
251
- slice_ = list (slice_ )
252
- libdeeplake_dataset = libdeeplake_dataset [slice_ ]
249
+ try :
250
+ if hasattr (hub2_dataset .storage , "cache_size" ):
251
+ libdeeplake_dataset ._max_cache_size = max (
252
+ hub2_dataset .storage .cache_size , libdeeplake_dataset ._max_cache_size
253
+ )
254
+ commit_id = hub2_dataset .pending_commit_id
255
+ libdeeplake_dataset .checkout (commit_id )
256
+ slice_ = hub2_dataset .index .values [0 ].value
257
+ if slice_ != slice (None ):
258
+ if isinstance (slice_ , tuple ):
259
+ slice_ = list (slice_ )
260
+ libdeeplake_dataset = libdeeplake_dataset [slice_ ]
261
+ except INDRA_API .api .NoStorageDatasetViewError : # type: ignore
262
+ pass
253
263
return libdeeplake_dataset
0 commit comments