Loading lib/galaxy/datatypes/binary.py +4 −9 Original line number Diff line number Diff line Loading @@ -603,8 +603,7 @@ class BamNative(CompressedArchive, _BamOrSam): return f"Binary bam alignments file ({nice_size(dataset.get_size())})" def to_archive(self, dataset: DatasetProtocol, name: str = "", trans=None) -> Iterable: if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) rel_paths = [] file_paths = [] rel_paths.append(f"{name or dataset.file_name}.{dataset.extension}") Loading Loading @@ -703,15 +702,12 @@ class BamNative(CompressedArchive, _BamOrSam): headers = kwd.get("headers", {}) preview = util.string_as_bool(preview) if offset is not None: if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) return self.get_chunk(trans, dataset, offset, ck_size), headers elif to_ext or not preview: return super().display_data(trans, dataset, preview, filename, to_ext, **kwd) else: if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) column_names = dataset.metadata.column_names if not column_names: column_names = [] Loading Loading @@ -2114,8 +2110,7 @@ class H5MLM(H5): to_ext = to_ext or dataset.extension return self._serve_raw(dataset, to_ext, headers, **kwd) if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) out_dict: Dict = {} try: Loading lib/galaxy/datatypes/blast.py +1 −2 Original line number Diff line number Diff line Loading @@ -236,8 +236,7 @@ class _BlastDb(Data): **kwd, ) if data.dataset.object_store: data.dataset.object_store.update_cache(data.dataset, trans=trans) dataset.sync_cache(trans=trans) if self.file_ext == "blastdbn": title = "This is a nucleotide BLAST database" Loading lib/galaxy/datatypes/data.py +4 −10 Original line number Diff line number Diff line Loading @@ -452,15 +452,13 @@ class Data(metaclass=DataMeta): if dataset.datatype.composite_type or dataset.extension.endswith("html"): main_file = f"{name}.html" rel_paths.append(main_file) if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) file_paths.append(dataset.file_name) for fpath, rpath in self.__archive_extra_files_path(dataset.extra_files_path): rel_paths.append(os.path.join(name, rpath)) file_paths.append(fpath) else: if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) rel_paths.append(f"{name or dataset.file_name}.{dataset.extension}") file_paths.append(dataset.file_name) return zip(file_paths, rel_paths) Loading Loading @@ -542,10 +540,7 @@ class Data(metaclass=DataMeta): if filename and filename != "index": # For files in extra_files_path extra_dir = dataset.dataset.extra_files_path_name if dataset.dataset.object_store: dataset.dataset.object_store.update_cache( dataset.dataset, extra_dir=extra_dir, alt_name=filename, trans=trans ) dataset.sync_cache(extra_dir=extra_dir, alt_name=filename, trans=trans) file_path = trans.app.object_store.get_filename(dataset.dataset, extra_dir=extra_dir, alt_name=filename) if os.path.exists(file_path): if os.path.isdir(file_path): Loading Loading @@ -587,8 +582,7 @@ class Data(metaclass=DataMeta): raise ObjectNotFound(f"Could not find '{filename}' on the extra files path {file_path}.") self._clean_and_set_mime_type(trans, dataset.get_mime(), headers) if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) downloading = to_ext is not None Loading lib/galaxy/datatypes/isa.py +1 −2 Original line number Diff line number Diff line Loading @@ -284,8 +284,7 @@ class _Isa(Data): if not preview: return super().display_data(trans, dataset, preview, filename, to_ext, **kwd) if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) # prepare the preview of the ISA dataset investigation = self._get_investigation(dataset) Loading lib/galaxy/datatypes/protocols.py +3 −0 Original line number Diff line number Diff line Loading @@ -83,6 +83,9 @@ class DatasetProtocol( def get_converted_files_by_type(self, file_type): ... def sync_cache(self, **kwargs): ... def get_mime(self) -> str: ... Loading Loading
lib/galaxy/datatypes/binary.py +4 −9 Original line number Diff line number Diff line Loading @@ -603,8 +603,7 @@ class BamNative(CompressedArchive, _BamOrSam): return f"Binary bam alignments file ({nice_size(dataset.get_size())})" def to_archive(self, dataset: DatasetProtocol, name: str = "", trans=None) -> Iterable: if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) rel_paths = [] file_paths = [] rel_paths.append(f"{name or dataset.file_name}.{dataset.extension}") Loading Loading @@ -703,15 +702,12 @@ class BamNative(CompressedArchive, _BamOrSam): headers = kwd.get("headers", {}) preview = util.string_as_bool(preview) if offset is not None: if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) return self.get_chunk(trans, dataset, offset, ck_size), headers elif to_ext or not preview: return super().display_data(trans, dataset, preview, filename, to_ext, **kwd) else: if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) column_names = dataset.metadata.column_names if not column_names: column_names = [] Loading Loading @@ -2114,8 +2110,7 @@ class H5MLM(H5): to_ext = to_ext or dataset.extension return self._serve_raw(dataset, to_ext, headers, **kwd) if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) out_dict: Dict = {} try: Loading
lib/galaxy/datatypes/blast.py +1 −2 Original line number Diff line number Diff line Loading @@ -236,8 +236,7 @@ class _BlastDb(Data): **kwd, ) if data.dataset.object_store: data.dataset.object_store.update_cache(data.dataset, trans=trans) dataset.sync_cache(trans=trans) if self.file_ext == "blastdbn": title = "This is a nucleotide BLAST database" Loading
lib/galaxy/datatypes/data.py +4 −10 Original line number Diff line number Diff line Loading @@ -452,15 +452,13 @@ class Data(metaclass=DataMeta): if dataset.datatype.composite_type or dataset.extension.endswith("html"): main_file = f"{name}.html" rel_paths.append(main_file) if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) file_paths.append(dataset.file_name) for fpath, rpath in self.__archive_extra_files_path(dataset.extra_files_path): rel_paths.append(os.path.join(name, rpath)) file_paths.append(fpath) else: if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) rel_paths.append(f"{name or dataset.file_name}.{dataset.extension}") file_paths.append(dataset.file_name) return zip(file_paths, rel_paths) Loading Loading @@ -542,10 +540,7 @@ class Data(metaclass=DataMeta): if filename and filename != "index": # For files in extra_files_path extra_dir = dataset.dataset.extra_files_path_name if dataset.dataset.object_store: dataset.dataset.object_store.update_cache( dataset.dataset, extra_dir=extra_dir, alt_name=filename, trans=trans ) dataset.sync_cache(extra_dir=extra_dir, alt_name=filename, trans=trans) file_path = trans.app.object_store.get_filename(dataset.dataset, extra_dir=extra_dir, alt_name=filename) if os.path.exists(file_path): if os.path.isdir(file_path): Loading Loading @@ -587,8 +582,7 @@ class Data(metaclass=DataMeta): raise ObjectNotFound(f"Could not find '{filename}' on the extra files path {file_path}.") self._clean_and_set_mime_type(trans, dataset.get_mime(), headers) if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) downloading = to_ext is not None Loading
lib/galaxy/datatypes/isa.py +1 −2 Original line number Diff line number Diff line Loading @@ -284,8 +284,7 @@ class _Isa(Data): if not preview: return super().display_data(trans, dataset, preview, filename, to_ext, **kwd) if dataset.dataset.object_store: dataset.dataset.object_store.update_cache(dataset.dataset, trans=trans) dataset.sync_cache(trans=trans) # prepare the preview of the ISA dataset investigation = self._get_investigation(dataset) Loading
lib/galaxy/datatypes/protocols.py +3 −0 Original line number Diff line number Diff line Loading @@ -83,6 +83,9 @@ class DatasetProtocol( def get_converted_files_by_type(self, file_type): ... def sync_cache(self, **kwargs): ... def get_mime(self) -> str: ... Loading