Skip to content
Snippets Groups Projects
Commit a5649dc8 authored by Miroslav Kratochvil's avatar Miroslav Kratochvil :bicyclist:
Browse files

Merge branch 'jt-data-tranfer-atlas-hpc' into 'develop'

data transfer between HPC and Atlas

See merge request !562
parents 10236b9d 52afa157
No related branches found
No related tags found
2 merge requests!566[release] Regular merge of develop,!562data transfer between HPC and Atlas
Pipeline #85705 passed with warnings
---
layout: page
permalink: /external/integrity/transfer/atlas-hpc/
shortcut: integrity:transfer:atlas-hpc
redirect_from:
- /cards/integrity/transfer:atlas-hpc
- /external/integrity/transfer:atlas-hpc
- /integrity/transfer/atlas-hpc
- /external/external/integrity/transfer/atlas-hpc/
---
# Data transfer between Atlas and UL HPC Clusters
A recommended storage pattern is to have the master copy of data on Atlas (project folder) and only store data on the UL HPC Clusters temporarily for the required practical duration of computational analysis. The derived data and results should hereafter be transferred back to Atlas. This How-to Card describes the different methods to transfer data between Atlas and the UL HPC Clusters. The three recommended methods to transfer data are:
1. [Via laptop with ```scp``` or ```rsync```](#1. Via laptop using scp or rsync)
2. [Via dedicated Virtual Machine (VM)](#2. Via dedicated Virtual Machine (VM) using rsync)
3. [Via Large File Transfer (LFT)](#3. Via Large File Transfer (LFT))
Please refer to the dedicated knowledge bases to see how to [connect to UL HPC Clusters](https://hpc-docs.uni.lu/connect/access/) and to [mount Atlas](https://service.uni.lu/sp?id=kb_article_view&sysparm_article=KB0010233).
<img src="img/data-transfer-flow.png">
## 1. Via laptop using `scp` or `rsync`
When using the UL laptop to transfer data between UL HPC Clusters and Atlas, you must [mount Atlas via smb on laptop](https://service.uni.lu/sp?id=kb_article_view&sysparm_article=KB0010233) before using `scp` and `rsync` for the transfer. While both commands ensure a secure transfer of data between the UL HPC Clusters and Atlas, `rsync` may be much faster for handling large amounts of small files (which are transferred very quickly in batches), and for selective incremental updates of large datasets (it is capable to automatically transferring only the changed files, thus saving time).
* `scp`: transfers all files and directories.
* `rsync`: transfers only the files which differ between the source and the destination.
Please visit the [UL HPC documentation](https://hpc-docs.uni.lu/data/transfer/#data-transfer-tofromwithin-ul-hpc-clusters) to see how to use `rsync` and `scp`.
## 2. Via dedicated Virtual Machine (VM) using rsync
Data can be transferred via a dedicated VM, which can be requested via [ServiceNow](https://service.uni.lu/sp?id=sc_cat_item&table=sc_cat_item&sys_id=49956812db3fa010ca53454039961978).
Instead of transferring data between Atlas and UL HPC Clusters through the laptop as described above, the transfer will go through the dedicated VM. Once connected to the VM and mounted to Atlas, the ```rsync``` command can be used in the same way as described in the [UL HPC documentation](https://hpc-docs.uni.lu/data/transfer/#data-transfer-tofromwithin-ul-hpc-clusters). This method is recommended for **recurring transfers of very large datasets** that benefit from high-speed network connection between the VM and the HPC.
**Note**: For larger transfers between Atlas and UL HPC Clusters, you may want to run the operations in background using `screen` or `tmux`. These prevent interruption of data transfer in cases when your `ssh` connection gets interrupted.
## 3. Via Large File Transfer (LFT)
An alternative solution is to use LFT for transferring data between Atlas and UL HPC Clusters. This method can reliably transfer large data volumes (typically several terabytes). However, LFT can only be used if the data is already on LFT (e.g., received from external collaborators). In this case, you can make a copy of the data and directly download it to the UL HPC Clusters for computational analysis. Note that a **master copy of the data must still be manually uploaded to Atlas** for internal archival.
Please refer to the [dedicated How-to Card on LFT]({{'/?exchange-channels:lft' | relative_url }}) for detailed information.
**Note**: In cases when the analysis data are not already received via LFT, we strongly recommend to use one of the other (simpler) methods instead.
external/exchange-channels/atlas-hpc/img/data-transfer-flow.png

127 KiB

0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment