-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DIRT migration #301
Comments
A minimal version of this is working now. Triggered from the top-right dropdown in the UI. Remaining tasks:
It may also be worthwhile to let users select a folder to transfer their data to, or create a new one, rather than creating and using one with hard-coded path |
In as of 92d06ed |
Reopening because we should preserve collection names and collection & image metadata. This can be pulled from the DIRT database given an image path. Because files are stored by date on the SQL queriesThe Drupal CMS produces a fairly unwieldy database schema where every object is a
|
Depends on #312 |
Occasionally the Celery process running the migration gets killed for excess memory usage, e.g.: Might need to give the Celery container more memory Update: could be a Paramiko memory leak where the client and transport don't clean up after themselves properly. Currently we keep a client open for the entire migration and reuse it for every SFTP download. Might be able to resolve this by opening/closing a new client for each file transferred, at risk of slowing things down a bit due to overhead. |
Refactor in progress to use Celery's eventlet scheduler for non-blocking IO. This should dramatically speed up data transfer as we can perform a large number of file downloads/uploads in parallel instead of serially. |
Each DIRT user may have any number of image sets. We need to prompt them to transfer their image sets to correspondingly named folders in the data store, so that they can then use
plantit
to run DIRT. We can prompt the user via:We should also have a dedicated page for this in the documentation (walkthrough + screenshots).
When the user begins, we first detect if they have any DIRT image sets. If so, for each, transfer files from
tucco
’s attached NFS to a smaller temporary staging area (also an NFS) onportnoy
, then transfer to its own folder in the user’s home directory in the data store, with folder name as DIRT image set name and file names preserved. We should also decorate datasets with a metadata tag indicating DIRT origin, as well as any attached metadata. We should show some kind of progress monitor in the UI, then send an email notification to the user.Data transfer via:
iCommands container?IMPT: make a final backup of all the DIRT datasets before the migration period ends
The text was updated successfully, but these errors were encountered: