DataBoxDiskJobDetails Class
DataBox Disk Job Details.
Variables are only populated by the server, and will be ignored when sending a request.
All required parameters must be populated in order to send to Azure.
- Inheritance
-
azure.mgmt.databox.v2020_11_01.models._models_py3.JobDetailsDataBoxDiskJobDetails
Constructor
DataBoxDiskJobDetails(*, contact_details: _models.ContactDetails, shipping_address: _models.ShippingAddress | None = None, data_import_details: List[_models.DataImportDetails] | None = None, data_export_details: List[_models.DataExportDetails] | None = None, preferences: _models.Preferences | None = None, key_encryption_key: _models.KeyEncryptionKey | None = None, expected_data_size_in_tera_bytes: int | None = None, preferred_disks: Dict[str, int] | None = None, passkey: str | None = None, **kwargs: Any)
Keyword-Only Parameters
Name | Description |
---|---|
contact_details
|
Contact details for notification and shipping. Required. |
shipping_address
|
Shipping address of the customer. |
data_import_details
|
Details of the data to be imported into azure. |
data_export_details
|
Details of the data to be exported from azure. |
preferences
|
Preferences for the order. |
key_encryption_key
|
Details about which key encryption type is being used. |
expected_data_size_in_tera_bytes
|
The expected size of the data, which needs to be transferred in this job, in terabytes. |
preferred_disks
|
User preference on what size disks are needed for the job. The map is from the disk size in TB to the count. Eg. {2,5} means 5 disks of 2 TB size. Key is string but will be checked against an int. |
passkey
|
User entered passkey for DataBox Disk job. |
Variables
Name | Description |
---|---|
job_stages
|
List of stages that run in the job. |
contact_details
|
Contact details for notification and shipping. Required. |
shipping_address
|
Shipping address of the customer. |
delivery_package
|
Delivery package shipping details. |
return_package
|
Return package shipping details. |
data_import_details
|
Details of the data to be imported into azure. |
data_export_details
|
Details of the data to be exported from azure. |
job_details_type
|
Indicates the type of job details. Required. Known values are: "DataBox", "DataBoxDisk", and "DataBoxHeavy". |
preferences
|
Preferences for the order. |
copy_log_details
|
List of copy log details. |
reverse_shipment_label_sas_key
|
Shared access key to download the return shipment label. |
chain_of_custody_sas_key
|
Shared access key to download the chain of custody logs. |
key_encryption_key
|
Details about which key encryption type is being used. |
expected_data_size_in_tera_bytes
|
The expected size of the data, which needs to be transferred in this job, in terabytes. |
preferred_disks
|
User preference on what size disks are needed for the job. The map is from the disk size in TB to the count. Eg. {2,5} means 5 disks of 2 TB size. Key is string but will be checked against an int. |
copy_progress
|
Copy progress per disk. |
disks_and_size_details
|
Contains the map of disk serial number to the disk size being used for the job. Is returned only after the disks are shipped to the customer. |
passkey
|
User entered passkey for DataBox Disk job. |
Azure SDK for Python
Feedback
https://aka.ms/ContentUserFeedback.
Coming soon: Throughout 2024 we will be phasing out GitHub Issues as the feedback mechanism for content and replacing it with a new feedback system. For more information see:Submit and view feedback for