Share via


Sample workflow to prepare hard drives for an import job

This article walks you through the complete process of preparing drives for an import job.

Sample data

This example imports the following data into an Azure storage account named mystorageaccount:

Location Description Data size
H:\Video\ A collection of videos 12 TB
H:\Photo\ A collection of photos 30 GB
K:\Temp\FavoriteMovie.ISO A Blu-Ray™ disk image 25 GB
\\bigshare\john\music\ A collection of music files on a network share 10 GB

Storage account destinations

The import job will import the data into the following destinations in the storage account:

Source Destination virtual directory or blob
H:\Video\ video/
H:\Photo\ photo/
K:\Temp\FavoriteMovie.ISO favorite/FavoriteMovies.ISO
\\bigshare\john\music\ music

With this mapping, the file H:\Video\Drama\GreatMovie.mov will be imported to the blob https://mystorageaccount.blob.core.windows.net/video/Drama/GreatMovie.mov.

Determine hard drive requirements

Next, to determine how many hard drives are needed, compute the size of the data:

12TB + 30GB + 25GB + 10GB = 12TB + 65GB

For this example, two 8TB hard drives should be sufficient. However, since the source directory H:\Video has 12TB of data and your single hard drive's capacity is only 8TB, you will be able to specify this in the following way in the driveset.csv file:

DriveLetter,FormatOption,SilentOrPromptOnFormat,Encryption,ExistingBitLockerKey
X,Format,SilentMode,Encrypt,
Y,Format,SilentMode,Encrypt,

The tool will distribute data across two hard drives in an optimized way.

Attach drives and configure the job

You will attach both disks to the machine and create volumes. Then author dataset.csv file:

BasePath,DstBlobPathOrPrefix,BlobType,Disposition,MetadataFile,PropertiesFile
H:\Video\,video/,BlockBlob,rename,None,H:\mydirectory\properties.xml
H:\Photo\,photo/,BlockBlob,rename,None,H:\mydirectory\properties.xml
K:\Temp\FavoriteVideo.ISO,favorite/FavoriteVideo.ISO,BlockBlob,rename,None,H:\mydirectory\properties.xml
\\myshare\john\music\,music/,BlockBlob,rename,None,H:\mydirectory\properties.xml

In addition, you can set the following metadata for all files:

  • UploadMethod: Windows Azure Import/Export service
  • DataSetName: SampleData
  • CreationDate: 10/1/2013

To set metadata for the imported files, create a text file, c:\WAImportExport\SampleMetadata.txt, with the following content:

<?xml version="1.0" encoding="UTF-8"?>
<Metadata>
    <UploadMethod>Windows Azure Import/Export service</UploadMethod>
    <DataSetName>SampleData</DataSetName>
    <CreationDate>10/1/2013</CreationDate>
</Metadata>

You can also set some properties for the FavoriteMovie.ISO blob:

  • Content-Type: application/octet-stream
  • Content-MD5: Q2hlY2sgSW50ZWdyaXR5IQ==
  • Cache-Control: no-cache

To set these properties, create a text file, c:\WAImportExport\SampleProperties.txt:

<?xml version="1.0" encoding="UTF-8"?>
<Properties>
    <Content-Type>application/octet-stream</Content-Type>
    <Content-MD5>Q2hlY2sgSW50ZWdyaXR5IQ==</Content-MD5>
    <Cache-Control>no-cache</Cache-Control>
</Properties>

Run the Azure Import/Export Tool (WAImportExport.exe)

Now you are ready to run the Azure Import/Export Tool to prepare the two hard drives.

For the first session:

WAImportExport.exe PrepImport /j:JournalTest.jrn /id:session#1  /sk:************* /InitialDriveSet:driveset-1.csv /DataSet:dataset-1.csv /logdir:F:\logs

If any more data needs to be added, create another dataset file (same format as Initial dataset).

For the second session:

WAImportExport.exe PrepImport /j:JournalTest.jrn /id:session#2  /DataSet:dataset-2.csv

Once the copy sessions have completed, you can disconnect the two drives from the copy computer and ship them to the appropriate Azure data center. You'll upload the two journal files, <FirstDriveSerialNumber>.xml and <SecondDriveSerialNumber>.xml, when you create the import job in the Azure portal.

Next steps