How to upload all files of a specific type to S3 Bucket?

10,843

Solution 1

When you fill in missing parameters like that from the command line, you need to specify their literal string values. When I mimicked your issue locally:

PS C:\> Write-S3Object

cmdlet Write-S3Object at command pipeline position 1
Supply values for the following parameters:
BucketName: MyTestBucketNameHere
Key: $testName
File: C:/test.txt

I ended up with a file on S3 whose key was named $testName, because variables aren't evaluated in that context. Likewise, you're getting this "The file indicated by the FilePath property does not exist!" error because there is no file in your filesystem named $f[0].fullName.

An example to write a single file to S3:

PS C:> Write-S3Object -BucketName "MyTestBucketName" -Key "file.txt" -File "C:/test.txt"

To write all of your files to S3:

PS C:\> (Get-ChildItem -filter "*.flv") | % { Write-S3Object -BucketName "MyTestBucketName" -File $_ -Key $_.name}

This will first get all files with the flv file-type in your current directory, and for each object (represented by the percent sign) we will write the file (represented by $_) to MyTestBucketName with a Key that is the name property of the current file being iterated.

Solution 2

The parameter -CannedACLName PublicRead is not correct, correct parameter is

foreach ($f in (Get-ChildItem -filter "*.flv")){ Write-S3Object -BucketName bucket.example -File $f.fullName -Key $f.name -CannedACLName public-read }

This fixed the issue for me and should fix for you as well.

Solution 3

For me, none of these solutions worked so what I found is this.

When you're in Powershell ISE (4.0) it doesn't matter how you send the local filename, it will know its location, which means you can simply do this, in my example I'm trying to upload all the backups of a folder named e:\Backups to a bucket in S3:

$results = Get-ChildItem -Path E:\Backups

foreach ($f in $results) { 

  $filename = [System.IO.Path]::GetFileName($f)

  Write-S3Object -BucketName bi-dbs-daily-backup -File $f -Key $filename

}

If I run this in Powershell ISE everything works fine, but If I create a .ps1 and try to run it with the Task Scheduler I get: The file indicated by the FilePath property does not exist!

It turns out that when Windows tries to run the .ps1 it does it basically from the CLI, and when it does, it stops recognizing the path of the files you're sending.

So I added | % { $_.FullName } to GetChildItem to get the fullpath of each file and now everything works fine:

$results = Get-ChildItem -Path E:\Backups **| % { $_.FullName }** 

foreach ($f in $results) { 

  $filename = [System.IO.Path]::GetFileName($f)

  Write-S3Object -BucketName bi-dbs-daily-backup -File $f -Key $filename

}

Hope this helps someone out there!

Share:
10,843
Admin
Author by

Admin

Updated on June 04, 2022

Comments

  • Admin
    Admin almost 2 years

    When I do this:

    foreach ($f in (Get-ChildItem -filter "*.flv")){
        Write-S3Object -BucketName bucket.example -File $f.fullName -Key $f.name -CannedACLName PublicRead
    }
    

    I get this error:

    Write-S3Object :
    At line:1 char:51
    +  foreach ($f in (Get-ChildItem -filter "*.flv")){ Write-S3Object -BucketName xx. ...
    + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
        + CategoryInfo          : InvalidOperation: (Amazon.PowerShe...eS3ObjectCmdlet:WriteS3ObjectCmdlet) [Write-S3Objec
       t], InvalidOperationException
        + FullyQualifiedErrorId : Amazon.S3.AmazonS3Exception,Amazon.PowerShell.Cmdlets.S3.WriteS3ObjectCmdlet
    

    What am I doing wrong? Is there anything I can do to see more of the error, or is this just a syntax issue?

    How can I otherwise upload all of a certain filetype to a bucket using powershell?

    EDIT:

    I intentionally set Set-DefaultAWSRegion to a region that the bucket wasn't in, and got

    Write-S3Object : The bucket you are attempting to access must be addressed using the specified endpoint. Please send all future requests to this endpoint. 
    

    as an error message, as expected, so it looks like it can connect to the bucket and it knows that it isn't in a certain region.

    Also, if I enter the s3:// prefix before the bucket name, I get a message that the bucket couldn't be found, so it it looks like what I'm entering now is correct.

    I can do Get-S3Bucket and see all of the buckets on my account, so I know that it's configured correctly.

    EDIT2:

    If I do:

    > $f = Get-ChildItem -filter "*.flv"
    > Write-S3Object
    
    cmdlet Write-S3Object at command pipeline position 1
    Supply values for the following parameters:
    BucketName: bucket.name
    Key: $f[0].name
    File: $f[0].fullName
    Write-S3Object : The file indicated by the FilePath property does not exist!
    At line:1 char:1
    + Write-S3Object
    + ~~~~~~~~~~~~~~
        + CategoryInfo          : InvalidOperation: (Amazon.PowerShe...eS3ObjectCmdlet:WriteS3ObjectCmdlet) [Write-S3Objec
       t], InvalidOperationException
        + FullyQualifiedErrorId : System.ArgumentException,Amazon.PowerShell.Cmdlets.S3.WriteS3ObjectCmdlet
    

    If I do $f[0].fullName seperately, I get the full path to the object. However, this has spaces in it. Could this be a problem?

  • Der_Meister
    Der_Meister over 8 years
    Thank you. The wrong name presents in Amazon docs: docs.aws.amazon.com/powershell/latest/userguide/… Correct names are listed on this page: docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html