amazonka-s3-streaming

Provides conduits to upload data to S3 using the Multipart API https://github.com/Axman6/amazonka-s3-streaming#readme

Version on this page:0.2.0.3
LTS Haskell 9.18:0.2.0.3
Stackage Nightly 2017-12-08:0.2.0.3
Latest on Hackage:0.2.0.3
BSD3 licensed by Alex Mason
Maintained by Alex.Mason@data61.csiro.au

Module documentation for 0.2.0.3

amazonka-s3-streaming Build Status

Provides a conduit based streaming interface and a concurrent interface to uploading data to S3 using the Multipart API. Also provides method to upload files or bytestrings of known size in parallel.

The documentation can be found on Hackage.

Changes

Changelog - amazonka-s3-streaming

0.2.0.3

  • Make all library generated messages use Debug level not Info

0.2.0.2

  • Update to mmorph < 1.2

0.2.0.1

  • Fixed a bug with the printf format strings which would lead to a crash (Thanks @JakeOShannessy for reporting).

0.2.0.0

  • Fixed a potential bug with very large uploads where the chunksize might be too small for the limit of 10,000 chunks per upload (#6).
  • Change API to allow the user to specify a chunk size for streaming if the user knows more about the data than we do.
  • Allow the user to specify how many concurrent threads to use for concurrentUpload as as well as chunk size (#4).
  • Better specify cabal dependency ranges.
comments powered byDisqus