Upload a binary file to S3 using AWS SDK for Node.js

Source: https://stackoverflow.com/questions/13807339/upload-a-binary-file-to-s3-using-aws-sdk-for-node-js

https://www.mydatahack.com/uploading-and-downloading-files-in-s3-with-node-js/


You don't need to convert the buffer to a base64 string. Just set body to data and it will work.


AWS S3 is probably the most utilised AWS storage services. It is affordable, highly available, convenient and easy to use. To interact with any AWS services, Node.js requires AWS SDK for JavaScript.
Let’s first create a project folder called nodeS3 and install SDK. Then, create the main program file and data folder. In the data folder, drop any file you want. In this example, I am using a json file called data.json.
mkdir nodeS3
npm init -y
npm install aws-sdk
touch app.js
mkdir data
Next, you need to create a bucket for uploading a file (after configuring your AWS CLI). Let’s create a bucket with the s3 command.
aws s3 mb s3://your.bucket.name
Uploading File
First of all, you need to import the aws-sdk module and create a new S3 object. It uses the credentials that you set for your AWS CLI. Locking in API version for S3 object is optional. Here is the further document on the S3 class.
There are two methods you can use to upload a file, upload() and putObject(). Both methods are using different API calls. The major difference is upload() allows you to define concurrency and part size for large files while putObject() has lesser control. For a smaller file, both methods are fine. In general, I recommend to use upload().
Simple File Upload Example
In this example, we are using the async readFile function and uploading the file in the callback. As the file is read, the data is converted to a binary format and passed it to the upload Body parameter.
const AWS = require('aws-sdk');
const fs = require('fs');
var s3 = new AWS.S3();
const filePath = './data/data.json';
const bucketName = 'your.bucket.name';
const key = 'data/data.json';
const uploadFile = (filePath, bucketName, key) => {
fs.readFile(filePath, (err, data) => {
if (err) console.error(err);
var base64data = new Buffer(data, 'binary');
var params = {
Bucket: bucketName,
Key: key,
Body: base64data
};
s3.upload(params, (err, data) => {
if (err) console.error(`Upload Error ${err}`);
console.log('Upload Completed');
});
});
};
uploadFile(filePath, bucketName, key);
Downloading File
To download a file, we can use getObject().The data from S3 comes in a binary format. In the example below, the data from S3 gets converted into a String object with toString() and write to a file with writeFileSync method. Alternatively, you can create the stream reader on getObject method and pipe to a stream writer as described here.

const AWS = require('aws-sdk');
const fs = require('fs');
const filePath = './data/downloaded.json';
const bucketName = 'your.bucket.name';
const key = 'data/data.json';
var s3 = new AWS.S3();
const downloadFile = (filePath, bucketName, key) => {
const params = {
Bucket: bucketName,
Key: key
};
s3.getObject(params, (err, data) => {
if (err) console.error(err);
fs.writeFileSync(filePath, data.Body.toString());
console.log(`${filePath} has been created!`);
});
};
downloadFile(filePath, bucketName, key);

Không có nhận xét nào:

Is it okay to use both fetchpriority="high" and loading="eager" in img tag?

 https://stackoverflow.com/questions/77744344/is-it-okay-to-use-both-fetchpriority-high-and-loading-eager-in-img-tag Yes Fetchpriority and l...