Amazon S3 Connector
Connect your S3 buckets to AI pipelines for seamless data integration. Sync files and data from AWS S3.
Connect your S3 buckets to AI pipelines for seamless data integration.
Setup Instructions
1. Navigate to Data Integrations
Open your flow and go to the Data Integrations tab.
2. Select S3 Integration
Click Add Connector, then search for Amazon S3 and select it.
3. Configure Basic Settings
Fill in the following fields:
- Connector Name: Give your S3 connector a descriptive name
- Bucket Name: Enter the name of your S3 bucket (e.g.,
my-bucket-name) - Region: Select the AWS region where your bucket is located (e.g.,
eu-central-1) - Folder (Optional): Specify a destination folder in the file manager
- If left empty, data will be stored in the root directory
4. Create AWS User and Access Keys
To connect securely, you need to create an IAM user with appropriate permissions:
- Go to the AWS Console and navigate to IAM (Identity and Access Management)
- Click on Users in the left sidebar
- Click Create User
- Enter a username (e.g.,
aicuflow-s3-connector) - Click Next to proceed to permissions
5. Set Up Permissions
Instead of adding the user to a group, we'll create an inline policy:
- Skip the group assignment and click Next
- After creating the user, click on the username to open the user details
- Go to the Permissions tab
- Click Add Permissions > Create Inline Policy
- Switch to the JSON tab
- Paste the following policy (replace
YOUR-BUCKET-NAMEwith your actual bucket name):
{
"Version": "2012-10-17",
"Statement": [
{
"Effect": "Allow",
"Action": [
"s3:ListBucket",
"s3:GetBucketLocation"
],
"Resource": "arn:aws:s3:::YOUR-BUCKET-NAME"
},
{
"Effect": "Allow",
"Action": [
"s3:GetObject"
],
"Resource": "arn:aws:s3:::YOUR-BUCKET-NAME/*"
}
]
}- Click Review Policy
- Give the policy a name (e.g.,
S3ReadAccess) - Click Create Policy
6. Generate Access Keys
- In the user details page, go to the Security Credentials tab
- Scroll down to Access Keys
- Click Create Access Key
- Select Application running outside AWS as the use case
- Click Next
- Add a description tag (optional)
- Click Create Access Key
- Important: Copy both the Access Key ID and Secret Access Key
- The secret key will only be shown once
- Store it securely
7. Complete the Connection
- Return to the connector setup in your flow
- Paste the Access Key ID in the corresponding field
- Paste the Secret Access Key in the corresponding field
- Configure sync settings:
- File Pattern (Optional): Specify patterns to filter files (e.g.,
*.csv,data/*.json) - Recursive: Enable to include files in subdirectories
- File Pattern (Optional): Specify patterns to filter files (e.g.,
- Click Create Connection
8. Monitor Sync Status
- Navigate to Data Synchronization to see the import progress
- Once complete, go to File Manager
- Your S3 data will appear in the specified folder
- You can now use this data in your AI pipelines and flows
Best Practices:
- Use specific file patterns to avoid importing unnecessary files
- Keep your AWS credentials secure and never share them
- Regularly rotate access keys for enhanced security
- Monitor your S3 usage to avoid unexpected costs