Data Integrations
Connect to external databases, APIs, and file storage systems. Synchronize data automatically or import as a one-time operation.
Data Source: Integrations
Connect to external databases, APIs, and file storage systems. Data can be synchronized automatically or imported as a one-time operation.
There is the off the shelf connectors to external platforms, and there is client libraries that can be used to collect live data (eg. from sensors).
Available Connectors
We support a wide range of data sources with automatic synchronization capabilities. Click on any connector below to see detailed setup instructions:
General Purpose
- HTTP - Download data from any publicly accessible URL. Perfect for raw files, APIs, and web resources.
Databases & Storage
- MySQL - Connect MySQL databases directly to AI pipelines with automatic synchronization.
- PostgreSQL - Use Postgres data for AI pipelines with automatic synchronization.
- Amazon S3 - Connect S3 buckets for seamless data integration and file synchronization.
- Airtable - Use Airtable data for AI pipelines with automatic synchronization.
Code & Datasets
- GitHub - Connect repositories with automatic synchronization. Sync code, documentation, and repository data.
- Kaggle - Download datasets and competition data from Kaggle directly to your pipelines.
- Hugging Face - Access thousands of datasets from Hugging Face Hub for AI and ML projects.
Quick Start
To set up any connector:
- Navigate to Data Integrations in your flow
- Select an Integration from the available connectors
- Configure Connection with your credentials and settings
- Create Connection to start syncing data
- Monitor Sync Status in Data Synchronization
- Access Your Data in File Manager
Each connector has specific requirements and setup steps. Click on a connector above to see detailed instructions.
Data Sources
You can load data from different sources easily:
- Upload manually in the file browser (use the file manager tab in your flow)
- Use an integration to load data from another tool (see connector tab in your flow)
- Use sync jobs to continually synchronize data from integrations (see tab in your flow)
More technical methods:
- Use python code to write to a file on aicuflow
- Stream sensor data using the aicuflow C++ SDK
- Use the public api to send data via HTTP
- (experimental:) Use the custom code node to generate data
- Possibly sooner than you think: scraping
Responsible Developers: Julia, Maxim, Finn, previously Shivam.
File Manager
Store, preview, sync data securely with big data capabilities. The file manager module is where you upload your data before feeding it into your data flow.
GitHub Connector
Connect your GitHub repositories to AI pipelines with automatic synchronization. Sync code, documentation, and repository data.