Cloud & Enterprise Nodes

Integrate with AWS, Azure, GCP, and enterprise platforms

Cloud & Enterprise Nodes

Integrate with major cloud providers to build scalable, enterprise-grade automation. These nodes connect you to AWS, Azure, GCP, and orchestrate complex enterprise workflows.


Cloud Integration Node

When to use: When you need to interact with AWS, Azure, or GCP services—upload files, invoke functions, query cloud databases, etc.

The Cloud Integration Node provides access to cloud platform services. Configure the provider and service, then perform the operation you need.

Configuration

Configuration:
  provider: aws | azure | gcp
  service: s3 | lambda | ec2 | blob | functions | storage
  operation: service-specific

Example 1: Upload File to AWS S3

When to use: Store files in the cloud for archiving, backups, or distribution.

Incoming data:

{
  "file_content": "base64-encoded-file-data",
  "filename": "invoice_2025_02_10.pdf",
  "bucket": "company-documents"
}

Cloud Integration Configuration:

provider: aws
service: s3
operation: put_object
bucket: "{{ input.bucket }}"
key: "invoices/{{ input.filename }}"
body: "{{ input.file_content }}"
content_type: "application/pdf"

Workflow:

  1. Generate or receive a file
  2. Cloud Integration uploads to S3
  3. Respond with download URL: https://s3.amazonaws.com/company-documents/invoices/invoice_2025_02_10.pdf

Tip: Use S3 for long-term storage, backups, and file distribution. Much cheaper than storing files in a database!

Example 2: Invoke AWS Lambda Function

When to use: Run serverless code—image processing, document analysis, complex calculations, etc.

Incoming data:

{
  "image_url": "https://example.com/photo.jpg",
  "resize_width": 300
}

Cloud Integration Configuration:

provider: aws
service: lambda
operation: invoke
function_name: "image-resizer"
payload: |
  {
    "image_url": "{{ input.image_url }}",
    "width": {{ input.resize_width }},
    "format": "webp"
  }
invocation_type: RequestResponse

Response from Lambda:

{
  "resized_image_url": "https://s3.amazonaws.com/processed-images/photo-300w.webp",
  "original_size": 2457600,
  "resized_size": 145920,
  "compression_ratio": 0.94
}

Example 3: Read Files from Cloud Storage

When to use: Process files stored in the cloud.

Cloud Integration Configuration (Azure):

provider: azure
service: blob
operation: get_blob
container: "input-files"
blob_name: "{{ input.filename }}"

Returns the file content, which you can then:

  • Parse with Data Parser
  • Process with AI
  • Insert into database
  • Transform and re-upload

Example 4: GCP Cloud Functions

When to use: Run Google Cloud functions for complex processing.

Cloud Integration Configuration:

provider: gcp
service: functions
operation: call
function_name: "process-document"
region: "us-central1"
payload:
  document_url: "{{ input.doc_url }}"
  output_format: "json"

Integration Automation Node

When to use: When you need to coordinate between multiple enterprise systems—ETL (extract, transform, load), orchestration, event-driven integration.

The Integration Automation Node handles complex multi-system workflows—syncing data between systems, triggering workflows based on events, etc.

Configuration

Configuration:
  type: etl | orchestration | events
  source: system-a
  target: system-b

Example 1: ETL (Extract, Transform, Load)

Scenario: Daily sync of customer data from Salesforce → Data Warehouse

Workflow:

Start (scheduled daily)
  ↓
Integration Automation (Extract)
  ├─ Query Salesforce for new/updated customers
  │  Return: 500 customer records
  │
  ↓
Data Transform (Transform)
  ├─ Normalize fields
  ├─ Validate data
  ├─ Enrich with additional fields
  │
  ↓
Integration Automation (Load)
  └─ Insert/update in data warehouse
     Return: 500 records processed

Configuration:

type: etl
source: salesforce
target: data_warehouse
operation: sync_customers

Example 2: Event-Driven Integration

Scenario: New Slack message → Log to database → Notify team

Workflow:

Integration Automation (Listen for Slack events)
  ├─ Event: message_posted
  │
  ↓
Extract data (channel, user, message, timestamp)
  │
  ↓
Database Insert (log message)
  │
  ↓
AI Agent (check if urgent)
  │
  ├─ If urgent → Send email alert
  └─ If normal → Just log

Example 3: Multi-System Orchestration

Scenario: Customer signup → Create account in 3 systems

Workflow:

Start (new customer signup)
  ├─ Split into 3 parallel paths
  │
  ├─ Path 1: Create in Salesforce
  │
  ├─ Path 2: Create in billing system
  │
  ├─ Path 3: Send welcome email
  │
  Merge (wait for all 3 to complete)
  │
  Log: "Customer onboarding complete"

Distributed Execution Node

When to use: When you need to process large workloads in parallel or distribute tasks across multiple workers.

The Distributed Execution Node breaks large jobs into smaller pieces, processes them in parallel, and aggregates results.

Configuration

Configuration:
  distribution_strategy: map-reduce | parallel | batch
  num_workers: 4
  timeout: 300000

Example 1: Distributed Image Processing

Scenario: Process 10,000 images in parallel

Incoming data:

{
  "image_urls": ["url1", "url2", "url3", ... "url10000"],
  "operation": "resize_and_compress"
}

Distributed Execution Configuration:

distribution_strategy: map-reduce
num_workers: 8
timeout: 600000

map_task: |
  For each image:
  - Download
  - Resize to 300x300
  - Compress to JPEG
  - Upload to S3

reduce_task: |
  Aggregate all results
  - Count total processed
  - Report any errors
  - Return summary

Result:

{
  "total_images": 10000,
  "processed": 9998,
  "failed": 2,
  "total_time_minutes": 12,
  "avg_time_per_image_ms": 72
}

Example 2: Distributed Data Analysis

distribution_strategy: parallel
num_workers: 4

task: |
  Analyze customer data in chunks:
  - Worker 1: Analyze customers 1-25000
  - Worker 2: Analyze customers 25001-50000
  - Worker 3: Analyze customers 50001-75000
  - Worker 4: Analyze customers 75001-100000

  Each worker computes:
  - Average lifetime value
  - Churn risk score
  - Product preferences

aggregate: |
  Combine results from all workers:
  - Global statistics
  - Top customer segments
  - Churn risk distribution

Common Cloud Patterns

Pattern 1: File Upload to Cloud Storage

Start (file received)
  ↓
Cloud Integration (upload to S3/GCS/Azure Blob)
  │
  ├─ Success → Generate download link
  │           ↓
  │           Email (send link to user)
  │
  └─ Error → Log error, notify admin

Pattern 2: Serverless Processing Pipeline

Webhook Trigger (image upload)
  ↓
Cloud Integration (trigger Lambda)
  │
  ├─ Lambda processes image
  │
  ↓
Cloud Integration (store result in S3)
  ↓
Database (log processing)
  ↓
Email (notify user)

Pattern 3: Multi-Cloud Data Sync

Start (daily)
  ├─ Path 1: Extract from AWS
  ├─ Path 2: Extract from Azure
  └─ Path 3: Extract from GCP
     ↓
     Merge (combine data)
     ↓
     Data Transform (normalize)
     ↓
     Database (store in data warehouse)

Pattern 4: Parallel Processing with Aggregation

Start (large job)
  ↓
Distributed Execution (split and process)
  │
  ├─ Worker 1 processes chunk 1
  ├─ Worker 2 processes chunk 2
  ├─ Worker 3 processes chunk 3
  └─ Worker 4 processes chunk 4Merge (collect all results)
     ↓
     Aggregate (combine stats)
     ↓
     Report

Cloud Cost Optimization

AWS

  • S3: Use lifecycle policies to move old files to Glacier (saves 90%)
  • Lambda: Set appropriate timeout/memory. More memory = faster = cheaper
  • EC2: Use spot instances for non-critical workloads

Azure

  • Blob Storage: Use hot/cool/archive tiers based on access patterns
  • Functions: Reserve capacity during peak hours

GCP

  • Cloud Storage: Use class-based storage (standard/nearline/coldline)
  • Cloud Functions: Optimize runtime and memory allocation

Tip: Always set resource limits and timeouts. Runaway processes cost money!


Enterprise Considerations

Security

  • Use VPC/Private networks for sensitive workflows
  • Enable encryption for data in transit and at rest
  • Regularly rotate credentials stored in DeepChain vault

Compliance

  • Log all data access for audit trails
  • Ensure data residency meets legal requirements
  • Use encryption keys compliant with HIPAA/PCI-DSS if needed

Scalability

  • Use Distributed Execution for large datasets
  • Enable caching to reduce API calls
  • Monitor performance metrics

Next Steps