Skip to the content

Automate Everything !

πŸ€– Explore with AI: ChatGPT Perplexity Claude Google AI Grok

For Enterprises | Teams | Start-Ups

eZintegrations

eZintegrations – AI Workflows & AI Agents Automation Hub

Automate to Innovate

0
$0.00
eZintegrations

eZintegrations – AI Workflows & AI Agents Automation Hub

Automate to Innovate

Menu
0
$0.00
  • Categories
    • Workflow Automation
    • AI Workflow
    • AI Agent
    • Agentic AI
  • Home
  • Automate Now !
  • About Us
  • Contact
  • Blog
  • Pricing
  • Free AI Workflow
  • Free AI Agents

eZintegrations

  • eZintegrations Introduction
  • Integration Bridge
    • Rename Integration Bridge
    • Enable and Disable Integration Bridge
    • Integration Bridge Save
    • Integration Bridge Run Once
    • Clear Logs of An Integration Bridge
    • Integration Bridge Share Feature
    • Copy Operation
    • Integration Bridge Import/Export
    • Integration Bridge Auto Save Feature
    • View An Integration Bridge
    • Copy Integration Bridge
    • Streaming Logs of Integration Bridge
    • Download Logs of An Integration Bridge
    • Status of Integration Bridge
    • Refresh an Integration Bridge
    • Stop An Integration Bridge
    • Start An Integration Bridge
    • Frequency
  • Feedback
    • Feedback: Tell Us What You Think
  • Understanding Session Timeout
    • Understanding Session Timeout and the Idle Countdown Timer
  • Alerts
    • Alerts
  • Marketplace
    • Marketplace
  • DIY Articles
    • 60+ Transformations for Smarter Data: How eZintegrations Powers Operations
    • From SOAP to GraphQL: Modernizing Integrations with eZintegrations
    • Accelerate Growth with eZintegrations Unified API Marketplace
    • Collaborative Integrations: Sharing Bridges in eZintegrations to Foster Cross-Team Innovation
    • Unlocking Hidden Value in Unstructured Data: eZintegrations AI Document Magic for Strategic Insights
    • Workflow Cloning Wizardry: Replicating Success with eZintegrations Integration Duplication for Rapid Scaling
    • Time Zone Triumph: Global Scheduling in eZintegrations for Synchronized Cross-Border Operations
    • Parallel Processing Power: eZintegrations Multi-Threaded Workflows for Lightning Fast Data Syncs
    • From Data Chaos to Competitive Edge: How eZintegrations AI Syncs Silos and Boosts ROI by 40%
    • From Emails to Insights: eZintegrations AI Turns Chaos into Opportunity
    • Handling XML Responses in eZintegrations
    • Text to Action: Shape Data with Plain English or Python in eZintegrations
    • AI Magic: Send Data to Any Database with a Simple English Prompt in eZintegrations
    • Configuring Netsuite as Source
    • Configuring Salesforce as Source
    • Overcoming Upsert Limitations: A Case Study on Enabling Upsert Operations in APIs without Inherent Support
    • Connecting QuickBooks to Datalake
    • Connecting Salesforce to Netsuite
    • Connecting My-SQL to Salesforce Using Bizdata Universal API
    • Effortless Integration Scheduling: Mastering Biweekly Execution with eZintegrations
    • Connecting MS-SQL or Oracle Database to Salesforce Using Bizdata Universal API
    • Establishing Token-Based Authentication within NetSuite
    • Registering a Salesforce App and Obtaining Client ID / Secret (for API Calls / OAuth)
  • Management
    • Adding Users and Granting Organization Admin Privileges : Step-by-Step Guide
    • Security Matrix
    • Adding Users as an Organization Admin (Step-by-Step Guide)
  • Appendix
    • Pivot Operation Use Cases
    • Efficient Column Renaming in eZintegration Using Python Operation
    • Filter Operation Use Cases
    • Connecting any Database to Database
    • Connecting Data Targets
    • Connecting Data Sources
  • Release Notes
    • Release Notes
  • Accounting & Billing
    • Invoices
    • Billing Information
    • Payment Method
    • Current Plan
    • Plans
    • Dashboard
  • My Profile
    • My Profile
  • OnBoarding
    • Microsoft Login
    • Multi-Factor Authentication
    • Login for New Users
  • Pycode Examples
    • Extract Domain Name from Email using Split
    • Split String with Regular Expression
    • Bulk Rename of Keys
    • Form a JSON Object from array of array
    • URL Parsing
    • Form a JSON Object based on the key and values available in JSON Dataset
    • Convert Empty String in a JSON to a “null” value
    • Generate a OAuth 1.0 Signature or Store a Code Response in a User Defined Variable
    • Rename JSON Key based on other key’s value
  • Sprintf
    • Sprintf
  • Data Source Management
    • Data Source Management
  • Data Source API
    • Response Parameters: Text, XML, and JSON Formats
    • Environment Settings for Reusable and Dynamic Configuration
    • API Numeric Parameters for Pagination and Record Limits
    • API Time Parameters for Date and Time Filtering
    • How to test the Data Source API
    • Pre- Request Scripts
      • Pre- Request Scripts for Amazon S3
      • Pre- Request Scripts for Oracle Netsuite
      • Pre-Request Script for Amazon SP API
      • Pre-Request Scripts
    • API Pagination Methods
      • Custom Pagination
      • Encoded Next Token Pagination
      • Cursor Pagination
      • Pagination with Body
      • Total Page Count Pagination
      • Offset Pagination
      • Next URL Pagination
      • API Pagination Introduction
      • Pagination examples
        • SAP Shipment API Pagination
        • Amazon SP API Pagination
    • API Authorization
      • OAuth 2.0 Authorization
      • OAuth 1.0 Authorization
      • Basic Authentication Method
      • API Key Authorization Method
      • Different Types of API Authorization
  • Console
    • Console: Check Your Data at Every Step
  • eZintegrations Dashboard Overview
    • eZintegrations Dashboard Overview
  • Monitoring Dashboard
    • Monitoring Dashboard
  • Advanced Settings
    • Advanced Settings
  • Summary
    • Summary
  • Data Target- Email
    • Data Target- Email
  • Data Target- Bizintel360 Datalake Ingestion
    • Data Target- Goldfinch Analytics Datalake Ingestion
  • Data Target- Database
    • Data Target – Database SQL Examples
    • Database as a Data Target
  • Data Target API
    • Response Parameters
    • REST API Target
    • Pre-Request Script
    • Test the Data Target
  • Bizdata Dataset
    • Bizdata Dataset Response
  • Data Source- Email
    • Extract Data from Emails
  • Data Source- Websocket
    • WebSocket Data Source Overview
  • Data Source Bizdata Data Lake
    • How to Connect Data Lake as Source
  • Data Source Database
    • How to connect Data Source Database
  • Data Operations
    • Deep Learning
    • Data Orchestration
    • Data Pipeline Controls
    • Data Cleaning
    • Data Wrangling
    • Data Transformation

Goldfinch AI

  • Goldfinch AI Introduction

Bizdata API

  • Universal API for Database
    • API for PostgreSQL Database – Universal API
    • API for Amazon Aurora Database (MySQL/Maria) – Universal API
    • API for Amazon Redshift Database – Universal API
    • API for Snowflake Database – Universal API
    • API for MySQL/Maria Database – Universal API
    • API for MS-SQL Database-Universal API
    • API for Oracle Database- Universal API
    • Introduction to Universal API for Databases
  • SFTP API
    • SFTP API
  • Document Understanding APIs
    • Document Understanding API- Extract data from Documents
  • Web Crawler API
    • Web Crawler API – Fast Website Scraping
  • AI Workflow Testing APIs
    • Netsuite Source Testing API (Netsuite API Replica)
    • Salesforce Testing API (Salesforce API replica)
    • OAuth2.0 Testing APIΒ 
    • Basic Auth Testing APIΒ 
    • No Auth Testing API
    • Pagination with Body Testing API
    • Next URL Pagination Testing APIΒ 
    • Total Page Count Pagination Testing API
    • Cursor Pagination Testing APIΒ 
    • Offset Pagination Testing API
  • Import IB API
    • Import Integration service with .JSON file
  • Linux File & Folder Monitoring APIs
    • Monitor Linux Files & Folder using APIs
  • Webhook
    • Webhook Integration-Capture Events in Real Time
  • Websocket
    • Websocket Integration- Fetch Real Time Data
  • Image Understanding
    • Image Understanding API – Extract data from Images

Goldfinch Analytics

  • Visualization Login
    • Enabling Two Factor Authentication
    • Visualization login for analytics users
  • Profile
    • Profile
  • Datalake
    • Datalake
  • Discover
    • Discover
  • Widgets
    • Filter
    • Widget List
    • Widgets Guide
    • Creating Widgets & Adding Widgets to Dashboard
  • Dashboard
    • Dashboard
  • Views
    • Views
  • Filter Queries
    • Filter Queries for Reports and Dashboard
  • Alerts
    • Alerts
  • Management
    • Management
  • Downloading Reports with Filtered Data
    • Downloading Reports with Filtered Data in Goldfinch Analytics
  • Downloads
    • Downloads – eZintegrations Documents & Resources | Official Guides & Manuals
View Categories

How to Connect Data Lake as Source

Overview

The Data Lake in eZintegrations is a search engine–based NoSQL database designed to store and process massive volumes of structured and unstructured data for analytics, storage, machine learning, and deep learning.

A Data Source in eZintegrations acts as a connection pool that retrieves data from the Data Lake and delivers it in JSON format to the integration pipeline.

Responses from the Data Lake source are stored under the bizdata_dataset_response key for further processing.

When to Use

Use Data Lake as a Source when large-scale analytical or operational data needs to be retrieved and processed within an Integration Bridge.

  • Extracting analytical datasets
  • Processing historical records
  • Streaming operational data
  • Supporting reporting workflows
  • Feeding machine learning pipelines

How It Works

The Data Lake Source retrieves records using JSON-based queries.

Data is streamed in chunks based on the configured size and pagination settings.

Retrieved records are stored in the bizdata_dataset_response key and passed to downstream operations and targets.

When using Single Line to Multiline Operations, the Chop key must be set to:

[‘bizdata_dataset_response’]

Data Lake Source Parameters

Data Lake Version

Specifies the Data Lake name and version assigned to the organization.

Index / Table Name

Defines the index or table from which data is retrieved.

Available indices and tables can be found in the Datalake section of the Visualization product.

Pagination Wait Time

Controls how long the system waits for the next page of data.

  • Default: 2m (2 minutes)
  • Supports: m (minutes), h (hours), s (seconds)
  • Increase for large responses or high network congestion

Timeout

Defines the maximum wait time for receiving a response.

  • Default: 2m
  • Increase when Data Lake response is slow
  • May be required for small cluster sizes

Size

Controls the number of records streamed per batch.

  • Default: 1000
  • Maximum: 10000
  • Recommended: 1000 for optimal performance

Query

Defines the JSON-based query used to retrieve records from the Data Lake.

Query Examples

Get All Records

{
  "query": {
    "match_all": {}
  }
}

Get Specific Columns

{
  "_source": ["store_number", "customer_number"],
  "query": {
    "match_all": {}
  }
}

Filter by Field Value

{
  "query": {
    "match": {
      "employee_id": 130
    }
  },
  "_source": {
    "includes": ["employee_id", "employee_name"]
  }
}

Filter with Multiple Conditions

{
  "size": 50,
  "sort": [{}],
  "_source": ["Project", "title", "Assigned To", "Priority", "Created By", "createdDateTime", "dueDateTime"],
  "query": {
    "bool": {
      "must": [
        { "query_string": { "query": "*" }},
        { "query_string": { "query": "Project:\"Project ABC\" AND Priority:[* TO *] AND NOT percentComplete:100" }},
        { "bool": { "should": [] }}
      ],
      "must_not": []
    }
  }
}

Key Names with Spaces

{
  "size": 1000,
  "sort": [{}],
  "_source": ["ThreadId", "Ticket Created At"],
  "query": {
    "bool": {
      "must": [
        { "query_string": { "query": "*" }},
        { "query_string": { "query": "NOT Status:\"Closed\" AND Thread\\ Type: \"create\"" }},
        { "bool": { "should": [] }}
      ],
      "must_not": []
    }
  }
}

Check for NULL Values

{
  "query": {
    "bool": {
      "must_not": {
        "exists": {
          "field": "asn"
        }
      }
    }
  }
}

Dynamic Filter Using Sprintf

{
  "query": {
    "bool": {
      "must": [
        { "term": { "ipAddress": "{%ipAddress%}" }}
      ],
      "must_not": [
        { "exists": { "field": "asn" }}
      ]
    }
  }
}

Limit and Sort Results

{
  "_source": ["asn", "as"],
  "query": {
    "bool": {
      "must": [
        { "term": { "ipAddress": "{%ipAddress%}" }},
        { "exists": { "field": "as" }}
      ]
    }
  },
  "size": 1,
  "terminate_after": 1,
  "sort": [
    {
      "_doc": { "order": "asc" }
    }
  ]
}

Frequently Asked Questions

What is Data Lake Source in eZintegrations?

It is a source connector that retrieves structured and unstructured data from the Goldfinch Analytics Data Lake.

Where is the response stored?

All retrieved data is stored under the bizdata_dataset_response key.

What is the recommended batch size?

The recommended size is 1000 records for balanced performance and reliability.

Can I use dynamic values in queries?

Yes. Dynamic values can be passed using Sprintf placeholders.

When should I increase timeout and pagination time?

Increase these values when working with large datasets, slow networks, or small cluster sizes.

Notes

  • Always validate queries before production deployment.
  • Use secure filters to avoid unnecessary data loads.
  • Optimize size and pagination for performance.
  • Monitor cluster capacity for large workloads.
  • Maintain consistent query structures across integrations.
Updated on February 20, 2026

What are your Feelings

  • Happy
  • Normal
  • Sad

Share This Article :

  • Facebook
  • X
  • LinkedIn
  • Pinterest
How to Connect Data Lake as SourceHow to Connect Data Lake as Source
Table of Contents
  • Overview
  • When to Use
  • How It Works
  • Data Lake Source Parameters
    • Data Lake Version
    • Index / Table Name
    • Pagination Wait Time
    • Timeout
    • Size
    • Query
  • Query Examples
    • Get All Records
    • Get Specific Columns
    • Filter by Field Value
    • Filter with Multiple Conditions
    • Key Names with Spaces
    • Check for NULL Values
    • Dynamic Filter Using Sprintf
    • Limit and Sort Results
  • Frequently Asked Questions
    • What is Data Lake Source in eZintegrations?
    • Where is the response stored?
    • What is the recommended batch size?
    • Can I use dynamic values in queries?
    • When should I increase timeout and pagination time?
  • Notes
Β© Copyright 2026 Bizdata Inc. | All Rights Reserved | Terms of Use | Privacy Policy