Live Webinar | Jan 15, 11 AM EST — Search, Import, Automate: How Enterprises Launch AI Workflows in Minutes. Register Now !

Skip to the content

Automate Everything !

🤖 Explore with AI: ChatGPT Perplexity Claude Google AI Grok

For Enterprises | Teams | Start-Ups

eZintegrations – AI Workflows & AI Agents Automation Hub

eZintegrations – AI Workflows & AI Agents Automation Hub

Automate to Innovate

0
$0.00
eZintegrations – AI Workflows & AI Agents Automation Hub

eZintegrations – AI Workflows & AI Agents Automation Hub

Automate to Innovate

Menu
0
$0.00
  • Categories
    • Workflow Automation
    • AI Workflow
    • AI Agent
    • Agentic AI
  • Home
  • Automate Now !
  • About Us
  • Contact
  • Blog
  • Free AI Workflow
  • Free AI Agents

eZintegrations

  • eZintegrations Introduction
  • Integration Bridge
    • Rename Integration Bridge
    • Enable and Disable Integration Bridge
    • Integration Bridge Save
    • Integration Bridge Run Once
    • Clear Logs of An Integration Bridge
    • Integration Bridge Share Feature
    • Copy Operation
    • Integration Bridge Import/Export
    • Integration Bridge Auto Save Feature
    • View An Integration Bridge
    • Copy Integration Bridge
    • Streaming Logs of Integration Bridge
    • Download Logs of An Integration Bridge
    • Status of Integration Bridge
    • Refresh an Integration Bridge
    • Stop An Integration Bridge
    • Start An Integration Bridge
    • Frequency
  • Feedback
    • Feedback: Tell Us What You Think
  • Understanding Session Timeout
    • Understanding Session Timeout and the Idle Countdown Timer
  • Alerts
    • Alerts
  • Marketplace
    • Marketplace
  • DIY Articles
    • 60+ Transformations for Smarter Data: How eZintegrations Powers Operations
    • From SOAP to GraphQL: Modernizing Integrations with eZintegrations
    • Accelerate Growth with eZintegrations Unified API Marketplace
    • Collaborative Integrations: Sharing Bridges in eZintegrations to Foster Cross-Team Innovation
    • Unlocking Hidden Value in Unstructured Data: eZintegrations AI Document Magic for Strategic Insights
    • Workflow Cloning Wizardry: Replicating Success with eZintegrations Integration Duplication for Rapid Scaling
    • Time Zone Triumph: Global Scheduling in eZintegrations for Synchronized Cross-Border Operations
    • Parallel Processing Power: eZintegrations Multi-Threaded Workflows for Lightning Fast Data Syncs
    • From Data Chaos to Competitive Edge: How eZintegrations AI Syncs Silos and Boosts ROI by 40%
    • From Emails to Insights: eZintegrations AI Turns Chaos into Opportunity
    • Handling XML Responses in eZintegrations
    • Text to Action: Shape Data with Plain English or Python in eZintegrations
    • AI Magic: Send Data to Any Database with a Simple English Prompt in eZintegrations
    • Configuring Netsuite as Source
    • Configuring Salesforce as Source
    • Overcoming Upsert Limitations: A Case Study on Enabling Upsert Operations in APIs without Inherent Support
    • Connecting QuickBooks to Datalake
    • Connecting Salesforce to Netsuite
    • Connecting My-SQL to Salesforce Using Bizdata Universal API
    • Effortless Integration Scheduling: Mastering Biweekly Execution with eZintegrations
    • Connecting MS-SQL or Oracle Database to Salesforce Using Bizdata Universal API
    • Establishing Token-Based Authentication within NetSuite
    • Registering a Salesforce App and Obtaining Client ID / Secret (for API Calls / OAuth)
  • Management
    • Adding Users and Granting Organization Admin Privileges : Step-by-Step Guide
    • Security Matrix
    • Adding Users as an Organization Admin (Step-by-Step Guide)
  • Appendix
    • Pivot Operation Use Cases
    • Efficient Column Renaming in eZintegration Using Python Operation
    • Filter Operation Use Cases
    • Connecting any Database to Database
    • Connecting Data Targets
    • Connecting Data Sources
  • Release Notes
    • Release Notes
  • Accounting & Billing
    • Invoices
    • Billing Information
    • Payment Method
    • Current Plan
    • Plans
    • Dashboard
  • My Profile
    • My Profile
  • OnBoarding
    • Microsoft Login
    • Multi-Factor Authentication
    • Login for New Users
  • Pycode Examples
    • Extract Domain Name from Email using Split
    • Split String with Regular Expression
    • Bulk Rename of Keys
    • Form a JSON Object from array of array
    • URL Parsing
    • Form a JSON Object based on the key and values available in JSON Dataset
    • Convert Empty String in a JSON to a “null” value
    • Generate a OAuth 1.0 Signature or Store a Code Response in a User Defined Variable
    • Rename JSON Key based on other key’s value
  • Sprintf
    • Sprintf
  • Data Source Management
    • Data Source Management
  • Data Source API
    • Response Parameters: Text, XML, and JSON Formats
    • Environment Settings for Reusable and Dynamic Configuration
    • API Numeric Parameters for Pagination and Record Limits
    • API Time Parameters for Date and Time Filtering
    • How to test the Data Source API
    • Pre- Request Scripts
      • Pre- Request Scripts for Amazon S3
      • Pre- Request Scripts for Oracle Netsuite
      • Pre-Request Script for Amazon SP API
      • Pre-Request Scripts
    • API Pagination Methods
      • Custom Pagination
      • Encoded Next Token Pagination
      • Cursor Pagination
      • Pagination with Body
      • Total Page Count Pagination
      • Offset Pagination
      • Next URL Pagination
      • API Pagination Introduction
      • Pagination examples
        • SAP Shipment API Pagination
        • Amazon SP API Pagination
    • API Authorization
      • OAuth 2.0 Authorization
      • OAuth 1.0 Authorization
      • Basic Authentication Method
      • API Key Authorization Method
      • Different Types of API Authorization
  • Console
    • Console: Check Your Data at Every Step
  • eZintegrations Dashboard Overview
    • eZintegrations Dashboard Overview
  • Monitoring Dashboard
    • Monitoring Dashboard
  • Advanced Settings
    • Advanced Settings
  • Summary
    • Summary
  • Data Target- Email
    • Data Target- Email
  • Data Target- Bizintel360 Datalake Ingestion
    • Data Target- Goldfinch Analytics Datalake Ingestion
  • Data Target- Database
    • Data Target – Database SQL Examples
    • Database as a Data Target
  • Data Target API
    • Response Parameters
    • REST API Target
    • Pre-Request Script
    • Test the Data Target
  • Bizdata Dataset
    • Bizdata Dataset Response
  • Data Source- Email
    • Extract Data from Emails
  • Data Source- Websocket
    • WebSocket Data Source Overview
  • Data Source Bizdata Data Lake
    • How to Connect Data Lake as Source
  • Data Source Database
    • How to connect Data Source Database
  • Data Operations
    • Deep Learning
    • Data Orchestration
    • Data Pipeline Controls
    • Data Cleaning
    • Data Wrangling
    • Data Transformation

Goldfinch AI

  • Goldfinch AI Introduction

Bizdata API

  • Universal API for Database
    • API for PostgreSQL Database – Universal API
    • API for Amazon Aurora Database (MySQL/Maria) – Universal API
    • API for Amazon Redshift Database – Universal API
    • API for Snowflake Database – Universal API
    • API for MySQL/Maria Database – Universal API
    • API for MS-SQL Database-Universal API
    • API for Oracle Database- Universal API
    • Introduction to Universal API for Databases
  • SFTP API
    • SFTP API
  • Document Understanding APIs
    • Document Understanding API- Extract data from Documents
  • Web Crawler API
    • Web Crawler API – Fast Website Scraping
  • AI Workflow Testing APIs
    • Netsuite Source Testing API (Netsuite API Replica)
    • Salesforce Testing API (Salesforce API replica)
    • OAuth2.0 Testing API 
    • Basic Auth Testing API 
    • No Auth Testing API
    • Pagination with Body Testing API
    • Next URL Pagination Testing API 
    • Total Page Count Pagination Testing API
    • Cursor Pagination Testing API 
    • Offset Pagination Testing API
  • Import IB API
    • Import Integration service with .JSON file
  • Linux File & Folder Monitoring APIs
    • Monitor Linux Files & Folder using APIs
  • Webhook
    • Webhook Integration-Capture Events in Real Time
  • Websocket
    • Websocket Integration- Fetch Real Time Data
  • Image Understanding
    • Image Understanding API – Extract data from Images

Goldfinch Analytics

  • Visualization Login
    • Enabling Two Factor Authentication
    • Visualization login for analytics users
  • Profile
    • Profile
  • Datalake
    • Datalake
  • Discover
    • Discover
  • Widgets
    • Filter
    • Widget List
    • Widgets Guide
    • Creating Widgets & Adding Widgets to Dashboard
  • Dashboard
    • Dashboard
  • Views
    • Views
  • Filter Queries
    • Filter Queries for Reports and Dashboard
  • Alerts
    • Alerts
  • Management
    • Management
  • Downloading Reports with Filtered Data
    • Downloading Reports with Filtered Data in Goldfinch Analytics
  • Downloads
    • Downloads – eZIntegrations Documents & Resources | Official Guides & Manuals
View Categories

How to Connect Data Lake as Source

Data Source: Overview of Data Lake

Data Lake is a search engine-based NO-SQL database. Users can ingest petabyte, zettabyte, yottabyte of records both structured and unstructured for Analytics, Storage, Machine Learning and deep learning.

Data Source is a connection pool in eZintegrations platform to retrieve data in JSON format.

Response from Data Lake source is stored in key bizdata_dataset_response. If you are using Single Line to Multiline Operations as the next operation in your pipeline, then the Chop key will have value as

['bizdata_dataset_response']

Goldfinch Analytics Data Lake Source has the following parameters:

Data Lake Version: Data Lake Version is the Data Lake Name and its version assigned to the organization.

Index / Table Name: Index or Table name that you want to retrieve data from Data source. For the list of table names or indices, check Datalake section in the Visualization product.

Pagination Wait Time: By default, it is 2m, where m is minute. Pagination is a standard API capability. Data source retrieves data in a paginated way. This parameter is to set how long to wait for the next page. If the response of a Table/Index is very high (having 100+ keys in a single record), try increasing the Pagination Wait Time. In general, 2m is sufficient to stream data. Use this when you have high network traffic congestion.

Can also use h for hours and s for seconds

Timeout: By default, it is 2m, where m is minute. In general, 2m is high enough to get a response from Data sou. Increase this when response from Data Lake is slow. This can happen when the Data Lake Cluster size is small. Reach out to support team to increase cluster size of Goldfinch Analytics Data Lake.

Can also use h for hours and s for seconds

Size: By default, it is 1000. Size is number of streaming record count from Data Lake source. The source will stream the records inside pipeline in chunks of 1000 and move them to operations and finally to Data Target. This can be increased to max of 10,000 records in cases such as one-time historical data loads.

For better performance and durability 1000 is recommended. This size helps to process 1000 records in target faster and ensures efficiency in real-time data processing.

Query: JSON Body based query to retrieve data from tables/index of Data Lake.

Get all the Records from a table

This example retrieves all the records from a table. Similar to SQL: select * from table

{
    "query": {
        "match_all": {}
    }
}

Get all Records with specific columns/keys from a table

Example: retrieves only store_number and customer_number. Similar to SQL:

select store_number, customer_number from table

{
    "_source": ["store_number", "customer_number"],
    "query": {
        "match_all": {}
    }
}

Get specific records with specific column/keys from a table

Example: retrieves only employee_id and employee_name where employee_id = 130. Similar to SQL:

select employee_id, employee_name from table where employee_id=130

{
    "query": {
        "match": {
            "employee_id": 130
        }
    },
    "_source": {
        "includes": ["employee_id", "employee_name"]
    }
}

Get Specific Columns/keys with Filters from a table

Similar to SQL:

select Project,title,Assigned To,Priority,Created By,createdDateTime,dueDateTime 
from table
where Project='Project ABC' 
and Priority is not null
and percentComplete=100
{
    "size": 50,
    "sort": [{}],
    "_source": ["Project", "title", "Assigned To", "Priority", "Created By", "createdDateTime", "dueDateTime"],
    "query": {
        "bool": {
            "must": [
                { "query_string": { "query": "*" }},
                { "query_string": { "query": "Project:\"Project ABC\" AND Priority:[* TO *] AND NOT percentComplete:100" }},
                { "bool": { "should": [] }}
            ],
            "must_not": []
        }
    }
}

Get Specific Columns/keys with Filters when key name contains spaces

Similar to SQL:

SELECT ThreadId, Ticket Created At
FROM YourTableName
WHERE Status != 'Closed' AND Thread Type = 'create'

Note: Thread Type becomes Thread\\ Type in JSON.

{
    "size": 1000,
    "sort": [{}],
    "_source": ["ThreadId", "Ticket Created At"],
    "query": {
        "bool": {
            "must": [
                { "query_string": { "query": "*" }},
                { "query_string": { "query": "NOT Status:\"Closed\" AND Thread\\ Type: \"create\"" }},
                { "bool": { "should": [] }}
            ],
            "must_not": []
        }
    }
}

SELECT * FROM table WHERE asn IS NULL

{
  "query": {
    "bool": {
      "must_not": {
        "exists": {
          "field": "asn"
        }
      }
    }
  }
}

SELECT id,ipAddress FROM table WHERE asn IS NULL

{
  "_source": ["id", "ipAddress"],
  "query": {
    "bool": {
      "must_not": {
        "exists": {
          "field": "asn"
        }
      }
    }
  }
}

SELECT * FROM table WHERE asn IS NULL AND ipAddress = ‘{%ipAddress%}’

{
  "query": {
    "bool": {
      "must": [
        { "term": { "ipAddress": "{%ipAddress%}" }}
      ],
      "must_not": [
        { "exists": { "field": "asn" }}
      ]
    }
  }
}

SELECT asn, as FROM table WHERE as IS NOT NULL AND ipAddress = ‘{%ipAddress%}’

{
  "_source": ["asn", "as"],
  "query": {
    "bool": {
      "must": [
        { "term": { "ipAddress": "{%ipAddress%}" }},
        { "exists": { "field": "as" }}
      ]
    }
  }
}

SELECT asn, as FROM table WHERE as IS NOT NULL AND ipAddress = ‘{%ipAddress%}’ TOP 1

{
  "_source": ["asn", "as"],
  "query": {
    "bool": {
      "must": [
        { "term": { "ipAddress": "{%ipAddress%}" }},
        { "exists": { "field": "as" }}
      ]
    }
  },
  "size": 1,
  "terminate_after": 1
}

SELECT asn, as FROM table WHERE as IS NOT NULL AND ipAddress = ‘{%ipAddress%}’ order record by ascending TOP 1

{
  "_source": ["asn", "as"],
  "query": {
    "bool": {
      "must": [
        { "term": { "ipAddress": "{%ipAddress%}" }},
        { "exists": { "field": "as" }}
      ]
    }
  },
  "size": 1,
  "terminate_after": 1,
  "sort": [
    {
      "_doc": { "order": "asc" }
    }
  ]
}
Updated on December 29, 2025

What are your Feelings

  • Happy
  • Normal
  • Sad

Share This Article :

  • Facebook
  • X
  • LinkedIn
  • Pinterest
Table of Contents
  • Data Source: Overview of Data Lake
    • Get all the Records from a table
    • Get all Records with specific columns/keys from a table
    • Get specific records with specific column/keys from a table
    • Get Specific Columns/keys with Filters from a table
    • Get Specific Columns/keys with Filters when key name contains spaces
    • SELECT * FROM table WHERE asn IS NULL
    • SELECT id,ipAddress FROM table WHERE asn IS NULL
    • SELECT * FROM table WHERE asn IS NULL AND ipAddress = '{%ipAddress%}'
    • SELECT asn, as FROM table WHERE as IS NOT NULL AND ipAddress = '{%ipAddress%}'
    • SELECT asn, as FROM table WHERE as IS NOT NULL AND ipAddress = '{%ipAddress%}' TOP 1
    • SELECT asn, as FROM table WHERE as IS NOT NULL AND ipAddress = '{%ipAddress%}' order record by ascending TOP 1
© Copyright 2026 Bizdata Inc. | All Rights Reserved | Terms of Use | Privacy Policy