Product Docs
  • What is Dataworkz?
  • Getting Started
    • What You Will Need (Prerequisites)
    • Create with Default Settings: RAG Quickstart
    • Custom Settings: RAG Quickstart
    • Data Transformation Quickstart
    • Create an Agent: Quickstart
  • Concepts
    • RAG Applications
      • Overview
      • Ingestion
      • Embedding Models
      • Vectorization
      • Retrieve
    • AI Agents
      • Introduction
      • Overview
      • Tools
        • Implementation
      • Type
      • Tools Repository
      • Tool Execution Framework
      • Agents
      • Scenarios
      • Agent Builder
    • Data Studio
      • No-code Transformations
      • Datasets
      • Dataflows
        • Single Dataflows:
        • Composite dataflows:
        • Benefits of Dataflows:
      • Discovery
        • How to: Discovery
      • Lineage
        • Features of Lineage:
        • Viewing a dataset's lineage:
      • Catalog
      • Monitoring
      • Statistics
  • Guides
    • RAG Applications
      • Configure LLM's
        • AWS Bedrock
      • Embedding Models
        • Privately Hosted Embedding Models
        • Amazon Bedrock Hosted Embedding Model
        • OpenAI Embedding Model
      • Connecting Your Data
        • Finding Your Data Storage: Collections
      • Unstructured Data Ingestion
        • Ingesting Unstructured Data
        • Unstructured File Ingestion
        • Html/Sharepoint Ingestion
      • Create Vector Embeddings
        • How to Build the Vector embeddings from Scratch
        • How do Modify Existing Chunking/Embedding Dataflows
      • Response History
      • Creating RAG Experiments with Dataworkz
      • Advanced RAG - Access Control for your data corpus
    • AI Agents
      • Concepts
      • Tools
        • Dataset
        • AI App
        • Rest API
        • LLM Tool
        • Relational DB
        • MongoDB
        • Snowflake
      • Agent Builder
      • Agents
      • Guidelines
    • Data Studio
      • Transformation Functions
        • Column Transformations
          • String Operations
            • Format Operations
            • String Calculation Operations
            • Remove Stop Words Operation
            • Fuzzy Match Operation
            • Masking Operations
            • 1-way Hash Operation
            • Copy Operation
            • Unnest Operation
            • Convert Operation
            • Vlookup Operation
          • Numeric Operations
            • Tiles Operation
            • Numeric Calculation Operations
            • Custom Calculation Operation
            • Numeric Encode Operation
            • Mask Operation
            • 1-way Hash Operation
            • Copy Operation
            • Convert Operation
            • VLookup Operation
          • Boolean Operations
            • Mask Operation
            • 1-way Hash Operation
            • Copy Operation
          • Date Operations
            • Date Format Operations
            • Date Calculation Operations
            • Mask Operation
            • 1-way Hash Operation
            • Copy Operation
            • Encode Operation
            • Convert Operation
          • Datetime/Timestamp Operations
            • Datetime Format Operations
            • Datetime Calculation Operations
            • Mask Operation
            • 1-way Hash Operation
            • Copy Operation
            • Encode Operation
            • Page 1
        • Dataset Transformations
          • Utility Functions
            • Area Under the Curve
            • Page Rank Utility Function
            • Transpose Utility Function
            • Semantic Search Template Utility Function
            • New Header Utility Function
            • Transform to JSON Utility Function
            • Text Utility Function
            • UI Utility Function
          • Window Functions
          • Case Statement
            • Editor Query
            • UI Query
          • Filter
            • Editor Query
            • UI Query
      • Data Prep
        • Joins
          • Configuring a Join
        • Union
          • Configuring a Union
      • Working with CSV files
      • Job Monitoring
    • Utility Features
      • IP safelist
      • Connect to data source(s)
        • Cloud Data Platforms
          • AWS S3
          • BigQuery
          • Google Cloud Storage
          • Azure
          • Snowflake
          • Redshift
          • Databricks
        • Databases
          • MySQL
          • Microsoft SQL Server
          • Oracle
          • MariaDB
          • Postgres
          • DB2
          • MongoDB
          • Couchbase
          • Aerospike
          • Pinecone
        • SaaS Applications
          • Google Ads
          • Google Analytics
          • Marketo
          • Zoom
          • JIRA
          • Salesforce
          • Zendesk
          • Hubspot
          • Outreach
          • Fullstory
          • Pendo
          • Box
          • Google Sheets
          • Slack
          • OneDrive / Sharepoint
          • ServiceNow
          • Stripe
      • Authentication
      • User Management
    • How To
      • Data Lake to Salesforce
      • Embed RAG into your App
  • API
    • Generate API Key in Dataworkz
    • RAG Apps API
    • Agents API
  • Open Source License Types
Powered by GitBook
On this page
  • Step 1: Configure a New Salesforce Connection
  • Step 2: Establish Configuration between Dataworkz and Salesforce
  • Step 3: Write Back to Salesforce
  • Monitoring write back jobs:
  1. Guides
  2. How To

Data Lake to Salesforce

PreviousHow ToNextEmbed RAG into your App

Last updated 10 months ago

Are you grappling with the challenge of consolidating customer data from various sources and seamlessly incorporating it into Salesforce? If so, you’re not alone. The process can be complex, especially for those without a technical background. Fortunately, Dataworkz is here to simplify the entire procedure for business users.

We’ll guide you through the steps to effortlessly move data from a Data Lake like Google Cloud Storage (GCS) to Salesforce. By leveraging Dataworkz, you can efficiently combine customer data dumps from multiple sources and effortlessly push them into Salesforce. Better yet, by scheduling a recurring job, you’ll always have an up-to-date view of all your customer data within the Salesforce platform.

Let’s break down the process into three simple steps:

Step 1: Configure a New Salesforce Connection

  • Begin by configuring a new Salesforce connection to the GCS bucket from which you need to enable write backs.

  • Select your preferred authentication method.

  • Add the workspace and collection for the Salesforce app.

Upon completing the process, you will receive a prompt to sign in to your Salesforce developer account, and Dataworkz will securely save your configuration.

Once you return to the Dataworkz salesforce configuration page, you should see your salesforce connection:

Step 2: Establish Configuration between Dataworkz and Salesforce

Once connected, head to the Configuration tab. Click on the configuration name that has been created, and go to the writeback permissions tab. Here you can select the SFDC objects that you wish to support writeback. Follow these steps:

  • From the drop down select the SFDC objects for which you want to support writeback

  • Once you have configured the writeback permissions, click save

Step 3: Write Back to Salesforce

To write the datasets within the workspace and collection back to Salesforce, follow these simple steps:

  • Go to the GCS dataset within Dataworkz that you want to write back to Salesforce

  • Find the “Transform” option in the top right of the dataset screen, where you can perform any necessary transformation functions on the dataset before pushing it to SFDC

  • After defining all of the transformations, click “Execute” in the top right corner and define the target.

  • Find the SaaS destination workspace, and the collection will be the name of the salesforce configuration that was given, in this case ‘dw_writebacks_sfdc’.

  • Map the unique identifier key, and then map all existing columns from the dataset that you would like to move to salesforce

  • Choose the frequency on which you want the job to run – you can either run this as a one time job, or you can run this as a recurring job. For this example we set the recurring frequency to be once every 24 hours

  • After completing the process, click “submit,” and Dataworkz will write your data back into Salesforce. Because the job is configured as recurring, any new data that arrives will automatically be written back to the SFDC entity once per day.

Monitoring write back jobs:

  • Utilize the monitoring screen to retrieve details about all the records written to Salesforce. In the event of errors, Dataworkz captures the detailed error message along with the data passed to Salesforce for follow-up action. In this example, 5 out of the 9 records being updated in Salesforce failed, and you can view the input message along with the detailed error message for one of the failed records.

By following these steps, you can seamlessly integrate your data between GCS and Salesforce, ensuring a smooth and up-to-date flow of information. Dataworkz empowers business users to manage this process efficiently, making Salesforce data integration a breeze no matter the data source.