Total 161 Questions
Last Updated On :
Preparing with Data-Cloud-Consultant practice test is essential to ensure success on the exam. This Salesforce SP25 test allows you to familiarize yourself with the Data-Cloud-Consultant exam questions format and identify your strengths and weaknesses. By practicing thoroughly, you can maximize your chances of passing the Salesforce certification spring 2025 release exam on your first attempt. Surveys from different platforms and user-reported pass rates suggest Data-Cloud-Consultant practice exam users are ~30-40% more likely to pass.
A customer wants to use the transactional data from their data warehouse in Data Cloud. They are only able to export the data via an SFTP site. How should the file be brought into Data Cloud?
A. Ingest the file with the SFTP Connector.
B. Ingest the file through the Cloud Storage Connector.
C. Manually import the file using the Data Import Wizard.
D. Use Salesforce's Dataloader application to perform a bulk upload from a desktop.
Explanation:
Salesforce Data Cloud supports ingesting external data from a variety of sources using connectors, and one of the supported methods is via SFTP (Secure File Transfer Protocol).
When a customer can only export data to an SFTP site, the best and most scalable solution is to use the:
→ SFTP Connector
This connector:
1. Automates ingestion of flat files (like CSVs) hosted on an SFTP server.
2. Supports scheduled ingestion, meaning files can be picked up regularly.
3. Enables data to flow directly into Data Lake Objects (DLOs) or Data Model Objects (DMOs) within Data Cloud.
🚫 Why not the other options?
B. Ingest the file through the Cloud Storage Connector
❌ This is used for platforms like Amazon S3, Google Cloud Storage, or Azure Blob Storage, not SFTP servers.
C. Manually import the file using the Data Import Wizard
❌ The Data Import Wizard is part of Salesforce Core (CRM), not Data Cloud. It doesn’t support Data Cloud ingestion and is not meant for large or scheduled data loads.
D. Use Salesforce's Dataloader application to perform a bulk upload from a desktop
❌ Dataloader is also for Salesforce Core (standard objects like Contacts, Leads), and doesn’t support ingestion into Data Cloud’s Data Lake or Data Model Objects.
📘 Reference:
Salesforce Help: Use the SFTP Connector in Data Cloud
Data Cloud Ingestion Guide:
“Use the SFTP connector to ingest files from an external SFTP server into Data Cloud on a scheduled basis.”
A consultant wants to build a new audience in Data Cloud. Which three criteria can the consultant include when building a segment? Choose 3 answers
A. Direct attributes
B. Data stream attributes
C. Calculated Insights
D. Related attributes
E. Streaming insights
Explanation:
When building a segment in Salesforce Data Cloud, the consultant can use the following criteria to define the audience:
1. Direct Attributes (Correct - A)
Definition: Fields directly stored on a Data Cloud object (e.g., Individual.Email, Account.Industry).
Use Case: Filtering based on explicit values (e.g., Country = "USA").
2. Calculated Insights (Correct - C)
Definition: Derived metrics (e.g., "Customer Lifetime Value," "Predicted Churn Score").
Use Case: Segmenting based on AI/analytics outputs (e.g., CLV > $1000).
3. Related Attributes (Correct - D)
Definition: Fields from connected objects (e.g., Individual → Cases → Case.Status).
Use Case: Filtering based on relationships (e.g., "Individuals with open Cases").
Why Not the Others?
B. Data stream attributes → These are raw, unprocessed data fields from sources (e.g., Kafka streams). They must first be mapped to the Data Model before segmentation.
E. Streaming insights → Real-time metrics (e.g., "Current Session Duration") are not directly used in segment logic (segments rely on processed data).
Key Takeaway:
Segments are built using structured data (direct/related attributes) and precomputed insights.
Raw streams require transformation before segmentation.
Reference:
Salesforce Help - Segment Builder
Exam Objective: Audience Segmentation.
A segment fails to refresh with the error "Segment references too many data lake objects (DLOS)". Which two troubleshooting tips should help remedy this issue? Choose 2 answers
A. Split the segment into smaller segments.
B. Use calculated insights in order to reduce the complexity of the segmentation query.
C. Refine segmentation criteria to limit up to five custom data model objects (DMOs).
D. Space out the segment schedules to reduce DLO load.
Explanation:
The error "Segment references too many Data Lake Objects (DLOs)" occurs in Salesforce Data Cloud when a segment query exceeds the limit of 50 DLOs referenced in a single query. This typically happens due to complex segmentation criteria involving multiple filters, nested segments, or exclusion criteria that pull in numerous DLOs. Below are the two correct troubleshooting tips to remedy this issue, along with detailed explanations:
A. Split the segment into smaller segments.
Why it works:
By dividing a large, complex segment into smaller segments with fewer filters, nested segments, or exclusion criteria, you reduce the number of DLOs referenced in each segment query. This helps stay within the 50-DLO limit per query. The smaller segments can then be used independently or as nested segments within a larger segment, or activated separately, depending on the use case.
How to implement:
In the Salesforce Data Cloud Segmentation interface, review the segment's configuration and identify filters or criteria that reference multiple DLOs. Break the segment into multiple smaller segments, each focusing on a subset of the criteria. For example, if a segment combines purchase history, demographic data, and engagement metrics, create separate segments for each category and combine them as needed.
Reference:
Salesforce documentation highlights splitting segments as a solution to reduce DLO references and avoid this error.
B. Use calculated insights in order to reduce the complexity of the segmentation query.
Why it works:
Calculated insights allow you to pre-process data using formulas to create derived attributes, reducing the need for complex filters or nested segments in the segmentation query. By consolidating multiple DLO references into a single calculated insight, you simplify the query and decrease the number of DLOs referenced.
How to implement:
In Data Cloud, navigate to the Calculated Insights interface and create a new insight. For instance, instead of using multiple filters to segment customers based on purchase history (e.g., total purchases, last purchase date), create a calculated insight that computes a single metric, such as Customer Lifetime Value (CLV). Use this insight as a filter in the segment, reducing the query's complexity and DLO references.
Reference:
Salesforce documentation recommends using calculated insights to simplify segmentation queries by replacing multiple filters with a single attribute.
Why the Other Options Are Incorrect:
C. Refine segmentation criteria to limit up to five custom data model objects (DMOs).
Why it’s incorrect:
The error pertains to Data Lake Objects (DLOs), not Data Model Objects (DMOs). The limit of 50 DLOs applies to both standard and custom DMOs, so restricting to five custom DMOs is not a valid or relevant solution. Additionally, the issue is about the number of DLOs referenced in the query, not specifically custom DMOs.
Reference:
Salesforce documentation clarifies that the 50-DLO limit applies broadly, making this option ineffective.
D. Space out the segment schedules to reduce DLO load.
Why it’s incorrect:
The error is caused by the segment query referencing too many DLOs, not by the concurrent load or scheduling of segment refreshes. Spacing out schedules may help with performance issues related to processing load, but it does not address the root cause of exceeding the DLO limit in the query.
Reference:
Salesforce documentation confirms that the error is due to query complexity, not scheduling or DLO load.
Additional Context and Best Practices:
Understanding DLOs:
Data Lake Objects (DLOs) are storage containers in Salesforce Data Cloud that hold ingested data before it’s mapped to Data Model Objects (DMOs). A segment query referencing multiple DLOs (e.g., through complex filters or joins) can hit the 50-DLO limit, triggering the error.
Proactive Troubleshooting:
Before creating complex segments, review the number of DLOs involved by checking the Data Streams and Data Model configurations in Data Cloud. Use the Data Explorer to understand the schema and relationships, ensuring efficient query design.
Testing and Validation:
After applying the remedies (splitting segments or using calculated insights), test the segment refresh to confirm the error is resolved. Monitor the segment’s performance to ensure it meets the desired business outcomes.
References:
Salesforce Help: Troubleshoot Segment Errors
Salesforce Help: Create a Calculated Insight
Salesforce Help: Create a Segment in Data Cloud
The Salesforce CRM Connector is configured and the Case object data stream is set up. Subsequently, a new custom field named Business Priority is created on the Case object in Salesforce CRM. However, the new field is not available when trying to add it to the data stream.
Which statement addresses the cause of this issue?
A. The Salesforce Integration User Is missing Rad permissions on the newly created field.
B. The Salesforce Data Loader application should be used to perform a bulk upload from a desktop.
C. Custom fields on the Case object are not supported for ingesting into Data Cloud.
D. After 24 hours when the data stream refreshes it will automatically include any new fields that were added to the Salesforce CRM.
Explanation:
When using the Salesforce CRM Connector to bring objects like Case into Salesforce Data Cloud, the data is pulled using the Salesforce Integration User (a system user created automatically or assigned for integration).
If a new custom field (like Business_Priority__c) is added to a Salesforce object (e.g., Case), and it doesn't appear in the list of available fields in the Data Stream UI, the most common cause is:
→ The integration user lacks Read access to that field.
Field-level security (FLS) controls which fields are visible to a user, including the integration user. If the new field doesn’t have Read permission for the Integration User’s profile or permission set, it will not be exposed through the connector.
🔧 How to fix it:
Go to Setup in Salesforce CRM.
Locate the Business Priority field on the Case object.
Click Set Field-Level Security.
Ensure the Integration User's profile has Read Access checked.
Revisit the Data Stream configuration in Data Cloud — the field should now appear.
🚫 Why not the other options?
B. The Salesforce Data Loader application should be used to perform a bulk upload from a desktop
❌ Irrelevant. Data Loader is for manual data uploads to core Salesforce, not for Data Cloud ingestion or metadata field visibility.
C. Custom fields on the Case object are not supported for ingesting into Data Cloud
❌ False. Custom fields are fully supported — as long as the integration user has access.
D. After 24 hours when the data stream refreshes it will automatically include any new fields that were added to the Salesforce CRM
❌ Misleading. The data stream does not auto-include new fields. Fields must be manually selected, and the field must be visible first, which depends on permissions.
📘 References:
Salesforce Help: Salesforce CRM Connector Field Permissions
Note: “New fields added to CRM must be granted field-level access to the Integration User before they appear in the Data Stream configuration.”
A customer requests that their personal data be deleted. Which action should the consultant take to accommodate this request in Data Cloud?
A. Use a streaming API call to delete the customer's information.
B. Use Profile Explorer to delete the customer data from Data Cloud.
C. Use Consent API to request deletion of the customer's information.
D. Use the Data Rights Subject Request tool to request deletion of the customer's information.
Explanation:
When a customer requests personal data deletion (e.g., under GDPR's "Right to Erasure" or CCPA), the proper method in Salesforce Data Cloud is:
1. Data Rights Subject Request Tool (Correct - D)
Purpose:
Officially processes deletion requests while ensuring compliance with privacy laws.
Automates the deletion across all linked systems (e.g., CRM, Marketing Cloud).
How it Works:
Submit the request via the tool (or API).
Data Cloud identifies all instances of the customer’s data (including Unified Profiles).
Deletes or anonymizes the data based on org policies.
Why It’s Best:
Auditable, compliant, and covers all Data Cloud integrations.
Why Not the Other Options?
A. Streaming API call → Manual deletions risk incomplete removal (e.g., missing linked records) and violate compliance workflows.
B. Profile Explorer → Only views data; does not support bulk/compliant deletions.
C. Consent API → Manages opt-in/opt-out, but not data deletion.
Key Takeaway:
Data Rights Subject Request Tool is the only Salesforce-recommended method for lawful erasure.
Manual deletions (e.g., APIs) risk compliance violations.
Reference:
Salesforce Help - Data Deletion Requests
Exam Objective: Privacy and Compliance.
Which consideration related to the way Data Cloud ingests CRM data is true?
A. CRM data cannot be manually refreshed and must wait for the next scheduled synchronization,
B. The CRM Connector's synchronization times can be customized to up to 15-minute intervals.
C. Formula fields are refreshed at regular sync intervals and are updated at the next full refresh.
D. The CRM Connector allows standard fields to stream into Data Cloud in real time.
Explanation:
✅ D. The CRM Connector allows standard fields to stream into Data Cloud in real time.
True. Salesforce's CRM Connector for Data Cloud supports real-time data streaming for many standard fields from Salesforce CRM. This allows changes made in CRM (like updates to records) to be reflected in Data Cloud in near real-time, enabling up-to-date insights and customer profiles.
❌ A. CRM data cannot be manually refreshed and must wait for the next scheduled synchronization.
False. While syncs are often scheduled, manual refreshes of CRM data are possible via the Data Streams configuration. Admins can trigger a manual sync when needed.
❌ B. The CRM Connector's synchronization times can be customized to up to 15-minute intervals.
False. While some connectors allow scheduled sync intervals, real-time data streaming bypasses the need for such interval-based syncing. Also, the 15-minute limit is not a fixed constraint across all sync types.
❌ C. Formula fields are refreshed at regular sync intervals and are updated at the next full refresh.
Partially true, but misleading. Formula fields in Salesforce are calculated in real time on the CRM side, but in Data Cloud, they do not automatically update unless the record is otherwise updated or a full refresh occurs. So while the statement touches on a technical truth, it does not represent the most relevant or complete truth in this context.
A customer notices that their consolidation rate is low across their account unification. They have mapped Account to the Individual and Contact Point Email DMOs. What should they do to increase their consolidation rate?
A. Change reconciliation rules to Most Occurring.
B. Disable the individual identity ruleset.
C. Increase the number of matching rules.
D. Update their account address details in the data source
Explanation
To address the issue of a low consolidation rate in account unification within Salesforce Data Cloud, we need to understand what consolidation rate means and how it relates to the unification process. The consolidation rate in Data Cloud refers to the percentage of records successfully matched and unified into a single profile (e.g., an Individual profile) during the identity resolution process. A low consolidation rate indicates that many records are not being matched, resulting in fragmented profiles. The customer has mapped the Account to the Individual and Contact Point Email Data Model Objects (DMOs), but the unification process is not effectively consolidating records.
C. Increase the number of matching rules.
Why it works:
In Salesforce Data Cloud, identity resolution relies on matching rules to determine when records from different data sources (e.g., Account and Contact Point Email) represent the same individual. Matching rules use attributes like email, name, phone, or other identifiers to link records. A low consolidation rate suggests that the existing matching rules are too restrictive or insufficient to identify matches across the data sources. By increasing the number of matching rules (e.g., adding rules for additional attributes like phone numbers, addresses, or alternate emails), the system can identify more matches, thereby increasing the consolidation rate.
How to implement:
In the Data Cloud interface, navigate to the Identity Resolution section and review the existing identity ruleset for the Individual DMO. Add new matching rules to include additional attributes from the Account and Contact Point Email DMOs, such as:
1. Exact match on alternate email addresses.
2. Fuzzy match on names (e.g., to account for variations like "John" vs. "Jon").
3. Match on other identifiers like phone numbers or loyalty IDs, if available. Ensure that the rules are prioritized appropriately (e.g., exact matches before fuzzy matches) to avoid over-consolidation. After updating, run the identity resolution process and monitor the consolidation rate in the Identity Resolution dashboard.
Reference:
Salesforce documentation emphasizes that adding more matching rules with relevant attributes increases the likelihood of record matching, improving the consolidation rate (Salesforce Help: Identity Resolution in Data Cloud).
Why the Other Options Are Incorrect:
A. Change reconciliation rules to Most Occurring.
Why it’s incorrect:
Reconciliation rules determine how conflicting data is resolved after records are matched (e.g., selecting the most recent or most occurring value for an attribute like email). These rules do not affect the matching process itself, which is the root cause of a low consolidation rate. Changing reconciliation rules to "Most Occurring" only impacts how unified profile attributes are populated, not the number of records matched.
Reference:
Salesforce documentation clarifies that reconciliation rules apply post-matching and do not influence the consolidation rate.
B. Disable the individual identity ruleset.
Why it’s incorrect:
Disabling the individual identity ruleset would prevent the unification process entirely, as identity rulesets define how records are matched and unified into Individual profiles. This would result in no consolidation at all, worsening the issue rather than improving the consolidation rate.
Reference:
Salesforce documentation states that identity rulesets are essential for unification, and disabling them halts the process.
D. Update their account address details in the data source.
Why it’s incorrect:
While improving data quality (e.g., updating account address details) can enhance matching accuracy, the question does not indicate that address details are part of the matching rules or that poor address data is causing the low consolidation rate. Without evidence that addresses are used in the matching rules, updating them is unlikely to directly address the issue. Additionally, this approach focuses on data source changes rather than the identity resolution configuration, which is more relevant to the consolidation rate.
Reference:
Salesforce documentation notes that data quality improvements help but must align with the attributes used in matching rules.
Additional Context and Best Practices:
Understanding Consolidation Rate:
The consolidation rate is calculated as the percentage of source records successfully unified into a single Individual profile. For example, if 1,000 records are ingested and 800 are unified into 200 profiles, the consolidation rate is 80%. A low rate indicates many records remain unmatched, creating duplicate or fragmented profiles.
Identity Resolution Process:
In Data Cloud, identity resolution involves:
1. Mapping: Linking data sources (e.g., Account, Contact Point Email) to DMOs.
2. Matching Rules: Defining criteria (e.g., email, name) to identify when records represent the same individual.
3. Reconciliation Rules: Resolving conflicts for unified profile attributes. A low consolidation rate typically points to issues with matching rules rather than mapping or reconciliation.
Optimizing Matching Rules:
When adding matching rules, consider:
1. Diversity of Attributes: Include multiple identifiers (e.g., email, phone, name) to capture more matches.
2. Fuzzy Matching: Use fuzzy matching for attributes like names to account for variations or typos.
3. Data Quality: Ensure data sources have consistent, clean data for the attributes used in matching rules (though this is secondary to adding rules).
Testing and Monitoring:
After adding matching rules, run the identity resolution process and check the consolidation rate in the Data Cloud Identity Resolution dashboard. Use the Data Explorer to analyze unmatched records and identify gaps in the matching criteria.
Potential Risks:
Adding too many or overly broad matching rules (e.g., fuzzy matching on common names without additional criteria) can lead to false positives, where unrelated records are incorrectly unified. Balance the number and specificity of rules to maintain accuracy.
References:
Salesforce Help: Identity Resolution in Data Cloud
Salesforce Help: Create and Manage Matching Rules
Salesforce Help: Monitor Identity Resolution
Which two requirements must be met for a calculated insight to appear in the segmentation canvas? (Choose 2 answers)
A. The metrics of the calculated insights must only contain numeric values.
B. The primary key of the segmented table must be a metric in the calculated insight.
C. The calculated insight must contain a dimension including the Individual or Unified Individual Id.
D. The primary key of the segmented table must be a dimension in the calculated insight.
Explanation:
To use a Calculated Insight (CI) in the Segmentation Canvas in Salesforce Data Cloud, the CI must be structured in a way that aligns with person-level segmentation. That means it must link to the same entity you're segmenting on — typically the Individual or Unified Individual.
To qualify for segmentation use, a Calculated Insight must:
✅ C. Contain a dimension including the Individual or Unified Individual Id
This is crucial because segmentation is usually performed on the Unified Individual table (or similar).
The Calculated Insight must contain a dimension (not a metric) that maps to the ID of the segmentation table, so the segment can join and apply the metric at the person level.
✅ D. The primary key of the segmented table must be a dimension in the calculated insight
The primary key (e.g., Unified_Individual_ID__c) acts as the linking field between the segmentation canvas and the calculated insight.
It must be a dimension so that the segmentation engine can filter and apply logic at the individual level.
🚫 Why not the other options?
A. The metrics of the calculated insights must only contain numeric values
❌ Not required. While metrics often are numeric (counts, sums, etc.), it's not a requirement that only numeric metrics exist. CIs can have other data types too — what matters is that dimensions are properly aligned.
B. The primary key of the segmented table must be a metric in the calculated insight
❌ Incorrect. The primary key must be a dimension, not a metric. Metrics are aggregated values; dimensions are grouping identifiers — and you group by the primary key, not aggregate it.
📘 Reference:
Salesforce Help: Use Calculated Insights in Segments
"To use a calculated insight in segmentation, include the Unified Individual ID or the Individual ID as a dimension."
Northern Trail Outfitters wants to be able to calculate each customer's lifetime value (LTV) but also create breakdowns of the revenue sourced by website, mobile app, and retail channels. How should this use case be addressed in Data Cloud?
A. Nested segments
B. Flow orchestration
C. Streaming data transformations
D. Metrics on metrics
Explanation:
To calculate customer lifetime value (LTV) with channel-specific revenue breakdowns (website, mobile app, retail), Metrics on Metrics is the ideal solution in Data Cloud. Here’s why:
1. Metrics on Metrics (Correct - D)
What It Does:
Allows layered calculations, where one metric (e.g., total revenue) is broken down into sub-metrics (e.g., revenue by channel).
Example:
Base Metric: Total Revenue (sum of all purchases).
Sub-Metrics:
Website Revenue (filtered by source = website).
Mobile App Revenue (filtered by source = mobile_app).
Retail Revenue (filtered by source = retail_store).
Why It Fits This Use Case:
Enables LTV to be calculated per customer while preserving channel attribution.
Supports dynamic segmentation (e.g., "High-LTV Mobile App Users").
Why Not the Other Options?
A. Nested segments → Useful for hierarchical audiences (e.g., "Premium Customers + Frequent Buyers"), but doesn’t calculate numeric breakdowns.
B. Flow orchestration → Coordinates processes (e.g., triggering campaigns), not metric calculations.
C. Streaming data transformations → Processes real-time data streams, but doesn’t aggregate historical metrics like LTV.
Key Takeaway:
Metrics on Metrics is the only feature that lets you:
1. Calculate LTV (a composite metric).
2. Slice it by channel (sub-metrics).
Critical for multi-touch revenue analysis.
Reference:
Salesforce Help - Metrics on Metrics
Exam Objective: Calculated Insights and Analytics.
Northern Trail Outfitters wants to use some of its Marketing Cloud data in Data Cloud. Which engagement channel data will require custom integration?
A. SMS
B. Email
C. CloudPage
D. Mobile push
Explanation:
To determine which engagement channel data from Marketing Cloud requires a custom integration to be used in Salesforce Data Cloud, we need to understand how Marketing Cloud data is integrated with Data Cloud and which channels are supported natively versus those requiring custom integration. Northern Trail Outfitters (NTO) wants to leverage Marketing Cloud data, and the engagement channels listed are SMS, Email, CloudPage, and Mobile Push.
C. CloudPage
Why it requires custom integration:
In Salesforce Data Cloud, native integrations with Marketing Cloud are provided through connectors like the Marketing Cloud Connector, which supports standard engagement channels such as Email, SMS, and Mobile Push (e.g., push notifications). These connectors allow seamless ingestion of data like email sends, opens, clicks, SMS messages, and push notification interactions into Data Cloud as Data Model Objects (DMOs) such as Contact Point Email, Contact Point Phone, or Engagement DMOs. However, CloudPages (custom landing pages created in Marketing Cloud for web-based interactions) are not natively supported by the Marketing Cloud Connector. To ingest CloudPage interaction data (e.g., page views, form submissions), a custom integration is required, typically involving APIs (e.g., Marketing Cloud REST API) or custom data extraction processes to pull CloudPage data into Data Cloud as Data Streams.
How to implement:
To integrate CloudPage data:
1. Use the Marketing Cloud REST API to extract CloudPage interaction data (e.g., page views, form submissions).
2. Transform the data into a format compatible with Data Cloud (e.g., JSON or CSV).
3. Ingest the data into Data Cloud using a Data Stream configured for a custom data source (e.g., via Amazon S3, SFTP, or API ingestion).
4. Map the data to appropriate DMOs, such as Engagement or Custom DMOs, for use in segmentation or activation.
Reference:
Salesforce documentation indicates that the Marketing Cloud Connector supports Email, SMS, and Mobile Push natively but does not include CloudPages, requiring custom integration for such data (Salesforce Help: Marketing Cloud Connector for Data Cloud).
Why the Other Options Are Incorrect:
A. SMS
Why it’s incorrect:
SMS data (e.g., message sends, deliveries, responses) is natively supported by the Marketing Cloud Connector for Data Cloud. The connector automatically ingests SMS engagement data into Data Cloud, mapping it to DMOs like Contact Point Phone or Engagement DMOs. No custom integration is needed.
Reference:
Salesforce documentation confirms that SMS data is included in the Marketing Cloud Connector’s native integration capabilities.
B. Email
Why it’s incorrect:
Email data (e.g., sends, opens, clicks, bounces) is natively supported by the Marketing Cloud Connector. The connector syncs email engagement data directly into Data Cloud, mapping it to DMOs such as Contact Point Email or Engagement DMOs, without requiring custom integration.
Reference:
Salesforce documentation lists Email as a core component of the Marketing Cloud Connector.
D. Mobile Push
Why it’s incorrect:
Mobile Push data (e.g., push notification sends, opens, interactions) is also natively supported by the Marketing Cloud Connector. The connector ingests Mobile Push data into Data Cloud, mapping it to relevant DMOs like Engagement or Contact Point Push, without the need for custom integration.
Reference:
Salesforce documentation includes Mobile Push as part of the standard data integration via the Marketing Cloud Connector.
Additional Context and Best Practices:
Marketing Cloud Connector Overview:
The Marketing Cloud Connector for Data Cloud enables seamless data flow between Marketing Cloud and Data Cloud for standard engagement channels (Email, SMS, Mobile Push). It ingests data such as customer profiles, campaign interactions, and engagement metrics, mapping them to DMOs for use in segmentation, activation, or analytics.
CloudPages in Marketing Cloud:
CloudPages are custom web pages hosted in Marketing Cloud, used for landing pages, preference centers, or forms. Interaction data from CloudPages (e.g., page visits, form submissions) is stored in Marketing Cloud but requires custom extraction (e.g., via APIs or Data Extensions) to be used in Data Cloud.
Custom Integration Considerations:
When implementing a custom integration for CloudPage data:
1. API Usage: Use Marketing Cloud’s REST API endpoints (e.g., /interaction/v1/events for tracking data) to retrieve CloudPage interactions.
2. Data Transformation: Ensure the extracted data is formatted to align with Data Cloud’s schema requirements (e.g., mapping to Individual, Engagement, or Custom DMOs).
3. Automation: Set up automated processes (e.g., using Salesforce Flow or an ETL tool) to regularly ingest CloudPage data into Data Cloud.
4. Security: Secure API calls with OAuth 2.0 authentication and ensure compliance with data privacy regulations.
Monitoring and Validation:
After setting up the integration, verify that CloudPage data appears in Data Cloud’s Data Explorer and is correctly mapped to DMOs. Test segmentation or activation scenarios to ensure the data is usable.
Alternative Approaches:
If custom integration is not feasible, consider storing CloudPage interaction data in Marketing Cloud Data Extensions and syncing those to Data Cloud via the connector, though this may still require some custom configuration.
References:
Salesforce Help: Marketing Cloud Connector for Data Cloud
Salesforce Help: Data Cloud Data Model Objects
Salesforce Developer Documentation: Marketing Cloud REST API
Page 2 out of 17 Pages |
Data-Cloud-Consultant Practice Test Home |