Home
Europe/Paris
Work

Luxury CDP on GCP: 18M+ contacts, 32 countries, global marketing activation

CDP on GCP for 18M+ contacts across 32 countries: how to augment an existing Adobe Campaign repository with enrichment, advanced segmentation, and multi-country governance.
CDP & Platform
March 9, 2026· 10 min read
Luxury CDP on GCP: 18M+ contacts, 32 countries, global marketing activation
Tech stack
GCPBigQueryAirflowPythonSQLAdobe CampaignDataikuPower BI
In large organizations, the challenge isn't finding yet another tool. It's making systems work together. Systems that have accumulated over time, each with its own local logic, legacy usage patterns, and trade-offs that have become invisible. Customer data exists in abundance, but it often remains fragmented across channels, markets, and teams, without a foundation coherent enough to activate quickly and steer effectively. It's in this context that I help structure a global CDP platform for an international luxury beauty house. The scope covers more than 18 million contacts, 32 countries, over €500M in e-commerce, and marketing use cases tied to more than €1B in global sales. The goal isn't to rip and replace the existing stack overnight. It's to make customer data actionable, accelerate marketing activation, better govern data flows, and establish governance that holds at this scale. The guiding principle throughout: keep what works on the business side, and add the data depth and industrial rigor that's missing. A common challenge in retail is that a single customer can exist multiple times depending on their entry channel. E-commerce, in-store, marketing forms, event-driven collection, or any number of local touchpoints. On top of that, partial histories, consent managed differently across countries, and data quality that varies from one channel and market to the next. From one country to another, the legal bases for collection, opt-in mechanisms, and retention rules diverge. Forms were designed at different times, with uneven levels of granularity. Harmonizing these practices without breaking the existing setup or slowing down local teams requires deep collaboration with legal and DPO teams, particularly on the multi-country opt-in strategy. On the quality side, we don't pretend we can fix everything at the source. We also implement controls within the platform itself: volume consistency checks, anomaly detection on key attributes, and alerting when drift occurs. This automated data quality monitoring secures downstream usage without depending on a uniform level of maturity across all markets. So this isn't just about centralization. We have to work with a single customer repository already in place within Adobe Campaign, a repository that remains the operational backbone of marketing activation. Meanwhile, business expectations have shifted. Teams no longer just want to execute campaigns. They want to segment, enrich, personalize, measure, and do it fast. We have neither a truly homogeneous global CIAM nor a sufficiently robust data quality framework at the source across the full scope. The platform must therefore absorb some of this complexity, without pretending it doesn't exist. The key trade-off on the project is to acknowledge Adobe Campaign for what it truly is, the legacy customer repository and the central marketing activation engine. Google Cloud Platform isn't meant to replace that foundation. Its role is to augment it. GCP is where we structure the processing chain, centralize flows, enrich profiles, produce calculated attributes, prepare analytical data, and feed back into Adobe Campaign whatever needs to be activated on the marketing side. This approach preserves the existing business setup while adding the computing power, analytical depth, and industrialization discipline that Adobe Campaign can't carry alone. Adobe Campaign stays central for orchestration and activation. Asking it to also carry advanced consolidation logic, massive enrichment, analytical computation, and cross-channel structuring at this scale hits its architectural limits fast. The right call is to let it keep its role on the activation side, while offloading preparation, computation, and enrichment to GCP. On the technical side, the platform runs on Google Cloud Platform, with BigQuery as the analytical backbone, Airflow for orchestration, Python and SQL for processing, and a group-level framework capable of auto-generating DAGs from configuration files. At this scale, every need treated as a bespoke pipeline quickly becomes maintenance debt. Standardized DAG generation delivers consistency, faster execution, and the ability to scale. On a platform processing over 10 million events per day across 32 markets, a broken pipeline on Friday evening can block a campaign on Monday morning. The architecture follows a simple, readable logic. The RAW layer absorbs incoming flows without trying to clean them up too early. It guarantees traceability, historization, and the ability to replay processing. The ODS layer normalizes the operational layer: formats, naming conventions, consent fields, identifiers, and data structures that vary across markets. The Data Warehouse layer handles consolidation, analytical preparation, and enrichments. This is where data becomes truly actionable, because it gains stability, readability, and practical value. Finally, the datamarts expose use-case-oriented views, particularly for BI, marketing, and certain analytical use cases. This prevents business tools from plugging directly into overly complex technical layers and limits the proliferation of ad hoc exports, always charming until the day no one knows where the numbers come from. As tables, calculated attributes, and segments multiply, discoverability becomes a central concern. Implementing a data catalogue with descriptions, owners, lineage, and certification levels allows marketing users to find and understand available data without systematically going through the data team. GCP primarily enables a clear separation between source data and derived data. BigQuery provides the computing power to produce enrichments at scale without burdening the operational repository. Airflow ensures reproducibility and traceability of processing. In practice, this creates a distinction between what we collect and what we compute. A date of birth is collected data. Age is calculated data. A purchase history is source data. A VIC segmentation or a time-to-repurchase indicator are derived objects. Consent is collected and regulated information. A score or a behavioral segmentation results from analytical processing.
Collected data• Email · date of birth• Consents · transactions
Derived data• Age · RFM · VIC segments• Time to repurchase
Predicted data• Churn score · purchase propensity• Predictive LTV
This distinction changes a lot. On the governance side, it prevents mixing raw attributes, enrichments, scores, and activable segments. On the business side, it makes clear what comes from collection, what is calculated, and what is ready to be activated. Most importantly, it enables a clean enrichment model for the existing repository: Adobe Campaign keeps its operational role, while GCP produces the calculated attributes, enrichments, and new segmentations that are then fed back into Adobe Campaign for use in campaigns.
Collected data• Profile · email · date of birth• Transactions, consents, events
Calculated data• Identity resolution · RFM · VIC• Time to Purchase · LTV scoring
Activable data• 100+ segments · personalization• Triggers · multichannel campaigns
The platform also paves the way for more advanced segmentation and data science use cases, particularly through Dataiku. The goal isn't to produce models that feed committee slide decks, it's to create activable objects. The platform enables the calculation of new segmentations, the production of derived attributes, the generation of propensity scores and indicators, and the injection of these results back into Adobe Campaign for targeting, personalization, and orchestration. This is the context in which we work on enrichments such as VIC segmentations, time-to-purchase logic, time-to-send, and time-to-repurchase. Computation sophistication matters, but only if the output actually reaches the marketing engine. An enriched data point isn't worth much if it sits in a corner of the platform gathering dust. It gains value when it flows back into Adobe Campaign in a directly usable form. The platform isn't designed solely for marketing. The datamarts exposed in GCP also feed Power BI, with a more stable approach to reporting and steering. In practice, marketing teams, market directors, and the executive committee can view dashboards built on the same definitions used in campaigns and segmentations. An "active customer" in Power BI corresponds to the same scope as the "active customer" targeted in Adobe Campaign. The datamarts cover several dimensions: performance by market, cohort analysis, opt-in rate tracking by channel and country, and customer lifecycle management. In many organizations, BI, marketing automation, and analytical use cases evolve in parallel, with different datasets and sometimes diverging definitions. Steering meetings then turn into debates about the numbers rather than decisions about actions. One of the project's key contributions is precisely to connect these components so that the figures in BI, in campaigns, and in analytics tell the same story. The project's value doesn't rest solely on the stack. It also depends on the framework built around it. A significant part of my role involves structuring the platform's governance: documentation, naming conventions, construction rules, reusable delivery patterns, and a more industrial approach to evolving data flows. Before this, many processes were handled on an ad hoc basis, sometimes manually, with practices varying across teams and needs. This way of working can survive for a while. It doesn't hold up across 32 countries. Compliance follows the same logic. I work with DPO, legal, and business teams to evolve collection forms, prepare staff, and clarify data usage rules according to local contexts, as part of a privacy program designed as an operational discipline rather than a one-off exercise. In this type of project, well-designed compliance isn't a hindrance. It's primarily a way to avoid industrializing chaos. And, if you know the right levers and local specificities, compliance can even become an accelerator. The project established a robust foundation for:
18M+ contactsActivated across 32 countries
10M+ events/dayIngested and processed daily
50+ marketing usersCreating 100+ segments/month
< 24h time-to-marketDown from 3–7 days
The most visible change is in time-to-market. Previously, launching a new segmented campaign went through the data team: extraction, preparation, qualification, then manual transfer to Adobe Campaign. Count 3 to 7 days. Now, segments are directly available in the activation engine and business teams create them themselves. But beyond the KPIs, the real change lies elsewhere. The platform is no longer a stack of ad hoc flows and one-off processes. It's a documented and governed system that marketing, BI, and analytics rely on daily. In short, it no longer just exists. It delivers. If you're facing a similar challenge like structuring a customer data platform, industrializing marketing flows, or implementing governance that holds at scale, feel free to get in touch.
Share this post:
Related projects
Customer 360°, data pipelines, cross-channel activation, segmentation, and privacy. Delivered in production at scale for retail and luxury.

38M+

Contacts Managed

€5B+

Managed Revenue

32+

Countries

7+

Years Data Platform Expertise

Luxury Group Opt-in: Multi-Brand Data Sharing and Legal Coordination

Privacy & Compliance
Group-level data sharing opt-in optimization across luxury brands worldwide with language element challenges and legal/technical team coordination.

Data Quality Monitoring: Automated Testing and Time-to-Market Alerting

Data Engineering
Automated data testing for developer time-to-market optimization, automated verification of data engineer work, and real-time alerting on quality anomalies.

Data Team Workflow: Agile, CI/CD, and Time-to-Market Optimization

Data Engineering
Development workflows with data delivery manager and data engineers: agile/kanban processes, priority management, automated testing, CI/CD, and time-to-market optimization.

Data Enrichment Campaigns: Progressive Profiling and Consent-Aware Strategy

Marketing Activation
Marketing/service campaigns for customer data enrichment with targeting by country/brand, service vs. marketing message arbitration based on consents, and profile completion optimization.

Customer 360° Datamarts: Single Source of Truth and Executive Dashboards

BI & Analytics
Implementation of datamarts and BI dashboards for large luxury company with single source of truth, business domain architecture, and executive steering dashboards.

Data Catalogue: Discoverability, Governance, and Self-Service Adoption

Data Governance
Implementation of data catalogue for large luxury company with business/technical definition alignment, adoption strategy, and improved data discoverability and quality.

User Journey Optimization: Cross-Channel Campaigns and Behavioral Targeting

Marketing Activation
Behavioral segmentation engine and automated campaign triggers optimized for user journey.

Multi-Brand Governance Framework: RACI and Cross-Group Steering

Data Governance
Data governance framework for multi-brand and multi-country company with stakeholder RACI, roles/responsibilities, data management processes, and cross-house/technical team steering.

Multi-Country Marketing Opt-in: Opt-in/Opt-out Strategy by Country

Privacy & Compliance
Implementation of consent management framework across 32 countries with opt-in rate optimization (+18-22%) and 100% compliant marketing activation generating 70-85% addressable audience.

Multi-Brand Audience Sharing: Prioritization and Conversion Impact

Marketing Activation
Implementation of marketing audience sharing between brands via third-party (retail media platform) with audience prioritization, privacy-safe processes, and measurable impact on conversion and campaign ROI.

Multi-Brand Cannibalization Analysis: Optimizing Global Customer Value

BI & Analytics
Cross-brand customer cannibalization and complementarity analysis for retail group, with multi-brand CLV measurement and global customer value optimization strategies.

Multi-Brand CDP: 20M+ Contacts, €4B+ Unified Sales

CDP & Platform
Group CDP consolidating 20M+ contacts and algorithms across multiple brands (€4B+ sales) with customer deduplication graph and complete infrastructure as code.
On this page