Interviews

How lean e-commerce data teams prove their value | Census

Sylvain Giuliani
Sylvain Giuliani April 01, 2021

Syl is the Head of Growth & Operations at Census. He's a revenue leader and mentor with a decade of experience building go-to-market strategies for developer tools. San Francisco, California, United States

Qasim Ashraf has deep roots in e-commerce and data. He’s started his own e-commerce companies, led internal data teams at companies like Boll & Branch, served as CTO at Lygics, and is now working on a secret new project that applies data-driven insights to fitness.

This atypical experience has led Qasim to develop opinions as unique as his skill set. We sat down with him to hear why he thinks data teams need to embrace a people-centric approach and how e-commerce data teams can prove their value.

Can you give us a little background on your career thus far?

I started with doing my own products and developed a pretty strict process around data that led to consulting. But there's only so much value you can provide as a consultant because data is a people problem, not an engineering problem. You can't solve people problems as a consultant, so I went in-house to solve those people problems and then bounced around because the grass is evergreen.

Can you expand on that “data is a people problem” idea?

The value of a data team is to make the company data-driven. A data-driven organization makes sure people are empowered by data to make the right decisions. That's why data is a people problem.

It's much easier for me to engineer a very complicated technical solution than it is to get alignment on stakeholder needs because politics come into play.

I think Census and the rise of operational analytics help make companies more data-driven even if they don't realize it initially. Historically, the main value add of data is to resurface data on a dashboard that someone may or may not use.

With Census, you're taking the data and all the work you're doing with it, and you're pushing it back to tools that everyone knows and uses, like Salesforce or NetSuite. That makes it easier for people to make decisions with data.

That's how you become data-driven. You reduce the time to insights. If the data is already in the tool you're using, then there's zero time to insights.

What’s your approach to selecting data tools?

The reality is that tools don't solve problems explicitly. Someone needs to solve problems. If you think of an RFP (request for proposal) as a way to solve a business problem, you're doing it wrong. You're just kicking the can down the road. Someone has to build the data strategy, and tools can never replace that strategy. That again ties data back into a people problem.

If you think of an RFP as a way to solve a business problem, you're doing it wrong.

How has this modern stack changed how companies interact with data?

Historically, the main output for this stack is for people to interact with the data via dashboards. That allows people to ask the right questions and make business decisions.

But the reality is that dashboard fatigue is real. Dashboards don't actually solve problems, to be honest. They don't answer specific questions most of the time. And, at the end of the day, they require someone to have the willingness to look at the right dashboard and do that lift on their end. Historically, that's how data got proliferated across the organization.

A tool like Census helps data get proliferated back upstream automatically. It helps the stakeholders see data exactly when it's needed, where it's needed, as opposed to going to a separate dashboard.

Dashboards don't actually solve problems.

It reduces that friction point, right? Census lets the data exist where the decision is being made so that the stakeholder can take that into account as opposed to any extra lift on their end.

What are some difficulties smaller e-commerce data teams face when establishing this modern stack?

I think it goes into the fact that now you have to justify your ROI of the data team. And it puts in the question of “Is it worth spending X thousands of dollars on all these tools and a modern stack?”

And, again, engineering this, there is a solvable solution for like, let's say time stamping, which may be the most annoying part of changing tools. There's a technical solution for that. However, it's getting people enough education to see the value out of data to invest in data that's typically more difficult.

Why is it difficult for e-commerce data teams to justify their ROI?

I think the reason it's difficult to justify is because data is inherently a support role. The data itself is there to service the rest of the organization and amplify their business needs.

Data is the growth coefficient for the organization. We don't necessarily directly impact the bottom line as much as amplify other departments to impact the bottom line more efficiently.

So a data team starts off as a “nice to have” as opposed to a “must have” or people that are focused on short term. It’s difficult for some people to see the long term. And that’s because the goal of becoming data-driven is such a high goal, people don't really see a path to it.

Unless you use operational analytics, which makes data very actionable for people who are not data driven—actionable and thus quantifiable.

How can a lean e-commerce data team prove their value?

The top priority is to balance tactical solutions with overall strategy to increase how "data driven" the organization is.

That being said, tactics that are a top priority for e-commerce data teams are:

  • Marketing analytics, like attribution, media mix modeling, etc.
  • Experimentation, like A/B testing, conversion rate optimization, etc.
  • Loyalty and retention, like email/CRM strategy and personalization
  • Automating and optimizing processes, like demand planning and supply chain analytics

But don't try to boil the ocean. Focus on actual pain points. Try to get actionable results based on well-thought-out hypotheses, and then let that balloon itself.

Is there a common mistake that you see data teams make when they try to boil the ocean?

I think a lot of data teams try to do everything for everyone. What happens then is that they’re not solving issues that anyone's passionate about. And if you're not actionably making someone's life easier, no one's going to champion for you internally.

No one gives a shit that you save them five minutes every two weeks. It’s better to save one person a significant amount of time. That person will become your data champion. Now that you solved their needs, you can move on to the next person's needs.

A concrete example of solving issues no one is passionate about is internal executive KPIs and metrics. Hot take: I think those meetings and dashboards have too many metrics that people focus on too much and that no one actually cares about.

What is the biggest challenge facing the data industry right now?

The biggest problem is the actionability of the data. We as an industry have focused a lot on making sure data is available and making sure data is correct, accurate, and robust, but we haven't focused on making sure data is actionable.

The result is that no one's tracking the actual impact and value out of data. So we're tracking work done, not necessarily outcomes or results. We're tracking dashboards created as opposed to business questions solved.

I think that data actionability is the next step in moving towards this outcomes-based approach. And I think that Census and operational analytics solve for data actionability.

If you liked this piece, we also interviewed Jamie Quint, former head of growth at Notion, about his ideal analytics stack, which you can read about here, as well as what he thinks about where the data industry is headed.

While you’re here, schedule a Census demo to see why both Qasim and Jamie are so excited about the future of operational analytics.

Related articles

Product News
Sync data 100x faster on Snowflake with Census Live Syncs
Sync data 100x faster on Snowflake with Census Live Syncs

For years, working with high-quality data in real time was an elusive goal for data teams. Two hurdles blocked real-time data activation on Snowflake from becoming a reality: Lack of low-latency data flows and transformation pipelines The compute cost of running queries at high frequency in order to provide real-time insights Today, we’re solving both of those challenges by partnering with Snowflake to support our real-time Live Syncs, which can be 100 times faster and 100 times cheaper to operate than traditional Reverse ETL. You can create a Live Sync using any Snowflake table (including Dynamic Tables) as a source, and sync data to over 200 business tools within seconds. We’re proud to offer the fastest Reverse ETL platform on the planet, and the only one capable of real-time activation with Snowflake. 👉 Luke Ambrosetti discusses Live Sync architecture in-depth on Snowflake’s Medium blog here. Real-Time Composable CDP with Snowflake Developed alongside Snowflake’s product team, we’re excited to enable the fastest-ever data activation on Snowflake. Today marks a massive paradigm shift in how quickly companies can leverage their first-party data to stay ahead of their competition. In the past, businesses had to implement their real-time use cases outside their Data Cloud by building a separate fast path, through hosted custom infrastructure and event buses, or piles of if-this-then-that no-code hacks — all with painful limitations such as lack of scalability, data silos, and low adaptability. Census Live Syncs were born to tear down the latency barrier that previously prevented companies from centralizing these integrations with all of their others. Census Live Syncs and Snowflake now combine to offer real-time CDP capabilities without having to abandon the Data Cloud. This Composable CDP approach transforms the Data Cloud infrastructure that companies already have into an engine that drives business growth and revenue, delivering huge cost savings and data-driven decisions without complex engineering. Together we’re enabling marketing and business teams to interact with customers at the moment of intent, deliver the most personalized recommendations, and update AI models with the freshest insights. Doing the Math: 100x Faster and 100x Cheaper There are two primary ways to use Census Live Syncs — through Snowflake Dynamic Tables, or directly through Snowflake Streams. Near real time: Dynamic Tables have a target lag of minimum 1 minute (as of March 2024). Real time: Live Syncs can operate off a Snowflake Stream directly to achieve true real-time activation in single-digit seconds. Using a real-world example, one of our customers was looking for real-time activation to personalize in-app content immediately. They replaced their previous hourly process with Census Live Syncs, achieving an end-to-end latency of <1 minute. They observed that Live Syncs are 144 times cheaper and 150 times faster than their previous Reverse ETL process. It’s rare to offer customers multiple orders of magnitude of improvement as part of a product release, but we did the math. Continuous Syncs (traditional Reverse ETL) Census Live Syncs Improvement Cost 24 hours = 24 Snowflake credits. 24 * $2 * 30 = $1440/month ⅙ of a credit per day. ⅙ * $2 * 30 = $10/month 144x Speed Transformation hourly job + 15 minutes for ETL = 75 minutes on average 30 seconds on average 150x Cost The previous method of lowest latency Reverse ETL, called Continuous Syncs, required a Snowflake compute platform to be live 24/7 in order to continuously detect changes. This was expensive and also wasteful for datasets that don’t change often. Assuming that one Snowflake credit is on average $2, traditional Reverse ETL costs 24 credits * $2 * 30 days = $1440 per month. Using Snowflake’s Streams to detect changes offers a huge saving in credits to detect changes, just 1/6th of a single credit in equivalent cost, lowering the cost to $10 per month. Speed Real-time activation also requires ETL and transformation workflows to be low latency. In this example, our customer needed real-time activation of an event that occurs 10 times per day. First, we reduced their ETL processing time to 1 second with our HTTP Request source. On the activation side, Live Syncs activate data with subsecond latency. 1 second HTTP Live Sync + 1 minute Dynamic Table refresh + 1 second Census Snowflake Live Sync = 1 minute end-to-end latency. This process can be even faster when using Live Syncs with a Snowflake Stream. For this customer, using Census Live Syncs on Snowflake was 144x cheaper and 150x faster than their previous Reverse ETL process How Live Syncs work It’s easy to set up a real-time workflow with Snowflake as a source in three steps:

Best Practices
How Retail Brands Should Implement Real-Time Data Platforms To Drive Revenue
How Retail Brands Should Implement Real-Time Data Platforms To Drive Revenue

Remember when the days of "Dear [First Name]" emails felt like cutting-edge personalization?

Product News
Why Census Embedded?
Why Census Embedded?

Last November, we shipped a new product: Census Embedded. It's a massive expansion of our footprint in the world of data. As I'll lay out here, it's a natural evolution of our platform in service of our mission and it's poised to help a lot of people get access to more great quality data.