Events

There's more to a data career than technical skills | Census

Allie Beazell
Allie Beazell May 21, 2021

Allie Beazell is director of developer marketing @ Census. She loves getting to connect with data practitioners about their day-to-day work, helping technical folks communicate their expertise through writing, and bringing people together to learn from each other. Los Angeles Metropolitan Area, California, United States

If you’re trying to land a data career by focusing solely on technical data skills, you’re going to have a bad time. Here’s an important truth: A data career is just as much a people-focused job as it is a data-focused job.

The importance of so-called soft skills was one of the main focuses in our discussion about data careers during the fourth session of our Operational Analytics Conference. On our panel, we featured some folks who do a lot of thinking about which data experts to hire:

The board agreed that to land and build a career in data, you need to understand what data leaders look for in candidates today. Turns out, much of what they look for has little to do with the technical aspects of data and more to do with the people data helps and the problems it solves.

Learn how to work with people and to understand what they need

A common refrain in our discussion was that you can teach the technical nuances that come with working in data. Teaching soft skills—like managing expectations, knowing (and clearly communicating) how long a task might take, and understanding why you’re doing that task in the first place—isn’t as easy. So, to land that awesome data job, you need to demonstrate you’ve already developed these soft skills or are actively working on them.

“The hardest part of data is not actually writing code,” Rachel said. “The much harder part of data is dealing with people. So, we’re looking for soft skills and then are willing to teach technical skills.”

Sarah echoed this point by saying that “technical skills are much easier to teach than business context.”Business context is how the business uses data, and it’s vital that you know how your work fits into larger organizational objectives for the teams you support. If, for example, you’re segmenting audiences for the marketing team, the business context is better audience targeting for a more effective ad campaign.

Familiarity with how stakeholders use data quickly emerged as the most important trait, and it’s only possible to gain via second-hand gold or hard-won experience. If you don’t want to spend years running ad campaigns for the sake of context, the next best thing is to tap into the experience and knowledge of your coworkers who design those campaigns daily.  

“What I think made the data teams that I’ve worked on be successful is an understanding of what people are doing with the data,” Rachel said. “Then, I can build tools and systems based on how I know it’s going to be used. If you’re an analytics engineer, data analyst, or a data product manager, you need to understand why marketing is asking this question.”

So, what does this mean for a hypothetical interview? Be curious and easy to talk to. Don’t rely just on your data or coding skills; exhibit an understanding of how data impacts business.  If you have gaps in your knowledge, show a willingness to learn. All of this together proves to your potential employer that you’re easy—and possibly fun—to work with.

Spend time and be active in data communities

When asked about where they find new data hires, the panelists all pointed to data-focused communities and content. Participating in data communities and writing data content can start to build a professional network of data professionals that think the same way you do.

Every panelist mentioned two communities as a good source for data hires: the dbt Slack and Locally Optimistic. Emilie said that Netlify made five data job offers in September and December of 2020 and that “three to four of them came in through the dbt and Locally Optimistic communities. Those were really great because there’s a baseline level of like, ‘Oh, we’re thinking about data in similarish ways.’” Garegin also said that Fivetran has had a lot of luck with both communities.

Emilie also mentioned Outer Join, a data-focused job board, as a good source for data talent. Lastly, when it comes to senior-level roles, personal blogs are a huge plus. Joey also chimed in to mention that, especially for experienced and high-level roles, having a data-focused blog is a huge plus.

The overarching theme here is that the people hiring for data roles are active members of data communities. They participate in the same discussions you should participate in, and they read the same blog posts you read. So, actively participate in data communities and even produce your own content if you’re so inclined. Data job offers may start to land in your lap if you can build a following among other folks who love data.

Aim for either depth or breadth of knowledge

Every panelist said there were two tracks at their company: one for individual contributors and one for management. You won’t be stuck on one track throughout your career—it’s not uncommon for people to switch between tracks pretty often, the panelists said—but having an awareness of them will help you create the career you want.

Each track favors a different type of knowledge:

  • To be a high-level individual contributor, you need deep knowledge in your area of expertise.
  • To be a high-level manager, you need a broader perspective about data and how it impacts the business at large.

For individual contributors, Rachel used an anecdote based on the alphabet. If you’re given a task, how far you make that task down the alphabet without handholding denotes your seniority. So, a junior engineer may take a task from A to C before looping someone in for help, review, or approval. A senior engineer, on the other hand, can take a task from A to Z before looping someone in due to their expertise and the trust they’ve built.

A manager needs a broader skillset focused on securing buy-in for data initiatives and then using that buy-in to build systems, like operational analytics. Sarah put it this way:

“You’re not just solving a one-off problem. Instead, over time as you see these problems arise, you should be able to identify patterns of problems. What are the systems that you need to build in order to solve for these patterns of problems? That translates into an architectural way of thinking around not just writing a query to solve something, but actually thinking about the bigger soul of that problem or type of problem on an ongoing basis.”

No matter which track you chose, as Rachel says, “The more senior you are, the more proactive you should be.” As a high-level individual contributor, you’ll know (to follow Rachel’s metaphor) what it takes to get from A to Z and can plan accordingly, manage expectations, and possibly get ahead. As a high-level manager, you can foresee patterns of problems, pivoting systems and strategy to compensate for next month’s problems today.

In an interview setting, showcase this proactivity by not only understanding what’s required of someone at a high level in the position you’ve applied for but also the impact your decisions might have. As a data analyst, can you get ahead of any requests marketing may have without being prompted? As a data lead, can you secure buy-in for a slick new tool your analytics engineers have been asking for by creating a detailed implementation plan?

Future data leaders need to understand the future of data operations

Of course, in addition to all of this advice, to be a great data professional, you need to stay on top of data industry trends. One of the most important data trends of late is an emerging class of data tools and processes that are changing how organizations use data. At Census, these kinds of trends fascinate us endlessly, so subscribe to our podcast and newsletter to receive the latest insights. We can help you not just build the future of your data career but also the future of data as a whole.

Related articles

Product News
Sync data 100x faster on Snowflake with Census Live Syncs
Sync data 100x faster on Snowflake with Census Live Syncs

For years, working with high-quality data in real time was an elusive goal for data teams. Two hurdles blocked real-time data activation on Snowflake from becoming a reality: Lack of low-latency data flows and transformation pipelines The compute cost of running queries at high frequency in order to provide real-time insights Today, we’re solving both of those challenges by partnering with Snowflake to support our real-time Live Syncs, which can be 100 times faster and 100 times cheaper to operate than traditional Reverse ETL. You can create a Live Sync using any Snowflake table (including Dynamic Tables) as a source, and sync data to over 200 business tools within seconds. We’re proud to offer the fastest Reverse ETL platform on the planet, and the only one capable of real-time activation with Snowflake. 👉 Luke Ambrosetti discusses Live Sync architecture in-depth on Snowflake’s Medium blog here. Real-Time Composable CDP with Snowflake Developed alongside Snowflake’s product team, we’re excited to enable the fastest-ever data activation on Snowflake. Today marks a massive paradigm shift in how quickly companies can leverage their first-party data to stay ahead of their competition. In the past, businesses had to implement their real-time use cases outside their Data Cloud by building a separate fast path, through hosted custom infrastructure and event buses, or piles of if-this-then-that no-code hacks — all with painful limitations such as lack of scalability, data silos, and low adaptability. Census Live Syncs were born to tear down the latency barrier that previously prevented companies from centralizing these integrations with all of their others. Census Live Syncs and Snowflake now combine to offer real-time CDP capabilities without having to abandon the Data Cloud. This Composable CDP approach transforms the Data Cloud infrastructure that companies already have into an engine that drives business growth and revenue, delivering huge cost savings and data-driven decisions without complex engineering. Together we’re enabling marketing and business teams to interact with customers at the moment of intent, deliver the most personalized recommendations, and update AI models with the freshest insights. Doing the Math: 100x Faster and 100x Cheaper There are two primary ways to use Census Live Syncs — through Snowflake Dynamic Tables, or directly through Snowflake Streams. Near real time: Dynamic Tables have a target lag of minimum 1 minute (as of March 2024). Real time: Live Syncs can operate off a Snowflake Stream directly to achieve true real-time activation in single-digit seconds. Using a real-world example, one of our customers was looking for real-time activation to personalize in-app content immediately. They replaced their previous hourly process with Census Live Syncs, achieving an end-to-end latency of <1 minute. They observed that Live Syncs are 144 times cheaper and 150 times faster than their previous Reverse ETL process. It’s rare to offer customers multiple orders of magnitude of improvement as part of a product release, but we did the math. Continuous Syncs (traditional Reverse ETL) Census Live Syncs Improvement Cost 24 hours = 24 Snowflake credits. 24 * $2 * 30 = $1440/month ⅙ of a credit per day. ⅙ * $2 * 30 = $10/month 144x Speed Transformation hourly job + 15 minutes for ETL = 75 minutes on average 30 seconds on average 150x Cost The previous method of lowest latency Reverse ETL, called Continuous Syncs, required a Snowflake compute platform to be live 24/7 in order to continuously detect changes. This was expensive and also wasteful for datasets that don’t change often. Assuming that one Snowflake credit is on average $2, traditional Reverse ETL costs 24 credits * $2 * 30 days = $1440 per month. Using Snowflake’s Streams to detect changes offers a huge saving in credits to detect changes, just 1/6th of a single credit in equivalent cost, lowering the cost to $10 per month. Speed Real-time activation also requires ETL and transformation workflows to be low latency. In this example, our customer needed real-time activation of an event that occurs 10 times per day. First, we reduced their ETL processing time to 1 second with our HTTP Request source. On the activation side, Live Syncs activate data with subsecond latency. 1 second HTTP Live Sync + 1 minute Dynamic Table refresh + 1 second Census Snowflake Live Sync = 1 minute end-to-end latency. This process can be even faster when using Live Syncs with a Snowflake Stream. For this customer, using Census Live Syncs on Snowflake was 144x cheaper and 150x faster than their previous Reverse ETL process How Live Syncs work It’s easy to set up a real-time workflow with Snowflake as a source in three steps:

Best Practices
How Retail Brands Should Implement Real-Time Data Platforms To Drive Revenue
How Retail Brands Should Implement Real-Time Data Platforms To Drive Revenue

Remember when the days of "Dear [First Name]" emails felt like cutting-edge personalization?

Product News
Why Census Embedded?
Why Census Embedded?

Last November, we shipped a new product: Census Embedded. It's a massive expansion of our footprint in the world of data. As I'll lay out here, it's a natural evolution of our platform in service of our mission and it's poised to help a lot of people get access to more great quality data.