About our Data Environment
We’ve invested heavily in creating a modern data foundation that supports analytics at scale. Our stack combines:
-
Ingestion & Transformation: Multiple data sources flow into a central data lake and warehouse (via event tracking, ETLs, and transformation layers such as dbt).
-
Data Delivery: Insights are shared through BI tools, and reverse-ETL processes push relevant datasets back into operational platforms like CRMs.
-
Workflow Management: Automated pipelines and operational tasks are coordinated with an orchestration framework.
-
Versioning: Git-based practices ensure every step in the pipeline is tracked and reproducible.
Rather than focusing only on dashboards, our team standardizes metric definitions and equips business users to explore data independently.
What You’ll Be Doing
Your responsibilities will span across three main areas:
Data Pipeline (≈50%)
-
Build and maintain robust data pipelines and orchestration solutions.
-
Implement ETL processes to transform raw data into reliable data models.
-
Set up automated validation/testing to safeguard data quality and accuracy.
-
Document pipeline processes and metadata for long-term transparency.
Data Modeling & Business Analytics (≈40%)
-
Develop and maintain semantic layers for data modeling.
-
Collaborate with business stakeholders to translate processes into measurable metrics.
-
Provide workshops and training sessions to improve self-service analytics usage.
-
Handle occasional special data requests (e.g., custom usage reports).
Product & Research Contributions (≈10%)
-
Offer practical insights into how data practitioners interact with BI platforms.
-
Explore and evaluate emerging data tools, sharing findings to inspire product improvements.
What We Value in Team Members
We’re looking for someone who:
-
Is genuinely passionate about the data analytics and BI ecosystem.
-
Embraces a growth mindset and enjoys learning new things daily.
-
Has strong problem-solving skills and the ability to break down complex issues.
-
Understands how data, people, and business processes interconnect in real-world settings.
-
Communicates ideas clearly in written English.
-
Works well independently, but also collaborates effectively with teammates.
-
Learns new tools quickly (we don’t expect you to know our exact stack upfront).
Required Skills
-
Data Modeling: Familiar with dimensional modeling and other schema design techniques.
-
SQL: Able to write clean, efficient queries and optimize them for performance/cost.
-
Programming: Comfortable with Python or similar scripting languages for pipeline maintenance.
-
Version Control: Practical experience with Git for collaborative workflows.