Ready to be pushed beyond what you think you’re capable of?
At Coinbase, our mission is to increase economic freedom in the world. It’s a massive, ambitious opportunity that demands the best of us, every day, as we build the emerging onchain platform — and with it, the future global financial system.
To achieve our mission, we’re seeking a very specific candidate. We want someone who is passionate about our mission and who believes in the power of crypto and blockchain technology to update the financial system. We want someone who is eager to leave their mark on the world, who relishes the pressure and privilege of working with high caliber colleagues, and who actively seeks feedback to keep leveling up. We want someone who will run towards, not away from, solving the company’s hardest problems.
Our work culture is intense and isn’t for everyone. But if you want to build the future alongside others who excel in their disciplines and expect the same from you, there’s no better place to be.
The data engineering team develops and maintains robust data pipelines, builds trusted data sources, and creates analytics/data products that inject automation into data science processes, with a goal to empower users via self-serve analytics.
What we do:
- Trusted data sources: Build and maintain a foundational data layer (data marts), that serve as the single source of truth across Coinbase.
- Reliable data pipelines: Design and implement robust data pipelines, guaranteeing data quality and timely data delivery across our organization.
- Data science developer tools: Build developer tools that inject automation into data science processes, improving efficiency and productivity. For example, data transformation, data modeling, and data quality tooling.
- Self-serve analytics products: Deliver tailored data products, designed to empower users with self-serve capabilities and ensure accurate answers to their data inquiries.
What you’ll be doing:
Your primary responsibilities will include building tools and products that enhance data science productivity, enable self-serve analytics, and ensure data reliability and quality. Additional responsibilities may include contributing to the maintenance of critical data marts and data pipelines as needed..
More specifically:
- Enforce best practices: Introduce automation mechanisms to enforce data best practices and optimal design principles.
- Data transformation tooling: Enhance our data transformation layer by adding new features and migrating data pipelines from legacy tools.
- Data quality tooling: Build systems to automate monitoring, alerting and fixing data anomalies.
- Productivity tooling: Build internal tools aimed at streamlining workflows, minimizing manual interventions and driving operational efficiencies.
- Cross-Functional collaboration: Work alongside fellow data engineers and cross-functional partners from Data Science, Data Platform, Machine Learning, and various analytics teams to ensure alignment on priorities and deliverables.
- Reliable data pipelines: Develop new data pipelines and ETL processes, while enhancing the efficiency of existing ones to reduce costs, improve landing times, and ensure reliability. This may involve indexing, partitioning, and tuning the underlying data models.
- Semantic layer integration: Build the semantic layer to ensure it effectively translates complex data models from the data marts into business terms. This includes defining dimensions, measures, hierarchies, and business logic.
- Self-serve analytics: Contribute to the self-serve analytics chatbot and its infrastructure by adding features, fixing bugs and promoting its use across the company.
- Knowledge base: Maintain clear and comprehensive documentation of processes to facilitate smooth team collaboration and hand-offs.
What we look for in you:
- Python: Must be adept at scripting in Python, particularly for data manipulation and integration tasks, with strong experience in Object-Oriented Programming (OOP).
- SQL: Must have a solid grasp of advanced SQL techniques for querying, transformation, and performance optimization.
- ETL/ELT Processes: Experience in designing, building, and optimizing ETL/ELT data pipelines to process large datasets. Experience with both batch and streaming is a plus.
- Apache Airflow: Experience in building, deploying, and optimizing DAGs in Airflow or a similar pipeline orchestration tool.
- GitHub: Experience with version control, branching, and collaboration on GitHub is a must-have.
- Data Visualization: Experience with tools like Superset, Looker, or Python visualization libraries (Matplotlib, Seaborn, Plotly…etc)
- Data Modeling: Understanding of best practices for data modeling, including star schemas, snowflake schemas, and data normalization techniques
- Collaboration and Communication: Ability to work closely with data scientists, analysts, and other stakeholders to translate business requirements into technical solutions. Strong documentation skills for pipeline design and data flow diagrams.
- Fundamental DevOps Practices: Knowledge of unit testing, CI/CD, and repository management.
- Technologies: Familiarity with Docker, Kubernetes, and cloud platforms such as AWS or GCP.
- Prompt Engineering for LLMs: Expertise in crafting, testing, and refining prompts for effective communication with LLMs, like GPT, to generate accurate and contextually relevant responses is a plus.
Job #: P60857
#LI-Remote
Commitment to Equal Opportunity
Coinbase is committed to diversity in its workforce and is proud to be an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, creed, gender, national origin, age, disability, veteran status, sex, gender expression or identity, sexual orientation or any other basis protected by applicable law. Coinbase will also consider for employment qualified applicants with criminal histories in a manner consistent with applicable federal, state and local law. For US applicants, you may view the Know Your Rights notice here. Additionally, Coinbase participates in the E-Verify program in certain locations, as required by law.
Coinbase is also committed to providing reasonable accommodations to individuals with disabilities. If you need a reasonable accommodation because of a disability for any part of the employment process, please contact us at accommodations[at]coinbase.com to let us know the nature of your request and your contact information. For quick access to screen reading technology compatible with this site click here to download a free compatible screen reader (free step by step tutorial can be found here).
Global Data Privacy Notice for Job Candidates and Applicants
Depending on your location, the General Data Protection Regulation (GDPR) and California Consumer Privacy Act (CCPA) may regulate the way we manage the data of job applicants. Our full notice outlining how data will be processed as part of the application procedure for applicable locations is available here. By submitting your application, you are agreeing to our use and processing of your data as required. For US applicants only, by submitting your application you are agreeing to arbitration of disputes as outlined here.