Senior Data Engineer Job in United State | Yulys
×

Job Title: Senior Data Engineer

Company Name: Growth Protocol
Salary: USD 0.00
-
USD 0.00 Hourly
Job Industry: Program Development
Job Type: Full time
WorkPlace Type: remote
Location: United State, United States
Required Candidates: 1 Candidates
Skills:
Error Handling
Self-Reflection
Problem Solving
Job Description:

We are searching for an ambitious go-getter who welcomes the challenge of meeting the needs of a hyper-growth startup. As a Sr. Data Engineer, you will be at the heart of Growth Protocol’s data infrastructure, playing a foundational role in building the systems that power our AI platform. Your work will directly influence product features, client outcomes, and strategic business decisions.


You will collaborate with Data Scientists, Backend Engineers, Client IT, and business stakeholders to build and maintain scalable pipelines that serve billions of rows of structured and unstructured data weekly, enabling high-impact insights across multiple industries.


THE ROLE

Growth Protocol is hiring a Senior Data Engineer to play a foundational role in building the systems that power our AI platform. You will be at the heart of Growth Protocol's data infrastructure, directly influencing product features, client outcomes, and strategic business decisions.

You will collaborate with Data Scientists, Backend Engineers, Client IT, and business stakeholders to build and maintain scalable pipelines that serve billions of rows of structured and unstructured data weekly, enabling high-impact insights across multiple industries.

Ideal candidates are ambitious go-getters who welcome the challenge of meeting the needs of a hyper-growth startup and bring deep technical expertise across modern data infrastructure and ML operationalization.


OBJECTIVES OF THE ROLE

Collaboration

  1. Work closely with Data Scientists to translate business and ML requirements into robust data workflows
  2. Ensure timely delivery of clean, reliable data to support model development and production features

Technical Development

  1. Engineer and manage scalable ETL architecture using Airflow, Snowpark, Cloud Run, and Apache Beam
  2. Design and implement a high-performance data infrastructure for seamless processing and integration
  3. Extract data from diverse online platforms
  4. Operationalize machine learning models, focusing on deployment, reliability, and performance

Data Connectivity

  1. Partner with client IT teams to identify the most efficient and secure methods for data ingestion including Snowflake Sharing, Databricks Delta Sharing, Private Link, and VPN
  2. Work alongside the Platform Engineering team to define requirements for secure networking paths that support high-performance data transfers
  3. Perform end-to-end testing of client connections to ensure data integrity and connectivity
  4. Integrate customer databases with our platform

Monitoring and Reliability

  1. Create and manage real-time monitoring systems for data ingestion and transformation pipelines
  2. Proactively identify and resolve issues to maintain high levels of system reliability and data integrity


REQUIRED SKILLS AND QUALIFICATIONS

  1. 5+ years of experience in Data Engineering
  2. Bachelor's degree in Computer Science, Engineering, or a related technical field, or equivalent practical experience
  3. Experience building data pipelines with robust unit and integration testing
  4. Proficiency in distributed computing frameworks including Apache Beam and Spark
  5. Functional understanding of enterprise networking including VPC peering, Private Link, and VPNs, with the ability to troubleshoot connectivity in a cloud environment
  6. Hands-on experience operationalizing ML models in production
  7. Familiarity with ML/AI, NLP, and Data Science workflows including MLFlow
  8. Deep understanding of ETL workflows, data modeling, and data architecture
  9. Strong debugging and problem-solving skills
  10. Excellent communication skills and experience collaborating across teams


Preferred Qualifications

  1. Experience working on enterprise products serving Fortune 500 clients across Financial Services, Industrials, and Consumer Products
  2. Prior startup experience
  3. Interest in current events, market dynamics, and emerging technologies
  4. Experience creating Agent Skills
  5. Familiarity with APIs and web scraping for data collection
  6. Familiarity with Graph Databases


TECH STACK

  1. Languages: Python, TypeScript
  2. Frameworks: Apache Beam, Spark, FastAPI, Airflow
  3. Cloud: Google Cloud Platform
  4. Data: Elasticsearch, Snowflake, Databricks, Neo4J, PostgreSQL, MongoDB, GCS
  5. Infrastructure and DevOps: Docker, Terraform, GitHub Actions, Cloud Run
  6. Frontend: Next.js


PERKS

  1. Competitive compensation and equity in a rapidly growing company
  2. 100% company-paid health, dental, and vision insurance plus 401(k)
  3. Pet-friendly office

Are you looking for remote jobs near your area? At Yulys, thousands of employers are looking for exceptional talent like yours. Find a perfect job now.

Become a part of our growth newsletter