Senior Data Engineer Product
Company details
Company: Prokeep
Job type: Remote
Country: United States
City: Los Angeles
Experience: With previous experience
Description of the offer
Prokeep is a cutting-edge distributor revenue and relationship engine on a mission to empower distributors to win.
Trusted by over 1,000 distributors and 40,000 users, Prokeep makes it easy for over 3 million contractors, technicians, plumbers, electricians, mechanics, and other skilled professionals to order parts, troubleshoot issues, and generally, get what they need and get on with their day.
We serve the $100+ billion distributor market, enabling them to build stronger relationships, drive demand, and positively impact the communities they serve. By enhancing customer experiences and streamlining communication, we’re not just solving problems, we’re transforming industries and building stronger connections across the supply chain.
Having recently secured a $25M Series A investment, we’re poised for rapid growth and innovation. If you’re passionate about tackling complex challenges in a collaborative, fast-paced environment while contributing to an often-overlooked yet vital sector of the economy, we’d love to have you on board.
Let’s shape the future together, one connection at a time!
Our Values
Build Together. – Collaboration, camaraderie, and customer-focus are core to how we collectively work and win.
Own It. – Persistence and personal growth drive the results – and the rewards! – that we can all enjoy.
Sponsored ads
Be Humble. Be Human. – Respect and authenticity allow us to build meaningful and lasting relationships.
Have Fun! – Positivity, perseverance, and a bit of laughter all allow us to make an impact while enjoying the rollercoaster!
The Role
We’re building a small, high-impact team to create new data-driven capabilities and AI-powered features that enhance our existing SaaS products or serve as standalone offerings. We’re looking for a Data Engineer who’s excited to be part of an entrepreneurial, product-minded team where creativity, curiosity, and experimentation are core to how we work.
This isn’t a role for someone who wants to be handed requirements and optimize pipelines in isolation. We’re looking for someone who thrives in ambiguous environments, loves turning messy data into structured assets, and has a nose for product opportunities hiding in the data.
You’ll be joining a team led by a deeply technical founder, alongside a founding engineer focused on AI and machine learning. Your role will focus on designing and building robust, scalable data pipelines, shaping our storage layer, and collaborating on new data products and proofs of concept.
What You’ll Do
- Design, build, and maintain ETL/ELT pipelines using modern data tools and frameworks
- Own and optimize our data lake/data warehouse setup (e.g. Parquet, Iceberg, Delta Lake)
- Collaborate closely with AI/ML engineers to build and serve training datasets, inference pipelines, and feedback loops
- Partner with product and engineering to turn raw data into product-ready insights, APIs, or feature sets
- Prototype and iterate on new data product ideas quickly, focusing on speed-to-learning over polish
- Take initiative in identifying data-driven opportunities within our platform or market
- Contribute to defining best practices, architectural choices, and internal data standards
What We’re Looking For
Sponsored ads
- Strong engineering background with 5-7 years of hands-on experience in data systems and distributed processing (e.g., Apache Spark, dbt, Airflow, etc.)
- Deep understanding of file formats and table formats like Parquet, Apache Iceberg, or Delta Lake
- Experience building and operating data infrastructure in AWS or similar
- Proven ability to work independently and ship data products, features, or APIs—not just internal dashboards
- Familiarity with data modeling for analytics and product use cases
- Curiosity and creativity in exploring what’s possible with data + AI
- A founder’s mindset — someone who wants to shape not just the code, but the product and direction of the team
- Comfort navigating ambiguity, iterating quickly, and learning through experimentation
- Bonus Points For
- Experience with real-time/streaming data tools (e.g., Kafka, Spark Streaming, Flink)
- Experience with MLOps practices like deploying and monitoring models, managing feature stores, or automating retraining workflows
- Familiarity with building LLM-powered features, including pipelines, embeddings, vector databases, or semantic search
- Exposure to tools like Trino, DuckDB, or OLAP engines used for ad-hoc querying over data lakes
Why Join Us?
- You’ll be part of a small, senior team with the autonomy to move fast and explore big ideas
- You’ll have direct access to a production-scale SaaS platform, customer usage data, and real market needs
- You’ll help shape entirely new data products, not just pipelines behind the scenes
- You’ll work with a team that values craft, experimentation, and impact over red tape
Why Prokeep?
At Prokeep, we offer a dynamic, passionate, and collaborative work environment where innovation thrives. Here’s what you can expect:
- Competitive Compensation: ($125K-135K) Reflecting your expertise and impact.
- Equity Package: Your success is our success, share in the growth you’ll help create.
- Comprehensive Benefits: Health, dental, vision, life, short & long-term disability, 401(k), and employee assistance program (EAP).
- Flexible PTO: Recharge and refocus with the flexibility to manage your time with no preset limits
- Hybrid Work: Hybrid from one of our office in Portland
- Continuous Growth: Yearly education stipend to support your professional development.
Location of employment
How to apply?
Click on the button to get the company email or employment application form.
Apply with External LinkSponsored ads
