Data engineer - Remote (US)
Job Description:
Data engineer - Remote (US)
Are you a data engineer who thrives on building robust pipelines and solving complex data challenges?
In this role, youll own the critical data infrastructure that powers pricing models and marketplace insights by ingesting transaction and listing data from dozens of external marketplaces. Youll be responsible for ensuring data pipelines deliver fresh, accurate information that drives pricing decisions, market analytics, and business intelligence across the platform.
What Youll Do
- Design, optimize, and own data pipelines that scrape, process, and ingest transaction and listing data from major auction houses and marketplaces.
- Build comprehensive monitoring and alerting systems to track latency, uptime, and coverage metrics across all data sources.
- Continuously improve data infrastructure by modernizing storage and processing technologies, reducing manual interventions, and optimizing for cost, performance, and reliability.
- Partner with internal teams to understand data usage patterns and ensure pipelines deliver clean, standardized data that meets product requirements.
This Role Is a Perfect Fit If You
- Care deeply about data quality and know that extraction is only the beginningthe real value comes from delivering clean, usable data.
- Thrive in a startup environment where you can make an immediate impact and own critical systems end-to-end.
- Prefer an incremental approach to improvements, guided by evidence-based decisions rather than wholesale replacements.
- Are intellectually curious about how data flows through systems and passionate about automation opportunities.
- Want to work at the intersection of data engineering and product, seeing directly how your pipelines impact business outcomes.
What You Bring to the Table
Must-Haves
- 3+ years of experience in data engineering or related fields
- Strong Python proficiency with at least 3 years of hands-on experience
- Proven experience with large-scale data processing using dataframe technologies (Pandas, Polars, PySpark, or similar)
- Hands-on experience with pipeline orchestration tools (Airflow, Dagster, or similar DAG-based systems)
- Track record of owning at least one data pipeline end-to-end within the past 2 years
- Solid SQL skills for data analysis and transformation
- Previous startup experienceyou understand the pace and adaptability required in a fast-moving environment
- A pragmatic mindset focused on delivering value incrementally rather than chasing perfection
Nice-to-Haves
- Experience with web scraping technologies (Selenium, Puppeteer, Beautiful Soup)
- Familiarity with data infrastructure and cloud services (AWS preferred)
- Interest in or knowledge of trading cards, collectibles, or alternative asset markets
- Experience with LLM-based automation tools for data extraction and processing
Our Client is on a mission to unlock the true value of alternative assets starting with the $5B trading-card market. Collectors can buy, sell, vault, and even finance their cards all in one seamless platform, making it easier than ever to treat collectibles like investable assets. Backed by world-class investors such as Stripe and Coinbase along with legendary athletes like Tom Brady they are building the future of this market.
If you are looking for the next challenging and rewarding career step - apply today!
Required Skills:
Data Engineering Cards Pandas Data Quality Intelligence Data Processing Analysis Pipelines Trading Business Intelligence Adaptability AWS Selenium Metrics Reliability Storage Analytics Infrastructure Automation Data Analysis Finance Design Engineering SQL Python Business