Data Infrastructure
Upload My Resume
Drop here or click to browse · PDF, DOCX, DOC, RTF, TXT
Requirements
• Primarily, you like to make great things with passionate colleagues. You are someone that likes to own outcomes, not only inputs. You’re motivated by having responsibility and accountability. You’re eager to ‘do the work,’ big and small. • You’re motivated by the question, “How can I improve this?” and have a track record of doing so, even in ways adjacent to your role. Much of our current team is made up of former founders and thrive in the level of autonomy at Roboflow. Maybe you had a side hustle in high school or college. • Many Roboflowers have used our tools before joining. One of the best ways to stand out amongst other applicants is to write about something you have built with Roboflow or contribute to one of our open source projects. https://roboflow.com/open-source • You are an analytical builder who wants to be an important part of an exceptional team that focuses on using Roboflow's computer vision tools to impact and improve every industry. You have high agency and a bias toward action. • Strong foundation in data engineering principles and patterns; experience building scalable, maintainable data pipelines • Deep understanding of modern data warehouse architecture and analytical database design • Hands-on experience with BigQuery and GCP data services (or equivalent cloud data platforms) • Proficiency with data transformation frameworks (dbt strongly preferred) • Experience with data visualization tools (Hex preferred) • Track record of balancing technical excellence with pragmatic cost optimization • Ability to translate business requirements into robust data solutions • Experience enabling non-technical teams to self-serve analytics • Strong communication skills and comfort working across organizational boundaries • Experience with workflow orchestration tools beyond GitHub Actions (Airflow, Prefect, Dagster, etc.) • Knowledge of real-time data processing and streaming architectures • Background in data governance and quality frameworks • Experience with reverse ETL or operational analytics patterns • Familiarity with BI tools and semantic layers • Understanding of privacy and compliance considerations in data systems • WHERE YOU'LL WORK • Roboflow is distributed across the US and Europe. We currently have Hubs in New York City and San Francisco (and plan to open more as we grow density in new cities). We provide opportunities (like team onsites in different cities) and resources https://blog.roboflow.com/how-we-work-together-at-roboflow/ (like a $4000/yr travel stipend) to work in person with other team members as much as you'd like, while also supporting remote team members. You can work from one of our Hubs (we offer a relocation bonus), work from home, work at co-working spaces, etc. We want you to work where you work best! • WHAT YOU'LL RECEIVE • 📈 In addition to our cash compensation, we offer generous perks and benefits. Below are some of the highlights: • $4000/yr Travel Stipend to travel anywhere anytime to work alongside other Roboflowers • $350/mo Productivity stipend to spend on things that make your work environment more productive, like high-speed internet at home or a co-working space • Cover up to 100% of your health insurance costs for you and your partner or family • Remote first/flexible schedule allowing you to work collaboratively with other team members and asynchronously • Unlimited PTO- with an annual 2 week minimum, we encourage you to take time off for yourself • 12 weeks parental leave • Equity in the company so we are all invested in the future of computer vision • INTERVIEW PROCESS (~5 HOURS) • Below is the interview process you can expect for this role. • Before the Interview: • We’ll review your application, LinkedIn, Github, etc. • The best way to stand out is to write about something you’ve built with Roboflow or contribute to one of our open source projects https://roboflow.com/open-source. • We may send you a technical screen if applicable. • Introduction Phase: • [30m] Meet with hiring manager to assess for overall mindset and skillset • [45m] Technical Assessment, if applicable • Team Interview Phase: • [30m] Meet with another member of the team: • Ellis Allen, Finance Lead • [60m] Meet with hiring manager again • Use this time to review specifics about the job description • Begin working through your 30/60/90 projects • Ask questions! • Final Interview Stage: • [45m] Meet with Kate Wagner, Head of Operations for a culture discussion • [60m] Meet with Joseph Nelson, CEO • Note: you are welcome to request additional conversations with anyone you would like to meet and we will accommodate as best we can. • NOT SURE IF THIS IS YOU?
Responsibilities
• We're looking for a Data Infrastructure engineer to centralize our data engineering capabilities while empowering teams across the organization to leverage data for their specific domains. Reporting to the Business Infrastructure Lead, this role sits at the intersection of infrastructure engineering and platform enablement; you'll be responsible for both the technical excellence of our data systems and ensuring every team can effectively access and utilize data for their needs. • Our current data stack powers critical business functions but needs thoughtful stewardship and evolution. You'll own our point of view on data services, lead the buildout of our data pipeline architecture, and ensure we're maximizing value while optimizing GCP utilization and spend. • Own the data platform: Provide stewardship for our BigQuery and dbt infrastructure, improving reliability, performance, and developer experience • Enable teams: Build self-service capabilities and tooling that allow product, sales, marketing, and engineering teams to access and analyze data independently • Design data architecture: Define and implement our philosophy on data services, including pipeline patterns, data modeling standards, and integration approaches • Optimize infrastructure: Deeply understand GCP and BigQuery to architect cost-effective solutions that scale with our growth • Build collaboratively: Work with stakeholders across the company to understand data needs and build pipelines that serve multiple use cases • Establish best practices: Create documentation, templates, and standards that make it easy for teams to work with data correctly • Act as advisor to team who produce and consume data with the goal of ensuring best practices • Drive data quality: Implement monitoring, testing, and validation frameworks to ensure data reliability