Data Engineer (Mexico City)
Rise Interactive
·
Posted:
August 1, 2023
·
Onsite
About the position
We are seeking a detail-oriented and highly motivated individual to join our Innovation team as a Software Engineer focused on Data Management. The ideal candidate will have a strong aptitude for learning, a sense of urgency, and a passion for delivering the best possible product and customer experience. In this role, you will be responsible for building and optimizing data processing pipelines, integrating client data into our analytics platform, and designing and building our data warehouse and real-time data reporting systems. Additionally, you will be expected to promote good team practices, mentor and guide developers, and adhere to security policies and guidelines. A bachelor's degree in computer science or a related field is required, along with 1-3 years of on-the-job experience. Strong ETL programming skills using tools like Python, Spark, PySpark, Hadoop, MapReduce, and Kafka are preferred.
Responsibilities
- Implement and optimize data processing pipelines for megabytes to terabytes of data
- Onboard and integrate client data into our analytics platform
- Design and build our data warehouse as well as real-time data reporting systems
- Promote and nurture good team practices such as unit testing, code reviews, build/test automation, etc.
- Proactively mentor and guide developers to improve their quality and simplicity in design and code
- Design, build and use tools to understand our product platform behavior and performance
- Design and conduct experiments to test concepts, technologies, and algorithms
- Implement analytics tools to maximize the value of collected data
- Implement data tests in data quality frameworks to ensure data is clean and accurate
- Adhere to security policies and guidelines to ensure our data is protected and safe
- Embrace and assist in evolving our Agile (Scrum) team processes and developer role responsibilities
Requirements
- Bachelor's degree in computer science or equivalent field
- 1-3 years of on-the-job experience
- Strong ETL programming skills using tools like Python, Spark, PySpark, Hadoop, MapReduce, Kafka
- Ability to implement and optimize data processing pipelines for megabytes to terabytes of data
- Experience onboarding and integrating client data into analytics platforms
- Ability to design and build data warehouses and real-time data reporting systems
- Knowledge of unit testing, code reviews, build/test automation, etc.
- Ability to mentor and guide developers to improve quality and simplicity in design and code
- Experience designing, building, and using tools to understand product platform behavior and performance
- Ability to design and conduct experiments to test concepts, technologies, and algorithms
- Familiarity with implementing analytics tools to maximize the value of collected data
- Knowledge of data quality frameworks and ability to implement data tests to ensure data cleanliness and accuracy
- Adherence to security policies and guidelines to protect data
- Familiarity with Agile (Scrum) team processes and developer role responsibilities
Benefits
- Clean and accurate data
- Adherence to security policies and guidelines
- Evolution of Agile (Scrum) team processes and developer role responsibilities
- Smart, high aptitude to learn new things and sense of urgency to get things done
- Extremely strong ETL programming skills using tools like Python, Spark, PySpark, Hadoop, MapReduce, Kafka
- Experience building data access layers via cubes, data marts, APIs, or visualization tools like Tableau or D3
- Practical experience with Big data and NoSQL technologies desired
- Comfortable working with several large, complex SQL databases and SQL queries
- Experience working in cloud-only infrastructure, especially Amazon Web Services
- 1+ years of experience working with large amounts of real data
- A strong passion for empirical research and answering hard questions
- Team player – demonstrated experience on a few teams that have shipped a product
- Professional developer – experienced with source control (Git) and bug tracking
- Practical-minded – chooses stability/reliability/maintainability over shiny new objects
- Passionate about technology – ideally you build things outside of work for fun
- Scripting skills – must be totally comfortable at the Linux command line
- Nice to have: Expertise in applied statistics or machine learning
- Experience delivering data products in the marketing, ad tech space is also a plus
- Excellent English verbal and written communication skills