Overview
Explore the Nearsure experience!
Join our close-knit LATAM remote team: Connect through fun activities like coffee breaks, tech talks, and games with your team-mates and management.
Say goodbye to micromanagement! We champion autonomy, open communication, and respect for diversity as our core values.
Your well-being matters: Our People Care team is here from day one to support you with everything from time-off requests to wellness check-ins.
Plus, our Accounts Management team ensures smooth, effective client relationships, so you can focus on what you do best.
What we offer
- Competitive USD salary – We value your skills and contributions!
- 100% remote work – Work from anywhere, with opportunities to connect with teammates and grow your network at our LATAM coworking spaces.
- Paid time off – Take time off according to your country’s regulations with full salary.
- National Holidays celebrated – Take time off to celebrate important events and traditions with loved ones.
- Sick leave – Focus on your health and recovery.
- Refundable Annual Credit – Spend it on perks to enhance your work-life balance.
- Team-building activities – Coffee breaks, tech talks, and after-work gatherings to bond with the Nearsure family.
- Birthday day off – An extra day off during your birthday week.
About the project
As a Senior Data Engineer, you will work with modern data technologies such as Apache Spark (PySpark or Spark), Apache Iceberg, and AWS services (S3, EMR, Athena, Glue), ensuring data quality, scalability, and operational excellence. Familiarity with orchestration tools (Airflow), CI/CD practices, and the ability to leverage AI-powered development assistants (e.g., GitHub Copilot) is essential.
Responsibilities
- Design, develop, and maintain batch ETL/ELT pipelines and data workflows for large-scale datasets in AWS.
- Implement and optimize data lakehouse architectures using Apache Iceberg on S3, with schema evolution, partitioning strategies, and table maintenance.
- Build and tune distributed data processing jobs with Apache Spark (PySpark or Spark) for performance and cost efficiency.
- Orchestrate workflows using Apache Airflow, including DAG design, scheduling, and SLA monitoring.
- Apply best practices in code quality, version control (Git/GitHub), and CI/CD for data engineering projects.
- Ensure data quality, security, and compliance through validation, monitoring, and governance frameworks (Glue Catalog, IAM, encryption).
- Collaborate with cross-functional teams (data scientists, analysts, architects) to deliver scalable and reliable data solutions.
- Contribute to the development and optimization of analytics applications, ensuring they are powered by well-structured, high-quality data pipelines.
- Continuously evaluate and adopt emerging technologies and AI-powered tools to improve productivity and maintain technical excellence.
Qualifications
- Bachelor's Degree in Computer Science, Engineering, or a related field.
- 5+ Years of experience in data engineering and designing large-scale data solutions.
- 5+ Years of experience with Python for data manipulation, scripting, and integration tasks.
- 5+ Years of experience with SQL & DBMS (e.g., PostgreSQL, MySQL, SQL Server), data modeling and query tuning.
- 3+ Years with Apache Spark (PySpark preferred).
- 3+ Years building batch ETL/ELT pipelines for large datasets.
- 3+ Years with AWS data services (S3, Athena/Presto, Glue, Lambda, CloudWatch).
- 2+ Years with Apache Iceberg (table design, partitioning, schema evolution, maintenance).
- 2+ Years with AWS EMR as the execution platform for big data workloads.
- 2+ Years with orchestrating data workflows with Apache Airflow.
- 2+ Years using Git/GitHub (branching, PR reviews, CI/CD for data pipelines).
- Experience designing efficient ingestion pipelines into analytical systems.
- Proficiency in logging, auditing, and monitoring for data pipelines.
- Experience with data cleansing, validation, and transformation for analytical/reporting systems.
- Familiarity with data security and privacy practices.
- Solid understanding of cloud-native analytics architectures (data lake/lakehouse, ELT patterns).
- Proven ability to leverage AI-powered assistants as part of the engineering workflow.
- Advanced English Level is required for this role due to work with US clients.
Hiring process
- Let’s chat about your experience!
- Impress our recruiters, and you’ll move on to a technical interview with our top developers.
- Nail that, and you’ll meet our client - your final step to joining our amazing team!
Closing
At Nearsure, we’re dedicated to solving complex business challenges through cutting-edge technology and we believe in the power of tailored solutions. Whether you are passionate about transforming businesses with Generative AI, building innovative software products, or implementing comprehensive enterprise platform solutions, we invite you to be part of our dynamic team!
We would love to hear from you if you are eager to make an impact and join a collaborative team that values creativity and expertise.
Let’s work together to shape the future of technology!
Apply now!
By applying to this position, you authorize Nearsure to collect, store, transfer, and process your personal data in accordance with our Privacy Policy. For more information, please review our Privacy Policy.