Posted at: 17 March
Senior DataOps Engineer (V)
Company
Blue Coding is a New York-based B2B company specializing in nearshore software development staffing and custom software solutions for startups and SMBs across Latin America.
Remote Hiring Policy:
Blue Coding supports remote work for specific roles, primarily hiring software development talent from Latin America ('Remote Latam'), while also offering on-site positions in New York and Mexico City.
Job Type
Full-time
Allowed Applicant Locations
South America, Mexico
Job Description
This role offers the opportunity to build and shape a modern DataOps ecosystem from the ground up, working with a highly scalable AWS stack and real-time data processing.
You’ll operate with a high level of autonomy, directly influencing architectural decisions and best practices in a fast-moving, low-structure environment.
If you enjoy owning systems end-to-end, solving complex data challenges, and working on infrastructure that directly impacts business-critical workflows, this role will give you that exposure.
Here are some of the exciting day-to-day challenges you will face in this role:
- Design, develop, and maintain data pipelines and datastores that support enterprise analytics, data science, and operational workloads.
- Lead and support large-scale database migration initiatives, including on-premises to cloud migrations.
- Monitor, analyze, and optimize the performance and stability of data layer services and platforms.
- Ensure data integrity, quality, and compliance across pipelines and datasets.
- Collaborate closely with peers across engineering, analytics, and technology teams.
- Guide, coach, and mentor data engineers, BI developers, and analysts.
- Design and implement enterprise-scale data solutions with long-term business impact.
- Build and maintain data processing solutions using Python and/or Scala.
- Work with a variety of data ingestion patterns, including SFTP, APIs, streaming, and batch processing.
- Design and support database models optimized for analytical and reporting use cases.
- Implement monitoring, alerting, and observability for data pipelines and infrastructure.
- Maintain clear and comprehensive documentation of data architectures, pipelines, and processes.
- Work within an Agile environment, collaborating through tools such as Jira and Git
You will shine if you have:
-
5+ years of experience in DevOps or DataOps roles with a strong focus on AWS.
-
Hands-on experience with AWS services such as EMR (Spark), Redshift, RDS, Glue, Lambda, Kinesis, Step Functions, EventBridge, SNS/SQS, KMS, and CloudWatch.
-
Strong proficiency in Python and SQL.
-
Experience working with relational, NoSQL, and columnar databases.
-
Experience implementing Infrastructure as Code using Terraform or CloudFormation.
-
Experience designing and maintaining CI/CD pipelines.
-
Familiarity with data quality frameworks, observability tools, and data governance practices.
-
Knowledge of handling sensitive data (PII) and compliance standards such as HIPAA.
-
Bachelor’s degree in Computer Science, Data Science, or equivalent experience.
-
Proven ability to work autonomously and take ownership of infrastructure and data workflows in a production environment
It doesn’t hurt if you also have:
-
AWS certifications (e.g., Data Engineer – Associate, DevOps Engineer – Professional).
-
Experience with AWS DMS, Secrets Manager, SES, and containerization tools such as Docker.
-
Experience with BI tools such as Tableau or Power BI.
-
Experience with data observability platforms.
-
Advanced degree in a related field
Here are some of the perks we offer you:
- Salary in USD
- Flexible schedule (within US Time zones)
- 100% Remote
- Work with modern AWS stack (EMR, Kinesis, Glue, etc.)
- High-impact role with ownership over key technical decisions