This position is based in Mérida, Yucatán, México. If you do not live in/around Merida, and are still interested in this position, relocation will be required within 60 days of accepting this position. Verato will provide a relocation bonus of $42,500 MNX pesos to help with your move.
ABOUT VERATO
As digital transformation and AI progress at lightning speed, organizations find themselves data-rich and insights-poor. Digital transformation’s promise to drive better experiences and business performance is falling short. Data is often trapped in silos across disconnected systems of record, such as ERPs and EHRs, systems of engagement, such as CRMs, and systems of insight, such as cloud data platforms. These systems cannot integrate seamlessly without a single source of truth for identity, making it impossible to share and consume complete and trusted 360-degree views of people, organizations, and networks.
Verato, the identity intelligence experts, powers exceptional experiences everywhere by solving the problem that drives everything else — knowing who is who. The Verato MDM Cloudä, the next generation of MDM, delivers unprecedented identity intelligence by uniquely combining extraordinary identity resolution and enrichment with identity verification, AI-powered data governance, and advanced insights. Verato re-imagines MDM to be purpose-built and nimble to drive a complete and trusted 360-degree view of people, organizations, and networks across complex ecosystems with unmatched speed to value, enterprise-grade performance, and customer success. More than 75% of the US population flows through Verato, powering a single source of truth for identity across the critical industries of healthcare, life sciences, financial services, public sector, and beyond. For more information, visit verato.com.
Core to Verato’s strategy for sustained growth is our commitment to building a strong, people-first culture that attracts, develops, and retains top talent worldwide. Verato operates on the simple principle that a company must prioritize its employees first and foremost. In return, these employees will take care of the company’s customers, and in turn, those customers will support the company’s shareholders. Verato believes in empowering teams with the best tools and development opportunities available. Staff are given chances to expand their knowledge in areas like technology (e.g., big data, distributed/cloud computing, complex algorithms), healthcare, and organizational development. As Verato continues a path of high growth and significant impact, every team member gains an influential front-row seat as we execute our business strategy. Together, we can bring about a profound and positive transformation in healthcare as we know it today.
VERATO VALUES
We are committed to continually raising the standard of excellence throughout the organization, from marketing to engineering to customer service. Our guiding principles are to Make a Difference, to be Trustworthy, and to be Customer Obsessed.
Verato employees have a precise focus on proactively protecting the privacy and security of all systems while always ensuring they are following documented policies and procedures.
About the Position
We are seeking a Data Engineer who is interested in joining a highly dynamic and creative development team. This is a leadership role within the Technology Department. Verato’s SaaS software offering is a Master Data Management ( Verato MDM Cloud) platform that provides our customers with a complete and trusted 360-degree view of their patients, consumers, and providers. This technical software development position is focused on enhancing an established product based on client needs. This position reports to the Director of Data Platform within the Technology Department.
Essential Functions and Responsibilities
- Design data pipelines for API, streaming, and batch processing to facilitate data loads into the Snowflake data warehouse.
- Collaborate with other engineering and DevOps team members to implement, test, deploy, and operate data pipelines and ETL solutions.
- Develop scripts to Extract, Load and Transform data and other utility functions
- Optimize data pipelines, ETL processes, and data integrations for large-scale data analytics use cases
- Build necessary components to ensure data quality, monitoring, alerting, integrity, and governance standards are maintained in data processing workflows
- Able to navigate ambiguity and thrives in a fast-paced environment. Takes initiative and consistently delivers results with minimal supervision.
- Performs data profiling and analysis required to perform development work or troubleshoot/assist in the resolution of data issues.
Required Skills:
- Bachelor's or master’s degree in Computer Science, Information Systems, or related field
- 3+ years of experience in building and maintaining data pipelines and ETL/ELT processes in data-centric organizations
- Strong coding skills using Python. Familiar with Python libraries related to data engineering and cloud services including pandas, boto, etc.
- At least 1-2 years of experience with AWS S3, SQS, Kinesis, Lambda, AWS DMS, Glue/EMR, AWS Batch or similar services
- Hands-on experience building streaming and batch big data data pipelines
- Must have knowledge of building Infrastructure in AWS cloud using Cloud formation or Terraform
- 1+ years of working experience with Snowflake cloud data warehouse including Snowflake data shares, Snowpipes, Snow SQL, Tasks etc
- Must have working knowledge of various databases, SQL and NoSQL
- Must have working knowledge of various file formats like CSV, Json Avro, and Parquet.
- Hands-on experience with cloud platforms such as AWS and Google Cloud
- Strong experience with Apache Spark, especially using PySpark for large-scale data processing
- Experience working with agile development methodology
- Experienced in CI/CD and release processes, proficient in Git or other source control management systems, to streamline development and deployment workflows
Other Desired Skills:
- Minimum 2 years of designing and implementing operational production grade large-scale data pipelines, ETL/ELT and data integration solutions.
- Exposure to multi-tenant/multi-customer environments is a big plus.
- Hands on experience with productionized data ingestion and processing pipelines
- Strong understanding of Snowflake Internals and integration of Snowflake with other data processing and reporting technologies.
- Experience working with structured, semi-structured, and unstructured data
- Familiarity with MongoDB or similar NoSQL database systems
- Some familiarity with Apache Airflow and building/maintaining DAGs