ETL Pipeline Developer & Data Engineer
Fully Remote
Job Type
Part-time
Description

 eSimplicity is a modern digital services company that partners with government agencies to improve the lives and protect the well-being of all Americans, from veterans and service members to children, families, and seniors. Our engineers, designers, and strategists cut through complexity to create intuitive products and services that equip federal agencies with solutions to courageously transform today for a better tomorrow.
 

Purpose and Scope

You will work collaboratively with Product Managers, Designers, and Engineers to set up, develop, and maintain critical back-end integrations for the data and analytics platform that allows users to access data catalogs, request compute resources, and manage data access controls for an enterprise data and analytics platform. 


Responsibilities:

  • Create and maintain new and existing data pipelines, Extract, Transform, and Load (ETL) processes, and ETL features using Azure cloud services.
  • Build, expand, and optimize data and data pipeline architectures. Optimize data flow and collection for cross functional teams of database architects, data analysts, and data scientists.
  • Operate large-scale data processing pipelines and resolve business and technical issues pertaining to the processing and data quality.
  •  Assemble large, complex sets of data that meet non-functional and functional business requirements
  •  Implement large dataset engineering: data augmentation, data quality analysis, data analytics (anomalies and trends), data profiling, data algorithms, and (measure/develop) data maturity models and develop data strategy recommendations.
  • Identify, design, and implement internal process improvements including re-designing data infrastructure for greater scalability, optimizing data delivery, and automating manual processes.
  •  Develop and document standard operating procedures (SOPs) for new and existing data pipelines
  •  Build analytical tools to utilize the data pipeline, providing actionable insight into key business performance metrics including operational efficiency and customer acquisition.
  • Work with stakeholders including data, design, product, and government stakeholders and assisting them with data-related technical issues
  •  Write unit and integration tests for all data processing code.
  • Read data specifications and translate them into code and design documents.
  • Perform other duties as assigned. 
Requirements

Required Qualifications:


 All candidates must pass public trust clearance through the U.S. Federal Government. This requires candidates to either be U.S. citizens or pass clearance through the Foreign National Government System which will require that candidates have lived within the United States for at least 3 out of the previous 5 years, have a valid and non-expired passport from their country of birth and appropriate VISA/work permit documentation.

  •  Bachelor's degree in Computer Science, Software Engineering, Data Science, Statistics, or related technical field.
  • 8+ years of experience in software/data engineering, including data pipelines, data modeling, data integration, and data management.
  • Expertise in data lakes, data warehouses, data meshes, data modeling and data schemas (star, snowflake…).
  • Extensive experience with Azure cloud-native data services, including Synapse, Data Factory, DevOps, KeyVault, etc.
  • Expertise in SQL, T-SQL, and Python with applied experience in Apache Spark and large-scale processing using PySpark.
  • Proficiency with data formats: parquet, distributed snappy parquet, and .csv.
  • Understanding of common connection protocols, such as SFTP.
  • Proven ability to work with incomplete or ambiguous data infrastructure and design integration strategies.
  • Excellent analytical, organizational, and problem-solving skills
  •  Strong communication skills, with the ability to translate complex concepts across technical and business teams.
  • Proven experience working with petabyte-level data systems. 

 Preferred Qualifications:


  • Experience working with Personally Identifiable Information (PII) and Sensitive Information (SI).
  • Demonstrated success providing support within a Federal agency system environment, ensuring alignment with organizational goals and technical standards.
  •  Familiarity with data governance, metadata management, and data quality practices. 


Working Environment:
eSimplicity supports a remote (or hybrid depending on the role and program) work environment operating within the Eastern time zone so we can work with and respond to our government clients. Expected hours are 9:00 AM to 5:00 PM Eastern unless otherwise directed by manager.
Occasional travel for training and project meetings. It is estimated to be less than 5% per year.

Benefits:
We offer a highly competitive salary and full healthcare benefits.

Equal Employment Opportunity:
eSimplicity is an equal-opportunity employer. All qualified applicants will receive consideration for employment without regard to race, religion, color, national origin, gender, age, status as a protected veteran, sexual orientation, gender identity, or status as a qualified individual with a disability. 

Salary Description
$110/hr