Required Experience:10 Years
Seeking a skilled Data Engineer to support the design, development, and maintenance of data lake infrastructure and data pipelines. The successful candidate will leverage modern data architecture best practices to ensure the effective movement, integration, and cleansing of large-scale data sets. This role plays a critical part in enabling data-driven decision-making across state agencies and improving the delivery of public services.
Key Responsibilities:
• Design, build, and maintain scalable and secure data lakes and data pipelines.
• Ingest, transform, and clean data from multiple sources (internal and external).
• Collaborate with data analysts, data scientists, and business stakeholders to understand data requirements and ensure quality and usability of data assets.
• Implement data governance, security, and compliance measures in line with state and federal policies.
• Optimize data flows for performance, scalability, and cost-efficiency.
• Automate data workflows and support real-time and batch processing systems.
• Document technical processes, data schemas, and pipeline designs.
⸻
Required Qualifications:
• Bachelor's degree in Computer Science, Information Systems, Engineering, or a related field.
• 3+ years of experience in data engineering or a related role.
• Strong proficiency in SQL and scripting languages such as Python or Scala.
• Hands-on experience with cloud platforms (e.g., AWS, Azure, or GCP) and tools such as S3, Redshift, Glue, or Databricks.
• Experience with big data frameworks such as Apache Spark, Hadoop, or Kafka.
• Knowledge of data modeling, ETL/ELT design patterns, and data lake architecture.
• Familiarity with data governance and data quality best practices.
We provide our valued customers with high-quality Python development services through our centers of excellence in major American tech and delivery centers worldwide.
98% Customer Satisfaction
150+ Fortune 500, Large & SMB clients
16+ Years of Service