image
  • Snapboard
  • Activity
  • Reports
  • Campaign
Welcome ,
loadingbar
Loading, Please wait..!!

Lead Data Engineer

  • ... Posted on: Feb 03, 2025
  • ... MSR Cosmos
  • ... Greenfield, Indiana
  • ... Salary: Not Available
  • ... CTC

Lead Data Engineer   

Job Title :

Lead Data Engineer

Job Type :

CTC

Job Location :

Greenfield Indiana United States

Remote :

No

Jobcon Logo Job Description :

Job Title: Technical lead (Databricks, Azure Data Lake, Python)

Duration: Long term contract

Location: Greenfield, Indiana

Key Skills: (Databricks, Azure Data Lake, Python)

Hiring Manager Educational Qualification*

  • Any Bachelor's Degree good in Computer Science, Information Technology

Job Description

  • We are seeking a highly skilled Data Engineering Specialist to join our team. The ideal candidate will have extensive experience in cloud technologies, DevOps development practices and Data Engineering to support and enhance our RDAP initiatives.

Key Responsibilities:

  • Design, develop, and maintain Databricks Lakehouse solutions sourcing from cloud platforms such as Azure Synapse and GCP.
  • Implement and manage DevOps and CICD workflows using tools like GitHub.
  • Apply best practices in test-driven development, code review, branching strategies, and deployment processes.
  • Build, manage, and optimize Python packages using tools like setup, poetry, wheels, and artifact registries.
  • Develop and optimize data pipelines and workflows in Databricks, utilizing PYSpark and Databricks Asset Bundles.
  • Manage and query SQL databases (Unity Catalog, SQL Server, Hive, Postgres).
  • Implement orchestration solutions using Databricks Workflows, Airflow, and Dagster.
  • Work with event-driven architectures using Kafka, Azure Event Hub, and Google+C4 Cloud Pub/Sub.
  • Develop and maintain Change Data Capture (CDC) solutions using tools like Debezium.
  • Extensive experience in design and implementation of data migration projects specifically involving Azure Synapse and Databricks Lakehouse
  • Manage cloud storage solutions, including Azure Data Lake Storage and Google Cloud Storage.
  • Configure and manage identity and access solutions using Azure Active Directory, including AD Groups, Service Principals, and Managed Identities.

Primary (Must have skills)

  • Python package builds (setup, poetry, wheels, artifact registries)
  • Specific technologies
  • Databricks (PYSpark, Databricks Asset Bundles)
  • Open File Formats (Delta/Parquet/Iceberg/etc.)
  • SQL Databases (Unity Catalog, SQL Server, Hive, Postgres)
  • Orchestration Tools (Databricks Workflows, Airflow, Dagster)
  • Azure Data Lake Storage
  • Azure Active Directory (AD groups, Service Principles, Managed Identities)
  • Experience Range 10-12 years

Secondary Skills (Good To have)

  • Kafka, Azure Event Hub, Cloud Pub/Sub
  • Change Data Capture (Debizum)
  • Google Cloud Storage

Soft skills/other skills (If any)

Communication Skills:

  • Ability to convey complex technical concepts in a clear and concise manner to both technical and non-technical stakeholders.
  • Strong documentation skills for creating process guidelines, technical workflows, and reports.

Problem-Solving and Analytical Thinking:

  • Capability to troubleshoot and resolve issues efficiently.
  • Analytical mindset for optimizing workflows and improving system performance.

Lead Responsibilities

  • Effective in Customer interactions for understanding requirements, participating in design discussions and translating requirements into deliverables by working with the development team at Offshore
  • Effective in collaborating with cross-functional teams across development, operations, and business units.
  • Strong interpersonal skills to build and maintain productive relationships with team members.

Thanks &

Jobcon Logo Position Details

Posted:

Feb 03, 2025

Employment:

CTC

Salary:

Not Available

Snaprecruit ID:

SD-CIE-15f349a552de9e298ba2672d8d635186ead3a2eec61a1d9ffa5d518e9326eaac

City:

Greenfield

Job Origin:

CIEPAL_ORGANIC_FEED

Share this job:

  • linkedin

Jobcon Logo
A job sourcing event
In Dallas Fort Worth
Aug 19, 2017 9am-6pm
All job seekers welcome!

Lead Data Engineer    Apply

Click on the below icons to share this job to Linkedin, Twitter!

Job Title: Technical lead (Databricks, Azure Data Lake, Python)

Duration: Long term contract

Location: Greenfield, Indiana

Key Skills: (Databricks, Azure Data Lake, Python)

Hiring Manager Educational Qualification*

  • Any Bachelor's Degree good in Computer Science, Information Technology

Job Description

  • We are seeking a highly skilled Data Engineering Specialist to join our team. The ideal candidate will have extensive experience in cloud technologies, DevOps development practices and Data Engineering to support and enhance our RDAP initiatives.

Key Responsibilities:

  • Design, develop, and maintain Databricks Lakehouse solutions sourcing from cloud platforms such as Azure Synapse and GCP.
  • Implement and manage DevOps and CICD workflows using tools like GitHub.
  • Apply best practices in test-driven development, code review, branching strategies, and deployment processes.
  • Build, manage, and optimize Python packages using tools like setup, poetry, wheels, and artifact registries.
  • Develop and optimize data pipelines and workflows in Databricks, utilizing PYSpark and Databricks Asset Bundles.
  • Manage and query SQL databases (Unity Catalog, SQL Server, Hive, Postgres).
  • Implement orchestration solutions using Databricks Workflows, Airflow, and Dagster.
  • Work with event-driven architectures using Kafka, Azure Event Hub, and Google+C4 Cloud Pub/Sub.
  • Develop and maintain Change Data Capture (CDC) solutions using tools like Debezium.
  • Extensive experience in design and implementation of data migration projects specifically involving Azure Synapse and Databricks Lakehouse
  • Manage cloud storage solutions, including Azure Data Lake Storage and Google Cloud Storage.
  • Configure and manage identity and access solutions using Azure Active Directory, including AD Groups, Service Principals, and Managed Identities.

Primary (Must have skills)

  • Python package builds (setup, poetry, wheels, artifact registries)
  • Specific technologies
  • Databricks (PYSpark, Databricks Asset Bundles)
  • Open File Formats (Delta/Parquet/Iceberg/etc.)
  • SQL Databases (Unity Catalog, SQL Server, Hive, Postgres)
  • Orchestration Tools (Databricks Workflows, Airflow, Dagster)
  • Azure Data Lake Storage
  • Azure Active Directory (AD groups, Service Principles, Managed Identities)
  • Experience Range 10-12 years

Secondary Skills (Good To have)

  • Kafka, Azure Event Hub, Cloud Pub/Sub
  • Change Data Capture (Debizum)
  • Google Cloud Storage

Soft skills/other skills (If any)

Communication Skills:

  • Ability to convey complex technical concepts in a clear and concise manner to both technical and non-technical stakeholders.
  • Strong documentation skills for creating process guidelines, technical workflows, and reports.

Problem-Solving and Analytical Thinking:

  • Capability to troubleshoot and resolve issues efficiently.
  • Analytical mindset for optimizing workflows and improving system performance.

Lead Responsibilities

  • Effective in Customer interactions for understanding requirements, participating in design discussions and translating requirements into deliverables by working with the development team at Offshore
  • Effective in collaborating with cross-functional teams across development, operations, and business units.
  • Strong interpersonal skills to build and maintain productive relationships with team members.

Thanks &

Loading
Please wait..!!