Snowflake Data Engineer (Principal Level) Apply
Role: Data Engineer Location: McLean VA(Onsite Hybrid Model) Duration: 9-12 Months || Extension Possible Job Responsibilities: Must haves include Pyspark, Hadoop, Snowflake. Must have Programming experience in Python, Spark for data processing and analysis. Must have Strong SQL experience is a must Experience with AWS Cloud is preferred. Cleanse, manipulate and analyze large datasets (Structured and Unstructured data – XMLs, JSONs, PDFs) using Hadoop platform. Experience in processing large volumes and variety of data (Structured and unstructured data, writing code for parallel processing, XMLS, JSONs, PDFs) Experience – using Hadoop platform and performing analysis. Develop Python, PySpark, Spark scripts to filter/cleanse/map/aggregate data. Manage and implement data processes (Data Quality reports) Develop data profiling, deduping logic, matching logic for analysis Programming Languages experience in Python, PySpark and Spark for data ingestion Programming experience in Big Data platform using Hadoop platform Present ideas and recommendations on Hadoop and other technologies best use to management Familiarity with Hadoop cluster environment and configurations for resource management for analysis work Experience with containerization and orchestration. Good to have Hands on experience with AWS, Kubernetes, Kubeflow, Docker etc.