Unfortunately, this job posting is expired.
Don't worry, we can still help! Below, please find related information to help you with your job search.
Some similar recruitments
Ansible Automation Engineer Jobs
Recruited by Deltacubes 8 months ago Address McLean, VA, United States
Data Engineer Jobs
Recruited by Deloitte 9 months ago Address Alexandria, VA, United States
Systems Engineer-Data Analysis (Sme)
Recruited by BAE Systems, Inc. 9 months ago Address Reston, VA, United States
Data Engineer Jobs
Recruited by BMK Careers 10 months ago Address Chantilly, VA, United States
Data Engineer Jobs
Recruited by Deloitte 10 months ago Address Richmond, VA, United States
Systems Engineer Jobs
Recruited by H2 Performance Consulting 10 months ago Address Norfolk, VA, United States
Network Engineer - Infrastructure
Recruited by ManTech 11 months ago Address Langley Forest, VA, United States
Data Engineer Jobs
Recruited by Narwal 11 months ago Address Reston, VA, United States
Data Engineer Jobs
Recruited by Deloitte 11 months ago Address Arlington, VA, United States

Data Engineer Jobs

Company

Technology Ventures

Address McLean, VA, United States
Employment type CONTRACTOR
Salary
Category IT Services and IT Consulting
Expires 2023-06-23
Posted at 1 year ago
Job Description

Qualifications:

  • Bachelor’s degree in Computer Science, Engineering, Data science or a related quantitative field.
  • 5-6 years of relevant experience in deign and development of data pipelines to processing large volumes and variety of data (Structured and unstructured data, writing code for parallel processing, XMLS, JSONs, PDFs)
  • Strong verbal and written communication skills.
  • Hands-on programming experience in Hadoop, Spark, Python and SQL for data processing and analysis.
  • Demonstrated ability to manage competing demands, prioritize work, and manage customer expectation.


Required Technical Skills

  • Advanced Python, SQL and Spark, very good familiarity with Bug data technologies like Hadoop, Scoop, Hive, Ambari
  • Prior experience working with AWS and Snowflake technologies
  • Unix Shell script, Autosys batch scheduling


Responsibilities

  • Use programming languages in Python, PySpark and Spark for data ingestion.
  • Be able to build Dashboards in R/Shiny for end user consumption
  • Develop programs in BigData platform using Hadoop platform.
  • Develop Python, PySpark, Spark scripts to filter/cleanse/map/aggregate data.
  • Cleanse, manipulate and analyze large datasets (Structured and Unstructured data – XMLs, JSONs, PDFs) using Hadoop platform.
  • Present ideas and recommendations on Hadoop and other technologies best use to management.
  • Manage and implement data processes (Data Quality reports)
  • Develop data profiling, deduping logic, matching logic for analysis