HADOOP, Spark, ETL Developer- Green Card or US Citizen. Alexandria, VA

Clearance Level
Alexandria, Virginia

REQ#: RQ90027

Travel Required: None
Public Trust: NACI (T1)
Requisition Type: Regular

We are GDIT.

  • Analyzes customer requirements and may provide technical knowledge on cloud cluster computing techniques and technologies of moderate complexity.
  • Assist with the research, evaluation, and deployment of cloud computing systems and large data analytics.
  • May assist customers with data integration and migration efforts, including mapping schemas and defining domain specific ontologies
  • Assists with the development of large data technologies, algorithms, and applications.
  • Maintains current knowledge of relevant technology as assigned.
  • DESIRED QUALIFICATIONS: BA/BS (or equivalent experience), 2 years of experience or less

Role & Responsibilities:

Design and implement Big Data analytic solutions on a Hadoop based platform. Design/develop/implement Big Data/Hadoop platforms related to data ingestion, storage, transformation and analytics. Refine data processing pipelines focused on unstructured and semi-structured data.

  • Design/develop/implement Big Data/Hadoop platforms related to data ingestion, storage, transformation and analytics.
  • Load data from disparate data sets, load Hive/HBase and RDBMS tables.
  • Import and export data using Sqoop from HDFS to RDBMS.
  • Preprocess data using Hive and Pig.
  • Develop shell/Scala/Python scripts to transform the data in HDFS.
  • Perform analysis of vast data stores and uncover insights.
  • Create scalable and high-performance web services for data tracking.
  • Create custom analytic and data mining algorithms for data extraction.
  • Experience in developing shell/python scripts to transform the data in HDFS.
  • Assist in resolution of infrastructure issues
  • Execute and troubleshoot Spark and Hive jobs including performance tuning.
  • #OpportunityOwned





  • 2+ years’ experience with Hive/HBase/MRV1/MRV2
  • Experience with Hadoop, HDFS, Hive, Apache Spark, Storm, and Kafka
  • Experience with Java, Scala, Python, Impala, R, Json, Knox, YARN, Tomcat, Ambari
  •  2+ Hands on experience with Apache Spark
  •  Strong  experience with ETL tools such as Sqoop and Pig
  • Experience with RDBMS, SQL, MongoDB, and hierarchical data management

Education Requirement:

A minimum of a Bachelor’s degree in computer science, computer information systems, information technology, or a closely related field, or a combination of education and experience equating to the U.S. equivalent of a Bachelor’s degree in one of the aforementioned subjects.


We are GDIT. The people supporting some of the most complex government, defense, and intelligence projects across the country. We deliver. Bringing the expertise needed to understand and advance critical missions. We transform. Shifting the ways clients invest in, integrate, and innovate technology solutions. We ensure today is safe and tomorrow is smarter. We are there. On the ground, beside our clients, in the lab, and everywhere in between. Offering the technology transformations, strategy, and mission services needed to get the job done.

GDIT is an Equal Opportunity/Affirmative Action employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, or veteran status, or any other protected class.