Bachelors degree in Engineering, Computer Science, Systems Engineering, or other computer or information technology-related field plus a minimum of 6 years relevant experience.
A TS/SCI security clearance with the ability to obtain a Polygraph is required at time of hire. Candidates must be able to obtain the Polygraph within a reasonable amount of time from date of hire. Applicants selected will be subject to a U.S. Government security investigation and must meet eligibility requirements for access to classified information. Due to the nature of work performed within our facilities, U.S. citizenship is required.
We have an immediate opening for a Python Developer on our award-winning Insider Threat Detection Team. A Python Developer will help us expand our insider threat capabilities in automating data integration and collection strategies. The successful candidate will help expand and optimize the data ingestion pipeline architecture, develop strategies for efficient ingestion, processing, storage, structuring, and access. In addition, the Python Developer will support data analysts, data scientists, and big data engineers in identifying data sources, performing exploratory data analysis, developing data models, ensuring data cleanliness and accuracy to provide new Insider Threat behavioral insights.
This position provides an opportunity to further advance the cutting-edge technology that supports some of our nations core defense/intelligence services and systems. Our employees work closely with esteemed customers to develop solutions that allow them to carry out high-stakes national security missions.
REPRESENTATIVE DUTIES AND TASKS:
We are seeking a Data Engineer to support the Insider Threat mission. Data Engineers work with various security system data owners to automate data integration and collection strategies. Work closely with the data science team to ensure data cleanliness and accuracy.
Support data science team by designing, developing, and implementing scalable ETL process for disparate datasets into a Hadoop infrastructure
Design, develop, implement, and maintain data ingestion process from various disparate datasets using StreamSets (experience with StreamSets not mandatory)
Develop processes to identify data drift and malformed records
Develop technical documentation and standard operating procedures
Improve responsiveness and overall performance of the data ingestion pipeline architecture
Assess, prioritize, and size features
Contribute to a cross functional agile team
KNOWLEDGE SKILLS AND ABILITIES:
Working knowledge of entity resolution systems
Experience with messages systems like Kafka
Experience with NoSQL and/or graph databases like MongoDB or ArangoDB
Any of the following databases: SQL, MongoDB, Oracle, Postgres
Working experience with ETL processing
Working experience with data workflow products like StreamSets or NiFi
Working experience with Python RESTful API services, JDBC
Experience with Hadoop and Hive/Impala
Experience with Cloudera Data Science Workbench is a plus
Understanding of pySpark Leadership experience
Ability to multi-task
Excellent use and understanding of data engineering concepts, principles, and theories
We are GDIT. The people supporting some of the most complex government, defense, and intelligence projects across the country. We deliver. Bringing the expertise needed to understand and advance critical missions. We transform. Shifting the ways clients invest in, integrate, and innovate technology solutions. We ensure today is safe and tomorrow is smarter. We are there. On the ground, beside our clients, in the lab, and everywhere in between. Offering the technology transformations, strategy, and mission services needed to get the job done.
GDIT is an Equal Opportunity/Affirmative Action employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, or veteran status, or any other protected class.