• Location: Chantilly, Virginia
  • Type: Contract
  • Job #2211

Data Engineer
Clearance:
TS/SCI with ability to obtain Polygraph within reasonable period of time
Location: Chantilly, VA

 

Company Overview:
Cornerstone Defense, in partnership with our military, intelligence, and civil government customers, supports U.S. operations worldwide through the use of many different types of intelligence, satellite, and cyber technologies. Cornerstone’s Intelligence Sector provides solutions to the United States Government for information collection, operations, exploitation and dissemination, and research activities. Our Team specializes in software development, cloud architecture, systems and network engineering, systems integration, agile management, as well as targeting operations and intelligence analysis. Our support to our mission customers includes cyber network operations, exploitation and defense, signals intelligence, human intelligence, and critical missions and networks.
 
We are looking for team-members with creative talent who are ready to take on the challenge of Data Engineer to work with collaborative teams to help expand and optimize the data ingestion pipeline architecture, develop strategies for efficient ingestion, processing, storage, structuring, and access. In addition, the Data Engineer will support data analysts, data scientists, and big data engineers in identifying data sources, performing exploratory data analysis, developing data models, ensuring data cleanliness and accuracy to provide new Insider Threat behavioral insights.
 
Roles and responsibilities potentially include:
  • Support data science team by designing, developing and implementing scalable ETL process for disparate datasets into a Hadoop infrastructure
  • Design, develop, implement, and maintain data ingestion process from various disparate datasets using StreamSets (experience with StreamSets not mandatory)
  • Develop processes to identify data drift and malformed records
  • Develop technical documentation and standard operating procedures
  • Leads technical tasks for small teams or projects
Required Experience and Qualifications:
  • Requires a Bachelor’s degree in Systems Engineering, or a related Science, Engineering or Mathematics field.
  • 5+ years of job-related experience, or a Master's degree plus 3 years of job-related experience.
Desired Experience and Qualifications:
  • Working knowledge of entity resolution systems
  • Experience with Hadoop and Hive/Impala
  • Experience with messages systems like Kafka
  • Experience with NoSQL and/or graph databases like MongoDB or ArangoDB
  • Any of the following databases: SQL, MongoDB, Oracle, Postgres
  • Working experience with ETL processing and Python
  • Working experience with data workflow products like StreamSets or NiFi
  • Working experience with Python RESTful API services, JDBC
  • Experience with Cloudera Data Science Workbench is a plus
  • Understanding of pySpark
  • Leadership experience
  • Creative thinker
  • Ability to multi-task
  • Excellent use and understanding of data engineering concepts, principles, and theories
Data Engineer
Clearance:
TS/SCI with CI Poly
Chantilly, VA
We are looking for team-members with creative talent who are ready to take on the challenge of Data Engineer to work with collaborative teams to help expand and optimize the data ingestion pipeline architecture, develop strategies for efficient ingestion, processing, storage, structuring, and access. In addition, the Data Engineer will support data analysts, data scientists, and big data engineers in identifying data sources, performing exploratory data analysis, developing data models, ensuring data cleanliness and accuracy to provide new Insider Threat behavioral insights.
Roles and responsibilities potentially include:
  • Support data science team by designing, developing and implementing scalable ETL process for disparate datasets into a Hadoop infrastructure
  • Design, develop, implement, and maintain data ingestion process from various disparate datasets using StreamSets (experience with StreamSets not mandatory)
  • Develop processes to identify data drift and malformed records
  • Develop technical documentation and standard operating procedures
  • Leads technical tasks for small teams or projects
Required Experience and Qualifications:
  • Requires a Bachelor’s degree in Systems Engineering, or a related Science, Engineering or Mathematics field.
  • 5+ years of job-related experience, or a Master's degree plus 3 years of job-related experience.
Desired Experience and Qualifications:
  • Working knowledge of entity resolution systems
  • Experience with Hadoop and Hive/Impala
  • Experience with messages systems like Kafka
  • Experience with NoSQL and/or graph databases like MongoDB or ArangoDB
  • Any of the following databases: SQL, MongoDB, Oracle, Postgres
  • Working experience with ETL processing and Python
  • Working experience with data workflow products like StreamSets or NiFi
  • Working experience with Python RESTful API services, JDBC
  • Experience with Cloudera Data Science Workbench is a plus
  • Understanding of pySpark
  • Leadership experience
  • Creative thinker
  • Ability to multi-task
  • Excellent use and understanding of data engineering concepts, principles, and theories
Attach a resume file. Accepted file types are DOC, DOCX, PDF, HTML, and TXT.

We are uploading your application. It may take a few moments to read your resume. Please wait!