LOCATION

McLean,VA

CLEARANCE

Required: TS/SCI Full Scope Poly Clearance

Description

Join our team and play a key role in shaping the future of data-driven decision-making! We are seeking a skilled Data Scientist to develop and enhance our suite of analytical tools. You will work closely with software developers and other data scientists in a fast-paced, agile environment, contributing to the entire development process. Your ability to communicate effectively with diverse stakeholders will be crucial to your success.

Responsibilities:

  • Support data engineering processes, including data exploitation and management.
  • Provide engineering support for data triage and assessment, as needed.
  • Directly support data ingestion processes to meet requirements.
  •  Document CONOPS (Concept of Operations), data workflows, and processes related to data engineering projects.
  •  Identify and evaluate new technologies, assessing their technical capabilities and performance.
  • Identify and implement agile approaches to automate manual and inefficient processes.
  • Provide insights into industry trends and make recommendations for the program’s future direction.

Mandatory Skills:

  • Bachelor’s degree in Computer Science, Software Engineering, or a related field.
  • 10+ years of experience in a professional role as a Data Scientist
    • Relevant Bachelors Degree: equivalent to 3 years of experience
    • Relevant Masters/Doctoral Degree: equivalent to 2 years of experience
    • Relevant Certifications: equivalent to 3 months of experience
  • Hands-on experience implementing, designing indexes for, and ingesting data into large Elasticsearch 7.x clusters.
  • Demonstrated ability to handle all aspects of identity resolution for massive datasets, including data ingestion, mapping, and integrating these capabilities into existing data engineering and systems workflows.
  •  Experience with distributed data processing technologies used in building ETL pipelines. This includes designing data flows, adapting the ETL process to handle new data types, and transferring data between systems or networks.
  • Experience in organizing and cataloging datasets, suggesting a need for data governance and metadata management skills.
  • Ability to communicate effectively with diverse stakeholders, both technical and non-technical, across different seniority levels.

Desired Skills & Certifications:

  • Experience with Hadoop technologies and integrating them into ETL (Extract, Transform, Load) data pipelines.
  • Familiarity with Apache Tika for metadata and text extraction from various document types.
  • Experience as a Data Layer Architect, focusing on fusing hybrid data sources into a common model.
  • A Master’s degree or equivalent experience in Computer Science is preferred
  • Proven ability to collaborate effectively within a team and work independently to deliver results.
  • Experience working with enterprise data catalogs, adhering to data governance policies, and handling sensitive data and information with appropriate discretion and security protocols.
  • Demonstrated organizational skills to manage complex projects and meet deadlines within established organizational processes, including data transfer protocols.

 

Certifications:

  • AWS solution developer or AWS solution architect.

About Us

Restless Software is an equal opportunity employer that provides highly skilled technologists and creative solutions to clients in both the private and public sectors.