Data Engineer

Recorded Future is the world's most advanced, and largest, intelligence company, with 1,000 intelligence professionals, over $300M in sales, and serving over 1,800 clients worldwide.
Data
Mid-Level Software Engineer
Hybrid
501 - 1,000 Employees
1+ year of experience
Cybersecurity

Description For Data Engineer

Recorded Future is seeking a Data Engineer to join their team working on building reliable data pipelines that ingest raw data and convert it into actionable analytics to identify and combat cyber security threats. The role involves developing, productionalizing, and maintaining high volume data extraction via custom harvesters, APIs, and data stores. You'll create well-monitored production data pipelines, maintain existing systems, and collaborate with senior engineers through best practices.

Key responsibilities include:

  • Developing and maintaining high volume data extraction systems
  • Creating production data pipelines to convert raw data into actionable information
  • Maintaining and upgrading existing data flow production systems
  • Collaborating with senior engineers through design reviews, code reviews, and pair programming

Requirements:

  • 1+ years of Python experience, with strong debugging skills
  • 1+ years of experience working with large datasets
  • Experience transforming large datasets in Python and querying modern datastores like MongoDB and RedShift
  • Passion for coding and eagerness to learn new skills and practices
  • Willingness to understand how your work impacts clients and the intelligence community

Recorded Future offers a diverse and inclusive work environment, with employees representing over 40 nationalities. They have a strong reputation, with a 4.8-star user rating from Gartner and more than 45 of the Fortune 100 companies as clients. The company is committed to empowering clients with intelligence to disrupt adversaries.

Benefits and perks are not explicitly mentioned in the job posting, but Recorded Future is known for its commitment to diversity, inclusion, and employee growth. The company offers opportunities to work on cutting-edge cybersecurity challenges and contribute to a rapidly evolving field.

Last updated 2 months ago

Responsibilities For Data Engineer

  • Develop, productionalize and maintain high volume data extraction via custom harvesters, APIs, and data stores
  • Create well monitored production data pipelines which convert raw data into well organized and actionable information
  • Maintain and upgrade existing data flow production systems
  • Collaborate with senior engineers through design reviews, code reviews, pair programming, and other software engineering best practices

Requirements For Data Engineer

Python
MongoDB
  • 1+ years of Python experience with strong debugging skills
  • 1+ years of experience working with large datasets
  • Experience transforming large datasets in Python
  • Experience querying modern datastores such as MongoDB and RedShift
  • Passion for coding and eagerness to learn new skills and practices
  • Willingness to understand how work impacts clients and the intelligence community

Interested in this job?

Jobs Related To Recorded Future Data Engineer

Data Engineer II, Attack Surface Module

Data Engineer II position at Recorded Future, focusing on building and maintaining data pipelines for security teams, requiring 3+ years of Python experience.

Data Engineer II, Data Pipelines

Data Engineer II at Recorded Future: Build and maintain high-volume data pipelines for cyber threat intelligence.

Business Intelligence Engineer, ORC (ORC- Operations Risk Compliance) Program Analytics

Business Intelligence Engineer role at Amazon focusing on ORC Analytics, combining statistical analysis, data engineering, and business intelligence expertise in London.

Data Engineer

Data Engineer position at WorldQuant focusing on developing data pipelines and engineering solutions for financial strategies.

Data Engineer

Data Engineer position at G-P, developing solutions for their Global Employment Platform, working with Python, SQL, and modern data technologies in a remote environment.