Senior Specialist, Data Engineer

Posted 2025-04-24
Remote, USA Full-time Immediate Start

Description: ? L3Harris Enterprise Data, Analytics, and Automation team is seeking a Data Engineer with experience in managing enterprise-level data life cycle processes. ? This role includes overseeing data ETL/ELT pipelines, ensuring adherence to data standards, maintaining data frameworks, conducting data cleansing, orchestrating data pipelines, and ensuring data consolidation. ? The selected individual will play a pivotal role in maintaining ontologies, building scalable data solutions, and developing dashboards to provide actionable insights for the enterprise. ? This position will support the company?s modern data platform, focusing on data pipeline development and maintenance, platform design, documentation, and user training. ? The goal is to ensure seamless access to data for all levels of the organization, empowering decision-makers with clean, reliable data. ? Design, build, and maintain robust data pipelines to ensure reliable data flow across the enterprise. ? Maintain data pipeline schedules, orchestrate workflows, and monitor the overall health of data pipelines to ensure continuous data availability. ? Create, update, and optimize data connections, datasets, and transformations to align with business needs. ? Troubleshoot and resolve data sync issues, ensuring consistent and correct data flow from source systems. ? Collaborate with cross-functional teams to uphold data quality standards and ensure accurate data is available for use. ? Utilize Palantir Foundry to establish data connections to source applications, extract and load data, and design complex logical data models that meet functional and technical specifications. ? Develop and manage data cleansing, consolidation, and integration mechanisms to support big data analytics at scale. ? Build visualizations using Palantir Foundry tools and assist business users with testing, troubleshooting, and documentation creation, including data maintenance guides. Requirements: ? Bachelor?s Degree and minimum 6 years prior Palantir experience or Graduate Degree and minimum 4 years of prior Palantir experience ? In lieu of degree, minimum 8 years of prior Palantir experience ? Experience with designing and developing data pipelines in PySpark, Spark SQL, SQL or Code Build. ? Experience in building and deploying data synchronization schedules and maintaining data pipelines using Palantir Foundry. ? Minimum of 4 years of experience with Data Pipeline development or ETL tools such as Palantir Foundry, Azure Data Factory, SSIS, or Python. ? Minimum of 4 years of experience in Data Integration. ? Strong understanding of Business Intelligence (BI) and Data Warehouse (DW) development methodologies. ? Hands-on experience with the Snowflake Cloud Data Platform, including data architecture, query optimization, and performance tuning. ? Proficiency in Python, PySpark, Pandas, Databricks, JavaScript, or other scripting languages for data processing and automation. ? Experience with other ETL tools such as Azure Data Factory (ADF), SSIS, Informatica, or Talend is highly desirable. ? Familiarity with connecting and extracting data from various ERP applications, including Oracle EBS, SAP ECC/S4, Deltek Costpoint, and more. ? Experience with AI tools such as OpenAI, Palantir AIP, Snowflake Cortex or similar. Benefits: ? Health and disability insurance ? 401(k) match ? Flexible spending accounts ? EAP ? Education assistance ? Parental leave ? Paid time off ? Company-paid holidays

Apply Job!

 

Similar Jobs

Back to Job Board