Data Engineer (Greenplum) jobs in United States
cer-icon
Apply on Employer Site
company-logo

CoreTek Labs · 12 hours ago

Data Engineer (Greenplum)

CoreTek Labs is seeking a Data Engineer with extensive experience in Big Data technologies. The primary responsibility of this role includes designing and building scalable data pipelines, ensuring data quality and compliance, and collaborating with stakeholders to translate business requirements into technical solutions.

Computer Software
badNo H1Bnote
Hiring Manager
Podila Satish
linkedin

Responsibilities

Designs and builds scalable data pipelines, integrates diverse sources, and optimizes storage/processing using Hadoop ecosystem and Greenplum
Ensures data quality, security, and compliance through governance frameworks
Implements orchestration, monitoring, and performance tuning for reliable, cost-efficient operations
Expertise in Hadoop ecosystem (HDFS, Hive, Spark, Kafka) and MPP databases like Greenplum for large-scale data processing and optimization
Collaborates with Data Owners and stakeholders to translate business rules into technical solutions
Delivers curated datasets, lineage, and documentation aligned with SLAs and regulatory standards
Subject matter expert having experience of interacting with client, understanding the requirement and guiding the team
Documenting the requirements clearly with defined scope and must play a anchor role in setting the right expectations and delivering as per the schedule
Design and develop scalable data pipelines using Hadoop ecosystem and Greenplum for ingestion, transformation, and storage of large datasets
Optimize data models and queries for performance and reliability, ensuring compliance with security and governance standards
Implement data quality checks, monitoring, and orchestration workflows for timely and accurate data delivery
Collaborate with Data Owners and business teams to translate requirements into technical solutions and maintain documentation and lineage

Qualification

GreenplumHadoopBig DataHDFSHiveSparkKafkaAgileData GovernanceData QualityDocumentationCollaboration

Required

12+ Years of experience
Only H1B visa holders – PP number is mandatory
Strong work experience in Data Engineering (Big Data – Hadoop, Greenplum, etc.)
Designs and builds scalable data pipelines
Integrates diverse sources
Optimizes storage/processing using Hadoop ecosystem and Greenplum
Ensures data quality, security, and compliance through governance frameworks
Implements orchestration, monitoring, and performance tuning for reliable, cost-efficient operations
Expertise in Hadoop ecosystem (HDFS, Hive, Spark, Kafka)
Expertise in MPP databases like Greenplum for large-scale data processing and optimization
Collaborates with Data Owners and stakeholders to translate business rules into technical solutions
Delivers curated datasets, lineage, and documentation aligned with SLAs and regulatory standards
Subject matter expert
Documenting the requirements clearly with defined scope
Plays an anchor role in setting the right expectations and delivering as per the schedule
Design and develop scalable data pipelines using Hadoop ecosystem and Greenplum for ingestion, transformation, and storage of large datasets
Optimize data models and queries for performance and reliability
Ensuring compliance with security and governance standards
Implement data quality checks, monitoring, and orchestration workflows for timely and accurate data delivery
Collaborate with Data Owners and business teams to translate requirements into technical solutions
Maintain documentation and lineage

Preferred

Agile environment preferred
Experience of interacting with client
Guiding the team

Company

CoreTek Labs

twittertwitter
company-logo
CoreTek Labs that specializes in Project Management, Product Engineering, Technology Consulting & Staff Augmentation.

Funding

Current Stage
Growth Stage

Leadership Team

leader-logo
Pradeep Kanneganti
CEO and Founder
linkedin
Company data provided by crunchbase