Role: Big Data DeveloperLocation: United States(Remote)
Responsibilities:Data Pipeline Development: Design, develop, and maintain scalable data pipelines to process and transform large datasets using big data technologies such as Hadoop, Spark, and Kafka.Data Integration: Integrate data from various sources into a unified data platform, ensuring data consistency, accuracy, and reliability.Performance Optimization: Optimize data processing and query performance, troubleshoot performance issues, and enhance system efficiency.Data Modeling: Develop and maintain data models, schemas, and data dictionaries to support analytics and reporting needs.Collaboration: Work closely with data scientists, analysts, and business stakeholders to understand data requirements and provide technical solutions.Documentation: Create and maintain comprehensive documentation for data pipelines, workflows, and system configurations.Monitoring and Maintenance: Monitor data processing workflows and systems, perform regular maintenance, and address any issues that arise.
Qualification:Proficiency in big data technologies (Hadoop, Spark, Hive, Pig, etc.).Experience with data processing frameworks and tools (e.g., Apache Kafka, Apache Flink).Strong programming skills in languages such as Java, Scala, Python, or SQL.Familiarity with cloud platforms and services (AWS, Azure, Google Cloud) is a plus.Experience with relational databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra).Relevant certifications in big data technologies or cloud platforms.
RegardsPraveen KumarTalent Acquisition Group – Strategic Recruitment Manager