Job Title:
Senior Data Engineer (Pyspark, Apache, Big data)
Job Description
We're Concentrix. The intelligent transformation partner. Solution-focused. Tech-powered. Intelligence-fueled.
The global technology and services leader that powers the world’s best brands, today and into the future. We’re solution-focused, tech-powered, intelligence-fueled. With unique data and insights, deep industry expertise, and advanced technology solutions, we’re the intelligent transformation partner that powers a world that works, helping companies become refreshingly simple to work, interact, and transact with. We shape new game-changing careers in over 70 countries, attracting the best talent.
In our Information Technology and Global Security team, you will deliver the latest technology infrastructure, transformative software solutions and industry-leading global security for our staff and clients. You will work with the best in the world to design, implement and strategize IT, security, application development, innovation, and solutions in today’s hyperconnected world. You will be part of the technology team that is core to our vision of develop, build and run the future of Integrated Services.
Our game-changers around the world have devoted their careers to ensuring every relationship is exceptional. And we’re proud to be recognized with awards such as "World's Best Workplaces," “Best Companies for Career Growth,” and “Best Company Culture,” year after year.
We embrace our game-changers with open arms, people from diverse backgrounds, who are curious and willing to learn. Your natural talent to help others and go beyond WOW for our customers will fit right in with what we do and who we are.
Join us and be part of this journey towards greater opportunities and brighter futures.
We are seeking a skilled Data Engineer to join our data team. The successful candidate will be instrumental in designing, developing, and maintaining our data pipelines and optimizing our data infrastructure to support analytics and operational insights.
Key Responsibilities:
- Data Pipeline Development: Design, build, and maintain ETL (Extract, Transform, Load) processes to facilitate the efficient flow of data.
- Big Data Processing: Utilize PySpark and Apache Spark as big data processing engines to handle large-scale data sets.
- Table Format Management: Work with Pyiceberg and Apache Iceberg for managing table formats and optimizing data storage.
- Cloud Services: Implement data solutions utilizing Google Cloud Platform services.
- Data Access and Analysis: Use Dremio for data querying and rapid data access.
- Data Manipulation: Apply other Python data libraries, such as Pandas, Numpy, and Scikit-learn, for data analysis and manipulation.
- SQL Proficiency: Leverage Oracle SQL for database management and complex queries.
Qualifications:
- Bachelor’s degree in Computer Science, Data Science, or a related field.
- Proven experience as a Data Engineer or in a similar role.
- Strong proficiency in PySpark, Apache Spark, Apache Iceberg, and Dremio.
- Experience with Google Cloud Platform.
- Expertise in Python data libraries (Pandas, Numpy, Scikit-learn).
- Strong knowledge of SQL, particularly Oracle SQL.
- Excellent analytical skills and problem-solving abilities.
- Ability to work in a team-oriented and collaborative environment.
Preferred Skills:
- Experience with data visualization tools.
- Familiarity with machine learning frameworks is a plus.
- Strong communication skills and the ability to explain complex technical concepts to non-technical stakeholders.
#LATAMCNX
#W@H
Location:
SLV Work-at-Home
Language Requirements:
Time Type:
Full time
If you are a California resident, by submitting your information, you acknowledge that you have read and have access to the Job Applicant Privacy Notice for California Residents