No more applications are being accepted for this job
- Develop and maintain data processing pipelines using Core Java and distributed data processing systems like Spark SQL.
- Implement distributed architectures, SOA, Micro-services, and PaaS solutions for efficient data handling.
- Utilize Hadoop ecosystem tools such as HDFS, Spark, Impala, and Hive for data storage and manipulation.
- Work on source code control using Git and perform Unix/Linux scripting as needed.
- Collaborate with teams to design and implement scalable data solutions for Big Data Analytics and Data Transformation.
- Conduct large-scale data crawling activities on cloud platforms, leveraging industry best practices.
- Apply computer science fundamentals in algorithm design, problem-solving, and complexity analysis to optimize data workflows.
- Proficiency in Core Java and Spark SQL with good SPARKS skills.
- Understanding of distributed architectures, SOA, Micro-services, and PaaS concepts.
- Hands-on experience with HDFS, Spark, Impala, Hive, and database technologies.
- Familiarity with source code control using Git and Unix/Linux scripting.
- Knowledge of Distributed Computing, Big Data Analytics, and Data Transformation.
- Experience in large-scale crawling on cloud platforms is a plus.
- Strong problem-solving skills and a background in Computer Science fundamentals.
1 alternative title for "Data engineer" using "en-US" language - Singapore - U3 INFOTECH PTE. LTD.
Description
Roles & ResponsibilitiesRequirements:
Git
Big Data
Pipelines
Hadoop
Data Transformation
Scripting
Big Data Analytics
ETL
Microservices
Data Engineering
SQL
Core Java
Python
SOA
Algorithm Design
Databases