Location : Metro Manila (1605) -
Salary : Competitive
Type : Permanent
Main Industry :
Job ID : 106963596
Employer : iScale Solutions
Posted On : 07 January 2021 06/02/2021
-Build, implement and optimize highly scalable batch and streaming data ingress/egress pipelines
-Collaborate with the company's Analytics / Machine Learning teams to build and implement data pipelines to feed machine learning algorithms within Company's Hadoop platform
-Develop tools and automate processes to aid in data collection, analysis and monitoring
-Collaborate with Product Engineering and Platform teams to make architecture design and implementation decisions atop the company's Hadoop platform
-Provide engineering, installation, configuration, maintenance and support in a highly transactional 24x7 environment
-Perform a mix of incident management and project work focused on automation, increasing scale and optimization of processes/system performance
-Monitor platform and applications and take corrective action to prevent or minimize system downtime on Hadoop platform
-Recommend best practices, and implementation strategies using Hadoop, Java, ETL tools.
-Assist Hadoop admins with incident global management/resolution - light on-call rotation
-Work with the global Engineering team to assess current platform configuration and make
-recommendations to achieve optimal performance and horizontal scalability within Hadoop ecosystem
-Lead periodic provisioning of environments, code deployments and maintenance patching in Hadoop environments globally
-Collaborate with the Product Engineering team to develop proof-of-concept solutions as well as custom solutions for customers that leverage the company's cloud-hosted Engagement Data platform
Requirements
-Bachelor's Degree in Computer Science or related field
-2-4 Years hands-on experience working with data at scale
-Self-starters who collaborate well with others and take ownership of their projects.
-Experience and desire to work in a Global delivery environment
-Hands-on experience working within the Hadoop ecosystem (Spark, Hive, HBase, Storm) ideally in a cloud environment (AWS, GCP or Azure).
-Experience with optimizing SQL/Hive queries for maximum throughput
-SQL/noSQL technology. Familiar with Databases like Oracle, SQL Server, MySQL, MongoDB, Redis etc. Ideally in a cloud setting (ie. AWS RDS)
-Experience operating web-scale deployments of distributed systems, e.g. Kafka, Flink, Storm, Cassandra, Kubernetes or Elasticsearch
-Experience with data warehouses and building ETL workflows / data pipelines
-Data application/platform instrumentation, measurement, log data processing, and monitoring.
-Fluency in Python, Java, Scala, or a similar language-familiarity with more than one is a plus
-Mastery of Unix/Linux system and shell scripting
-Experience with orc, parquet, avro and other data formats
-Excellent communication skills, both verbal and written
Nice to have
-Strong DevOps mindset and skill set
-Experience working with HDP Hive Interactive (LLAP)
-Experience in performance tuning for TEZ and Spark.
-Experience building data pipelines (ie. Apache Airflow, Spark Data Pipelines)
-Experience doing light data science in a Hadoop/Cloud setting (visualizations, clustering, classification, regression) to help predict and pro-actively optimize performance
-Experience creating dashboards with analytics tools like Looker, Tableau.
-Experience with Graph databases (ie. Neo4J, AWS Neptune, Cayley, Gremlin)
-Experience with Microsoft's SQL Server Integration Services (SSIS) or similar tools.
-Experience working with AWS services (ie. S3, EC2, Kinesis, Lambda, etc...)
-Amazon Web Services, Google Cloud Platform or Microsoft Azure certifications
-This is a remote position.
Tiptopjob
Salary : Competitive
Type : Permanent
Main Industry :
Job ID : 106963596
Employer : iScale Solutions
Posted On : 07 January 2021 06/02/2021
-Build, implement and optimize highly scalable batch and streaming data ingress/egress pipelines
-Collaborate with the company's Analytics / Machine Learning teams to build and implement data pipelines to feed machine learning algorithms within Company's Hadoop platform
-Develop tools and automate processes to aid in data collection, analysis and monitoring
-Collaborate with Product Engineering and Platform teams to make architecture design and implementation decisions atop the company's Hadoop platform
-Provide engineering, installation, configuration, maintenance and support in a highly transactional 24x7 environment
-Perform a mix of incident management and project work focused on automation, increasing scale and optimization of processes/system performance
-Monitor platform and applications and take corrective action to prevent or minimize system downtime on Hadoop platform
-Recommend best practices, and implementation strategies using Hadoop, Java, ETL tools.
-Assist Hadoop admins with incident global management/resolution - light on-call rotation
-Work with the global Engineering team to assess current platform configuration and make
-recommendations to achieve optimal performance and horizontal scalability within Hadoop ecosystem
-Lead periodic provisioning of environments, code deployments and maintenance patching in Hadoop environments globally
-Collaborate with the Product Engineering team to develop proof-of-concept solutions as well as custom solutions for customers that leverage the company's cloud-hosted Engagement Data platform
Requirements
-Bachelor's Degree in Computer Science or related field
-2-4 Years hands-on experience working with data at scale
-Self-starters who collaborate well with others and take ownership of their projects.
-Experience and desire to work in a Global delivery environment
-Hands-on experience working within the Hadoop ecosystem (Spark, Hive, HBase, Storm) ideally in a cloud environment (AWS, GCP or Azure).
-Experience with optimizing SQL/Hive queries for maximum throughput
-SQL/noSQL technology. Familiar with Databases like Oracle, SQL Server, MySQL, MongoDB, Redis etc. Ideally in a cloud setting (ie. AWS RDS)
-Experience operating web-scale deployments of distributed systems, e.g. Kafka, Flink, Storm, Cassandra, Kubernetes or Elasticsearch
-Experience with data warehouses and building ETL workflows / data pipelines
-Data application/platform instrumentation, measurement, log data processing, and monitoring.
-Fluency in Python, Java, Scala, or a similar language-familiarity with more than one is a plus
-Mastery of Unix/Linux system and shell scripting
-Experience with orc, parquet, avro and other data formats
-Excellent communication skills, both verbal and written
Nice to have
-Strong DevOps mindset and skill set
-Experience working with HDP Hive Interactive (LLAP)
-Experience in performance tuning for TEZ and Spark.
-Experience building data pipelines (ie. Apache Airflow, Spark Data Pipelines)
-Experience doing light data science in a Hadoop/Cloud setting (visualizations, clustering, classification, regression) to help predict and pro-actively optimize performance
-Experience creating dashboards with analytics tools like Looker, Tableau.
-Experience with Graph databases (ie. Neo4J, AWS Neptune, Cayley, Gremlin)
-Experience with Microsoft's SQL Server Integration Services (SSIS) or similar tools.
-Experience working with AWS services (ie. S3, EC2, Kinesis, Lambda, etc...)
-Amazon Web Services, Google Cloud Platform or Microsoft Azure certifications
-This is a remote position.
Tiptopjob
Other Info
Metro Manila
Permanent
Full-time
Permanent
Full-time
Submit profile
iScale Solutions
About the company
iScale Solutions jobs
Makati City, Metro Manila





Service Engineer / Sales Representative
Zion Diagnostics Solutions
MetroManila, Quezon, Quezon, ManilaAgreement





Position big Data Engineer recruited by the company iScale Solutions at MetroManila, Manila, Joboko automatically collects the salary of , finds more jobs on Big Data Engineer or iScale Solutions company in the links above
About the company
iScale Solutions jobs
Makati City, Metro Manila