hotskills.tech

HotSkills.Tech - Explore Job Opportunities

Explore Job Opportunities

Find the latest job openings from top companies around the world. Use the filters to refine your search and discover the perfect role for you.

Featured Jobs

TechCorp logo

Senior Python Developer

TechCorp

San Francisco, CA

DataCo logo

Data Scientist

DataCo

New York, NY

WebSolutions logo

Frontend Developer

WebSolutions

Remote

CloudTech logo

DevOps Engineer

CloudTech

Seattle, WA

Open Jobs

ADVANCED E-SOLUTIONS PTE. LTD. logo

Data Engineer(IDMC & Databricks)

ADVANCED E-SOLUTIONS PTE. LTD. - Singapore

nSearch Global logo

Kafka Developer

nSearch Global - Singapore

Government Technology Agency logo

Data Scientist (Computer Vision), AI Practice

Government Technology Agency - Singapore

EDGE logo

Radar Software Engineer

EDGE - Abu Dhabi - United Arab Emirates

ValueLabs logo

Rust Senior Engineer

ValueLabs - Dubai - United Arab Emirates

BlackStone eIT logo

Senior Data Engineer (2-Months) Contract

BlackStone eIT - Dubai - United Arab Emirates

Nouryon logo

Technical Service Manager-Oilfield Applications

Nouryon - Dubai - United Arab Emirates

ADVANCED E-SOLUTIONS PTE. LTD. logo

ADVANCED E-SOLUTIONS PTE. LTD. - Singapore

Job Type: Not specified

Experience: Not specified

About Job

POSITION GENERAL DUTIES AND TASKS : Key Skills: Expertise in AWS services, Databricks, and/or Informatica IDMC. Python, Java, or Scala SQL and NoSQL databases AWS certifications, Databricks certifications, and Informatica certifications • Design and architect data storage solutions, including databases, data lakes, and warehouses,using AW S services such as Amazon S3, Amazon RDS, Amazon Redshift, and Amazon DynamoDB, along with Databricks' Delta Lake. Integrate Informatica IDMC for metadata management and data cataloging. • Create, manage, and optimize data pipelines for ingesting, processing, and transforming data using AWS services like AWS Glue, AWS Data Pipeline, and AWS Lambda, Databricks for advanced data processing, and Informatica IDMC for data integration and quality. • Integrate data from various sources, both internal and external, into AWS and Databricks environments, ensuring data consistency and quality, while leveraging Informatica IDMC for data integration, transformation, and governance. • Develop ETL (Extract, Transform, Load) processes to cleanse, transform, and enrich data, making it suitable for analytical purposes using Databricks' Spark capabilities and Informatica IDMC for data transformation and quality. • Monitor and optimize data processing and query performance in both AWS and Databricks environments, making necessary adjustments to meet performance and scalability requirements. Utilize Informatica IDMC for optimizing data workflows. • Implement security best practices and data encryption methods to protect sensitive data in both AWS and Databricks, while ensuring compliance with data privacy regulations. Employ Informatica IDMC for data governance and compliance. • Implement automation for routine tasks, such as data ingestion, transformation, and monitoring, using AWS services like AWS Step Functions, AWS Lambda, Databricks Jobs, and Informatica IDMC for workflow automation. • Maintain clear and comprehensive documentation of data infrastructure, pipelines, and configurations in both AWS and Databricks environments, with metadata management facilitated by Informatica IDMC. • Collaborate with cross-functional teams, including data scientists, analysts, and software engineers, to understand data requirements and deliver appropriate solutions across AWS, Databricks, and Informatica IDMC. • Identify and resolve data-related issues and provide support to ensure data availability and integrity in both AWS, Databricks, and Informatica IDMC environments. • Optimize AWS, Databricks, and Informatica resource usage to control costs while meeting performance and scalability requirements. • Stay up-to-date with AWS, Databricks, Informatica IDMC services, and data engineering best practices to recommend and implement new technologies and techniques. Requirements: Bachelor’s or master’s degree in computer science, data engineering, or a related field. • Minimum 5 years of experience in data engineering, with expertise in AWS services, Databricks, and/or Informatica IDMC. • Proficiency in programming languages such as Python, Java, or Scala for building data pipelines. • Evaluate potential technical solutions and make recommendations to resolve data issues especially on performance assessment for complex data transformations and long running data processes. • Strong knowledge of SQL and NoSQL databases. • Familiarity with data modeling and schema design. • Excellent problem-solving and analytical skills. • Strong communication and collaboration skills. • AWS certifications (e.g., AWS Certified Data Analytics - Specialty, AWS Certified Data Analytics - Specialty), Databricks certifications, and Informatica certifications are a plus. Preferred Skills: • Experience with big data technologies like Apache Spark and Hadoop on Databricks. • Knowledge of containerization and orchestration tools like Docker and Kubernetes. • Familiarity with data visualization tools like Tableau or Power BI. • Understanding of DevOps principles for managing and deploying data pipelines. • Experience with version control systems (e.g., Git) and CI/CD pipelines. • Knowledge of data governance and data cataloguing tools, especially Informatica IDMC. job perk: Posted At: 15 hours ago, job perk: Schedule Type: Contractor

Enhance Your Job Prospects

  • Access exclusive job listings in specialized industries.
  • Optimize your CV with our AI-powered tools.
  • Define salary and equity expectations from the start.
  • Find the perfect fit with personalized job filters.
  • Get top founders and recruiters reaching out to you.

Showing 400 to 406 of 2411 results