Junior Big Data Engineer

May 30, 2024
Application deadline closed.

Job Description

DATASPARK HIRE is Hiring for Junior Big Data Engineer

We’re in search of a dynamic Big Data Engineer to join our data engineering squad. This role suits recent graduates brimming with enthusiasm for big data tech and eager to put their skills into practice in a professional capacity. The triumphant candidate will aid in crafting, erecting, and nurturing our big data infrastructure and pipelines.

Primary Duties:

  • Assist in Crafting Data Pipelines:

    • Participate in the design, construction, and upkeep of scalable data pipelines for handling hefty data loads.
    • Aid in the deployment of ETL (Extract, Transform, Load) procedures to safeguard data integrity and excellence.
  • Data Storage and Administration:

    • Master the art of managing and refining big data storage solutions (e.g., HDFS, NoSQL databases).
    • Contribute to the establishment and upkeep of data warehouses and data lakes.
  • Data Processing:

    • Acquire hands-on familiarity with big data processing frameworks like Apache Spark, Hadoop, and Kafka.
    • Contribute to the creation of real-time and batch data processing resolutions.
  • Data Integration:

    • Assist in weaving together data from diverse origins, ensuring coherence and precision.
    • Collaborate with data scientists, analysts, and other stakeholders to grasp data requisites.
  • Performance Enhancement:

    • Learn the ropes of optimizing data processing efficiency and tackling pipeline glitches.
    • Assist in embedding best practices for data security and conformity.
  • Collaboration and Learning:

    • Forge close alliances with multifaceted teams to amass requisites and deliver data resolutions.
    • Partake in educational sessions and mentorship initiatives to sharpen technical prowess.
  • Continuous Enhancement:

    • Keep abreast of cutting-edge industry trends and big data technologies.
    • Contribute to the ongoing enhancement of data engineering processes and utilities.

Essential Criteria:

  • Academic Background:

    • Bachelor’s degree in Computer Science, Information Technology, or a related discipline.
  • Technical Proficiency:

    • Foundational grasp of big data technologies (e.g., Hadoop, Spark, Kafka) via coursework or projects.
    • Familiarity with ETL processes and utilities.
    • Basic coding proficiency in Java, Scala, or Python.
    • Understanding of SQL and NoSQL databases.
    • Familiarity with cloud platforms (AWS, Azure, Google Cloud) and their big data services is advantageous.
  • Experience:

    • Prior internships or academic ventures involving big data technologies are advantageous, though not mandatory.

Soft Skills:

  • Proficient problem-solving and analytical abilities.
  • Effective communication and teamwork aptitude.
  • Swift learning curve and adaptability to fresh technologies.
  • Attentive to detail with an emphasis on data quality and precision.