For a hands-on learning experience to develop LLM applications, join our LLM Bootcamp today.
Early Bird Discount Ending Soon!

Associate Data Engineer – I

Remote | Pakistan

Full Time | Pacific Time

About the product and job role:

Our Product: Ejento, built by Data Science Dojo, is a cutting-edge RAG platform that enables LLM application developers to build retrieval-augmented generation (RAG)-based applications efficiently. As we scale our platform, ensuring security, privacy, and regulatory compliance is a top priority. We are committed to implementing robust security frameworks to protect data, ensure compliance, and mitigate risks in AI-driven environments.

Job Role: We are seeking an Associate Data Engineer – I to design, build, and optimize data pipelines that power our AI and analytics solutions. In this role, you will be responsible for configuring and managing large-scale data processing systems, developing robust data extraction and transformation workflows, and collaborating with cross-functional teams to meet evolving data needs. Your contributions will directly support our AI automation initiatives and strengthen our data infrastructure to enable smarter insights and decision-making.

What you will do:

  • Configuring components for big data pipelines to address diverse data science challenges.
  • Building and managing data pipelines for AI Automation, including data preprocessing tasks.
  • Creating robust data pipelines to facilitate data analysis and insights generation.
  • Contributing to product development using cutting-edge AI tools and cloud services.
  • Independently designing, building, and launching new data extraction, transformation, and loading processes in production.
  • Collaborating with cross-functional teams to meet data requirements and communicate insights effectively.
  • Designing and implementing data transfer solutions for various systems.

What we are looking for:

  • Pursuing or having completed a degree in Computer Science, Mathematics, Software Engineering, Computer Engineering, or a related field
  • Basic understanding of data engineering concepts: data modeling, ETL, data warehousing.
  • Familiarity with relational and NoSQL databases (MySQL, PostgreSQL, MongoDB).
  • Knowledge of SQL or similar query languages.
  • Awareness of cloud platforms and services for data engineering (AWS, Azure, Google Cloud).
  • Basic proficiency in Python, or any other programming language, for data manipulation and visualization (NumPy, Pandas, Matplotlib, Seaborn)
  • Knowledge of version control (Git).
  • Strong problem-solving and analytical thinking.
  • Robust organizational skills for multitasking and attention to detail; ensuring high-quality work.
  • Quick learning and adaptability to new technologies.
  • Effective communication of technical concepts and collaboration with team members.

Nice to have: 

  • Certifications in Azure Fundamentals (AZ-900), Azure Data Fundamentals (DP-900), or Azure Data Engineer Associate (DP-203) 
  • Knowledge of data processing technologies (Apache Spark, Hadoop, AWS Glue). 
  • Understanding of data visualization tools (Tableau, Power BI). 
  • Familiarity with data integration techniques and tools. 
  • Familiarity with workflow automation (Apache Airflow). 
  • Basic knowledge of containerization (Docker, Kubernetes). 
  • Knowledge of data serialization formats (JSON, Parquet, XML, Avro). 

Apply Now