fbpx
Learn to build large language model applications: vector databases, langchain, fine tuning and prompt engineering. Learn more

Data Engineer

Remote | Pakistan

Full-Time

About the role:

As the world becomes increasingly digitized, data is being generated at an unprecedented rate. That’s why at Data Science Dojo, we’re on the lookout for a Data Engineering Intern who is passionate about working with big data. 

In this role, you’ll be creating and evaluating data pipelines, investigating new technologies, and presenting insights understandably. You’ll also create tutorials and blog posts on trending topics in the world of data science. You’ll work closely with our data science, development, and social media teams to use trends from credible sources to add value to your work.  

What you will do:

  • Developing reliable data pipelines to convert data into valuable insights 
  • Configuring end-to-end components of big data pipelines for various data science problems 
  • Exploring and documenting new technologies in coordination with technical writers and content developers 
  • Developing, documenting, and launching products using the latest AI tools from cloud providers 
  • Sharing data engineering insights with a learning community 
  • Coming up with innovative ways to explain data engineering processes through tutorials, workshops, and training programs 
  • Curating, researching, designing, and writing tutorials and blog posts for various data science, big data, and predictive analytics tools 
  • Researching, analyzing, and evaluating new, big data technologies 
  • Defining metrics of success for products and services and keeping track of progress

What we are looking for:

  • An undergraduate degree in Mathematics, Computer Science, Electrical and Computer Engineering, or a related field. 
  • Knowledge and proven experience in working with at least one programming language, virtual machines, and virtual networks 
  • Experience working with both SQL and NoSQL databases 
  • Familiarity with data tools and services in Azure, AWS, and/or GCP eco-system 
  • Experience using quantitative and qualitative data to make decisions, devise strategies, and measure the progress of projects over time 
  • Excellent English verbal and written communication skills 
  • Ability to work in a fast-paced and highly collaborative environment 
  • Willingness to work as an individual and in a team as per the needs of a project 
  • Experience working with tools like chatbots, computer vision, NLP, and spatial analysis 
  • Technical writing experience 
  • Prior experience working in the development team at a reputed organization 
  • Development experience on Azure Cloud Platform, IntelliJ, and/or Visual Studio 
  • Proven experience working with MapReduce, Hadoop, and Spark 
  • Experience with Redis 
  • Experience with data validation (Great Expectations, Amazon Deequ) 
  • Experience with Azure Databricks 
  • Experience with PowerShell scripting and/or Azure CLI

Apply Now