Data Engineer with Python Internship/Training is a self paced 4 week Industrial Training program with mentorship support.
In this Internship/Training you’ll discover how to build an effective data architecture, streamline data processing, and maintain large-scale data systems. In addition to working with Python, you’ll also grow your language skills as you work with Shell, SQL, and Scala, to create data engineering pipelines, automate common file system tasks, and build a high-performance database.
Through hands-on exercises, you’ll add cloud and big data tools such as AWS Boto, PySpark, Spark SQL, and MongoDB, to your data engineering toolkit to help you create and query databases, wrangle data, and configure schedules to run your pipelines. By the end of this Internship/Training, you’ll have mastered the critical database, scripting, and process skills you need to progress your career.
⇩ What we cover in 4 weeks.
Week 1 – 3
Data Engineering for Everyone
Introduction to Data Engineering
Streamlined Data Ingestion with PANDAS
Writing Efficient Python Code
Writing Functions in Python
Introduction to Shell
Data Processing in Shell
Introduction to Bash Scripting
Unit Testing for Data Science in Python
Object-Oriented Programming in Python
Introduction to Airflow in Python
Introduction to PySpark
Introduction to AWS Boto in Python
Introduction to Relational Databases in SQL
Introduction to Scala
Big Data Fundamentals with PySpark
Cleaning Data with PySpark
Introduction to MongoDB in Python
Certificate of Completion
Letter of Recommendation
Complete this course while you work
Rigorous curriculum designed by industry experts
Best performers will also be offered a job within the company.