Job Description
Primary Responsibilities:
- Design and implement scalable and efficient data pipelines to collect, process, and store data from various sources.
- Strong understanding of data modeling concepts and able to design various components of data model and data engineering solution.
- Develop and maintain data warehouses, databases, and data lakes for storage and retrieval.
- Implement ETL (Extract, Transform, Load) processes to clean, transform, and load data into appropriate formats.
- Ensure data quality, consistency, and accuracy through data validation and cleansing techniques.
- Performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
- Provide technical support and guidance to data users and stakeholders.
Qualifications
Minimum Qualifications:
- Bachelor's degree in Computer Science, Information Technology, or related field
- Minimum 2+ years as a Data Engineer.
- Proficient in cloud technologies (AWS)
- Experience with Data engineering tools (Python, Spark (Scala or Python), Hadoop).
- Experience with relational databases (MySQL)
- Experience with data pipeline and workflow management tools
Preferred Qualifications:
- Proven experience as a Data Engineer, ETL Developer, or a related role.
- Strong knowledge of ETL tools and processes.
- Experience with data modeling, database design, and data warehousing concepts.
- Proficiency in at least one programming language (Python, Java, Scala, etc.)
- Familiarity with data visualization tools (Tableau, Power BI, etc.) is a plus.
- Strong problem-solving skills and attention to detail.
- Excellent communication and collaboration skills.
- Ability to work in a fast-paced and dynamic environment.