Responsibilities
- Python programmers/developers who done extensive hands-on work in data engineering space.
- Willingness to quickly learn and adapt.
- Experience in solutioning and implementing data pipelines, data curation, data modeling and implementing data solutions.
- Strong understanding of different type of data and the life-cycle of data.
- Design, develop, and launch extremely efficient and reliable data pipelines using Python frameworks to move data and to provide intuitive analytic to our partner teams.
- Collaborate with other engineers and Data Scientists to Client for the best solutions.
- Diagnose and solve issues in our existing data pipelines and envision and build their successors.
Required Qualifications
- Bachelor's degree in Computer Science or equivalent work experience
- Minimum 10+ years experience in IT
- 8+ years Proficiency working with Python specifically related to data processing with proficiency in Python Libraries – Pandas, NumPy, PySpark, PyOdbc, PyMsSQL, Requests, Boto3, SimpleSalesforce, Json.
- 4+ years Strong SQL (SQL, performance, Stored Procedures, Triggers, schema design) skills and knowledge of one of more RDBMS like MSSQL/MySQL.
- 3+ years Strong AWS skills using, Athena, Lambda, S3, EC2, EMR, AWS Console, IAM, STS AWS Data Exchange, Cloud Formation
- 3+ years on Data Warehouse technologies like Snowflake/Redshift
- 3+ years on Databricks/TalenD/Informatica
- 3+ years on ETL tools like Apache Airflow/AWS Glue/Azure Data Factory/Alteryx
- 3+ year in Hadoop, Hive
- Excellent verbal communication skills.
- Knowledge of DevOps/Git for agile planning and code repository