Hiring: Data Engineer (PySpark):
Location: Bangalore/ Kolkata
Package: Open to Discuss
Contact: suguna@bullhornconsultants.com
Looking for immediate to 30 Days of notice period candidate
• 5+ years of experience in Data Analytics role with at least 2 years of
development experience in PySpark
• Very strong in SQL
• Expertise in programming with Python with experience in Data Cleaning,
Feature Engineering, Transformation and building statistical/ML models.
• Experience working on Unix based platforms with basic knowledge of shell
scripting, writing Cron jobs etc.
• Knowledge of big data ecosystem with knowledge of Hadoop, Hive & GitHub
version management
• Knowledge of Cloud Computing (GCP), Data Modelling, exposure to Data
Quality assessment and control.
• Exposure to working on data pertaining to banking and financial services
domain
• Highly adaptable in quickly changing technical environments with strong
organizational and analytical skills
Mandatory Skill Set
1. Hive, Python, Spark/PySprak, on Bigdata platform
2. Python or Unix Scripting
3. GCP Cloud concepts along Airflow development.
Secondary Skill Set ( Not Mandatory Good to have)
1. Dev Op’s with GCP cloud

Hiring : Data Engineer (PySpark).docx

  • 1.
    Hiring: Data Engineer(PySpark): Location: Bangalore/ Kolkata Package: Open to Discuss Contact: suguna@bullhornconsultants.com Looking for immediate to 30 Days of notice period candidate • 5+ years of experience in Data Analytics role with at least 2 years of development experience in PySpark • Very strong in SQL • Expertise in programming with Python with experience in Data Cleaning, Feature Engineering, Transformation and building statistical/ML models. • Experience working on Unix based platforms with basic knowledge of shell scripting, writing Cron jobs etc. • Knowledge of big data ecosystem with knowledge of Hadoop, Hive & GitHub version management • Knowledge of Cloud Computing (GCP), Data Modelling, exposure to Data Quality assessment and control. • Exposure to working on data pertaining to banking and financial services domain • Highly adaptable in quickly changing technical environments with strong organizational and analytical skills Mandatory Skill Set 1. Hive, Python, Spark/PySprak, on Bigdata platform 2. Python or Unix Scripting 3. GCP Cloud concepts along Airflow development. Secondary Skill Set ( Not Mandatory Good to have) 1. Dev Op’s with GCP cloud