Primary Skillset: Python PySpark with ETL AWS background.
· 6 years working experience in data integration and pipeline development.
· BS degree in CS CE or EE
· 2 years of Experience with AWS Cloud on data integration with Apache Spark EMR Glue Kafka Kinesis and Lambda in S3 Redshift RDS MongoDB Dynamo DB ecosystems.
· Strong real life experience in python development especially in pySpark in AWS Cloud environment.
· Design develop test deploy maintain and improve data integration pipeline.
· Experience in Python and common python libraries Strong analytical experience with database in writing complex queries query optimization debugging user defined functions views indexes etc.
Strong experience with source control systems such as Git Bitbucket and Jenkins build and continuous integration tools Databricks or Apache Spark Experience is a plus
All your information will be kept confidential according to EEO guidelines.