We are looking for a strong Data Engineer with our direct clients in San Francisco Bay Area, CA. The positions are long term and remote until the offices are open.
Designs, develops, and implements Hadoop eco-system based applications to support business requirements. Follows approved life cycle methodologies, creates design documents, and performs program coding and testing. Resolves technical issues through debugging, research, and investigation. Experience/Skills Required:1. Bachelors degree in Computer Science, Information Technology, or related field and 5 years experience in computer programming, software development or related2. 3+ years of solid Java and 2+ years experience in design, implementation, and support of solutions big data solution in Hadoop using Hive, Spark, Drill, Impala, HBase3. Hands on experience with Unix, Teradata and other relational databases. Experience with cloud based technologies, preferably Google Cloud Platform .4. Strong communication and problem-solving skills Additional Job Details • 4+ years of experience in software development, focusing on distributed systems • A self-starter who enjoys reading well-written code, and writes code that others enjoy reading • Understands, articulates, and applies principles of the defined strategy to routine business problems that involve a single function. • Excellent command of design patterns, data structures and algorithms • Expert knowledge of software design, distributed systems • Extracts data cloud based and on-prim systems, creates data pipelines and transform data to a structure that is relevant by selecting appropriate techniques • Advanced coding skills in Python, Java, Scala, or similar • Willingness and ability to learn and apply new technologies quickly • Can navigate complex systems and deliver highly scalable and reliable production-ready code Nice to Have: • Experience with big data technology (Hadoop, Spark, Storm, Presto etc.) and large-scale analytics systems that are capable of ingesting, managing, storing, and analyzing hundreds of terabytes of data • Excellent data analytics skills and ability to explore and identify data issues • Familiarity with CI/CD processes and tools such as Jenkins and Maven • Experience with Google Cloud Platform technologies, especially BigQuery, GCS and DataProc • Experience with workflow management systems like airflow, oozie and digdag • Exposure to analytical trends, machine learning and data science • Experience in AdTech and advertising measurement