Job Description
Lead Data Engineer
Primary Skills
MapReduce, HDFS, Spark - Pyspark, ETL Fundamentals, SQL, SQL (Basic + Advanced), Spark - Scala, Python, Data Warehousing, Hive, Modern Data Platform Fundamentals, Data Modelling Fundamentals, PLSQL, T-SQL, Stored Procedures, Oozie Specialization
Big Data Engineering: Lead Data Engineer Job requirements
Job Description:
Remote (for California based team) we’re looking for a Senior Data Engineer to join our team. What we’re looking for: You’re a talented, creative, and motivated engineer who loves developing powerful, stable, and intuitive apps – and you’re excited to work with a team of individuals with that same passion. You’ve accumulated years of experience, and you’re excited about taking your mastery of Big Data and Java to a new level. You enjoy challenging projects involving big data sets and are cool under pressure. You’re no stranger to fast-paced environments and agile development methodologies – in fact, you embrace them. With your strong analytical skills, your unwavering commitment to quality, your excellent technical skills, and your collaborative work ethic, you’ll do great things here.
What you’ll do:
As a Senior Data Engineer, you’ll be responsible for building high performance, scalable data solutions that meet the needs of millions of agents, brokers, home buyers, and sellers. You’ll design, develop, and test robust, scalable data platform components. You’ll work with a variety of teams and individuals, including product engineers to understand their data pipeline needs and come up with innovative solutions. You’ll work with a team of talented engineers and collaborate with product managers and designers to help define new data products and features. Skills, accomplishments, interests you should have: BS/MS in Computer Science, Engineering, or related technical discipline or equivalent combination of training and experience. 7+ years core Scala/Java experience: building business logic layers and high-volume/low latency/big data pipelines. 5+ years of experience in large scale real-time stream processing using Apache Flink or Apache Spark with messaging infrastructure like Kafka/Pulsar. 7+ years of experience on Data Pipeline development, ETL and processing of structured and unstructured data. 5+ years of experience using NoSQL systems like MongoDB, DynamoDB and Relational SQL Database systems (PostgreSQL) and Athena. Experience with technologies like Lambda, API Gateway, AWS Fargate, ECS, CloudWatch, S3, DataDog. Experience owning and implementing technical/data solutions or pipelines. Excellent written and verbal communication skills in English. Strong work ethic and entrepreneurial spirit.
Know more about:
Brillio is an equal opportunity employer to all, regardless of age, ancestry, colour, disability (mental and physical), exercising the right to family care and medical leave, gender, gender expression, gender identity, genetic information, marital status, medical condition, military or veteran status, national origin, political affiliation, race, religious creed, sex (includes pregnancy, childbirth, breastfeeding, and related medical conditions), and sexual orientation.