We have a need for one Presto Data Engineer/BigData Developer to be on site in San Francisco.
Responsibilities
- Build and operate large-scale data infrastructure programs (performance, reliability, monitoring)
- Write well-crafted, well-tested, readable, maintainable code
- Participate in code reviews to ensure code quality and distribute knowledge, including Open-Source projects
- Share your knowledge by giving brown bags, tech talks, and evangelizing appropriate tech and engineering best practices
Experience
- 3+ years of relevant professional experience
- Deep understanding of distributed systems principles (consistency and availability, liveness and safety, durability, reliability, fault-tolerance, consensus algorithms)
- Experience bringing open source software to production at scale (Yarn, HDFS, Hive, Spark, Presto, ZooKeeper, Airflow)
- Experience designing, implementing and debugging distributed systems that run across thousands of nodes
- Hands-on experience with Hadoop (or similar) ecosystem - Yarn, Hive, HDFS, Spark, Presto, Parquet
- Experience configuring, identifying performance bottlenecks and tuning MPP databases
- Experience thinking through long-term impacts of important design decisions and handling failure scenarios