We're sorry, but this job has been placed on hold. See other open jobs at Bitcasa

Big Data Engineer

San Mateo, CA | Analytics

Job Description

Come join the Bitcasa team!

Take advantage of an amazing opportunity to be a part of a fast-growing start up company. Work in a laid back environment that offers magnificent perks including daily catered lunch, gym memberships to Crunch Fitness, a prime location in San Mateo within walking distance of Caltrain (Caltrain pass included) and ten minutes from SFO, and of course, an all-star team.

About Us

Bitcasa is the personal app that offers access to an Infinite Drive across all devices with support for Mac, Windows, iOS, Android, and Windows RT.

Users never have to worry about running out of disk space again. They can save, access, and share an unlimited number of documents, pictures, music, and movies from any device. Everything is encrypted locally and backed up in the cloud, so user data is always safe and secure.

Bitcasa was founded by former Mastercard and Mozy employees and backed by top-tier investors, including Horizons Ventures, First Round, CrunchFund, Pelion Venture Partners, Samsung Ventures, and Andreessen Horowitz.

Big Data Engineer

Bitcasa is seeking a Big Data Engineer to design, develop, and scale our big data platform that is vital to the core of our business. You will be a part of the big data and analytics team responsible for developing and refining tools that allow Bitcasa to leverage data effectively for both internal/external reporting, advanced analysis and applications, and enabling a data driven decision making culture. The ideal candidate has a passion for big data and is excited to join Bitcasa as it revolutionizes the way consumers and enterprises store data in the cloud.


  • Responsible for designing, developing, and implementing Bitcasa’s big data platform-- including big data infrastructure, data warehouse, data processing pipelines, and system integrations
  • Data modeling and metadata management
  • Streaming data processing and analysis 
  • Pay meticulous attention to end-to-end data quality, validation, and consistency
  • Ensure scalable, highly available, and robust big data platform architecture to meet service level agreements 
  • End-to-end data processing, trouble shooting, and problem diagnosis
  • Performance benchmark, load testings, and code reviews
  • Cross team collaboration and effective communication


  • 6+ years of software development experience on large scale distributed systems
  • Hands on experience with Kafka required
  • Expertise with SQL scripting, data warehousing, business intelligence, and ETL development
  • Familiarity with Hadoop open source stack including YARN, Kafka, Hive, Pig, Sqoop and RDBMS such as MySQL
  • Hands on experience with AWS EMR, Dynamo, RDS, Redshift, cloud storage such as S3, Glacier, EBS, etc.
  • Proficiency in programming languages like Python, Perl, Ruby, Scala, etc.
  • Experience building large-scale distributed applications and services
  • Experience with other cloud storage provider platform such as Google Cloud Engine, Azure bonus points.
  • Excellent written/oral communication skills
  • Proficiency in statistical analysis and data mining techniques a plus
  • BS/MS - Computer or relevant engineering discipline strongly desired
Position Filled
Not the right job?
Describe your perfect job
Join our Talent Network »