- BS in Computer Science or related area
- 10-12 years software development experience
- Minimum 2 Year Experience on Big Data Platform
- Proficiency with Java, Python, Scala, HBase, Hive, MapReduce, ETL, Kafka, Mongo, Postgres, Visualization technologies etc.
- Flair for data, schema, data model, how to bring efficiency in big data related life cycle
- Understanding of automated QA needs related to Big data
- Understanding of various Visualization platform (Tableau, D3JS, others)
- Proficiency with agile or lean development practices
- Strong object-oriented design and analysis skills
- Excellent technical and organizational skills
- Excellent written and verbal communication skills
Top skill sets / technologies in the ideal candidate:
* Programming language -- Java (must), Python, Scala, Ruby
* Batch processing -- Hadoop MapReduce, Cascading/Scalding, Apache Spark
* Stream processing -- Apache Storm, AKKA, Samza, Spark streaming
* NoSQL -- HBase, MongoDB, Cassandra, Riak,
* ETL Tools – Data Stage, Informatica,
* Code/Build/Deployment -- git, hg, svn, maven, sbt, jenkins, bamboo