Job description / Responsibilities - 5-7 years of experience in Big Data stacks: Spark/Scala/Hive/Impala/Hadoop Strong Expertise in Scala The resource should have good hands-on experience in Scala programming language. Should be able to model the given problem statement using Object Oriented programming concepts. Should have the basic understanding of the Spark in-memory processing framework and the concept of map tasks and reduce tasks. Should have hands-on experience on data processing projects. Should be able to frame sqls and analyze data based on the given requirements Advanced SQL knowledge Git hub or bit bucket Primary Skill Spark Scala. The resource should have good hands-on experience in Scala programming language Secondary Skill SQL, Python, Hive, Impala, AWS