Assemble large, complex data sets that meet functional / non-functional business requirements.
Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
Keep our data separated and secure.
Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
Build analytics tools that utilize the data pipeline to provide actionable insights into key business performance metrics.
Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.
Work with data and analytics experts to strive for greater functionality in our data systems.
Required Skills:
Experience in developing Big Data applications using Spark, Hive, Sqoop, Kafka, and Map Reduce.
Experience with stream-processing systems: Spark-Streaming, Strom etc.
Experience with object-oriented/object function scripting languages: Python, Scala etc
Experience in designing and building dimensional data models to improve accessibility, efficiency, and quality of data
Should be proficient in writing Advanced SQLs, Expertise in performance tuning of SQLs. Experience with data science and machine learning tools and technologies is a plus
Experience with relational SQL and NoSQL databases, including Postgres and Cassandra.