Best in Industry Salary as per experience & previous CTC
Full-Time
years of experience
2+
technical skills
Solid Computer Science fundamentals, excellent problem-solving skills and a strong understanding of distributed computing principles.
Proven track record of building scalable and performant data infrastructure.
Expert SQL knowledge and deep experience working with relational and NoSQL databases.
Advanced knowledge of Apache Kafka and demonstrated proficiency in Hadoop v2, HDFS,and MapReduce.
Experience with stream-processing systems (e.g. Storm, Spark Streaming), big data queryingtools (e.g. Pig, Hive, Spark) and data serialization frameworks (e.g. Protobuf, Thrift, Avro).
Bachelor’s or Master’s degree in Computer Science or related field from a top university.
soft skills
Good communication skills
excellent problem-solving skills and a strong understanding of distributed computing principles.
Ability to work methodically and meet deadlines
High degree of flexibility, and can shoulder additional workload when necessary
Ability to influence and align the stakeholders
your role
your responsibilities
Design, develop and maintain an infrastructure for streaming, processing and storage of data. Build tools for effective maintenance and monitoring of the data infrastructure.
Contribute to key data pipeline architecture decisions and lead the implementation of major initiatives.
Work closely with stakeholders to develop scalable and performant solutions for their data requirements, including extraction, transformation and loading of data from a range of data sources.
Develop the team’s data capabilities - share knowledge, enforce best practices and encourage data-driven decisions.
Develop data retention policies, backup strategies and ensure that the firm’s data is stored redundantly and securely.