This Opportunity is with an IT Product Company.
Responsibilities :
- Collaborate extensively with engineers, data scientists, and analysts to deliver the next generation of our data infrastructure.
- On a daily basis, you'll work with members of the team to update our data engineering roadmap and execute upon those initiatives
- Mentor and guide the professional and technical development of your team members.
- Build plans for prioritizing technical and resourcing challenges in your organization.
- Work closely with the recruiting team to hire high potential candidates from diverse backgrounds.
- Guide the team building a new technology stack for highly scalable and available data pipelines using the latest technologies
- Design canonical data models for various business domains. Formulate a vision to connect all data in the technology partner ecosystem
- Establish SLOs for Data Availability and Performance working with cross-functional stakeholders
- Establish standards and processes for incremental changes to data models
- Establish best practices to monitor availability, security, scale and cost-effectiveness of the data platform
- Ensure our user's privacy with data security automation
Requirements :
- 6+ years of data platform/engineering experience
- 2+ years of people/team management experience
- As a passionate and confident leader, you take pride and ownership in the work you and your team do.
- You have a bias for action, using the best available data to justify your actions and perspectives and make decisions that move the business forward.
- You can effectively hire, mentor, and develop career plans for engineers of all levels.
- You lead by example, holding yourself to the same standards you set for your team.
- You are a strong communicator: sharing knowledge upward, downward, and across the organization. You relish a regular cadence of one-on-one discussions with each team member.
- Should have worked with distributed data processing frameworks - Apache Spark, Apache Flink, or Hadoop.
- Should have an in-depth understanding & experience in big data technologies (Kafka, Cassandra, HBase, Spark).
- Have knowledge of frameworks & distributed systems, are good at algorithms, data structures, and design patterns.
- You have a deep understanding of Data Modeling, ETL processes and have a clear vision for a scalable data platform
- You have successfully built and deployed a streaming or batch processing data pipelines with a focus on its performance, scalability and reliability using big data tools such as Airflow and Spark
- You bring extensive experience with data modeling, data integration, database design, data analysis, and data migration
- You have a deep understanding of the AWS ecosystem including Redshift, DMS and RDS
- You have familiarity with the GCP Data Solutions including BigQuery.
- You have strong programming experience. Preferably Python.
Didn’t find the job appropriate? Report this Job