Responsibilities:
Contribute to the development of a big data platforms using pipeline technologies such as Spark, Kafka, Airflow
Design, develop and automate, Self-serviced and Executive dashboards with Tableau and Grafana.
Expert in normalization/de-normalization techniques for data analysis and optimum performance in big data environment using Presto, Redshift.
Expert in data warehouse, data lake design, data modeling and data management.
Own the design, development, and maintenance of ongoing metrics, reports, analyses, dashboards, etc. to drive key business decisions.
Deliver near-real-time and batch processed data and applications to a team of analysts and data scientists who create insights and analytics applications for our stakeholders.
Participate in data strategy and road map exercises, business intelligence/data warehouse product selection, design, and implementation
Qualifications
*Bachelor/master's in computer science, Engineering, Mathematics, Statistics or related field
*7+ years of work experience with ETL, data modeling, and business intelligence big data architectures.
*Proven experience and expertise with Tableau and Grafana
*Complete understanding of Tableau tools and processes.
*Experience with Scheduling and orchestration tools such as Airflow
*Proven experience with Python and Pyspark.
*Expert in at least one SQL language such as ANSI SQL, T-SQL or PL/SQL.
*Experience developing and managing data warehouses on a terabyte or petabyte scale.
*Strong experience in distributed and columnar databases.
*Advanced ability to draw insights from data and communicate them to the stakeholders and senior management as required
*Familiarity with AWS solutions such as S3, Parquet/ORC, Redshift
*Working with unstructured or semi-structured datasets
*Streaming data systems such as Kafka and Spark Streaming
*Distributed processing using tools such as Spark
*Container-based deployments using Docker and Kubernetes
*Experience working directly with business users and executives to build reports.
