Role: Principal Data Architect Experience: 12+ years” experience Job Description: *Lead and inspire large scale data engineering teams (50+) *Understand the latest Data engineering trends and best practices and implement it in client projects *Use one or more of the Azure/AWS/GCP data and analytics services in conjunction with third parties, such as Spark, EMR, DynamoDB, RedShift, Kinesis, Lambda, Glue, Snowflake, and Data bricks, to consult, design, create, and execute large-scale business data solutions. *Analyze, re-architect, and re-platform on-premises data stores/ Databases to modern data platforms on Azure/AWS/GCP using cloud or 3rd party services. *Design and build production data pipelines from ingestion to consumption within a big data architecture, using Java, Python, and Scala. *Design and optimize data models on Azure/AWS/GCP Cloud using cloud data stores such as SQL Database, Data factory, Redshift or Big query *Design and implement data engineering, ingestion, and curation functions on Azure/AWS/GCP cloud using cloud native or custom programming *Perform detailed assessments of current state data platforms and create an appropriate transition path to cloud as part of customer consultation and business proposals. *Participate in client design workshops and provide trade-offs and recommendations for building solutions Personal attributes: *Excellent verbal and written communication skills to display and communicate information to all levels of employees, management, and clients. *Strong ability to build stakeholder relationships.
Job Requirements *Extensive expertise of design and solutioning in Cloud – Azure, AWS, GCP *Sound Knowledge of different cloud services and components in AWS, Azure, and GCP *Expertise in Data transformations, Data modelling, Data Migrations, Creating Pipelines and Data Governance *Experience in working with Big Data technologies like HDFS, HBase, Hive, Kafka, Sqoop, and Flume *Good knowledge of performance tuning of Spark and T-SQL