Designing and implementing the architecture of ETL using PySpark and SQL language in AWS Glue.
Designing and developing transactional and analytical data structures.
Aligning the Data technical capabilities to business objectives.
Developing reporting solutions to support Business Intelligence, including resolving problems related to data analysis and reporting services.
Establishing and maintaining policies, operational procedures, and associated documentation for user interaction with the database environment.
Following client’s standards and best practice for the deployment of AWS components and code using source code controls such as BitBucket, and utilizing DevOps tools (e.g., Terraform), apps, and CLI as appropriate.
Recommending improvements to existing approaches and practices within Data Governance forums and other appropriate meetings.
Minimum of 4 years’ experience in Python
Experience working with Python virtual environments
At least 3 years’ experience in a similar role, preferably in financial services or a multinational with digital capabilities
Experience working with databases and analyzing data insights
Proficient with BI software such as JasperSoft and Tableau for creating interactive web-based reports, dashboards, and analyses
ETL/ELT experience within a big data environment
Experience with both SQL and NoSQL databases
Strong coding skills with source control management (e.g., Git) and a focus on unit and integration testing
Proficient in SQL (ANSI, T-SQL, PL/pgSQL)
Experience deploying AWS components remotely using AWS CLI
Skilled in database modeling, design, development, debugging, and troubleshooting, including stored procedures and functions
Knowledgeable in Linux with the ability to write shell scripts
Experience developing and maintaining software solutions within the AWS ecosystem (S3, Redshift, DynamoDB
PythonAmazon Web ServicesShellCloud ComputingJavaAlgorithms & Data StructuresMiddlewareLinuxDocumentationReports