| 20,000+ Fresh Resumes Monthly | |
|
|
| | Click here or scroll down to respond to this candidateCandidate's Name
Cell:PHONE NUMBER AVAILABLEEMAIL AVAILABLEUS CITIZENProfessional Summary:Around 5 years of IT industry experience in Data-Engineer from gathering requirements to Developing, Testing, Implementation and support in Data Extraction, Transforming and Loading (ETL) and Data migrationExtensive experience on major components in Hadoop Ecosystem like Hadoop, Map Reduce, HDFS, Spark, HIVE, Python, Parquet and Avro1+ years of experience in Databricks and SnowflakeWorked with large datasets using Partitions, Spark in Memory capabilities, Broadcasts in Spark, Efficient Joinsand Transformations.Implemented Partitioning, Dynamic Partition, Buckets, Cubes, Windowing & Analytics functions in HIVEExperience in Benchmarking Cluster & SQL optimizations and setting the right Spark configuration parameters to the job and obtaining the best performance out of it.Extensively used AWS EC2, EMR, S3 AND AthenaExtensive experience with SQL extensively by using Joins, Sub Queries, Set Operations and Advanced OLAP functions.Good knowledge in Data Warehousing concepts and Dimensional modeling like Star Schema and Snowflake schema.Having good communication skills, Interpersonal relations, hardworking and result- oriented as an Individual and in Team.EducationBA Degree of business management 2011Technical Skills:TECHNOLOGYTOOLDatabaseHDFS, Hive, Spark, MYSQLOperating SystemsLinux, Windows,,Git,VS,VSCProgrammingLanguagespython, SQL,AWS StackEMR, S3, Athena,ETL(GLUE)Lambda Function, TerraformProfessional Experience:Data Engineer, AIG Jersey City NJJune 2021May 2022ResponsibilitiesResponsible for building scalable distributed data solutions with Apache Spark using Python/Scala.Developed aggregated & bi tables in AWS EMR by using hive, spark and load them into Data Lake and Athena.Developing and scheduling the ETL pipelines by using Databricks.Implemented Partitioning, Dynamic Partition, Buckets in HIVE.Load and transform large sets of structured, semi-structured data.Developed Airflow workflows for orchestrationDeveloped the Python automation framework & scripts for Data integrity and validationsStrong problem-solving experience on test, identify, address, debug and then resolve technical issues that affect the integrity of the application.Deutsche Bank, Manhattan, New York May 2020 to May 2021Responsibilities:Responsible for building scalable distributed data solutions with Apache Spark using Python.Developed aggregated & bi tables in AWS EMR by using hive, spark and load them into Data Lake and Athena.Developed Airflow workflows for orchestrationDeveloping and scheduling the ETL pipelines by using Databricks, snowflake.
Implemented snowflake cache design for building real-time aggregates in snowflakeLoad and transform large sets of structured, semi-structured data.Developed the Python automation framework & scripts for Data integrity and validationsStrong problem-solving experience on test, identify, address, debug and then resolve technical issues that affect the integrity of the application.Data AnalystCredit Agricole, New York, New York February 2015 to March 2016Responsibilities:Created customized Table Calculations, Conditions and Filters for various analytical reports and dashboards.Created and modified Interactive Dashboards and Creating guided navigation links within Interactive Dashboards.Created navigation page which had all the links to various dashboards.Scheduled frequency and time to refresh data for the when sources are published or extracted to server.Used SQL tools like Toad, Pl/SQL to run SQL queries and validate the data in the warehouse and data marts.Used Performance recording on Desktop and on Tableau Server and Tuned the Performance of Dashboards.Followed Best practices while developing DashboardsProvided UAT assistance and post implementation production support.Used Dual Axes Charts-Shown multiple measures with dollar numbers on left Y axis and Percentages on the right Y axisCreated Tableau Server backups using Tableau admin utilitySet up Users, Groups, Permissions, Schedules on Tableau ServerDone unit testing of Dashboards for data quality and done Interface testing like Quick filters, parameters, Actions. |