Quantcast

Big Data Processing Resume Austin, TX
Resumes | Register

Candidate Information
Name Available: Register for Free
Title Big Data Processing
Target Location US-TX-Austin
Email Available with paid plan
Phone Available with paid plan
20,000+ Fresh Resumes Monthly
    View Phone Numbers
    Receive Resume E-mail Alerts
    Post Jobs Free
    Link your Free Jobs Page
    ... and much more

Register on Jobvertise Free

Search 2 million Resumes
Keywords:
City or Zip:
Related Resumes

Senior big data engineer San Marcos, TX

Data Engineer Big San Antonio, TX

Data Engineer Big Austin, TX

Big Data Developer(Scala, Python, Kafka, Databricks) Austin, TX

Scrum Master Big Data Cedar Park, TX

Data Engineer Processing Leander, TX

Big Data Software Engineer Austin, TX

Click here or scroll down to respond to this candidate
Candidate's Name  EMAIL AVAILABLE PHONE NUMBER AVAILABLE LINKEDIN LINK AVAILABLE EDUCATION:Stevens Institute of Technology  Hoboken, USA August Street Address -May 2021 Master of Science, Engineering ManagementUniversity of Mumbai  Mumbai, India August 2013-May 2017 Bachelor of Engineering, Electronics & Telecommunication TECHINCAL SKILLS:Big Data: Apache Spark, Apache Airflow, PySpark, Spark Scala, Apache Hadoop, MapReduce, Pig, Hive, HDFS, HBase, Databricks.Cloud services: ACI (Apple Cloud Infrastructure), Azure Data Factory (ADF), Azure Datalake (ADLS), AWS (Amazon Web Services), EMR (Elastic Map Reduce), S3 (Simple Storage Service), Lambda (serverless), ECS (Elastic Container Service), SNS (Simple Notification Service), SQS (Simple Queue Service), Amazon Redshift, AWS Glue. Data Warehousing/BI Tools: Informatica, Snowflake, Tableau, Power BI Operating Systems: MacOS, Windows, Linux Kernels.Programming languages: Python, Scala, PL/SQLDatabases: MySQL, PostgreSQL, AWS RedShift, Azure SQL DB, Snowflake, Oracle, MongoDB. Monitoring: AWS CloudWatch, Splunk, Crontab, Apache Airflow, Kafka. Data Analysis: Databricks, Jupiter Notebooks, Microsoft Excel. Container and CI/CD: Docker, Jenkins.Project Management: Radar, JIRA, ConfluenceVersion Control: Git, GitHub, GitLab, Bitbucket.PROFESSIONAL EXPERIENCE:Apple Inc., Austin, TX May 2022-PresentSpark & Big Data Developer Implemented complex data transformations and processing logic using Apache Spark and Scala, ensuring efficient data processing and analysis. Collaborated with cross-functional teams to design, develop, and maintain scalable and high-performance big data pipelines, enabling data-driven decision-making processes. Utilized Hadoop ecosystem components, such as HDFS, YARN, and Hive, to store and manage vast amounts of structured and unstructured data. Designed and optimized data ingestion processes, enabling real-time and batch data loading from various sources into Hadoop clusters. Managed and executed data warehouse plans, Implemented Snowflake for robust data warehousing, ensuring secure and scalable storage and retrieval of structured and semi-structured data, optimizing analytical processes. Leveraged HBase as a NoSQL database for storing and retrieving large-scale, semi-structured data, ensuring high availability and low-latency access. Developed custom Spark Scala applications to process and analyze large datasets, improving data quality and enhancing business intelligence capabilities. Designed and implemented Scala jobs, incorporating Spark queries and data modeling techniques for enhanced data processing and analysis. Employed Tableau for creating interactive and insightful data visualizations and dashboards, enabling stakeholders to gain valuable insight. Continuously monitored and optimized data processing performance, improving job execution times and resource utilization. Implemented version control and code management using GitHub, facilitating collaboration and code review processes. Extracted and transformed data from Oracle databases, ensuring data accuracy and integrity, and integrated it into data processing workflows, contributing to effective data analysis, and reporting. Utilized Apache Airflow for orchestrating and scheduling data workflows, ensuring data pipelines run efficiently and reliably. Configured and managed Crontab schedules for periodic batch job execution and data processing tasks. Implemented RADAR as the Agile methodology for project management, ensuring streamlined workflows, efficient collaboration, and effective project tracking, resulting in improved project outcomes and team productivity. Wayfair, Boston, MA Feb 2021-Apr 2022Big Data Engineer Designed aggregate data models for complex data ingestion, consolidating multiple streams of credit card data. Orchestrated data ingestion using Apache Airflow and Directed Acyclic Graphs (DAGs). Developed PySpark code to process raw data from various sources into S3 buckets. Implemented CI/CD pipelines using Jenkins and wrote unit tests. Designed, developed, and deployed data pipelines using AWS services (S3, Redshift, ECS, EMR, Glue) and integrated real-time monitoring with AWS CloudWatch and CloudWatch Alarms. Fidelity Investments, Boston, MA Jun 2020-Feb 2021 Big Data Engineer Extracted and ingested data from Teradata to Azure Blob storage using Azure Data Factory. Managed Databricks clusters and set up secure Key Vaults. Designed and processed datasets, handling data transformations. Extracted data from various sources and developed Logic Apps for custom transformations. Created Azure Databricks notebooks for data analysis and reporting. Built complex stored procedures in Azure SQL DWH and led legacy application migrations to Azure. Participated in testing, defect resolution, and code review meetings. Anthem, Mumbai, IndiaData Engineer Jun 2017-May 2019 Migrated legacy Hadoop MapReduce and HDFS-based data processing to a cloud-native AWS solution. Configured Hadoop nodes and Utilized Kafka to architect and build high-performance streaming data pipelines, enabling real-time data ingestion and processing for mission-critical business applications. Improved data organization and query optimization with Apache Hive partitioning and bucketing. Transitioned Hadoop MapReduce and HDFS to real-time processing using AWS S3, Lambda and SQS. Managed access with AWS IAM, SQS policies and performed real-time data processing and migration. VOLUNTEERING:BAPS Swaminarayan Organization, India & USA Jan 2007-Present Engaged in community outreach, including food drives, health camps, and anti-addiction campaigns. Supported cultural and educational initiatives to preserve Indian traditions. Assisted in event planning and management and participated in disaster relief and fundraising efforts.

Respond to this candidate
Your Message
Please type the code shown in the image:

Note: Responding to this resume will create an account on our partner site postjobfree.com
Register for Free on Jobvertise