Quantcast

Senior Data Analyst Resume Austin, TX
Resumes | Register

Candidate Information
Name Available: Register for Free
Title Senior data analyst
Target Location US-TX-Austin
Email Available with paid plan
Phone Available with paid plan
20,000+ Fresh Resumes Monthly
    View Phone Numbers
    Receive Resume E-mail Alerts
    Post Jobs Free
    Link your Free Jobs Page
    ... and much more

Register on Jobvertise Free

Search 2 million Resumes
Keywords:
City or Zip:
Related Resumes

Senior Quality Service Analyst Leander, TX

Data Analyst Power Bi Austin, TX

Data Analyst Project Management San Antonio, TX

Data Analyst Team Member San Antonio, TX

Power Bi Data Analyst Round Rock, TX

Senior Decision Support Analyst San Antonio, TX

Data Analyst Austin, TX

Click here or scroll down to respond to this candidate
 Candidate's Name
Contact: PHONE NUMBER AVAILABLE
Email: EMAIL AVAILABLELinkedIn: https://LINKEDIN LINK AVAILABLEProfessional Summary
Professional experience  7+ years  experience building data models and leading database migration initiatives. Organized and detailed database architect able to integrate and manage SQL / NOSQL databases including Redshift, mysql, and MongoDB, to meet growing data handling. Worked on Bigdata, professional designing in Enterprise Data Warehouse, data mart, and BI solutions with various integration areas, such as data integration, application integration, business process integration, and UI integration.      Excellent knowledge of Hadoop cluster architecture and its key concepts - Distributed file systems, Parallel processing, High availability, Fault tolerance and Scalability.      Obtained and processed data from Enterprise applications, Clickstream events, API gateways, Application logs and database updates.      Extensive experience in development of Bash scripting, T-SQL, and PL/SQL Scripts.      Sustaining the BigQuery, PySpark and Hive code by fixing the bugs and providing the enhancements required by the Business User.      Installed application on AWS EC2 instances and configured the storage on S3 buckets.
Work Experience

Applied Materials - Austin, TX				                             2023 JAN   Till Now
Sr Data Analyst

Responsibilities      Designed Power BI dashboards and performed code search feature with project managers and peers to optimize operations, standardize data, perform complex SQL operations to validate data.      Loading the data from multiple data sources like (SQL, DB2, and Oracle) into HDFS using Sqoop and loading into Hive tables.      Developing Data Extraction, Transformation, and Loading jobs from flat files, Oracle, SAP, and Teradata Sources into Teradata using BTEQ, Fast Load, Fast Export, Multiload, and stored procedures.      Implemented Cluster for NoSQL & writing SQL scripts with NoSQL databases like HBase in creating HBase tables to load large sets of semi-structured data coming from various sources.
      Experienced with Kafka-Storm on HDP 2.2 platform for analysis on live-data      Developed and optimized Spark jobs, Databricks Notebooks, and SQL queries for data processing, analytics, machine learning tasks, leveraging technologies like PySpark, Scala, SQL.      Experience in extracting files from MongoDB through Sqoop and placed in HDFS and processing.      Experienced with Kafka-Storm on HDP 2.2 platform for analysis on live-data      Created APIs using Apache Kafka and node.js and Consuming event data from Kafka using      Ensuring cluster security by deploying and managing authentication using AD Kerberos. Troubleshooting Kerberos related access issues, creating and managing local keytabs for service accounts using ktutil tools.      Build a data pipeline for getting data and storing data models in Cassandra.      Implemented cloud solutions utilizing a range of AWS services, such as EC2, VPC, S3, Glacier, EFS, AWS Kinesis, IAM, Lambda, Directory Services, Security-Groups, CloudFront, Cloud Formation.      Developed Tableau data visualization using Heat maps, Bubble charts, Scatter Plots, Geographic Map, Pie Charts and Bar Charts and Density Chart.      Alteryx advanced analytics Macros, Predictive Analytics, Spatial Data, Connectors, Tableau Hyper Extracts, In-Database tools, Transform & Join, recommending metrics, measures, reporting and solutions for Tableau Visualization.      Responsible for design and delivering IoT Data Model for PI Tag stream data, design and architecture delivering Time Series Instance for Hierarchy, Types, and instances.      Expertise in building pipelines to migrate existing On-Prem cloud data/data warehouse on Snowflake cloud using FiveTran and ADF.
      Experience in DBT to build transformations on the data loaded in Snowflake.      Worked and applied Pandas, umPy, and PyTorch for advanced data analysis, statistical modeling, and machine learning tasks.      Collaborated on designing and implementing data solutions in Azure Cloud, utilizing services such as Azure SQL Database, Azure Data Factory (ADF), and Databricks.      Developed and maintained data pi      pelines using Hadoop, Spark, and MapReduce for large-scale data processing and analytics.      Leveraged Informatica for data integration, transformation, and loading processes, ensuring reliable and efficient data movement across diverse sources and targets.      Excellent understanding of NOSQL databases like HBASE, Cassandra, MongoDB.      Experienced with Apache Hive, Apache HBase, MongoDB, and other NoSQL databases for managing large volumes of data with data mesh and event streaming architectures using Azure Event andHub/Kafka.

Connectwise - Tampa, Florida							2020 Aug   2023 JanSr Database Engineering
 Responsibilities
      Involved in the complete software development life cycle (SDLC) process by analysing business requirements and understanding the functional workflow of information from source systems to destination systems.      Hands-on experience working with Amazon Web Services (AWS) using Elastic Map Reduce (EMR), Redshift, and EC2 for data processing.      Worked on Snowflake Schemas and Data Warehousing and processed batch and streaming data load pipeline using Snow Pipe and Matillion from data lake Confidential AWS S3 bucket.      Used cloud shell SDK in GCP to configure the services Data Proc, Storage, BigQuery.      Automated and scheduled recurring reporting processes using UNIX shell scripting and Teradata utilities such as MLOAD, BTEQ, and Fast Load.      Experience in GCP Dataproc, GCS, Cloud functions, BigQuery in moving data between GCP and Azure using Azure Data Factory.      Conducted validation, reviewing, cleaning, and querying of data using SQL to identify data discrepancies, missing values, and outliers.      Design and implement data cleansing, transformation, and enrichment processes using Alteryx tools and  workflows for data blending, predictive analytics, and spatial analysis.      Applied AWS Lambda functions to automate tasks within the data pipeline, enhancing scalability.      Developed and maintained big data pipelines using PySpark on AWS, employing Data and Analytics services, such as S3, Athena, EMR, Glue      Experienced with Apache Hive, Apache HBase, MongoDB, and other NoSQL databases for managing large volumes of data with data mesh and event streaming architectures using Azure Event andHub/Kafka.      Built a generic data ingestion framework to extract data from multiple sources like SQL server, delimited flat files, XML, and JSON, using it to build redshift tables.
      Execution of TSM commands over Tableau server for day to day server admin tasks like log file generation backup file generation and restore of the backup file.      Develop spark-Scala batch applications to process the data and transfer the data into insight stores.      Proficient in Design and Development of process required to Extract, Transform and Load data into the data warehouse using Snowflake Cloud Database, AWS S3 and Azure BLOB/ADLS Gen 2.0.      Design and develop a data pipeline using Spark and Ingest data from Oracle Database and Visualize the current trends on Tableau dashboards.      Developed and maintain Kafka CI/CD pipelines using AWS DevOps tools, including creating build and release pipelines, managing environments, and configuring deployment strategies.      Containerization using Docker and Kubernetes, with proficiency in managing Helm charts for package release and created pipelines in ADF to Extract, Transform from different sources like Azure SQL, Blob Storage to Azure Synapse.      Migrating data from source systems such as mainframe, RDBMS, files into a data lake solution.      Data curation done using azure data bricks, Pyspark, HDInsight, u-SQL, tSQL, spark SQL, Azure ADW and hive used to load and transform data in utilized machine learning libraries/frameworks like Keras, TensorFlow, XGBoost, and AdaBoost for predictive modelling tasks.State Farm - Bloomington, Illinois 	                                                                May 2016       Dec  2018Database Resource Administration Responsibilities
      Built scalable structured and unstructured data models using RDS(Redshift, Aurora) and NoSQL (Elastic Search, DynamoDB), for fast and efficient access by data applications.      Translated complex data into clear, data-driven, for business improvements operationalization large-scale data and analytics solutions on Snowflake Data Warehouse.      Firm understanding of Hadoop architecture and various components including HDFS, Yarn, MapReduce, Hive, Pig, HBase, Kafka etc.      Migrated SQL database to Azure Data Lake, Data Lake Analytics, Azure SQL Database, Data Bricks, Azure SQL Data warehouse, controlled & granted database access, migration On-premises databases to Azure Data Lake store using Azure Data Factory.      Developed and performed SQL and PL/SQL scripts for data extraction, transformation, and loading operations in databases such as DB2, ensuring data integrity and optimal.      Developing ETL pipelines in and out of the data warehouse using a combination of Python and Snow SQL.      Worked on creating different Power BI (Dashboard reports), Crystal reports, and SSRS using MS SQL server. Integration of SSRS Reports with Visual Studio .NET and writing queries with grouping, sub-grouping, formatting, conditional formatting, and creating complex formulas.      Used Pandas in Python for Data Cleansing and validating the source data.
      Hands on experience on Unified Data Analytics with Databricks, Databricks Workspace User Interface, Managing Databricks Notebooks, Delta Lake with Python, Delta Lake with Spark SQL.
      Execution of TSM commands over Tableau server for day to day server admin tasks like log file generation, Backup file generation and restore of the backup files.      Developed Complex database objects like Stored Procedures, Functions, Packages, and Triggers using SQL and PL/SQL. Developed PL/SQL triggers and master tables for the automatic creation of primary keys.      Design and implement data cleansing, transformation, and enrichment processes using Alteryx tools and workflows for data blending, predictive analytics, and spatial analysis.      Created additional Docker Slave Nodes for Jenkins using custom Docker Images and pulled them to ECR, working on all major components of Docker like Docker Daemon, Hub, Images, and RegistryEducation Details

Masters in Data Analytics from Concordia University St Paul: St Paul, Minnesota.
Bachelors from Westin College. INDIA

Respond to this candidate
Your Email «
Your Message
Please type the code shown in the image:
Register for Free on Jobvertise