Quantcast

Data Engineer Analyst Resume Prescott va...
Resumes | Register

Candidate Information
Title Data Engineer Analyst
Target Location US-AZ-Prescott Valley
Email Available with paid plan
Phone Available with paid plan
20,000+ Fresh Resumes Monthly
    View Phone Numbers
    Receive Resume E-mail Alerts
    Post Jobs Free
    Link your Free Jobs Page
    ... and much more

Register on Jobvertise Free

Search 2 million Resumes
Keywords:
City or Zip:
Related Resumes

Data Analyst Sciences Flagstaff, AZ

Office Assistant Data Entry Flagstaff, AZ

Data Entry Office Receptionist Flagstaff, AZ

Data Entry Front Desk Camp Verde, AZ

Medical Assistant Data Entry Flagstaff, AZ

Medical Assistant Data Entry Prescott Valley, AZ

Senior Writer Cave Creek

Click here or scroll down to respond to this candidate
KRISHNASENIOR INTEGRATION ENGINEER / BIG DATA ENGINEEREmail: EMAIL AVAILABLE phone: PHONE NUMBER AVAILABLELinked ln: LINKEDIN LINK AVAILABLEPROFESSIONAL SUMMARY:With over 9 years of professional experience, I am a highly skilled Senior Integration Engineer and Senior Data Engineer IT professional, specializing in designing and implementing scalable, maintainable, and secure integration solutions.Developed, tested, and maintained robust backend services and APIs using C# and .NET Core. Ensured high performance and scalability for various enterprise applications, utilizing advanced features such as asynchronous programming, dependency injection, and Entity Framework Core.Designed and implemented integration solutions for diverse systems such as ERP, CRM, and other enterprise applications. Leveraged the power of C# and .NET to build seamless interconnectivity and data flow, using technologies such as ASP.NET Core, Web API, and WCF (Windows Communication Foundation).Designed and implemented RESTful APIs and managed them using API gateways. Utilized C# and .NET to ensure secure and efficient access to backend services, incorporating best practices for API security, versioning, and documentation with tools like Swagger and API Management (APIM).Designed, developed, and optimized data pipelines and workflows for processing large volumes of data using Azure Data Factory, Azure Synapse Analytics, and Azure Databricks. Ensured data integrity, quality, and security across all data integration processes with tools such as Azure Data Lake Storage and Azure SQL Database.Implemented data transformation and cleansing processes using Azure Data Flow and Azure Functions. Utilized Azure Monitor and Azure Log Analytics to monitor data pipeline performance and troubleshoot issues. Developed and maintained CI/CD pipelines using Azure DevOps to automate the build, test, and deployment processesWorked with Informatica PowerCenter, offering a robust ETL toolset to extract, transform, and load data from various sources to target systems. Utilized Informatica Cloud for seamless integration of cloud-based applications and data sources with on-premises systems, facilitating real-time data synchronization.Leveraged Informatica Intelligent Cloud Services (IICS) to automate data integration and management tasks, improving efficiency and reducing errors. Employed Informatica Data Quality to assess data quality, cleanse, and standardize data, ensuring accuracy and consistency.Collaborated with data scientists, analysts, and other stakeholders to understand data requirements and provide appropriate solutions using Azure Machine Learning and Azure Cognitive Services. Integrated on-premises systems with cloud-based solutions, ensuring seamless data flow and interaction. Implemented event-driven architectures for real-time data processing and analytics.Troubleshot and resolved integration issues to maintain system stability and performance. Worked closely with cross-functional teams, including software engineers, data engineers, DevOps, and product managers, to understand integration requirements and deliver effective solutions.Utilized Power BI and Tableau for creating insightful dashboards and interactive reports. Proficient in DAX for data analysis and reporting. Developed custom data quality rules and metrics to proactively identify and resolve issues, enforcing data governance policies using Master Data Management (MDM).I have worked extensively with various programming languages, including Python, SQL, Scala, Java, R, Pig, HiveQL, C, and C++. My experience encompasses both NoSQL databases, such as MongoDB, DynamoDB, HBase, and Cassandra, and relational databases, including SQL Server, MySQL, and PL/SQL. I have utilized monitoring and reporting tools such as Power BI, Tableau, and custom shell scripts to ensure effective data visualization and analysis.EDUCATIONAL DETAILS (Bachelors): GUDLAVALLERU ENGINEERING COLLEGE, VIJAYAWADA, INDIATECHNICAL SKILLS:Cloud Platforms: Amazon AWS, Microsoft Azure, GCPBig Data Technologies: Hadoop ecosystem, Cloudera CDP, Hortonworks HDP, Spark, KafkaProgramming/Scripting Languages: Python, SQL, Scala, Java, .NET Core, TypeScript, JavaScript, R, C, C++, C#NoSQL Databases: Azure Cosmos DB, MongoDB, DynamoDB, HBase, CassandraIntegration Platforms: Informatica, MuleSoft, Apache Kafka, RabbitMQ, Apache NIFI, Oracle EBS, SAP ECC, S/4HANA, SAP OCM, SOMWeb Services: RESTful APIs, SOAP, Web HooksRelational Databases: Azure SQL Database, SQL Server, Oracle, MySQL, PL/SQLDevelopment Environments: Local development setups (Windows, macOS, Linux), Virtual Machines, Dockerized environmentsCI/CD: Jenkins, Azure DevOps, GitHub Actions, TeamCityMonitoring/Reporting Tools: Power BI, Tableau, Azure Monitor, Log AnalyticsJava Technologies: Spring, Hibernate, StrutsWeb Development Technologies: HTML, CSS, JavaScript, jQueryDevelopment/Build Tools: PyCharm, Anaconda, Eclipse, IntelliJ, Maven, GradleOperating Systems: Linux, Unix, Windows, macOSVersion Control Systems: Git, Bitbucket, SVNMethodologies: Agile, WaterfallPROFESSIONAL EXPERIENCE:Client: CIGNA JAN 2023  PresentRole: Senior Integration Engineer / Big Data EngineerResponsibilities:Developed, tested, and maintained backend services and APIs using C# and .NET Core. Ensured high performance and responsiveness of backend systems. Designed and implemented real-time integration solutions to support various business processes. Utilized message brokers like Azure Service Bus or Kafka for real-time data processing.Applied strong understanding of Informatica to support data integration tasks, ensuring data quality and efficiency. Collaborated with data engineering teams to integrate Informaticaworkflows with backend services. Developed scalable backend solutions that could handle high loads and ensure efficient data processing.Optimized code and database queries for performance and efficiency. Ensured seamless data flow between systems in real-time, minimizing latency and maximizing data accuracy. Implemented and maintained data synchronization mechanisms to keep data consistent across systems.Designed, implemented, and optimized complex SQL queries for efficient data retrieval and manipulation. Performed database tuning and troubleshooting to ensure optimal performance. Implemented unit tests, integration tests, and performance tests to ensure the reliability and quality of backend services.Developed and managed data pipelines using Azure Data Factory, Azure Databricks, and Azure Synapse Analytics. Implemented data storage solutions using Azure SQL Database, Azure Data Lake, and Cosmos DB. Leveraged Informatica tools to support ETL processes, ensuring data quality and consistency.Collaborated with data analysts and developers to integrate Informatica workflows into Azure data pipelines. Designed and implemented real-time data ingestion and processing solutions using Azure Stream Analytics and Event Hubs. Ensured seamless data flow and integration across various data sources and destinations.Architected scalable data solutions to handle large volumes of data efficiently. Optimized data storage and retrieval processes to improve overall system performance. Ensured compliance with data governance policies and security standards. Implemented data encryption, masking, and anonymization techniques as needed.Integrated various backend systems and third-party services to support business operations. Developed and maintained integration documentation and diagrams. Worked closely with front-end developers, data engineers, and business stakeholders to deliver integrated solutions. Provided technical support and troubleshooting for integration issues.Environment: Microsoft Azure  Datalake storages, Azure Data Factory (ADF), Azure SQL Database, Cosmos DB, Azure Function apps, Azure Logic Apps, Azure Web apps, Azure Blob, Azure DevOps, Git, Azure Databricks, C#, Microsoft .NET Core Framework, Informatica, MuleSoft, Apache Kafka, RabbitMQ, Apache NIFI, Oracle EBS, SAP ECC, S/4HANA, SAP OCM, SOM.Client: JOHNDEERE SEP 2019 to DEC 2022Role: Senior Integration Engineer / Big Data EngineerResponsibilities:Developed, tested, and maintained backend services and APIs using C# and .NET Core, ensuring high performance and responsiveness of backend systems. Designed and implemented real-time integration solutions to support various business processes. Utilized message brokers like Google Cloud Pub/Sub or Apache Kafka for real-time data processing.Applied strong understanding of Informatica to support data integration tasks, ensuring data quality and efficiency. Collaborated with data engineering teams to integrate Informatica workflows with backend services. Developed scalable backend solutions that could handle high loads and ensure efficient data processing.Optimized code and database queries for performance and efficiency. Ensured seamless data flow between systems in real-time, minimizing latency and maximizing data accuracy. Implemented and maintained data synchronization mechanisms to keep data consistent across systems.Designed, implemented, and optimized complex SQL queries for efficient data retrieval and manipulation. Performed database tuning and troubleshooting to ensure optimal performance. Implemented unit tests, integration tests, and performance tests to ensure the reliability and quality of backend services.Developed and managed data pipelines using Google Cloud Dataflow, Dataproc, and Big Query. Implemented data storage solutions using Google Cloud Storage, Bigtable, and Firestore. Leveraged Informatica tools to support ETL processes, ensuring data quality and consistency.Collaborated with data analysts and developers to integrate Informatica workflows into GCP data pipelines. Designed and implemented real-time data ingestion and processing solutions using Google Cloud Pub/Sub and Dataflow. Ensured seamless data flow and integration across various data sources and destinations.Architected scalable data solutions to handle large volumes of data efficiently. Optimized data storage and retrieval processes to improve overall system performance.Integrated various backend systems and third-party services to support business operations. Developed and maintained integration documentation and diagrams. Worked closely with front-end developers, data engineers, and business stakeholders to deliver integrated solutions. Provided technical support and troubleshooting for integration issues.Environment: Google Cloud Platform (GCP)  Google Cloud Storage, Google Cloud Dataflow, Big Query, Bigtable, Firestore, Google Cloud Functions, Google Cloud Run, Google Kubernetes Engine (GKE), Google Pub/Sub, Apache Kafka, RabbitMQ, Apache NIFI, Oracle EBS, SAP ECC, S/4HANA, SAP OCM, SOM, Azure DevOps, Git, C#, Microsoft .NET Core Framework, Informatica, MuleSoft.CLIENT: LEIDOS SEP 2017 to AUG 2019Role: Big Data Engineer / Junior API Integration EngineerResponsibilities:Developed batch and stream processing applications, implementing functional pipelining through Spark APIs. Created Databricks notebooks for data preparation, including cleansing, validation, and transformations. Contributed to building a data pipeline and conducting analytics using the AWS stack (EMR, EC2, S3, RDS, Lambda, SQS, Redshift).Developed and managed data pipelines using AWS Data Pipeline, EMR, and Redshift. Implemented data storage solutions using Amazon S3, DynamoDB, and RDS. Leveraged Informatica tools to support ETL processes, ensuring data quality and consistency. Implemented unit tests, integration tests, and performance tests to ensure the reliability and quality of backend servicesArchitected scalable data solutions to handle large volumes of data efficiently. Optimized data storage and retrieval processes to improve overall system performance. Ensured compliance with data governance policies and security standards. Implemented data encryption, masking, and anonymization techniques as needed.Developed, tested, and maintained backend services and APIs using C# and .NET Core, ensuring high performance and responsiveness of backend systems. Designed and implemented real-time integration solutions to support various business processes, utilizing message brokers like AWS SNS and SQS or Apache Kafka for real-time data processing.Optimized code and database queries for performance and efficiency. Implemented and maintained data synchronization mechanisms to keep data consistent across systems. Designed, implemented, and optimized complex SQL queries for efficient data retrieval and manipulation. Performed database tuning and troubleshooting to ensure Performance.Applied strong understanding of Informatica to support data integration tasks, ensuring data quality and efficiency. Collaborated with data engineering teams to integrate Informatica workflows with backend services. Developed scalable backend solutions that could handle high loads and ensure efficient data processing.Collaborated with data analysts and developers to integrate Informatica workflows into AWS data pipelines. Designed and implemented real-time data ingestion and processing solutions using AWS Kinesis and Lambda. Ensured seamless data flow and integration across various data sources and destinations.Integrated various backend systems and third-party services to support business operations. Developed and maintained integration documentation and diagrams. Worked closely with front-end developers, data engineers, and business stakeholders to deliver integrated solutions. Provided technical support and troubleshooting for integration issues.Environment: Amazon Web Services (AWS)  Amazon S3, AWS Data Pipeline, Amazon Redshift, DynamoDB, RDS, AWS Lambda, AWS Kinesis, AWS EMR, AWS SNS, AWS SQS, Apache Kafka, RabbitMQ, Apache NIFI, Oracle EBS, SAP ECC, S/4HANA, SAP OCM, SOM, Azure DevOps, Git, C#, Microsoft .NET Core Framework, Informatica.CLIENT: DIRECT TV JUNE 2015 to JULY 2017 Role: Big Data EngineerResponsibilities:Enhanced the performance and scalability of data processing jobs by optimizing Spark configuration parameters, including executor memory, executor cores, and shuffle partitions. Implemented data partitioning and bucketing techniques in Spark to distribute and organize data across clusters, facilitating parallel processing and reducing latency.Improved query performance and minimized resource consumption by fine-tuning Spark SQL queries, optimizing joins, aggregations, and data filtering operations. Utilized Spark's caching mechanisms, such as RDD or Data Frame caching, to persist intermediate results in memory, preventing redundant computations and expediting data processing.Developed a centralized data engineering framework using technologies like Apache Airflow or Luigi, providing a standardized approach for defining, scheduling, and monitoring ETL workflows across teams. Designed and implemented reusable data transformation modules or libraries to promote code reusability and streamline development across multiple ETL projects.Integrated version control systems like Git into the data engineering workflow for proper code management, collaboration, and versioning across different teams and projects. Collaborated with Data Architects to establish data modeling best practices and guidelines, ensuring consistency and scalability in data structures across various ETL processes.I gained experience working with the Microsoft .NET Core Framework (version 6 or higher), Angular, Entity Framework, Typescript, JavaScript, and Bootstrap.Explored and implemented cloud-based big data technologies such as AWS EMR, Google Cloud Dataproc, or Azure Databricks to leverage scalability and elasticity for large-scale data processing. Utilized containerization technologies like Docker or Kubernetes to create portable and scalable data processing environments, ensuring efficient resource utilization and easy deployment across different clusters. Familiar with Apache NIFI.Implemented data quality monitoring and validation mechanisms, ensuring the integrity and accuracy of processed data using tools like Apache Griffin or custom validation scripts. Kept abreast of emerging big data technologies and frameworks, continuously exploring and evaluating new tools and techniques to enhance the scalability and performance of data pipelines.Environment: Apache Spark, Apache Kafka, AWS Glue, AWS EMR, Amazon S3, Amazon Redshift, AWS Lambda, AWS CloudWatch, AWS CloudFormation, Microsoft .NET Core Framework (version 6 or Higher), Angular, Entity Framework, SAP ECC, S/4HANA, SAP OCM, SOM, Apache NIFI, SSIS, CUI, Oracle EBS, Typescript, JS, Bootstrap, Apache Airflow, Terraform, Apache Nifi, Teradata, Snowflake

Respond to this candidate
Your Message
Please type the code shown in the image:

Note: Responding to this resume will create an account on our partner site postjobfree.com
Register for Free on Jobvertise