Quantcast

Big Data Developer Scala Python Kafka Da...
Resumes | Register

Candidate Information
Name Available: Register for Free
Title Big Data Developer(Scala, Python, Kafka, Databricks)
Target Location US-TX-Austin
Email Available with paid plan
Phone Available with paid plan
20,000+ Fresh Resumes Monthly
    View Phone Numbers
    Receive Resume E-mail Alerts
    Post Jobs Free
    Link your Free Jobs Page
    ... and much more

Register on Jobvertise Free

Search 2 million Resumes
Keywords:
City or Zip:
Related Resumes

Senior big data engineer San Marcos, TX

Big Data Processing Austin, TX

Sql Developer Data Modeler Austin, TX

Data Engineer Big San Antonio, TX

Scrum Master Big Data Cedar Park, TX

Big Data Software Engineer Austin, TX

Software Developer Data Science Austin, TX

Click here or scroll down to respond to this candidate
Candidate's Name
PHONE NUMBER AVAILABLEEMAIL AVAILABLEDeveloper & Architect - Java, Scala, Python, Kafka, Databricks, Spark, Snowflake, DBT18+ years of overall ExperiencePROFESSIONAL SUMMARY: 11+years of experience in Java, J2EE, Javascript, Spring, Struts, Hibernate, Oracle, MySql, DB2 7+ years of experience in Scala, Kafka, Kafka Stream(KStream, KTable, KSql), Kafka Connect, KSqlDB, MirrorMaker 2, Akka, Akka Streams 5+ years of experience in Hadoop, Map/Reduce, Hive, HBase, Pig, Sqoop, Oozie, Flume, Avro, Yarn, Spark(Scala), Spark Streaming 2+ years of experience in Snowflake, DBT 2+ years of experience in GCP(Dataflow, BigQuery, PubSub) 5+ years of experience in Python 3.9+, FastApi 2+ years of experience in Kafka Administration with Prometheus, Graphana 5+ years of experience in creating microservices using Spring Boot Knowledge in Golang, RUST, TypeScript 6+ years of experience in Docker, Kubernetes, Openshift, Terraform, CI/CD pipeline using Gitlabs & Jenkins, Github Actions 2+ years of experience in Apache Nifi, Apache Camel, Airflow 5+ years of experience in AWS - S3, EC2, EMR, EKS, Glue, Step Functions, Athena, Lambda, EBS 2+ years of experience in GCP - Dataflow, Apache Beam, Scio 2+ years of experience in CassandraDB, MongoDB 4+ years of experience in Git. Experience with Agile Methodology.Current Hobbies:Playing with Micro:bit using RustLearning more about Rust microservicesCERTIFICATIONS:Oracle Certified Associate, Java SE 7 Programmer.EDUCATION:MCA First class from Pondicherry University.BSc (CS) First class from Periyar University, Salem.PROFESSIONAL EXPERIENCE:TEID Client ProfileNov 2023 to till dateClient: ThriventDesignation: Senior Scala EngineerBuilding Scala micro services using Spring BootDesign and code ETL using Spark(Databricks) with Scala and PySpark in batch processingStreaming data using KafkaData Warehousing using Snowflake DBT(Snow SQL, Snow Pipe, Tasks, Streams, Time travel, Zero Copy Cloning, Optimizer, Metadata Manager, data sharing and stored procedures)Snowflake connector(SnowPipe Streaming) to feed KafkaManaging Kafka Connect and built pipeline on Kafka Streams(KStream, KTable) includes CDC(Change Data Capture)Built CI/CD pipeline using Gitlabs, Docker and KubernetesAWS stack: AWS - S3, EC2, EMR, EKS, Glue, Athena, Lambda, Github ActionsFormat InquiryDec 2022 to Nov 2023Client: ExperianDesignation: Senior EngineerBuilding micro services using Scala, Spring BootDesign and code ETL using AWS Glue/Spark with Python in batch processingStreaming data using KafkaCoded Validation tool using Python to compare the output of AWS Glue/Spark job vs mainframes outputData Warehousing using Snowflake DBT(Snow SQL, Snow Pipe, Tasks, Streams, Time travel, Zero Copy Cloning, Optimizer, Metadata Manager, data sharing and stored procedures)Managing Kafka Connect and built pipeline on Kafka Streams(KStream, KTable) includes CDC(Change Data Capture)Design and developed AWS Step functions to automate the overall flowBuilt CI/CD pipeline using Gitlabs, Docker and KubernetesAWS stack: AWS - S3, EC2, EMR, EKS, Glue, Step Functions, Athena, Lambda, Github ActionsApache Beam Jobs using ScioJan 2021 to Dec 2022Client: Twitter(Google)Designation: Principal EngineerDesign and implement ETL using Apache Beam Job using Scio(Scala wrapper of Beam code)Migrated C++ project to PythonDeploy the Beam job in GCP(Dataflow), Google Big QueryTesting the Job using Parquet file by storing it in Google StorageMigrating existing Apache Beam jobs from Scalding to ScioBuilt CI/CD pipeline using Gitlabs, Docker and KubernetesAutomating Terraform using GolangDiff Correlation & Automating SDLC process for faster delivery, Orpine IncJan 2019 to Dec 2020Client: Morgan StanleyDesignation: Sr. Scala EngineerDesk strategists run the regressions across various books to validate the pricing with the help of proprietary tools which results in differences between BAS version and current (REG) versions. Differences are investigated further to find the correlation between the tickets.Automating the above Diff correlation makes desk strategist job easier to focus on the root cause and fix it faster. Apart from Diff correlation, in the processes of pricing tool, some of the SDLC processes are manual and have less performance over treadmill(similar to Kubernetes) which leads to delivery to PROD in ~2 weeks. The goal is to automate the release process, shortening the feature release time from ~2 weeks to less than 2 days with an aim to reach continuous deployment, which is more productive for desk strategist.Spread the load across multiple Akka actors using Routers with related supervision and mailbox.Using Akka Reactive streams, the Kafka Streams were implemented with offset management and backpressure.Using Akka-Http, build the web app and route the request to related Actors.Roles and Responsibilities:Create applications using Python and Scala with Optimus(extended from Scala - proprietary)Automate the SDLC process reactively using Kafka and Akka-typedWork with Bitemporal Store - DAL(proprietary storage architecture and will never access DB directly) and MongoDBAnalyse the Pricing results using Scala & Python AppsPython scripting for automationBuild CI/CD pipeline using JenkinsData Processing using Kafka StreamsWriting business logic with Kafka consumers & producers using PythonKafka Administration - Add/remove nodes, partition rebalance, scalingKafka monitoring and alerting setup on brokers, topics and Consumers using burrowKafka Components setup  Kafka Connect and Schema registryKafka Stream Datapipeline & KSqlDBKafka Security  SSL and SASL.Kafka offset reset toolKafka Administration - Add/remove nodes,partition rebalance, monitoring, scalingBuilt CI/CD pipeline using Jenkins and deployed using Docker, KubernetesAt least an hour spend everyday for code review for the team membersFollowing agile methodology and involved in Backlog refinement, Sprint planning, Retrospection.Technologies Java, Scala, Python, Kafka, Spark, Spark Streaming, Akka-typed, Hadoop, Hive, Devops, Kubernetes, Jenkins, Git, AWS stack: AWS - S3, EC2, EMR, EKS, Glue, Step Functions, Athena, LambdaTrade Journaling, Orpine IncMay 2017 to Dec 2018Client: Point72 Asset ManagementDesignation: Platform Engineer(Scala Lead/Architect)Trade Journaling receives Portfolio Managers trade fills, process & transform the data into trade, positions, trading area, trading books and preparing profit & loss reports for each perspective. This project also automated reconciliation process between the received fill vs Prime Broker report.Roles and Responsibilities:Writing Spark(Python and Scala) stream jobs for processing Kafka Real Time trade fill eventsAnalyzing data using Hive Query and writing UDF functionsStoring & Retrieving the trade fill data after processing stages such as positions, profit and loss, PM perspectiveProvide control configuration service to manage environment variable and other openshift parameters against app or project, which automatically update the pod when deployed.Implemented microservice using Spring Boot, Akka-typed,, Akka Stream, Scala and Kafka as messaging layerBuilt and deployed microservices using Docker, Kubernetes, OpenshiftBuilt CI/CD pipeline using JenkinsProvisioned & managed software using Ansible configurationsKafka Administration - Add/remove nodes, partition rebalance, scalingKafka monitoring and alerting setup on brokers, topics and Consumers using burrowKafka Components setup  Kafka Connect and Schema registryKafka Stream Datapipeline & KSqlDBKafka Security  SSL and SASL.Writing business logic with Kafka consumers & producers using Scala/JavaDesign and code with CassandraDBCreating Nifi custom processor and building the flow according. Built using scala and Kafka as messaging layerPOC on camunda workflow to meet the error management mechanism across multiple project using scala and Kafka as messaging layerCreated tool for resetting Kafka offset based on predicate for consumer groupsBuilt CI/CD pipeline using JenkinsPerforming Code ReviewManaging a team of 6 memberUsing agile methodology and involved in Backlog refinement, Sprint planning, Retrospection.Technologies Java, Scala, Spring Boot, Spark, Spark Streaming, Hive, HBase, Kafka, Akka-typed, Akka Stream, Camunda, Redis, AWS, Devops, Docker, Kubernetes, OpenShift, Jenkins, Git, AWS stack: AWS - S3, EC2, EMR, EKS, Glue, Step Functions, Athena, LambdaCentralized Data Store, MphasisSep 2016 to Apr 2017Client: JP Morgan ChaseDesignation: Project LeadA platform to dump data from multiple data sources into HBase and workflow managed by Nifi after fetching the values from Kafka.Roles and Responsibilities:Lead Developerinvolved in Backlog refinement, Sprint planning, Retrospection.Implemented Akka Actor for better concurrency and stateBuilt and deployed microservices using Docker, KubernetesImplemented microservice using Spring Boot, Akka, Akka Stream, Scala and Kafka as messaging layerCreating Nifi custom processor using scala and building the flow accordinglyHadoop/Hive AdministrationSetting up and managing Kafka for stream processingBroker and topic configuration and creationRebalance replica leader electionSecuring by Kerberos authenticationConfiguring Producer and Consumer based on the requirementUnit testing through VerifiableConsumer and VerifiableProducer to test configurationsSpark(Scala) for analyzing data and transformingDesigning HBase schemaPerforming Code ReviewTechnologies Scala, Akka, Java, Nifi, Hadoop, Spark(Scala), Hadoop/Hive Administration Map/Reduce, YARN, Kafka, HBase, Zookeeper, Sqoop, Avro, HortonWorks, Linux, GitRate Management (Viewership), Experis IncNov 2015 to Sep 2016Client: Sabre incDesignation: Sr. Programmer AnalystA platform between Suppliers and Agencies to assign rate which they accepted for contract. Currently this process is TPF command basis which interacts with TPF systems. This project will eliminate the use of TPF by providing the API which expose the services to be invoked by UI and manages the Rate between the supplier and agencies.Roles and Responsibilities:Lead Developer and Kafka administrationInvolved in Backlog refinement, Sprint planning, Retrospection.Creating Camel Routes and building the flow accordinglySetting up and managing Kafka for stream processingHadoop/Hive AdministrationSpark(Scala) for analyzing data and transformingCommunicating BA for requirement clarification, Analyzing dataPerformed Code ReviewTechnologies Java, J2EE, Spring, Hadoop/Hive Administration, Apache Camel, JAXB, SOAP, Design Patterns, Web Services, S2 Container, Unix, SVN, Crucible, Jenkins, Hadoop, Map/Reduce, Spark, Scala, Python, YARN, Kafka, Hive, Pig, HBase, Zookeeper, Sqoop, Oozie, Flume, AvroVehicle Data Analysis, Ford ITNov 14 to September 2015Client: Ford ITDesignation: Sr. ConsultantWork closely with teams across the portfolio to identify and solve business challenges utilizing large structured, semi-structured, and unstructured data in a distributed processing environment. Develop a reporting analytical strategy for last 15 years of data of vehicle based customer purchases which correctly predicted the resurgence of particular vehicle giving us a jump on the competition.Working on Automobile Purchasing domain.Designing and Development effort the project requirements as Technical Lead for different projects on different platforms such as web Services (SOAP) DB links and J2EE.Evaluating the tools and utilities for the project and participating in scalability assessment.Project planning, task assignments, monitoring and tracking.Managed and reviewed Hadoop log files.Lead programmers and helped them to understand coding and functionality of each module in detail for the project related to functional and technical perspectives.Design and develop applications from J2EE end, writing designing patterns, during the development process. Testing & optimizing programs for achieving optimum performance.Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System from Sqoop which pre-process the data with Hive respectively.Hadoop AdministrationProvided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.Tested raw data and executed performance scripts.Roles and Responsibilities:Communicating BA for requirement clarification, Analyzing dataPerformed Code ReviewWorking with HiveProgramming with Map/ReduceTechnologies Java, Camel, Hadoop, Hadoop Administration, Map/Reduce, Hive, Pig, HBase, Zookeeper, Sqoop, Oozie, FlumeGlobal Purchasing Program Management, Ford ITOctober 2010 to Nov 2014Client: Ford ITDesignation: Sr. Lead Developer / Designer / Team LeadThe goal of the GPPM application is to establish a common Global Purchasing Program Management business process and reporting tool to ensure program launches are supported efficiently and with quality. This includes the automation of manual tasks performed by Global Purchasing Program Management for the following five work streams:Production Purchase OrdersProduction Tool OrdersPrototype Purchase OrdersCapacity StudiesUncommitted FundsModules involved are:Customer ReportData Summary ReportRegion - Web FocusPMT - Web FocusWorkstream - Web FocusProgram - Web FocusAdminRoles and Responsibilities:Designer & Developer, Hands on Java, J2EE Developerinvolved in Backlog refinement, Spring planning, Retrospection.Communicating BA for requirement clarificationProviding high level design to developers including the design pattern and UMLLeading the team of 8 membersMentoring the team for their low level designPerformed Code ReviewTechnologies Java 1.5, J2EE, Struts 1.1, Toplink, Web-Services, Sql Server 2008Environment IBM RSA, AccuRev, WebSphere 8Morcom ID Management, VirtusaOct 2009 to Oct 2010Client: JP Morgan ChaseDesignation: Sr. Developer / Designer / Team LeadThis system deals with KOPS User ID creation and its capabilities for Broker Dealers. The KOPS ID creation performs with various manual step approval based on the capability selected and finally persisted into an external system (KOPS/COSMOS).Development methodology  AgileModules involved are, Inbox, Request Status, User Search & Reports, Template, AdminRoles and Responsibilities:Hands on Java, J2EE DeveloperCommunicating BA for requirement clarificationProviding high level design to developer including the design patternLeading a team of 6 members.Mentoring the team for their low level designPerformed Code ReviewDeployment in DEV and QA using UnixPreparing release notes for QA and PROD deliverablesTechnologies Java 1.5, J2EE, Spring 2, Web-Services, Oracle 10g, Oracle BPM, Spring MVCEnvironment WebLogic 10.3, SVNOnline Work Request, Compunet ConnectionsMarch 2008 to Oct 2009Client: Photon infotechThis system developed for online work request which is sent to Super user, Managers and Special users and they will assign the project for Technicians based on their department. All the process for the request will be based on status and the status are managed by the managers on the project# assigned to the specific department.Main features and modules of this project include.Make new request, Assign Project, Outside AgencyRefine SearchManage SQL DataRoles and Responsibilities:Hands on Java, J2EE DeveloperDeveloped the Refine search and Manage SQL Data modules.Used Hibernate ORM to integrate Spring Framework.Implemented business logic using SpringPerformed Unit Testing of the developed components.Technologies JSP, Servlets, Spring 2.0, Hibernate 2.0, Oracle 10g, Web ServicesEnvironment WebLogic 8.1, SVN, EclipseHealth Record Tracking System, Compunet ConnectionsApril 2007 to March 2008Client: Photon infotechThis project was aimed to generate the health record ratios quarter based, and maintain the generated ratios for certain period. This was split into 5 modules, User maintenance, Uploading of files to calculate, calculation part, maintenance of generated ratios and report generation.Roles and Responsibilities:Hands on Java J2EE developerCustomized Data Access Module using DAO patterns, for entire transaction with database.Coded business logic with session bean, hibernate for User maintenance and calculation part.Technologies JSP, Servlets, EJB 2.0, Spring 1.2.6, Hibernate, Oracle 9i, WebServices, WebLogic 8.1, SVN, EclipseE-Accounting System, Compunet ConnectionsJune 2006 to March 2007E-Accounting System is basically used to maintain all account details of telecom organization. It will provide user friendly system maintaining accounting details. It consists of operation cash and bankbook, collection cash and bankbook; unpaid cases trail balance details and monthly reports. Operation cashbook contained all cash paid details from the company which include salary payments and advances, whereas bank book consists of payments paid in form of drafts. Collection cashbook consists of details regarding receipts, which include all deposits paid in cash form, Whereas Collection bank book consist of details of receipts in drafts. Deposits of contract work; bills monthly rentals will be added in this collection bankbook.Roles and Responsibilities:Hands on Java, J2EE DeveloperDesign, code, implementation and maintenance of scalable infrastructure for Web.Developed Servlet programs to transfer the control to other pages and programs depending on the business logics.Creating result pages using JSP.Processing the client form using Servlet.Technologies JSP, Servlets, Struts 1.3, EJB 2.0, Sql Server 2000, HibernateEnvironment WebLogic 8.1, SVN, Eclipse

Respond to this candidate
Your Message
Please type the code shown in the image:

Note: Responding to this resume will create an account on our partner site postjobfree.com
Register for Free on Jobvertise