| 20,000+ Fresh Resumes Monthly | |
|
|
| Related Resumes Sap Mdg Bods Charlotte, NC SAP BASIS HANA Charlotte, NC Regional Sales, Project Manager, Consultant, company planning Sherrills Ford, NC Sap Fico Accounts Payable Kannapolis, NC RPA Lead Consultant Kannapolis, NC Sap Abap Developer Charlotte, NC Systems Analyst SAP Tega Cay, SC |
| Click here or scroll down to respond to this candidate HariPHONE NUMBER AVAILABLE EMAIL AVAILABLEProfessional Summary 14 years of experience in IT, SAP MDG on S/4HANA 2022, SAP BODS, Info Steward, BW HANA, Data Services, Tableau and 2years in Microsoft .NET and Software Testing SAP MDG Data Modeling, UI, Process Modelling, DRF, BRF+, CBA, VALIDATIONS, WORKFLOWS SAP MDG EDITION, MASS CHANGE, FILE UPLOAD, EXTENDING DATA MODELS, CHANG REQ
Expertise in SAP MDG 9.2/9.1 Data, Process, UI modelling and Data Replication in MDG-M/S/C/F Domains S/4HANA 1809/1909/2022 SAP MDG DATA DOMAINS MATERIAL MASTER/BUSINESS PARTNER (CUST, SUPPL)/FINANCE
Involved in Analysis, Design, Development, Testing, Implementation and Support phases of the Projects for both SAP MDG, SAP BW, BO, HANA, BODS and Microsoft .NET Extensively worked on BW Composite Provider, ADSO, Info Cubes, DSO Objects, Info Objects, Info Sources, Data Sources, Transformations, DTP s, Multiproviders, Transfer rules and Update rules Good Knowledge in Generic extraction and LO Extraction Extensively worked on Web Analyzer and BEx Query Development for Reports creation. Experience in CO-PA extraction process and have reports built on it Proficient in handling errors errors occurred during Profitability Analysis (CO-PA) data loads Familiar loading data to CO-PA data targets COPAo01B and COPA_T02A Involved in Business Objects Administration creating users and groups, scheduling reports, publishing reports, planning security, authentication and single sign-on in for the reports developed using Business Objects 4.2/4.0. Used Open Hub Services to push data out of SAP BI using BODS Worked with Match Transforms, Address Cleanse and Data Cleanse Transforms for US Data Cleansing in Data Quality in BODS. Used Data Integrator and Platform Transforms extensively in BODS Worked extensively with Sales & Distribution team to design data migration from Legacy system to SAP using BODS Exporting and Importing meta data in Native HANA and expertise in managing Native Hana modeling concepts, Hana Live (RDS) views
Good experience on PL/ SQL(DDL, DML, DCL, Stored procedures) using Native HANA as a Database Good experience with all the Data Provisioning techniques (SDI,SLT and SAP BODS), Modeling constructs (Attribute, Analytic, and Calculation) and Reporting solutions (BOBJ Suite of tools) with Native HANA information models as a source. Experience in handling the Business ObjectsModules Designer, Web Intelligence, Central Management Console and InfoView, Crystal Reports, Lumira and Design Studio Experience in Data Visualization tools LUMIRA, TABLEAU Good knowledge in Universe designing, developing and generating complex, Adhoc Reports, Business views, List of values, reporting functionalities such as Master Detailed, Slice and Dice, drilling methodology, Filters, Ranking, Sections, Graphs and Breaks. SME in Dashboards, Best Practices in Dashboard Design, KPI Formulation & Web Reporting, developing information spaces and exploration sets in BOBJ Exposure in Oil & Gas, Pharmaceutical, Retail, Telecommunication, Production and Sales Data Warehouses. S/4 HANA migration cocpit (LTMC) - Migration Object Modeler (LTMOM) Legacy Master Data (Business Partner/Material/Financials) & TD migration to S/4 HANA 1909 Good knowledge in Eclipse based BW/4HANA & S/4 HANA (1809/1909)Technical Skill Set ERP Packages : SAP BW 7.0, 7.3, 7.4, ECC 6.0, Master Data Management : SAP MDG 9.2, 9.19.0, DRF, DM, PM, UI Modeling, BRF+, DQ SEARCHBusiness Intelligence Tools : SAP Business Objects (WebI, Crystal reports2013), AO, OBIEEData Visualization : BO Lumira 1.31,2.0 Design Studio 1.6, Tableau 10,10.3ETL Tools : BODS 4.1, 4.2, BODI, Talend 7.1.1, 7.2.1, SAP RDM s/4 hanaData Replication Technique : SLT, SDI,SDADatabases : HANA1.0/2.0, Oracle 10g/11g, SQL Server 2008/2012/2016 Microsoft Technologies : .NET 2005, C#
Programming Languages : ABAP/4-BWTrainings SAP MDG on S/4 HANA 2022/2021/1909/1809 (Material, BP, Customer, Supplier, Financial) Master Data BIG Data Insight/ Talend Open Studio Data Integrator, Job Design,Joins,Filter,Oracle/Flat file load SAP BPC 10.0 Lumira Discovery 2.0Assignment SummaryClient : HALLMARK CARDS/KANSAS CITY, MOEnvironment : SAP MDG on S/4HANA 2022, SAP BODS 4.3-14.3, SQL SERVER 2016, Teradata, Mainframe systemsDuration : Nov 2022 to PresentKey Roles and Responsibilities: Extending DM s (Material/Business Partner) in SAP MDG on S/4Hana Streamline approval process design with SAP MDG workflow s Replicating & Troubleshooting data into SAP Data Services, Ariba SLP & 3rd party systems with DRF IDOC s SOA in SAP MDG on S/4HANA SAP MDG-C integration & Replication to SAP C4C (Cloud for Customer) Setup MDG DRF filters to 3rd party systems Governing UIBB s at Workflow step level in MDG on S/4 hana 2022 Understand BADi for Parallel WF/Dynamic selection of Agents in Sap MDG Inbound: migration from Mainframe files to CSG Merchandise planning tool using BODS 4.3 Move file location data from MicroStrategy to CSG Merchandise planning tool Run stored procedures from SAP DS/BODS IDOC S/Bapi load, ABAP Dataflow using SAP BODS Perform Data Conversions and SAP DS code migration
OutBound: Load Data from CSG Merchandise planning tool to RDW (Teradata) for MicroStrategy reportingClient : ENERGIZER/SAINT LOUISEnvironment : SAP MDG on S/4HANA 2021, SAP BODS 4.2, SQL SERVER 2018, S/41909, SAP MDG9.2,FioriDuration : Feb 2022 to August 2022Key Roles and Responsibilities: Understand SAP Data Services, MDG HUB & CO-Deployment landscape Design MDG BRF+ workflows based on User Agent approval SAP MDG-S integration & Replication to SAP Ariba SLP Define Parallel WF for Multi distribution center approval using BaDI in SAP MDG on S/4HANA Authorization and Governance settings in SAP MDG for Business Users
Data migration from LEGACY to SAPS42021 using BODS 4.2 Migrate Data using IDOC interfaces/Bapi/Function Modules
Master Data: BP/C/S, Material DRF: Material master Replication to ERP using ALE IDOCs Perform Data Conversions SAP MM data migration including Plant/storagelocation/taxclassification/SalesorgnizationClient : MOOG INC/NYEnvironment : SAP MDG on S/4HANA 1909, SAP BODS 4.2, HANA2.0, S/41909, Fiori
Duration : July 2021 to Feb 2022Key Roles and Responsibilities: Change Request setup for Finance OG domain using Edition in MDG 1909 Setup BRF+ Validations for MM data Perform Data Services, Mass Change and File Upload in SAP MDG 1909
Activate BC Sets in MDG for Change request SAP MDG-S Data Replication to SRM Data migration from LEGACY to S4HANA1909 using BODS 4.2 Migrate Data using IDOC interfaces
Master Data: BP/C/S, Routings, Document Info Record, BoM, Material, Equipment, Purchase Info Record migration Perform Data Conversions Migrate your data using Fiori App (Flat file/Staging Tables) Knowledge in SAP SYNITI ADM field mapping, Datasource connections, executions SAP MDG/MDQ ConfigureMDQ (Remediation) on Product data with BRF+ validation rules
Client : TENNECO/MIEnvironment : SAP BODS 4.2, BW 7.4, SAP ECC/CRM/SCM, AZURE, HANA 2.0, S/4Hana 1809Duration : Dec 2020 to May 2021Key Roles and Responsibilities: Implementation/AMS
Data migration from SAP ECC & Flatfiles to Data Services, SAP HANA with BODS 4.2 Data migration from SAP ECC & Flatfiles to Microsoft Azure Data lake Defect fix/Data Reconciliation Data Services Production system Improvement measures Perform Adhoc loads &Archive source file once data load complete to HANA 2.0 S/4 HANA migration cocpit (LTMC) - Migration Object Modeler (LTMOM)/Material/Costcenter etc CPS redwood work load automations/Scheduling parameters/Pre conditions/Mappings similar to UC4Client : ECOLAB/MNEnvironment : MDG 9.1 on S/4HANA, SAP BODS 4.2, BW 7.4, SAP ECC/CRM/SCM, BIG DATA, AZURE, CLOUDERA,Duration : July 2019 to July 2020Key Roles and Responsibilities: Implementation/AMS
Involved in SAP MDG Functional (Data, Process, UI) Modeling, Data Services, Data Replication Framework Flex /Reuse modes (HUB & Co-Deployment) layout architecture Used Hub deployment method for MDG implementation, data resides in MDG systems, approved data replicated to ECC for Customer/Vendor and Material. Finance data replicated using the SOA and ALE services Material/Business Partner/Customer/Supplier/Financials Master Data Replication to S/4HANA
Create/Change/Block Unblock/Mass Change Material/Business Partner/Customer/Supplier Understanding Type 1/Type3/Type 4 Entities in Data Modeling Configuring Business Objects /Business activity/Logical actions for Change Request BRF+ and Static Workflows steps in Decision Tables for Change Request Troubleshoot FPM page errors in UI modeling Perform Initial Load using SAP BODS in HUB-Reuse storage mode from ECC Perform Data Quality Search & Duplicate check validation and setup HANA SEARCH for all Domains Design and Monitor the Cutover checklist and Go-Live plans Understand BADI s/Interfaces/Methods/ Access Class for Material/Business Partner Troubleshoot SAP MDG Change requests & WorkFlows during AMS activity
Designed and developed data ingestion (Replicate/ETL/ELT), conceptual, logical, physical data models, data marts, analytics and visualization, performance Ingest Data Services, SAP ECC SCM Full & Delta loads to cloud-MS AZURE BLOB container Ingest SAP CRM/ECC BW EXTRACTOR data & flatfiles to cloud-MS AZURE BLOB Master Data Attribute, Text & Transactional Data Build DSO/OpenHub and Ingest Nalco system data to AZURE BLOB Collect and import BW transports to Production Build Batchjobs and perform Delta Initialization in Sap Data Services Automate Batch jobs with loadtype variable for Full, Delta and Adhoc loads Expose custom extractors to ODP API to consume in Data Services Import extractors in Query & CDC mode Perform archive operation, MD/TD Data reconciliation with source Define Primary keys, Delta type &Metadata for extractors (ECC, SCM , CRM) Assist HADOOP team for Schema creation, schema enhancement, FL, DL, Data Validation Automate the SAP DS jobs in UC4 Data migration LTMC LTMOM to S/4 HANA using Files & Staging Tables with BODS/Data Services Troubleshooting Data Service BW extractors Delta monitor Recovering files from Archive to Azure blob Analyze data in AZURE DATA BRICKS/AZURE DATA LAKE/AZURE DATA FACTORY Recovering failed Delta from BW extractors (SCM) in ODQMON Data Services Production system Improvement measures Handling Data Reconciliation issues Azure DATA LAKE and ECC/CRM/BW Setup Data Store connection to ATTUNITY and extract FULL load using BODS to Azure Knowledge in PYTHON scriptsClient : DuPont/DEEnvironment : SAP MDG 9.1, SAP BODS 4.2 , Tableau, Sql server 2012, SAP HANA 2.0,BI 4.2, BWDuration : December 2018 to May 2019Key Roles and Responsibilities: Implementation/Enhancements Involved in Configuring SAP MDG Material, Supplier, Customer and Finance Master data Extend Material/Customer/Supplier by new entity in SAP MDG 9.1 Involved in Data Services, Troubleshooting Data Replication DRF and Workflow issues
Define Change Request actions and understand scenario for Parallel Workflows
Data Replication and Monitoring Change Requests in SAP MDG
Troubleshoot BODS connectivity issues with source systems, Sql server and Hana Develop code migration methodology for SAP BODS across Dev, Test, Prod landscape Migrating Specialty Products Data(SPECCO),CORTEVA(AGCO) MD/TD
Master Data includes Cust Master/GL account/Profit center/Material Master/Trade/Plan/Company Code/Cost Element/Cost Center etc Transactional Data Includes OpenOrders /Costcenter /Daily Sales/GL Sources include SAP & Non Sap (Flatfiles/Json/Excel) Worked on Data Quality Transforms Address Cleanse/Match/Data Cleanse
Read Inbound shared point files and batch numbers and load automatically to extraction layer 3 Layer Architecture - Extract data from source and load to Output Layer Further load data from Staging Layer to Data Inbound/Data Hub Target system data is posted to HANA 2.0 Perform Adhoc loads &Archive source file once data load complete to HANA 2.0 Analyze the reports built on Target System with Tableau
Written PL/SQL queries to extract source data to Data Services staging Managed Error Handling and Success results to trace data migration Worked on incremental loading for all the dimensions after initial load from stage to Data Mart (Star schema Data migration to SAP HANA Enterprise and build Information models (Cal Views), SQL scripting Involved in Unit and Integration Testing on reports and then exported them to Schedule Manager for Scheduling and Refreshing. Performed Data Conversions/functions before migrating data Worked on Data integration transforms History preserving, Table comparison, XML_pipeline Setting up of Data store connection to read source data from multiple source systems in BODS Extensively used BODS Management Console to schedule and execute jobs, manage repositories and perform metadata reporting. Extracted larger volumes of data(SDI) using ABAP data flow transform into HANA DB in BODS Created BODS Jobs, Work Flows, Data Flows, and Scripts using various Transforms (Integrator, Quality and Platform) to successfully load data from multiple sources into a desired target Retrieved data using SQL Select statements as part of performance tuning in Data services 4.2 Worked on scripts while reading files from a folder using WHILE_Loop in Data servicesClient : Canon/NYEnvironment : BODS 4.2, BI 4.2(WebI), Lumira,Oracle 11g, XML, WebService, SAP HANA 2.0Duration : April 2018 to November 2018Key Roles and Responsibilities: Handling priority/problem tickets/incidents/ housekeeping activities Troubleshooting Data Services, BODS connectivity issues with source systems and Soap Webservice Develop code migration methodology for SAP BODS across Dev, Test, Prod landscape Migrating Internal, Partner, Admin, Principle, Marketing, Promotion, Service, Pricing and Super users Master Data, User core attributes, Roles and Sellto Billto Information Sources include CNA UAM Oracle schema and LZ(flatfile)
Target system Soap webservice and intern data is posted to Oracle db
Validate against HR database for internal or external user classification Generating unique identity for each user using a function Convert userlevel codes for internal and Partner users based on requirement Worked on SQL transforms, DBlinks and Stored Procedures Written PL/SQL queries to extract source data to Data Services staging Consolidate data from source systems and build XML structure Managed Data Services, Error Handling and Success results to trace every user migration Worked on incremental loading for all the dimensions after initial load from stage to Data Mart (Star schema Data migration to SAP HANA Enterprise and build Information models (Cal Views), SQL scripting Sending and Receiving IDOC s from SAP through BODS Integration systems Involved in Unit and Integration Testing on reports and then exported them to Schedule Manager for Scheduling and Refreshing. Written & Performed Data Conversions/functions in PL/SQL Oracle 11g before Data migration Worked on BO universes IDT, WEBI, Crystalreports as per Business requirements. Information Steward for data profiling, validating rules, scorecards, metadata management Worked on Data integration transforms History preserving, Table comparison, XML_pipeline Setting up of Data store connection to read source data from multiple source systems in BODS Worked on scripts while reading files from a folder using WHILE_Loop in Data Services Extensively used BODS Management Consoleto schedule and execute jobs, manage repositories and perform metadata reporting. Created BODS Jobs, Work Flows, Data Flows, and Scripts using various Transforms (Integrator, Quality and Platform) to successfully load data from multiple sources into a desired target Extracted larger volumes of data(SDI) using ABAP data flow transform into HANA DB in BODS Retrieved data using SQL Select statements as part of performance tuning in Data services 4.2 Extensively worked on data validation and reconciliation BW, ECC, BO, HANA, DATASERVICESClient : AMGEN (Life sciences)/NJ &BangaloreEnvironment : SAP HANA 1.0, BODS 4.2, BI 4.2, Tableau, BW 7.4, IPDuration : July 2015 to Feb 2018Key Roles and Responsibilities: Handling priority/problem tickets/incidents/ housekeeping activities Monitoring and Scheduling process chains. Handling RFC s through Service now tool Created Native HANA information models Attribute, Analytical and Calculation Views Worked on Explain Plan Tables, Visual Plan, Sql Tracing as part of Performance tuning in Native Hana Configured and Replicated the data from SAP / Non-SAP sources to SAP HANA using SAP Landscape Transformation (SLT), extracted external system data using SDI/SDA Built graphical Calculation views using Table functions as datasource for complex calculation in Native HANA Worked on performance optimization of SQL Query/Models using Explain Plan, Visualization Plan and Expensive Statements trace Good experience in working on HANA optimized Calculation Engine functions (CE functions datasource access& relational operators) Built Calculation views on virtual tables with Smart Data Access(SDA) mechanism in HANA Converted Attribute and Analytic view to Calculation views to reduce data transfer between engines in HANA Created Sql script based calculation views
Analyzed and fixed real time replication server SLT delta failures in HANA
Followed Best practices/Performance tuning in designing and building HANA information models Transport HANA model objects using HANA ALCM Source to Target systems Functional verification activities of Development, Quality and Integration systems Enhancing reports in LUMIRA and Webi, working on Lumira 2.0 proof of concept source as HANA
Worked on linking Datasets and Building and Publishing Lumira Stories to BI Platform Created reports including formatting options such as Chart, Grouping, Sorting, Alerter, Ranking, Breaks, Sections, parameterized prompts. Involved in Unit and Integration Testing on reports and then exported them to Schedule Manager for Scheduling and Refreshing. Modified existing BO universes and WEBI reports as per Business requirements. Created Crystal reports when needed and exported to enterprise repository in order to make Crystal reports accessible. Support Business objects WEBI reports, Universe handling issues through incidents Information Steward for data profiling, validating rules, scorecards, metadata management
Extensively used BODS Management Console to schedule and execute jobs, manage repositories and perform metadata reporting. Worked on scripts while reading files from a folder using WHILE_Loop in Data servicesClient : STATOIL (Oil & Gas)/Capgemini/BangaloreEnvironment : SAP BODS, BO 4.0, Tableau, HANASPS 11, IP, BPC 10.0, BW7.0Duration : Jan 2014 to June 2015Key Roles and Responsibilities:
Involved in Statoil BO BW transition from IBM to CapGemini Handling priority/problem tickets/incidents/ housekeeping activities Monitoring and Scheduling process chains.
Handling RFC s through Service now tool
Classic and Sql script Analytic privileges and assign users to restrict accessing data native hana
SAP process area s involved SD, CRM, HR, Treasury & Payments, Accounting & Controlling, Plant O&M
Business Objects reports on HANA modeling constructs (analytic, Attribute & Calculation views) Created Data Visualization dashboards using embedded Excel sheet integrated with BEX Queries. Worked on complex reports that require running totals, sub-totals, sub-reports, cross-tab and conditional display of groups, fields, and messages. Created, managed & formatted standard reports, master/detail reports and cross tab reports using Report Manager Developed crystal reports and performed unit testing
Migration of BO reports, folders and universes from one repository to other repository Support Business objects WEBI reports, Universe handling issues through incidents Developed and executed unit test and integration test plan for BO reports Data load experience to SAP BW as Target, Success Factor using Data services Extensively worked on Data Services (Code Migration, Error Handling, Recovery Mechanism,IDOCS) Making use of SQL transform to read data from source system i.e Oracle in BODS Setting up of Projects, Jobs, workflow, dataflow and make them uniquely grouped while moving across different environment in BODS Extensive use of Data integrator and platform transforms such as Key_generation, Table_comparision, Case, Merge, Validation, sql, Query for transformation Created Profile and Central repositories to manage data consistency in BODS Migrate data for SAP FICO and HCM module objects such as Vendors, Customer, Materials, Purchase orders, Cost Centers, Profit Centers, Transactional data such as AP, AR, GL to single system ECC 6.0 in BODS Performed Full pushdown, Source & Target based performance tuning, Bulkloading in BODS
Worked on Data Services Management console to create users and grant privilege to access repositoriesClient : Philadelphia Energy Solutions (Energy Oil & Gas)/Capgemini/BangaloreEnvironment : SAP BODS, BI7.3,Sql serverDuration : June 2013 December 2013Key Roles and Responsibilities: Involved in designing technical specification document for PES_BW Involved in build phase of the project Developed BI data flows using FLAT file extraction Responsible for creation of test scripts and documentation Involved in Data Loading & Process chain Monitoring
Involved in unit testing, test case reviews and documentation Created Crystal reports when needed and exported to enterprise repository in order to make Crystal reports accessible Design and Develop BODS batch jobs for Material Master, Purchase orders, Asset Master and Batch Determination Use of Try-catch and while- loop transforms during splitting records in multiple files in BODS Use BODS conditional clauses such as where, group by, order by while restricting records
Generating Tab delimited output file as per load program Publishing Data Quality report for quickly status meeting and decision making As per data quality report assist business users to modify and correct the data Following are Objects worked as part of data reconciliation activity: PO vendor master, Asset PO, WBSmaster, asset location, cost center in BODSClient : Target (Retail)/Capgemini/BangaloreEnvironment : SAP BI7.3, BOXI R3, Sql serverDuration : December 2012 - June 2013Key Roles and Responsibilities:
Involved in requirements gathering, designing technical specification document for Project Systems
Involved in BUILD and UAT phase of Project Systems Module (Master Data and Transaction Data)
Responsible for creation test scripts and documentation Involved in planning and co-ordination of tasks Involved in Report testing, test case reviews and documentation Involved in unit testing, test case reviews and documentation Migration of existing in SSIS to Crystal reports from scratch Migrated the reports using LCM and import wizard Developed and executed unit test and integration test plan for BO reports Creation and maintenance of BO Universe and reports Worked on CMC module in creating user groups and folders and managing security in BOClient : Centrica (Gas & Energy)/Capgemini/BangaloreEnvironment : SAP BODS,BI7.3, SAPCRM, SAP ISU,OracleDuration : June 2012 December 2012Key Roles and Responsibilities: Involved in designing technical specification document for British Gas and Dataware House Involved in build phase of BGDW (Master Data and Transaction Data)
Developed BI data flows using LSA methodology Involved in enhancing CRM & ISU datasources Responsible for creation test scripts and documentation Involved in Data Loading &Process chain Monitoring Test case reviews and documentation Involved in unit testing, test case reviews and documentation Understanding the client business setup and structuring Object repository in Business object data services Developed and implemented solutions with data warehouse, ETL, data analysis, and BI reporting technologies. Most of the BODStransforms used Case, Merge, and Sequence Generator, Query Transform, Map Operations, Table Comparison, SQL, and Validation Transforms Created Data Flows to load data from flat file, CSV files with various formats into Data Warehouse. Involved in creating batch jobs for data cleansing and address cleansing. History preserving, capturing slowly changing dimension data at every interval, script writing, preparing the data from initial load perspective in BODSClient : Fossil (Lifestyle brands)/Capgemini/BangaloreEnvironment : SAP BI7.3, OracleDuration : December 2011 May 2012Key Roles and Responsibilities: Involved in requirements gathering technical pre-upgrade and technical post-upgrade activities Responsible technical pre-upgrade and post-upgrade activities Responsible for creation test scripts and documentation Involved in planning and co-ordination of tasks Co-ordination in India with respect to testing Involved in unit testing, test case reviews and documentationClient : Vodafone (Telecom)/IBM/BangaloreEnvironment : SAP BI7.0, BO, ECC 6.0, OracleDuration : Jan 2010 Nov2011Key Roles and Responsibilities: Interacting with Business Analysts for Requirements Gathering and clarifications. Provided the weekly status update on Project Progress. Created Cubes, DSOs, Multiproviders, Infoset, Transformations, InfoPackages and DTPs. Involved in implementing logic at Start Routine level
Created Process Chains for automating data load process Created Open hub for extracting data from info cube into Flatfile/table. Involved in developing Interaction Analysis Queries using Restricted Key figures, Calculated Key figures, Filters, Free Characteristics, and Variables. Worked on Report to report interfaces(RRI) to drilldown to detailed Report
Involved in Transportation of all Objects like: InfoObjects, Query s and DSO objects, Routines, Programs, etc. into BI test system after validation and testing the same in development. Prepared Project documents like Transport Checklist, Go-Live check list, Support document and Provided the Support Training to the Support Teams Created Unit Test Cases and Performed Unit testing Extensively Involved in System Testing and User AcceptanceTesting Created crystal reports including graphical reports, formula based and well formatted reports according to user requirements Enhancements in BO universes and reports Interacted with users for analyzing diverse BO reports and scheduling them Involved in creation of webi,crystal and dashboards according to user requirementsClient : Vodafone (Telecom)/IBM/BangaloreRole : Application Developer
Environment : C#, WINDOWS APPLICATION, SQL SERVER 2005Duration : March 2007 Dec 2009Description: The unified Front End (UFE) has been built as a unique interface for Vodafone stores and customer care services. The Legacy System which existed prior to UFE used Siebel and SAP applications to support Call Centre and Stores activities. These applications even when optimized were not process oriented, not allowing the optimal service time for repetitive easy operations, implicating substantial training period and requiring extensive software licensing.The main objectives behind the creation of UFE are: To have a single interface which can be customized according to the user profile An interface which is independent of backend systems (Smart Client) Easy to use (reduced training for users and short handling time)Key Roles and Responsibilities: Creation of (HLD) High Level Document from the Feasibility document Involved in preparing Detailed Design Document(LLD)
Preparing Unit Test Plan and Unit Test Result document Decision making on UI design of new applications; Suggesting changes to customer Coding of the functionality from the Detailed Design Documents(LLD) document Database related activities and implementation Preparing Release Notes using open source software Wiki Writing scripts for database related changes Developing Windows Services, Web ServicesClient : Royal Bank of Canada (Banking - CANADA) /IGate/BangaloreEnvironment : C#, ASP.NET, SQLSERVER 2000, JAVA SCRIPTDuration : July 2006 Jan 2007Key Roles and Responsibilities: Designing UI, Coding, Unit Testing. Safekeeping is banking project to generate fee file for specific range of Canadian regions. Ontario, International and Quebec Developed on OPS Tools framework and modules are Excluded Accounts,Feefile Generation,TaxMaster,Confirmation Page ,Fee file Generation Report. OPS Tools framework is designed with XML and WEB SERVICES using YDO0_DEV (sqlserver) as Database.
Work flow of project: UI interface is user controls Excluded Accounts screen will display the previously excluded accounts; at the same time we can include these accounts for fee file generation. From Account Filtration page/Fee file generation page list of accounts are checked/unchecked with available checkboxes.Feefile will be displayed on Confirmation Screen as well as you can generate PDF report.TaxMaster which updates tax details GST,QST,HST for Ontario, International and Quebec regions
EducationBachelor s Degree (B.Tech),CSIT, (India) |