0% found this document useful (0 votes)
42 views10 pages

Yethishag Data Analyst Resume

Uploaded by

iamtyagi14
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
42 views10 pages

Yethishag Data Analyst Resume

Uploaded by

iamtyagi14
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 10

Sr.

Data Analyst
Yethisha
Ph: +1 313-314-8903
Email: [email protected]
Employer : [email protected], +1 980-304-8251
PROFESSIONAL SUMMARY:
 Highly analytical and process-oriented data analyst with 10+ years of experience in data analysis and data
management and proven ability to work efficiently in independent and teamwork environments.
 Experienced in Requirement Analysis, Test Design, Test Preparation, Test Execution, Defect Management,
and Management Reporting.
 Experience AWS Cloud technology for building and testing Data Pipelines and performing ETL/ELT
process.
 Experience with Data Ingestion, Data Storage, Data Analysis, and Visualization using cloud technologies and
building Data pipelines.
 Extensive knowledge in all phases of Data Acquisition and Data Warehousing, Data Modeling
 Experiences with Data Processing and Data Transformations such as Mapping, Monitoring, Debugging,
Performance Tuning, and Troubleshooting Hadoop clusters.
 Experience in large-scale application development using Big Data technologies like Hadoop (HDFS,
MapReduce, Yarn), Spark, Hive, Impala, HBase, Sqoop, and AWS Cloud
 Expert in Amazon EMR, Spark, Kinesis, S3, Boto3, Bean Stalk, ECS, Cloud watch, Lambda, ELB, VPC,
Elastic Cache, Dynamo DB, Redshift, RDS, Athena, Zeppelin & NIFI.
 Experience in using AWS cloud services like Kinesis, S3 Data Lake, AWS Glue, Athena, and Amazon
Quick Sight.
 Experience in transferring data from the data resources to Amazon Redshift, AWS S3 using AWS data
pipelines.
 Led end-to-end data migration from legacy systems to SAP S/4HANA, ensuring data integrity and achieving
a 98% migration success rate.
 Experience in creating, managing, analyzing, and reporting the internal business client’s data using AWS
services like Athena, Redshift, EMR, and Quick Sight.
 Performed root cause analysis on supply chain disruptions, identifying bottlenecks in transportation and
recommending cost-effective rerouting strategies.
 Designed and optimized multi-layered ETL frameworks in Databricks using Delta Live Tables (DLT) for
real-time data processing, automated schema enforcement, and change data capture (CDC).
 Experience building applications using PySpark and AWS Lambda.
 Responsible for storing data on S3 files using Lambda Functions and AWS Glue using Python and
PySpark
 Experience in designing and implementing modern data platforms such as Data Fabric, Data Mesh, Data
Hubs, and data-driven environments.
 Collaborated with marketing and product teams to extract and model data from Adobe Analytics
Warehouse, delivering actionable insights that improved campaign ROI by 25%.
 Executed full end-to-end analytics projects using Waterfall—conducted rigorous requirement gathering,
designed data models, developed ETL pipelines, performed batch testing, and deployed finalized
dashboards—ensuring each phase was completed before progressing to the next (Requirement → Design →
Build → Test → Deploy).
 Performed advanced performance tuning using adaptive query execution (AQE) and caching strategies in
Databricks to reduce job execution time on billion-row datasets by 60%.
 Built customized dashboards in Looker/Tableau to enable stakeholders to interact with live A/B test
metrics and derive on-demand insights.
 Led the migration of on-premise ETL processes to GCP, utilizing native tools like BigQuery, Cloud DataProc,
and Cloud Composer, which enhanced data processing capabilities and reduced operational costs.
 Developed and executed complex data models and CDS views in SAP S/4HANA to extract real-time
business insights across finance, logistics, and supply chain modules.
 Skilled in designing data catalogs and dictionaries driven by active metadata, enabling efficient data
management and governance.
 Facilitating migration of R code from on-prem to AWS, showcasing familiarity with packages such as
ggplot2, caret, and dplyr.
 Led cross-functional collaboration between IT, legal, and business stakeholders to develop adaptive data
ethics policies for AI applications.
 Worked closely with cross-functional teams to gather requirements and translate business needs into
technical specifications within Alteryx workflows.
 Mastery over SQL, with an ability to pivot R code for AWS data source integrations.
 Proficient in RStudio, leveraging its IDE features for efficient script development, debugging, and
visualization.
 Built analytics dashboards and reports by integrating SharePoint with Power BI, Excel Power
Query/Pivot, and SSRS. Enabled real-time tracking of site usage, data quality, and business performance.
 Developed & maintained interactive BI reports by converting business input into technical specs—
enhanced dashboard-driven decision-making.
 Extensive experience in developing and deploying interactive web applications using RShiny to represent
complex datasets in an intuitive manner, enabling stakeholders to derive actionable insights.
 Facilitated migration of R code to AWS, Qlik Sense, and Power BI, leveraging SQL for data source
integrations.
 Extensive experience in developing interactive web applications using RShiny, Qlik Sense, and Power BI.

TECHNICAL SKILLS:

Programming: SQL, Python, R


Cloud: AWS, Google Cloud Platform
Reporting Tools: Business Objects, Tableau, Microsoft Power BI, Qlik Sense, DOMO
Data Modeling: Star-Schema Modeling, Snowflake-Schema Modeling, FACT and dimension tables,
Pivot Tables,
Testing Tools: Win Runner, Load Runner, Test Director
Databases: Microsoft SQL Server 2014/2012/2008 R2/2008, MySQL, Oracle, DB2, Teradata, MS
Access.
Environment: Windows (95, 98, 2000, NT, XP), R Studio, Rshiny
Other Tools: MS-Office suite (Word, Excel, Project, and Outlook).

PROFESSIONAL EXPERIENCE:

Client: Walgreens, Wayne, MI May 2022- Present


Role: Senior Data Analyst
Responsibilities:
 Proficient in developing SQL and Snowflake Queries, extracting data seamlessly from multiple databases.
 Experienced in supporting the migration of data processes, including R code, from on-prem solutions to
AWS.
 Set up and managed data tasks using Apache Airflow, ensuring automation and reliability.
 Developed custom workflows in Airflow to sequence tasks efficiently.
 Built dashboards in Tableau/Power BI to monitor key supply chain KPIs including on-time delivery, order
fulfillment rate, and inventory turnover.
 Leveraged Adobe Analytics to track and analyze web and mobile user behavior, driving actionable insights
that increased campaign performance by 30%.
 Conducted in-depth analysis of consumer behavior using transactional, loyalty, and clickstream data to
identify trends and drive data-backed marketing strategies.
 Segmented customers based on demographic and purchase data using k-means clustering and RFM
analysis, resulting in a 20% lift in targeted campaign engagement.
 Developed custom dashboards and reports in Adobe Workspace to visualize KPIs and customer journeys
for executive and marketing stakeholders.
 Designed and executed custom queries using Adobe Data Warehouse to extract large volumes of user-level
behavioral data for in-depth analysis and reporting.
 Collaborated with cross-functional HLS teams to analyze patient engagement, prescription adherence, and
pharmacy operations data, driving actionable insights for improved care delivery.
 Checked the health of data tasks using Airflow's tools and promptly resolved any issues.
 Deployed Apache Airflow on cloud platforms like AWS for continuous availability.
 Automated the delivery of data warehouse reports to stakeholders by configuring scheduled exports in
various formats (CSV, Excel, FTP), reducing manual effort by 40%.
 Used SQL and Python to automate daily sales reporting across stores and e-commerce, saving 12+
hours/week in manual reporting.
 Created detailed design documentation, including ER diagrams, data flow maps, and schema definitions, to
guide development and ensure alignment with stakeholder requirements .
 Leveraged Delta Lake on Databricks for versioned, reliable, and performant data lakes to ensure data
consistency and ACID compliance.
 Interpreted A/B test outcomes to deliver data-driven recommendations, resulting in a 15–30% increase in
conversion rates and improved user experience.
 Leveraged tools like Google Optimize, Optimizely, and Adobe Target to execute controlled experiments and
measure impact on KPIs such as conversion rate, bounce rate, and user engagement.
 Expertise in testing, including end-to-end evaluation of process flows and ensuring reliability and accuracy.
 Extracted customer transaction data using SQL-based systems and performed data cleansing and
preprocessing.
 Integrated Databricks with enterprise data warehouses (Snowflake, Redshift) and BI tools (Power BI,
Tableau, Looker) for unified and governed data consumption across departments.
 Managed the end-to-end lifecycle of data analytics projects on GCP, from requirements gathering and data
modeling to deployment and post-implementation support
 Applied dimensional and relational data modeling techniques, including star and snowflake schemas, to
structure data warehouses in BigQuery, facilitating more efficient data analysis.
 "Developed and optimized complex SQL queries in Teradata SQL Assistant to extract, transform, and
analyze multi-terabyte datasets for [specific business area, e.g., customer behavior, sales trends]."
 Integrated Excel with external systems, such as Google Sheets and APIs, enabling real-time data
collaboration and cutting update time by 35%.
 Collaborated with cross-functional teams and senior stakeholders to deliver clear, actionable insights
through interactive dashboards, executive summaries, and tailored data stories.
 Designed, implemented, and analyzed A/B and multivariate tests to evaluate feature changes, UI
modifications, and marketing strategies across web and mobile platforms.
 Leveraged Adobe Data Warehouse to build multi-dimensional data segments and funnel visualizations for
tracking customer journeys across web and mobile channels.
 Generated and maintained system documentation and policies, standardizing CMS procedures and
controls, which contributed to enhanced SOX compliance and operational effectiveness
 Utilized Databricks notebooks to perform large-scale data analysis using PySpark, SQL, and Python,
enabling faster insights and real-time business reporting.
 Built interactive dashboards in Tableau and Power BI to track product performance, consumer retention
metrics, and promotional ROI in real time.
 "Wrote efficient BTEQ scripts to automate data extraction from Teradata, reducing manual effort by X%."
 Conducted data audits and compliance assessments to ensure adherence to data governance standards and
regulations.
 Integrated A/B testing results into broader marketing attribution models, improving cross-channel
strategy alignment and ROI measurement accuracy.
 Developed and executed data migration strategies, including data mapping, transformation, validation, and
loading using tools like SAP Data Services, LSMW, and SAP Migration Cockpit.
 Collaborated with cross-functional teams to gather requirements and ensure seamless integration of
migrated data into SAP S/4HANA.
 Analyzed and tuned Teradata SQL queries, resulting in a [X]% reduction in query execution time and
improved reporting efficiency.
 Implemented real-time data processing frameworks using GCP services to analyze terabytes of data,
increasing data accuracy by 45%
 Integrated diverse data sources, including SQL Server, Oracle, APIs, and flat files, using Alteryx Designer to
create unified datasets for comprehensive analysis.
 Developed and maintained ETL pipelines in Databricks to ingest and transform data from structured and
unstructured sources, improving data availability for analytics.
 Developed statistically significant test hypotheses and defined clear success metrics using knowledge of
hypothesis testing, confidence intervals, and p-values.
 Designed a data staging layer approach, collaborating with Master Data and Migration Leads to simplify
complex S/4HANA structures—mapping legacy data to target schemas in streamlined staging tables to
improve iteration speed and minimize errors
 Designed and maintained critical BI reports by gathering user requirements and translating them into
technical specifications—ensured accurate, timely insights from CMS/ERP data for order management, EDI,
pricing, and customer modules
 Designed and executed complex SQL queries for data extraction, transformation, and validation across
large datasets, ensuring high data accuracy and integrity during migration.
 Utilized Power BI and Qlik Sense for data visualization and dashboard creation, enabling stakeholders to
derive actionable insights.
 Implemented Dataform for version-controlled, collaborative, and testable SQL-based data transformations
within BigQuery, improving development efficiency and data quality.
 Developed and maintained complex Alteryx workflows and analytical applications, automating data
preparation and ETL processes, resulting in a 40% reduction in manual data handling time.
 Collaborated with Data Engineers and Scientists within Databricks Workspaces using Git-integrated
notebooks and shared clusters, streamlining cross-functional data workflows.
 Translated analytics tasks into Agile artifacts, creating and refining user stories with clear acceptance
criteria and using JIRA/Confluence for backlog management—ensuring transparency and traceability
 Implemented ETL and data quality processes, performing comprehensive data cleansing, mapping,
exception reporting, and exception resolution to ensure CMS data accuracy and consistency
 Automated data transformation workflows using scheduled queries and Dataform integration with
BigQuery.
 Worked on real-time streaming data analysis using Structured Streaming in Databricks for live operational
dashboards.
 Designed and implemented scalable ETL pipelines using GCP services such as BigQuery, Dataflow, and
Cloud Storage, resulting in a 30% improvement in data processing efficiency.
 Architected ETL pipelines in Excel with Power Query to ingest, clean, normalize, and merge data from SQL,
CSV, XLSX, XML, JSON, and web APIs—reducing prep time by 40%+
 Designed interactive dashboards in Power BI and Qlik Sense, incorporating dynamic visualizations for
enhanced data exploration.
 Integrated SAP S/4HANA with SAP Analytics Cloud and Power BI to deliver real-time, self-service BI and
visualizations to business users.
 Proficient in AWS, Power BI, Qlik Sense, SQL, Snowflake, Apache Airflow, PySpark, R, and data
analysis techniques.
Environment: R, SQL, AWS, Snowflakes, Python, Power BI, Qlik Sense, Apache Airflow, CMS, PySpark.

Client: Mayo Clinic Rochester, MN June 2019-April 2022


Role: Data Analyst
Responsibilities:
 Conducted data exploration using both Python's SciPy and R's tidyverse packages to analyze patterns and
select features effectively.
 Responsible for creating Data Engineering solutions for Business related problems
 Experience with Batch Processing using AWS Cloud technologies .
 Performed deep-dive analysis using Adobe Customer Journey Analytics to uncover cross-channel
behavioral trends and funnel drop-offs.
 Responsible for maintaining a Hadoop cluster on AWS Elastic MapReduce
 Performed end-to-end architecture and implementation evaluations of different AWS services such as
EMR, Redshift, S3, Athena, Glue, and Kinesis.
 Used AWS Athena to import structured data from S3 Data Lake into other systems such as RedShift to
generate reports
 Integrated Adobe Analytics data with BI tools (Power BI/Tableau) to create multi-source visualizations
supporting customer segmentation and ROI tracking.
 Used AWS services such as EC2 and S3 for dataset processing and storage.
 Developed and implemented ETL pipelines using Amazon Cloud technologies.
 Designed and implemented data quality metrics and dashboards for better decision-making and
monitoring of data governance KPIs.
 Implemented strategic partitioning and clustering on high-cardinality fields in multi-terabyte BigQuery
tables, reducing average query runtime by 40% and lowering monthly query costs by 25%
 Automated data extraction and integration processes with GCP Data Fusion and orchestrated workflows
using Cloud Composer, reducing manual intervention by 20%.
 Integrated Agile into DataOps pipelines by applying Agile principles to ETL development and deployment—
automated test/validation steps and introduced data collaboration practices
 Supported data warehouse development: contributed to requirement gathering, schema mapping, and
rigorous testing, integrating CMS application data into enterprise-level analytics
 Maintained a centralized A/B test knowledge base to share outcomes, insights, and learnings across
product and business teams for future strategic decisions.
 Partnered with ETL developers to define data transformation logic and participated in ETL workflow
design and validation using tools like Informatica and Talend.
 Set up BigQuery external tables to query data directly in Cloud Storage, enabling cross-environment joins
and reducing data duplication and ETL workload by 30%
 Utilized breakdown dimensions and calculated metrics in Adobe Data Warehouse to support cohort
analysis and churn prediction for high-value customer segments.
 Applied lift analysis and cohort analysis post-test to assess long-term impact of A/B variants on retention,
CLV (Customer Lifetime Value), and ROI.
 Implemented appropriate indexing strategies (e.g., Primary Indexes, Secondary Indexes) and collected
statistics on Teradata tables to enhance query performance.
 Automated job scheduling and orchestration using Databricks Jobs, reducing manual workloads and
ensuring data freshness for daily reporting.
 Mentored junior analysts on best practices in A/B testing design, interpretation, and ethical
experimentation principles.
 Established data governance protocols and security measures on GCP, mitigating security risks by 75% and
ensuring compliance with industry standards.
 Implemented phase-gate reviews at the end of each stage (requirements, design, build, test) with
stakeholders and QA teams to confirm objectives were met and sign-offs were documented .
 Built Factor Analysis and Cluster Analysis models in RStudio leveraging both Python's SciPy and R's
factoextra package to segment customers into distinct target groups.
 Conducted comprehensive data analysis on large datasets within Teradata to identify key trends, patterns,
and anomalies, informing strategic business decisions for [specific department/project.
 Used RShiny to develop an interactive dashboard showcasing the customer segments derived from
clustering for easy stakeholder interpretation.
 Integrated Databricks with cloud storage systems (e.g., AWS S3, Azure Data Lake, GCP Cloud Storage) to
source data in real time for analysis.
 Built ETL pipelines using SAP Data Services, LSMW, LTMC, BODS, or Migration Cockpit, supporting over
100+ objects (vendors, materials, GL masters, cost centers) and achieving 99%+ migration accuracy
 Provided help desk assistance and end-user support in a multi-plant ERP/CMS environment, improving
system reliability and empowering users with best-practice workflows
 Designed and optimized complex SQL queries on HANA tables for real-time data processing, minimizing
latency for large-scale financial and logistics reports.
 Built end-to-end data pipelines using Cloud Composer (Apache Airflow) to orchestrate workflows across
GCP services.
 Translated complex data models and statistical outcomes into business-friendly language, enabling
informed decision-making for C-level executives and non-technical stakeholders.
 Created and maintained complex views and macros in Teradata to provide streamlined access to data for
business users and reporting tools (e.g., Tableau, Power BI).
 Developed and maintained optimized BigQuery data pipelines to process structured and semi-structured
data from sources like Cloud Storage and Cloud SQL.
 Generated ad-hoc and scheduled reports from Teradata using SQL, providing actionable insights into
specific business metrics, e.g., customer churn, product performance.
 Collaborated with business stakeholders to develop interactive dashboards in Tableau and Power BI,
leveraging Alteryx-prepared data to provide real-time insights.
 Developed and validated data integrity checks on Teradata datasets, ensuring data accuracy for critical
business reporting
 Enhanced data performance and post-migration reporting, leveraging HANA-based calculated and analytic
views, and integrating with SAP BW/BODS ETL processes
 Integrated Adobe Data Warehouse output with SQL-based BI tools such as Tableau and Power BI to
develop dynamic dashboards and cross-platform performance visualizations.
 Automated supply chain data pipelines using Python and scheduled ETL jobs, improving reporting
frequency from weekly to daily.
 Led formal quality assurance and system testing, including unit, integration, and user acceptance testing
(UAT), to ensure data accuracy and business logic integrity before final rollout .
 Conducted training sessions for staff on data governance best practices, enhancing organizational
understanding and adherence to data policies.
 Performed deep-dive analysis using raw data exported from Adobe Warehouse to uncover traffic
anomalies, page performance bottlenecks, and attribution gaps.
 Utilized RMarkdown in RStudio for comprehensive report generation, merging code, results, and
commentary, which was shared with business teams for deeper insights.
 Connected Airflow with other data tools like Spark for more powerful data processing.
 Kept Airflow updated with the latest features and safety updates.
 Created partitioned and clustered tables in BigQuery to improve query performance and reduce costs.
 Trained my team on how to use Airflow effectively.
 Leveraged Regression, Tree-based ensemble methods, and other techniques using both Python and R to
build robust predictive models.
 Participated in Databricks performance tuning and cluster optimization to enhance execution times for
high-volume data queries.
 Integrated ERP and warehouse management system (WMS) data for end-to-end supply chain visibility.
 Supported migration projects across hybrid environments including cloud platforms (AWS, Azure) and on-
premise systems.
 Automated data cleaning of mixed structured and unstructured data from diverse sources using R and
Python scripts.
 Managed large-scale data read/writes from csv and Excel files, extensively using both pandas in Python
and readr in R.
Environment:- Python (SciPy, NumPy, Pandas, StatsModel, Plotly), GCP, CMS, MySQL, Star Schema,
Snowflake Schema, Excel, Google Cloud Platform, Tableau 9.x, D3.js, SVM, Random Forests, Naïve Bayes
Classifier, A/B experiment, Git 2.x, regression, logistic regression, Hadoop, NoSQL, Teradata, OLTP, random
forest, R, RStudio, Rshiny.

Client: ConocoPhillips – Houston, TX Jan 2017-May 2019


Data Analyst/Informatica Developer
Responsibilities:
 Extracted, transformed, and analyzed large volumes of operational and financial data using SQL, Python,
and Oracle ERP Cloud, generating actionable insights for forecasting, budgeting, and enterprise
performance strategies.
 Designed and implemented interactive Power BI and Tableau dashboards to visualize asset performance,
supply chain efficiency, and resource utilization, enabling real-time, data-driven decisions across
operations.
 Developed and maintained ETL pipelines using SSIS, Informatica, and Python scripts, ensuring
standardized, automated ingestion of structured and semi-structured data for enterprise reporting.
 Automated data warehouse refresh cycles and orchestrated complex ETL workflows using Apache
Airflow, dbt, and cloud-native tools, significantly improving data pipeline reliability and uptime.
 Implemented BigQuery slots management and reservation planning to ensure consistent performance for
critical analytical workloads.
 Validated Adobe Data Warehouse exports with raw server logs and data layer audits to ensure data
integrity and trust in analytical insights.
 Implemented data provisioning and ETL workflows using SAP BODS, SLT, SDA/SDI to feed both S/4HANA
and BW on HANA environments; optimized data ingestion and replication processes
 Integrated third-party data sources such as Salesforce, Google Analytics, and NetSuite into the enterprise
data warehouse, enabling a single source of truth for company-wide reporting and analytics.
 Conducted regression analysis in Excel and built SQL-based models to forecast demand, production
bottlenecks, and performance anomalies across field operations.
 Designed and implemented Bayesian A/B tests in Python for dynamic decision-making and faster
experimentation in high-traffic environments.
 Utilized Adobe Experience Platform (AEP) to unify customer data from multiple sources and build
predictive models for personalization and churn reduction.
 Created data visualizations using built-in Databricks SQL dashboards and exported insights to Power
BI/Tableau for executive reporting.
 Assisted in designing and implementing ETL processes to load data from diverse sources into the Teradata
data warehouse using FastLoad and MultiLoad utilities.
 Conducted regular security audits of GCP data environments, identifying and remediating vulnerabilities to
maintain a strong security posture.
 Facilitated stakeholder workshops to walk through analytical models, root cause analysis, and predictive
insights, ensuring alignment on outcomes and next steps.
 Conducted spatial analysis using Alteryx's spatial tools to optimize regional sales strategies, contributing to
a 10% boost in regional performance.
 Maintained extensive documentation—requirements docs, test plans, change logs, and support guides—for
auditability and future maintenance in post-deployment phase.
 Managed and monitored data loading jobs into Teradata, ensuring timely availability of data for reporting
and analysis.
 Developed automated workflows and custom forms using SharePoint Designer, Power Automate, InfoPath,
and third-party form tools—cutting manual data entry time by up to 40% and ensuring high data integrity.
 Facilitated training and knowledge transfer, developing user guides and conducting sessions for CMS/ERP
best practices—bolstered adoption and reduced support tickets
 Architected and deployed an enterprise-wide data governance framework for a Fortune 100 financial
services firm, reducing regulatory compliance risks by 47%.
 Applied machine learning models on Adobe Analytics export data to predict customer lifetime value (CLTV)
and inform retention strategies.
 Optimized query performance and data processing speed by applying tuning techniques in Snowflake,
BigQuery, and SQL Server, ensuring efficient use of cloud resources and reduced latency.
 Ensured data security and compliance with internal governance policies and industry standards by
implementing AWS IAM policies, data encryption, and audit frameworks.
 Used PowerShell scripting to automate SharePoint admin tasks such as permissions management, site
provisioning, and data migration—reducing site setup time by ~30%.
 Employed Alteryx's predictive analytics tools to build models for customer segmentation, leading to a 15%
increase in targeted marketing effectiveness.
 Managed JIRA product backlogs, facilitated Agile ceremonies, and collaborated with cross-functional teams
— including engineering, finance, and IT — to align data initiatives with strategic business goals.
 Created and maintained dynamic project documentation using Confluence and visual workflows in
Lucidchart, promoting transparency and knowledge sharing across stakeholders.
 Utilized Alteryx's in-database tools to process large datasets efficiently, reducing data processing time by
25%.
 Developed custom DOMO applications, extended Dynamics 365 reporting functionality, and utilized
Smartsheet Gantt charts for resource tracking and project planning.
 Utilized BigQuery ML to build and evaluate predictive models directly within the data warehouse
environment.
 Integrated BigQuery with visualization tools such as Looker Studio and Tableau for building interactive
dashboards and real-time analytics.
 Built and managed metadata repositories, data governance documentation, and lineage tracking to support
MDM frameworks involving OLTP, ODS, and OLAP systems.
 Mentored junior analysts and developers on SQL, Tableau, and Informatica best practices, while driving
iterative improvements through rapid prototyping and stakeholder feedback loops.
Environment: SQL Server, Python (Pandas, NumPy), Tableau, Power BI, Informatica, AWS (IAM, S3,
CloudShell SDK), JIRA, Excel (Power Query, Regression Tools).

Client: Apollo Pharmacy, Hyderabad, India April 2015- May 2016


Junior Data Analyst
Responsibilities:
 Gathered, cleaned, and integrated large-scale healthcare datasets from EHR systems, insurance claims,
and patient surveys, ensuring consistency and high data quality.
 Collaborated with clinicians and hospital administrators to align data insights with real-world clinical
workflows and healthcare protocols.
 Built and refined predictive models to reduce hospital readmissions, incorporating feedback from clinical
teams to improve model accuracy based on evolving patient outcomes.
 Ensured compliance with HIPAA and healthcare data privacy standards while analyzing sensitive
patient records.
 Developed and maintained interactive dashboards and KPIs using Power BI, providing visual insights into
patient readmissions, treatment efficiency, and departmental performance.
 Utilized BigQuery Omni to analyze data across multiple cloud environments (AWS, Azure) without data
movement, enabling cross-cloud analytics and unified reporting
 Designed and implemented data quality metrics and dashboards for better decision-making and
monitoring of data governance KPIs.
 Designed and implemented ETL pipelines using Python and SQL to load, transform, and structure patient
data for analytics.
 Mapped data to HL7 and FHIR standards to ensure interoperability between systems, supporting care
coordination and regulatory compliance.
 Performed data validation and quality assurance testing, creating test cases for inbound/outbound data
flows and reporting metrics.
 Automated data refresh schedules, file handling, and pipeline monitoring using Python scripting, improving
data timeliness and operational efficiency.
 Participated in Agile ceremonies (daily standups, sprint reviews) and used JIRA to track tasks, identify
bottlenecks, and deliver iterative improvements.
 Managed data security and access controls using BigQuery IAM roles and column-level security for
sensitive datasets.
 Developed and enforced data governance policies that improved data accuracy by 30% and reduced
inconsistencies across business units.
 Facilitated Joint Application Design (JAD) sessions to capture cross-functional requirements and
prioritize reporting needs with end-users.
 Migrated legacy data from on-premise systems and other cloud databases into BigQuery using Data
Transfer Service and custom ETL scripts.
 Collaborated with cross-functional teams to establish data stewardship programs, reducing data-related
incidents by 25%.
 Conducted statistical analysis to identify trends, root causes of readmissions, and high-risk patient cohorts,
enabling targeted clinical interventions.

Environment: Windows 10/11, SQL Server, MySQL, PostgreSQL, Python (Pandas, NumPy, Matplotlib,
Seaborn), Power BI, Tableau, Excel (Pivot Tables, Power Query), JIRA, Git/GitHub, Shell Scripting,
Agile/Scrum, HIPAA-compliant data platforms

Client : Recro, Bangalore, India June 2014- March 2015


BI Reporting Consultant/ Data Analyst
Responsibilities:
 Led full-cycle implementation of Data Warehousing projects, including requirement gathering, system
analysis, ETL design, data modeling, development, testing, deployment, documentation, and post-
production support.
 Designed and developed Power BI dashboards using DAX, custom filters, drill-throughs, and visualizations
(funnel, donut, scatter plots) to deliver interactive and data-rich insights across departments.
 Migrated Tableau reports to Power BI, streamlining analytics processes and improving performance for
supply chain and log data visualization teams.
 Optimized query performance by monitoring slot usage and query execution plans via BigQuery's Query
Plan Explanation tool.
 Conducted data audits and compliance assessments to ensure adherence to data governance standards and
regulations.
 Analyzed data quality issues and provided actionable insights leading to a 15% improvement in data
reliability.
 Built robust ETL pipelines with Informatica PowerCenter, transforming and loading data from DB2,
Oracle, SQL Server, and flat files into staging, ODS, EDW, and data marts for Sales, Finance, and Marketing.
 Developed SCD Type 1 and Type 2 mappings, designed complex transformation logic, and implemented
metadata-driven pipelines to support scalable data warehousing solutions.
 Created real-time data pipelines using Spark and SQL for user activity tracking and ad profile generation;
implemented duplicate removal logic with Spark Data Frames and “Upset” logic.
 Managed and automated SQL Server upgrades, patch installations, and maintenance activities while
ensuring database integrity and high performance.
 Developed and validated test cases for ETL workflows and reporting using transformation rules and
source-to-target mapping; executed both manual and automated testing strategies.
 Performed extensive data validation using SQL queries in UNIX environments, ensuring end-to-end data
integrity and alignment with business rules.
 Conducted black-box and white-box testing (unit, integration, regression, and data-driven) and tracked
defects using Quality Center to ensure software reliability and performance.
 Created and managed role-based access controls in Power BI, embedding reports and maintaining secure,
scalable analytics environments for end users.
 Supported collaborative Excel-based data modeling and validation, enabling multi-user access for efficient
review, data preparation, and project coordination.
 Executed performance and load testing using LoadRunner, assisting in capacity planning by estimating
required virtual user licenses and identifying system bottlenecks.

o Environment: Python (Pandas, NumPy, Matplotlib, Seaborn), Power BI, Tableau, Excel (Pivot
Tables, Power Query), JIRA, Git/GitHub, Windows 10/11, SQL Server, MySQL, PostgreSQL, Shell
Scripting, Agile/Scrum, HIPAA-compliant data platforms

Education details: -
 Bachelor of Technology Computer Science and Engineering Jawaharlal Nehru Technological University in
2014.

You might also like