profile-pic
Vetted Talent

Mohan Vankudoth

Vetted Talent

9 Years of experience in Data mining, Cleaning, Validation, Modelling, Analysis, Data Ware housing, Visualization, Statistical modelling with large data sets of Structured and Unstructured data with technologies Power BI, Tableau, Azure SQL, Azure Analysis service, SSAS, Dataflow, DataMart, Datahub, Microsoft Fabric.

  • Role

    Sr Power BI Consultant & Fabric Developer

  • Years of Experience

    9 years

Skillsets

  • Matplotlib
  • Plotly
  • Performance analyzer
  • Oracle
  • OLTP
  • OLAP
  • Object level security
  • Normalization techniques
  • MySQL
  • MS SQL
  • Mongo DB
  • Microsoft Fabric
  • MDX
  • Power bi mobile
  • JSON
  • Implementation
  • Git
  • Development
  • DAX Studio
  • Datamart
  • Datahub
  • Database diagrams
  • Database Design
  • Data modelling tools
  • Data Modelling
  • SSAS
  • PowerBI - 9 Years
  • Power bi deployment pipeline
  • Report page level
  • Workspace level
  • Tenant level
  • Vertipaq analyzer
  • validation
  • text
  • Teradata
  • support
  • Statistical Modelling
  • Ssas tabular models
  • D3
  • Security
  • Sap bw hana
  • Salesforce
  • Row Level Security
  • Report level
  • R
  • QlikView
  • Python - 3 Years
  • Power BI Report Server
  • Power bi pro
  • Power BI Premium
  • Statistical Modelling
  • SQL - 9 Years
  • Github - 3 Years
  • Git - 3 Years
  • Excel - 4 Years
  • Databricks - 2 Years
  • CI/CD - 3 Years
  • Azure Synapse
  • Azure SQL Server
  • azure pipelines - 3 Years
  • Azure Data Lake - 2 Years
  • Dashboard - 9 Years
  • MDX - 5 Years
  • Tableau - 4 Years
  • Visualization - 9 Years
  • Modelling - 9 Years
  • Data Warehousing - 5 Years
  • Data Mining - 3 Years
  • Azure Analysis Services - 6 Years
  • SSRS - 5 Years
  • Datamart - 3 Years
  • DAX - 9.3 Years
  • Power BI Service - 9 Years
  • Dataflow - 3 Years
  • PowerBI - 9.3 Years
  • QlikView - 4 Years
  • CSV
  • Crystal Reports
  • Cleaning
  • CI/CD
  • Bitbucket
  • Azure sql
  • Azure Logic Apps
  • Azure analysis service
  • Analysis
  • Cloud - 5 Years
  • Data Analysis - 9 Years
  • Azure Data Factory - 2 Years
  • Plotly - 2 Years
  • D3
  • Microsoft Fabric - 2 Years
  • Power bi mobile - 4 Years
  • Power BI Premium - 9 Years
  • Power bi pro - 9 Years
  • Power BI Report Server - 5 Years
  • SQL Server 2017 - 9 Years
  • Datahub - 2 Years
  • SSAS - 9 Years
  • Azure analysis service - 4 Years

Vetted For

19Skills
  • Roles & Skills
  • Results
  • Details
  • icon-skill_image
    Big Data Engineer with Streaming Experience (Remote)AI Screening
  • 81%
    icon-arrow-down
  • Skills assessed :Spark, CI/CD, Data Architect, Data Visualization, EAI, ETL, Hive, PowerBI, PySpark, Talend, AWS, Hadoop, JavaScript, 組込みLinux, PHP, Problem Solving Attitude, Shell Scripting, SQL, Tableau
  • Score: 73/90

Professional Summary

9Years
  • Jul, 2023 - Present2 yr 2 months

    Senior Power BI Consultant

    Y&L Consultancy - US Contract
  • Oct, 2021 - Jul, 20231 yr 8 months

    Sr. Power BI Data Analyst

    Indium Software Pvt Ltd
  • Jun, 2020 - Oct, 20211 yr 4 months

    Sr. Power BI Developer

    Crafsol Technology Solutions Pvt Ltd
  • Sep, 2015 - Nov, 20194 yr 2 months

    BI Developer

    KPIT Technologies Ltd
  • Nov, 2019 - May, 2020 5 months

    BI Developer

    Cap Gemini Software Pvt Ltd

Applications & Tools Known

  • icon-tool

    Microsoft Power BI

  • icon-tool

    Power BI Service

  • icon-tool

    DAX

  • icon-tool

    SSAS

  • icon-tool

    SSIS

  • icon-tool

    Dataflow

  • icon-tool

    DataMart

  • icon-tool

    Azure Data Factory

  • icon-tool

    Azure Data Lake Storage Gen2 (ADLS)

  • icon-tool

    T-SQL

  • icon-tool

    Microsoft Azure SQL Database

  • icon-tool

    MySQL

  • icon-tool

    SQL Server Reporting Services

  • icon-tool

    Python

  • icon-tool

    R Programming Language

  • icon-tool

    Azure Analysis Services

  • icon-tool

    MDX

  • icon-tool

    Git

  • icon-tool

    Azure DevOps Server

  • icon-tool

    Data Warehousing

  • icon-tool

    Data Analysis

  • icon-tool

    Data Modelling

  • icon-tool

    Power BI

  • icon-tool

    SQL Server

  • icon-tool

    Hadoop

  • icon-tool

    Hive

  • icon-tool

    Sqoop

  • icon-tool

    PowerShell

  • icon-tool

    Power BI Mobile

  • icon-tool

    Tableau

  • icon-tool

    QlikView

  • icon-tool

    Power BI Desktop

  • icon-tool

    SSAS

  • icon-tool

    Tableau

  • icon-tool

    Azure Data Lake

  • icon-tool

    Azure Synapse

  • icon-tool

    DAX

  • icon-tool

    SQL Server

  • icon-tool

    My SQL

  • icon-tool

    Hive

  • icon-tool

    Oracle

  • icon-tool

    MongoDB

  • icon-tool

    Teradata

  • icon-tool

    Salesforce

  • icon-tool

    SSIS

Work History

9Years

Senior Power BI Consultant

Y&L Consultancy - US Contract
Jul, 2023 - Present2 yr 2 months
    Implementation and delivery of Power BI platform solutions to develop and deploy analytical, reporting and using Power BI, Power BI Report Builder, Paginated Reports. Utilized Power Query in Power BI to Pivot and Un-pivot the data model for data cleansing and data massaging. Version controlling by integrating with GitHub, SharePoint and deploying the reports with the help of Deployment pipeline. Generated ad-hoc reports, sub-reports, drill-down reports, drill-through reports and parameterized reports to provide visible data for data analysts and business. Built Dataflow, DataMarts to reduce redundant data modelling and transformations. Used DAX Studio, Vertipaq Analyzer, Performance Analyzer to optimize power bi reports visuals rendering, DAX query performance and modelling optimization. Implemented Securities such as Tenant level, Workspace, Reports, Report Pages, Object level Security, Row Level Security. Analysed and mapped Qlik Sense data models and transformations to Power BI, ensuring data integrity and seamless transition. Developed equivalent visualizations and dashboards in Power BI, optimizing for improved performance and user experience. Collaborated with cross-functional teams to validate data accuracy and business logic during migration, reducing downtime and ensuring reporting consistency. Built Power BI App to share the reports with outside tenant users. Worked on huge semantic models which are more than 5+GB in size and optimized for the better performance. Built daily, weekly, monthly graphs and charts of loan data using R and Stored Procedures. Involved in designing, developing and testing of the ETL strategy to populate the data from various source systems. Developed high level daily summary reports (Flash reports), and monthly business summary reports (Monthly Management Report) and setup different email subscriptions for different business units as needed.

Sr. Power BI Data Analyst

Indium Software Pvt Ltd
Oct, 2021 - Jul, 20231 yr 8 months
    Imported data from SQL Server DB, Azure SQL DB to Power BI to generate reports. Created DAX Queries to generate computed columns in Power BI. Generated computed tables in Power BI by using DAX. Utilized Power Query in Power BI to Pivot and Un-pivot the data model for data cleansing and data massaging. Creating and managing Power Bi Gateway and workspaces, Migrating dashboards from one environment to another. Created power bi reports and deployment them in shared workspaces, along with automating them using both gateway and share point (cloud services). Implemented several DAX functions for various fact calculations for efficient data visualization in Power BI. Utilized Power BI gateway to keep dashboards and reports up to date with on-premises data sources. Experience in working with Power BI Admin Portal, and azure active directory in Azure portal. Good Knowledge with DAX to create measures, calculated columns, transform various data sets into enterprise data models and application specific solutions. Designed for Power - BI report designs including backgrounds, buttons and logos etc. Creating functions in Power BI report by creating toggles, Navigation panes, drill through filters, dill down reports, data hierarchies and conditional formatting visuals. Delivered on-demand reports in multiple rendering formats with standard and data-driven subscriptions. Led all sprint planning, sprint reviews, sprint retrospectives and daily scrums with the development team resulting in DEV, QA and Prod. Developed Data Mapping, Performance Tuning, and Identifying Bottlenecks of sources, mappings, targets, and sessions.

Sr. Power BI Developer

Crafsol Technology Solutions Pvt Ltd
Jun, 2020 - Oct, 20211 yr 4 months
    Developed Rich looking Analytical Dashboards and Reports in Power BI for Business executives to take better decision Designed and documented the entire Architecture of Power bi Poc AutomatedPowerQuery refresh usingpowershell script and windows task scheduler. Scheduled Automatic refresh and scheduling refresh in power bi service. Expertise with DAX Functions such as Time Intelligence, Filter, Conditional, Parent & Child, Aggregation, Date & Time, Statistical functions. Created Drilldowns, Drill through and Bookmarks in Power BI Reports for easy navigation and storytelling. Created and Managed Parameters in Power BI Desktop for changing the source with different environments such as development, test and production. Created Dynamic Dropdown measures in slicer with the help of DAX in Power BI Desktop. Developed dynamic Measures, Calculated Columns, Conditional Columns and Custom Columns in Power BI. Created and developed Dynamic and Static Row-Level-Security in Power BI Desktop according to the roles assigned to the users Installed and Configured Power BI gateway (Personal/Enterprise) to schedule daily jobs to refresh reports/dashboards in Power BI Service. Wrote SQL Queries in Teradata & Azure SQL Data warehouse to test the data and verify the quality of data. Designed Data Models in Power Bi Designer by using schemas such as STAR, SNOWFLAKE and GALAXY. Developed and deployed reports for the internal users in Power BI Report Server. Worked with both the modes of connection in power bi such as Direct/Import Query Modes.

BI Developer

Cap Gemini Software Pvt Ltd
Nov, 2019 - May, 2020 5 months
    Developed PowerBImodel used for financial reportings. Expertise in writing complex DAX & LOD functions inPower BI and Power Pivot & Tableau. AutomatedPowerQuery refresh usingpowershell script and windows task scheduler. Used various sources to pull data intoPower BI such as Sql Server, Oracle, SQL Azure etc. Installed and configured Enterprise gateway and Personal gateway in Power bi service. Created Workspace and content packs for business users to view the developed reports. Scheduled Automatic refresh and scheduling refresh in power bi service. Wrote calculated columns, Measures queries in power bi desktop to show good data analysis techniques. Worked on all kind of reports such as Yearly, Quarterly, Monthly, and Daily. Extensively usedJoins and Sub-Queriesto simplify complex queries involving multiple tables. Developed tabular queries for efficient analysis of report usingPivot/Un pivotinT-SQL. Installed SQL 2017 SQL server Database engine, SSIS & SSRS features in the Development environment as needed. Performed DB ADMIN activities on the Server by taking Back-up & restoring the DB, providing user access privileges, creating linked servers. Experience in creating and managing SSAS Tabular models, creating Dimension and Fact Tables Worked in Creating DAX Expressions and implement Partitions in Tabular models Created shared dimension tables, measures, hierarchies, levels, cubes, and aggregations on MS OLAP/ OLTP/Analysis Server (SSAS) in Tabular Model.

BI Developer

KPIT Technologies Ltd
Sep, 2015 - Nov, 20194 yr 2 months
    Over 4.3 years of experience as a Data Engineer, Data Transforming, Data Modelling, Data Exploring, Statistical Analysis on Data and Software Engineer using Python. Experienced in reading and writing different data sources, Reshaping, Pivoting, slicing, Fancy Indexing, and sub setting of data sets, Grouping, Merging, Joining, data Filtration, Handling of missing data Using Pandas, Created Data Pipelines for keeping transactional data from various Databases (MongoDB, MySQL, MS SQL) and store in Hadoop databases like Hive. Designed, Developed and Implemented Power BI Dashboards, Scorecards & KPI Reports, Power BI Mobile Reports, Power BI Report Server configuration, administration. used for financial reporting of P & L and Headcount. Expertise in writing complex DAX & MDX functions inPower BI and Power Pivot. Worked on all types of transformations that are available in Power bi query editor Experienced in Software Development Life Cycle (Analysis, Design, Development, Testing and maintenance) using agile methodologies.

Testimonial

MGAE

MGAE

Recognizing Exceptional Performance

At MGAE Entertainment, we believe in recognizing and celebrating exceptional performance, and Mohan Vankudoth has exemplified these qualities in every aspect of their work. As Sr Power BI Consultant, Mohan Vankudoth has consistently demonstrated a level of dedication, expertise, and professionalism that has truly set them apart.

Mohan Vankudoth has been instrumental in the success of our Power BI project, playing a pivotal role in its conception, implementation, and ongoing optimization. Their expertise in data analytics, coupled with their strategic mindset, has been invaluable in unlocking insights and driving impactful outcomes for our organization.

What truly sets Mohan Vankudoth apart is their exceptional ability to translate complex data into actionable insights that inform decision-making at all levels of the organization. Whether it's analyzing sales performance, identifying growth opportunities, or optimizing inventory management, Mohan Vankudoth approaches every challenge with a meticulous attention to detail and a commitment to excellence.

Moreover, Mohan Vankudoth has been a catalyst for driving innovation and fostering a culture of continuous improvement within our team. Their proactive approach, willingness to explore new ideas, and collaborative spirit have inspired their colleagues to raise the bar and push the boundaries of what's possible.

Beyond their technical expertise, Mohan Vankudoth embodies the core values of MGAE Entertainment creativity, passion, and integrity. Their unwavering dedication to delivering exceptional results, coupled with their humility and team-first attitude, make them a true asset to our organization.

In recognition of Mohan Vankudoth's exceptional performance and contributions, we extend our sincerest gratitude and appreciation. Their leadership, dedication, and commitment to excellence serve as an inspiration to us all, and we look forward to continuing our journey of success together.

Chris Martin

Business Intelligence Manager

MGA Entertainment, Chatsworth

Major Projects

9Projects

MGAE - Power BI Modernization

MGAE Entertainment
Jul, 2023 - Present2 yr 3 months

    MGAE Entertainment aims to leverage Power BI to gain actionable insights into its sales data across regions, product lines, and distribution channels. The project's primary objectives are to enhance decision-making, identify growth opportunities, optimize inventory management, and ultimately increase profitability.

    Key Components:

    1. Data Integration: Integrate sales data from various sources including POS systems, online sales platforms, and distribution partners. Cleanse and transform the data to ensure consistency and accuracy.
    2. Dashboard Development: Design interactive dashboards and reports to visualize key sales metrics such as revenue, units sold, average order value, and sales growth over time. Incorporate filters and slicers for dynamic exploration of data.
    3. Regional Analysis: Analyze sales performance across different regions, countries, and markets. Identify top-performing regions, emerging markets, and areas for improvement. Compare sales trends and seasonality patterns.
    4. Product Line Analysis: Segment sales data by product lines including LOL Surprise!, Bratz, Little Tikes, and others. Evaluate the performance of each product line in terms of revenue contribution, popularity, and market share.
    5. Channel Analysis: Evaluate sales performance across various distribution channels such as retail stores, e-commerce platforms, and wholesale partners. Identify the most profitable channels and optimize channel strategies accordingly.
    6. Inventory Management: Analyze inventory levels, stock movement, and product demand to optimize inventory management processes. Identify slow-moving inventory, stockouts, and excess inventory to minimize carrying costs and improve supply chain efficiency.
    7. Customer Segmentation: Segment customers based on purchasing behavior, demographics, and preferences. Identify high-value customers, repeat buyers, and customer acquisition opportunities. Personalize marketing strategies to enhance customer engagement and loyalty.

Implementation and delivery of Power BI platform solutions to develop and deploy analytical, reporting and using Power BI, Power BI Report Builder, Paginated Reports

Jul, 2023 - Nov, 20241 yr 3 months
    Utilized Power Query in Power BI to Pivot and Un-pivot the data model for data cleansing and data massaging. Version controlling by integrating with GitHub, SharePoint and deploying the reports with the help of Deployment pipeline. Generated ad-hoc reports, sub-reports, drill-down reports, drill-through reports and parameterized reports to provide visible data for data analysts and business. Built Dataflow, DataMarts to reduce redundant data modelling and transformations. Used DAX Studio, Vertipaq Analyzer, Performance Analyzer to optimize power bi reports visuals rendering, DAX query performance and modelling optimization. Implemented Securities such as Tenant level, Workspace, Reports, Report Pages, Object level Security, Row Level Security. Analyzed and mapped Qlik Sense data models and transformations to Power BI, ensuring data integrity and seamless transition. Developed equivalent visualizations and dashboards in Power BI, optimizing for improved performance and user experience. Collaborated with cross-functional teams to validate data accuracy and business logic during migration, reducing downtime and ensuring reporting consistency. Built Power BI App to share the reports with outside tenant users. Worked on huge semantic models which are more than 5+GB in size and optimized for the better performance. Built daily, weekly, monthly graphs and charts of loan data using R and Stored Procedures. Involved in designing, developing and testing of the ETL strategy to populate the data from various source systems. Developed high level daily summary reports (Flash reports), and monthly business summary reports (Monthly Management Report) and setup different email subscriptions for different business units as needed.

Imported data from SQL Server DB, Azure SQL DB to Power BI to generate reports

Oct, 2021 - Jul, 20231 yr 8 months
    Created DAX Queries to generate computed columns in Power BI. Generated computed tables in Power BI by using DAX. Utilized Power Query in Power BI to Pivot and Un-pivot the data model for data cleansing and data massaging. Creating and managing Power Bi Gateway and workspaces, Migrating dashboards from one environment to another. Created power bi reports and deployment them in shared workspaces, along with automating them using both gateway and share point (cloud services). Implemented several DAX functions for various fact calculations for efficient data visualization in Power BI. Utilized Power BI gateway to keep dashboards and reports up to date with on-premises data sources. Experience in working with Power BI Admin Portal, and azure active directory in Azure portal. Good Knowledge with DAX to create measures, calculated columns, transform various data sets into enterprise data models and application specific solutions. Designed for Power - BI report designs including backgrounds, buttons and logos etc. Creating functions in Power BI report by creating toggles, Navigation panes, drill through filters, dill down reports, data hierarchies and conditional formatting visuals. Delivered on-demand reports in multiple rendering formats with standard and data-driven subscriptions. Led all sprint planning, sprint reviews, sprint retrospectives and daily scrums with the development team resulting in DEV, QA and Prod. Developed Data Mapping, Performance Tuning, and Identifying Bottlenecks of sources, mappings, targets, and sessions.

LAM Research - BI Enhancements

LAM Research Pvt Ltd
Jan, 2023 - Jul, 2023 6 months

    Project Overview:

    LAM Research, a leader in semiconductor manufacturing equipment, is embarking on a Power BI project to optimize its supply chain operations. The project aims to integrate data from various sources across the supply chain, analyze key performance indicators, and visualize insights to improve decision-making and efficiency.

    Key Components:

    1. Data Integration: Integrate supply chain data from ERP systems, manufacturing databases, logistics platforms, and supplier databases. Clean, transform, and consolidate data to ensure consistency and accuracy.
    2. Dashboard Development: Develop interactive dashboards and reports to visualize critical supply chain metrics such as inventory levels, lead times, order fulfillment rates, and supplier performance. Incorporate drill-down capabilities for deeper analysis.
    3. Inventory Management: Analyze inventory levels across warehouses, production facilities, and distribution centers. Optimize inventory levels to minimize stockouts, reduce carrying costs, and improve working capital efficiency.
    4. Supplier Performance: Evaluate supplier performance based on metrics such as on-time delivery, quality, lead times, and cost. Identify top-performing suppliers and areas for improvement. Establish supplier scorecards to track performance over time.
    5. Demand Forecasting: Utilize historical sales data, market trends, and demand forecasts to predict future demand for components and materials. Optimize procurement and production planning to meet demand while minimizing excess inventory and stockouts.
    6. Production Planning: Analyze production schedules, capacity utilization, and throughput to optimize production efficiency. Identify bottlenecks, schedule optimizations, and resource allocation strategies to maximize output and minimize lead times.
    7. Logistics Optimization: Optimize transportation routes, modes of transportation, and shipment consolidation to reduce freight costs, transit times, and carbon footprint. Utilize real-time tracking and analytics to monitor shipments and improve delivery performance.

Inteva - QlikView, Tableau, Cognos Migration to Power BI

Inteva Products
May, 2022 - Jan, 2023 8 months

    Project Overview:

    Inteva Products, a global automotive supplier, is initiating a project to migrate its existing business intelligence (BI) platforms from QlikView, Tableau, and Cognos to Microsoft Power BI. The migration aims to consolidate BI tools, standardize reporting processes, and leverage Power BI's advanced analytics capabilities to drive better decision-making across the organization.

    Key Components:

    1. Assessment and Planning: Conduct a comprehensive assessment of existing BI infrastructure, data models, reports, and user requirements for QlikView, Tableau, and Cognos. Develop a migration strategy and roadmap outlining timelines, resources, and deliverables.
    2. Data Migration: Extract data models, datasets, and reports from QlikView, Tableau, and Cognos environments. Transform and migrate data to Power BI-compatible formats while ensuring data integrity, consistency, and security.
    3. Dashboard and Report Conversion: Convert existing dashboards, reports, and visualizations from QlikView, Tableau, and Cognos to Power BI. Redesign and optimize reports for Power BI's capabilities, including interactive visuals, drill-down functionality, and mobile responsiveness.
    4. Integration with Data Sources: Integrate Power BI with existing data sources, including enterprise data warehouses, operational databases, cloud-based applications, and external data sources. Ensure seamless connectivity and real-time data refresh capabilities.
    5. User Training and Adoption: Provide training sessions and workshops for end-users to familiarize them with Power BI's interface, features, and functionalities. Promote user adoption through hands-on demonstrations, best practices, and knowledge sharing sessions.
    6. Governance and Security: Implement governance policies, access controls, and data security measures to ensure compliance with regulatory requirements and protect sensitive information. Define roles and responsibilities for BI development, administration, and user access management.
    7. Performance Optimization: Optimize Power BI performance by fine-tuning data models, optimizing DAX queries, and leveraging caching and incremental data refresh techniques. Monitor performance metrics and address any bottlenecks or issues proactively.

PepsiCo Supply Chain Optimization through Power BI

PepsiCo, Inc - Food Company
May, 2021 - May, 20221 yr

    Project Overview:

    PepsiCo is embarking on a Power BI project aimed at enhancing its supply chain management processes. The project focuses on leveraging data analytics to optimize inventory management, streamline production, and improve distribution efficiency.

    Key Components:

    1. Data Integration: Integrate data from various sources including ERP systems, manufacturing databases, warehouse management systems, and transportation logs. Ensure data quality and consistency through cleansing and normalization.
    2. Inventory Analysis: Analyze inventory levels across warehouses, distribution centers, and retail outlets. Identify trends, seasonality patterns, and stock-out risks to optimize inventory levels and improve product availability.
    3. Production Efficiency: Analyze production schedules, equipment utilization, and downtime to identify bottlenecks and opportunities for efficiency improvements. Optimize production planning and scheduling to meet demand while minimizing costs.
    4. Supplier Performance: Evaluate supplier performance based on metrics such as on-time delivery, quality, and cost. Identify top-performing suppliers and areas for improvement to enhance supply chain reliability and resilience.
    5. Demand Forecasting: Utilize historical sales data, market trends, and demand forecasts to predict future demand for PepsiCo products. Improve demand forecasting accuracy to reduce stockouts, minimize excess inventory, and optimize production planning.
    6. Logistics Optimization: Analyze transportation routes, modes of transportation, and delivery schedules to optimize logistics operations. Reduce transportation costs, improve delivery accuracy, and minimize transit times to enhance customer satisfaction.
    7. Visualization and Reporting: Develop interactive dashboards and reports in Power BI to visualize key supply chain metrics such as inventory turnover, fill rates, on-time delivery performance, and transportation costs. Enable stakeholders to drill down into data for deeper insights and analysis.

Developed Rich looking Analytical Dashboards and Reports in Power BI for Business executives to take better decision

Jun, 2020 - Oct, 20211 yr 4 months
    Designed and documented the entire Architecture of Power bi Poc AutomatedPowerQuery refresh usingpowershell script and windows task scheduler. Scheduled Automatic refresh and scheduling refresh in power bi service. Expertise with DAX Functions such as Time Intelligence, Filter, Conditional, Parent & Child, Aggregation, Date & Time, Statistical functions. Created Drilldowns, Drill through and Bookmarks in Power BI Reports for easy navigation and storytelling. Created and Managed Parameters in Power BI Desktop for changing the source with different environments such as development, test and production. Created Dynamic Dropdown measures in slicer with the help of DAX in Power BI Desktop. Developed dynamic Measures, Calculated Columns, Conditional Columns and Custom Columns in Power BI. Created and developed Dynamic and Static Row-Level-Security in Power BI Desktop according to the roles assigned to the users Installed and Configured Power BI gateway (Personal/Enterprise) to schedule daily jobs to refresh reports/dashboards in Power BI Service. Wrote SQL Queries in Teradata & Azure SQL Data warehouse to test the data and verify the quality of data. Designed Data Models in Power Bi Designer by using schemas such as STAR, SNOWFLAKE and GALAXY. Developed and deployed reports for the internal users in Power BI Report Server. Worked with both the modes of connection in power bi such as Direct/Import Query Modes.

Developed PowerBImodel used for financial reportings

Nov, 2019 - May, 2020 5 months
    Expertise in writing complex DAX & LOD functions inPower BI and Power Pivot & Tableau. AutomatedPowerQuery refresh usingpowershell script and windows task scheduler. Used various sources to pull data intoPower BI such as Sql Server, Oracle, SQL Azure etc. Installed and configured Enterprise gateway and Personal gateway in Power bi service. Created Workspace and content packs for business users to view the developed reports. Scheduled Automatic refresh and scheduling refresh in power bi service. Wrote calculated columns, Measures queries in power bi desktop to show good data analysis techniques. Worked on all kind of reports such as Yearly, Quarterly, Monthly, and Daily. Extensively usedJoins and Sub-Queriesto simplify complex queries involving multiple tables. Developed tabular queries for efficient analysis of report usingPivot/Un pivotinT-SQL. Installed SQL 2017 SQL server Database engine, SSIS & SSRS features in the Development environment as needed. Performed DB ADMIN activities on the Server by taking Back-up & restoring the DB, providing user access privileges, creating linked servers. Experience in creating and managing SSAS Tabular models, creating Dimension and Fact Tables Worked in Creating DAX Expressions and implement Partitions in Tabular models Created shared dimension tables, measures, hierarchies, levels, cubes, and aggregations on MS OLAP/ OLTP/Analysis Server (SSAS) in Tabular Model.

Designed, Developed and Implemented Power BI Dashboards, Scorecards & KPI Reports, Power BI Mobile Reports, Power BI Report Server configuration, administration

Sep, 2015 - Nov, 20194 yr 2 months
    used for financial reporting of P & L and Headcount. Expertise in writing complex DAX & MDX functions inPower BI and Power Pivot. Worked on all types of transformations that are available in Power bi query editor Experienced in Software Development Life Cycle (Analysis, Design, Development, Testing and maintenance) using agile methodologies.

Education

  • Pre-University course

    Rajiv Gandhi University of Knowledge and Technologies (2011)
  • B-Tech in CSE

    Rajiv Gandhi University of Knowledge and Technologies (2015)
  • SSC

    APRS Nagarjuna Sagar (2009)
  • Pre-University

    Rajiv Gandhi University of Knowledge and Technologies (2011)

Certifications

  • Power BI

    Microsoft (Jan, 2023)

Interests

  • Badminton
  • Watching Movies
  • Dance
  • AI-interview Questions & Answers

    Hi. Myself, Mohan, and I'm from Hyderabad. I have overall 8.6 years experience as a COBRA consultant. I've been to 4 organizations in my, uh, 8.6 years of experience with a different different domain knowledge. I've been working on different different skills like Power BI, Power BI report server, Power BI desktop, and then writing task queries, writing SQL queries, MSSQL, MySQL, Oracle, and then connecting with different different data sources like the SharePoint, Azure SQL, Azure Analysis Services. And I do have to have a understanding on Azure environment like Azure Data Factory, Azure, uh, data pipelines, and then Microsoft Fabric, Synapse. And in this in this, uh, 4 point in in this, uh, 8.6 years of experience, I I've been to different different domains like health care, automobiles, and then incidents, retail, banking. These different different domains that I knowledge that I would have. Currently, I'm not working. I'm looking for an opportunity where I can join immediately. My previous submission was, uh, my internal consultancy with the client of Farm America, where in this project, my roles and responsibilities are connected with the clients, understand their requirements, their own data, and sit with the data engineering team to get the data from one source to another source to have a reports implementation to do the analysis in terms of a Power BI desktop. Here, we we did the migration of QuickView, Tableau, and QuickSens published reports migration from to Power BI Desktop by understanding their data modeling, data transformations, the features that they have implemented, everything from end to end.

    How would you integrate a Python based machine learning model with the big data pipeline on AWS? How would you integrate a Python based machine learning model with a big data pipeline on AWS? Prepare the machine learning model. That is the first thing where develop and train your machine learning model using Python libraries such as scikit learn, TensorFlow, or PyTorch. Serialize and save the trained model to a file using a format such as pickle or joblib. The second step is set up an AWS big data pipeline. Choose AWS services for building your big data pipeline such as Amazon S3 for storage, Amazon EMR for data processing, and AWS Glue for ETL, which is extract, transform, and load. Design and implement the data workflow for data ingestion processing and storage according to your requirements. Display, deploy the machine learning model. That's the that's the third part where set up an environment on AWS where you can deploy your machine learning model. This could be an EC 2 instance, ew, AWS Lambda, or an AWS SageMaker endpoint. Depending on your requirement for scalability and latency, install the necessary dependencies and libraries for running the machine learning model. The 4th part would be integration. Implement a mechanic mechanism to trigger AWS step functions, or custom scripts. Then define the input and output data formats for machine learning model. Ensure that the data processed by your big data pipeline can be efficiently fed into the model for interface. Retrieve the input from your big data pipeline. Preprocess if it is necessary and pass it to deployed machine learning model for prediction interference. Capture the output from the model and store it in suitable locations such as s three bucket or a database as a part of your big data pipeline output. 5th part would be testing and monitoring. Test the end to end integration of the machine learning model with your big data pipeline to ensure behaves as expected. Implement monitoring Implement monitoring and logging to track the performance of the model interference process and identify. Deploying and scaling. Deploy the integrated solution into your production environment and configure auto scaling mechanisms if it is necessary to handle variations in workload and data volume, monitor the performance and scalability of the integrated system, and make adjustments as needed to optimize resource utilization and maintain readability or reliability.

    What would be your approach to design a resilient stream processing system using Kafka and Spark on AWS? What would be your approach to design a resilient stream processing system using Kafka and Spark on AWS? What would be your approach to design a resilient stream processing system using Kafka and Spark on AWS? So the first thing would be architecture design, define the overall architecture of your stream processing systems, including the components, data flow, and interactions between different services. Choose AWS services such as Amazon, Kinesis, Amazon EMR, and Amazon EC2 to build the core components of your system. Second thing would be data ingestion, set up data processes to ingest streaming data into Kafka topics. This could involve using AWS Lambda functions, Kinesis data streams, or custom applications. Configure Kafka Connect to stream data from Kafka topics to Spark for processing. Stream processing with Spark, that would be third step, where deploy a Spark cluster on Amazon EMR to process the streaming data from Kafka topics, use Spark streaming or structured streaming APIs to read data from Kafka topics, perform real-time processing, and write the results to downstream systems for storage. Fault tolerance and resilience, implement fault tolerance mechanisms in your Spark streaming application to handle failures gracefully. This could involve checkpointing right-hand logs and handling transient errors. Configure Spark to automatically recover from failures, restart processing from the last checkpoint. Monitor and alerting, that would be the fifth thing, and scaling and performance of optimization and data processing and storage, choose appropriate storage solutions for storing both raw and processed streaming data. This could include Amazon S3, Amazon DynamoDB, or Amazon Redshift. And then security and access control, implement security measures such as encryption, authentication, and authorization to protect sensitive data, and ensure compliance with security standards.

    How would you handle schema evaluation in I tables being populated by ongoing UTM jobs? How to handle schema evaluation in 5 tables being populated by ongoing ETL jobs. Okay? So the first step would be the first approach is schema on grid. Design your data pipeline to follow a schema on grid approach where data is ingested into high tables without enforcing a strict schema upfront. This allows flexibility in handling changes to the data data schema over a time. Then partitioning. Partitioning your, uh, high tables based on time or other relevant attributes to segregate data and facilitate efficient query. When schema changes occurs, you can create new partitions with the updated schema without affecting existing partitions. External tables is a third approach where Using external tables, use, uh, this, uh, big how to decouple the storage data storage from the schema definition. This allows you to alter the schema without modifying the underlying data files. Schema evaluation support. Leverage HIVE's built in support for schema evolution, which allows you to add new columns to existing tables or modify column data types without requiring a table effect. This feature helps minimizing downtime and disruption to ongoing ATM jobs. AGRO or PORCAT formats. Store data in hard tables using avro or parquet file formats which supports schema evaluation by design. This file format store schema information along with the data, making it easier to handle changes schema changes without impacting existing data. Versioning. Implement versioning mechanisms of your hand tables to track changes to the schema over time. This allows you to maintain a history of schema changes and roll back to previous versions if needed. ETL jobs, versioning, and then data validation, and then we can do the, uh, communication and coordination between the team members with data engineers and data scientists and other stakeholders involved in the schema evolution.

    How would you design a talent job to not only process data transformations, but also handle error logging and recover gracefully? How would you design a talent job to not only process data transformations, but also handle error logging and recover gracefully? So the first thing is data transformations. First use talent's built-in components like TMAP, T-join, T-filter row, etc. to perform the necessary data transformations. Ensure the data quality by using T-data quality components to validate and standardize the data. And then error logging. So implement a T-log catcher to capture any Java exceptions or talent errors. Use T-stat catcher to gather statistics about job performance and T-flow meter for real-time monitoring. Direct error logs to a file or database using a T-file output delimiter or TDB output. Then graceful recovery can be done using a utilizer T-die and T-wall components to manage error handling and warnings. Set up checkpoints using T-checkpoint to save the state of a job at specific points. And then test use T-run job to modularize your job and allow for easier reruns of a specific sub-jobs in case of failure. So job orchestration would be like employ T-pre-job and T-post-job for initializing resources and cleaner tasks. Use T-context variables and T-context load for dynamic job configuration. Transaction management for database operations use T-begin, T-commit and T-rollback to manage transactions and ensure data integrity. Retry mechanism we can with T-loop or T-wait for a file to handle transient issues.

    What strategy would you apply for ETL testing to ensure data integrity across different storage systems like S3 and HDFS? Data validation performs source-to-target count checks to ensure that the number of records loaded into the target system matches the source. Use data profiling to understand the data and identify anomalies or patterns that need attention. Transformation rules testing validates that business rules are correctly applied during the transformation process. Data type consistency and format correctness to ensure data conforms to the target schema requirements. End-to-end testing conducts that recover the ETL entire process for extracting data from source to loading into the target source system. Automated regression testing to quickly identify issues after changes into the ETL process. Error handling and recovery. Error handling mechanisms to ensure they capture and log errors accurately. Recover processes to confirm that the system can recover from failures without data loss or corruption. Performance tuning of the ETL process to ensure it can handle expected data volumes within the required timeframes. Data integrity checksums or hash totals to verify that the data has not been altered during the ETL process. And comparative analysis between the source and target systems to ensure consistency. Source and compliance testing with data security measures are effective throughout the ETL process. Compliance with relative data protection regulations and standards check value. Testing automation. Utilize ETL testing tools to automate such as for the testing process of possible reducing and manual effort and the potential of illuminate.

    In a talent job designed to process large datasets using a tmap component. You observe that the job is failing due to an out of memory error. What could be the cause of this error, and how would you debug this issue to identify the root cause and propose a solution. First thing is insufficient heat space. That is the first, which is, uh, the Java virtual machine might not have enough free histories allocated to, which is necessary for processing large volumes of data. And the data volume might process to exceed the volume available memory, especially if the team app is performing complex transformations or lookups, insufficient data processing where the job design might not be optimized for memory usage, leading to insufficient processing and memory overflow. To debug this issue, identify the root cause. You can increase the JVM memory, adjust the xmx parameter to increase the, uh, mix maximum chip size available to the job. Then optimize TMAP settings were enabled, store temp data to disk option in TMAP to offload some of the data processing to disk, reducing memory consumption, Review job design and then use ETL components such as pushing processing down to the database level and reducing the memory load to talent. Monitor memory usage to recover to observe memory usage during job execution to identify when and where the memory issue occurs. Test with sample data, and then increase the JVM. So the solution, uh, to address the out of memory issue is optimized with a team app, job refractory, and then database side filtering where apply filters directly in the database query to reduce the amount of data retrieved and processing entire. Streaming options would be like if using a database input components enable streaming options to process a row 1 by row instead of loading all data into memory.

    In a talent big data job, the developer has used a file input delimited component to read a CSC file and a THDFS output component to write the processed data to HDFS. However, the job is producing incomplete files on HDFS without any errors. What can you infer from this, and how would you debug and resolve this issue? The first thing would be buffer data not flush, which is data might still be in the buffer and not flush to HDFS when the job ends and then incorrect file handling. The job might not be handling file, uh, file writing correctly, especially if it's being stopped or caused unexpectedly. And then configuration issues, like there could be a misconfiguration in the connection setup with HDFS, such as incorrect replication factor or block size. To debug and resolve this issue, check job configurations, verify the HDFS collection is correctly configured with the right host port and user credentials, review component settings like the THVFS output component, ensure that the merge result to single file option is not causing conflicts, Examine job logs and then validate data flow using a t log rule, and then test with a smaller dataset and ensure proper job termination. Make sure that the job is allowed to finish completely and isn't terminated prematurely. Use t HDFS put component. If you're using a talent for big data, consider using t HDFS put component as a alternative to write to HDFS. Check h HDFS and ensure that HDFS is running correctly and then that there are no underlying issues with the h and Hadoop cluster monitoring Hadoop HDFS during job executions directly HDFS directly while the job is running to check if the files are being written and updated correctly or not.

    Design a method to orchestrate the deployment of a new version of an ETL pipeline, minimizing downtime and data inconsistency. First thing is, uh, you can follow the, you know, steps such as, like, uh, version control. Use a version control system to manage changes to the ETL pipeline too, and then testing environment, set up a separate testing environment that mirrors production to test the new version. Then continuous integration, continuous deployment, which is CICD, implements CICD practices to automate the testing and deployment process. Bluegreen Deployment, which uses the bluegreen deployment strategy to switch between 2 identical production environments. Feature toggles. Implement feature toggles to enable or disable features without deploying a new port. Data migration scripts. Pip prepare data migration scripts to handle changes in data structures or schemas. Monitoring and alerts. Set up monitoring and alerts to quickly identify issues during deploy deployment and have a rollback plan in case the new version, uh, introduces unexpected issues. Perform data validation checks to ensure consistency between the old and new versions. Gradually roll out the new version to a subset of users before a full deployment. Update documentation to reflect changes in the new version, and then train the t team on new features that changes the changes in the pipeline, and then utilize orchestration tools like Apache Airflow and, uh, Azure Data Factory to manage the workflow.

    How would you integrate ETL processes with S3 event notifications for real-time alerting on data ingestion issues? So first thing is, there are several steps like configure S3 event notifications where set up Amazon S3 to publish events such as S3 object created star to capture when new data is ingested and then AWS Lambda function which is to create an AWS Lambda function that is triggered by the S3 event notification and this function will process the event and perform initial checks on the ingested data then error detection logic where implement logic within the Lambda function to direct detect any issues with the data such as format errors or incomplete files and then use Amazon simple notification service which is SNS or Amazon simple queue service which is SQS to send alerts if the Lambda function detects any issues integration with ETL tools like tool or platform is capable of receiving these alerts and can respond accordingly either by triggering or a corrective workflow or notifying an administrator and then set up a cloud watch to monitor the ETL process and long way log any events or errors for further analysis and then depending on the ETL tool you are using you may be able to set up automated responses to certain types of errors such as retrying the ingestion or running a cleanup job

    What strategies can you utilize for reducing the time it takes to visualize complex datasets with Power BI? So to reduce the time it takes to visualize complex datasets with Power BI, uh, we can follow the, like, you know, uh, certain strategies like performance analyzer, which is, you know, used to analyze the identify and locate any bottlenecks in your reports. Like, as, you know, it gives us the 3 parameters for each individual where in the tags, visual, and other. So based on that, you know, we can take, um, understanding on, you know, whether it is having the issue with the or at a time of visual rendering because of the data points that we can move. And then optimize the data modeling where, you know, speed up the data model, which is often main source of performance issues. So this includes minimizing the use of calculated column and avoiding complex relationships such as many to many relationships and the many to many relationships with both sides filter directions or cross filter directions in both side. These will cause a lot of issues with the, you know, optimize, uh, with the with the performance of a report. And then simplify visuals where limit the number of visuals on a report page and evaluate the performance of a custom visuals so that, uh, we we do not overburden a particular report page to load too much of a data with the too many visuals, um, based on the filters that we're applying. Handling, uh, data handling techniques such as implement a proper data ingestion and summarizing techniques to avoid loading unnecessary data into memory of a PowerStore. DirectQuery mode is for large datasets considering using a direct query mode for incremental refresh to improve responsiveness and reduce data loads. And then filtering filter or filter and aggregate data early. Like, that means, um, at the time of model creation process to reduce the amount of data stored in the memory, we can, you know, directly use the, uh, power query or SQL queries to, uh, filter and aggregate the data.