Analytics??
- Anand Nerurkar
- Nov 13
- 20 min read
šÆ Why Do We Use Analytics?
In one line:
We use analytics to transform raw data into actionable insights that drive better business decisions, operational efficiency, and intelligent automation.
š§© 1ļøā£ Business Perspective ā Turning Data into Decisions
Analytics helps organizations move from:
Data ā Information ā Insight ā Action ā Outcome
Type | What it Answers | Example in Banking |
Descriptive Analytics | What happened? | Loan default rate last quarter |
Diagnostic Analytics | Why did it happen? | Defaults spiked due to job loss in SME sector |
Predictive Analytics (ML) | What will happen? | Which customers are likely to default next |
Prescriptive Analytics (AI) | What should we do? | Offer restructuring plan to high-risk customers |
Cognitive / GenAI Analytics | How can we automate decisions? | AI assistant summarizes risk reports or drafts emails to clients |
So, analytics is not just about dashboards ā itās about data-driven decision-making at every level.
š§© 2ļøā£ Technology Perspective ā The Foundation for AI/ML
Analytics is the bridge between data and AI.Before you can build an ML or GenAI model, you need:
Clean, curated, feature-enriched data
Exploratory Data Analysis (EDA) to understand patterns
Historical trends and statistical validation
Without analytics, models are blindĀ ā they canāt learn meaningfully or perform accurately.
Example ā Credit Scoring Flow:
Stage | Role of Analytics |
Data Collection | Aggregate income, repayment, employment data |
Data Cleansing | Handle missing values, remove outliers |
Feature Engineering | Create income-to-debt ratio, credit utilization score |
Model Building | Train logistic regression / random forest |
Insights | Which parameters contribute most to risk |
Decision | Approve / reject / manual review loan applications |
So analytics provides the intelligence layerĀ between data engineering and machine learning.
š§© 3ļøā£ Operational Perspective ā Analytics + AIOps
In AIOps, analytics plays a real-time diagnostic and predictiveĀ role:
Function | Analytics Use |
Monitoring | Time-series analysis of logs, CPU, latency |
Anomaly Detection | Statistical or ML models detect deviations |
Root Cause Analysis | Correlation analytics across systems |
Predictive Maintenance | Forecast failures before they happen |
Optimization | Trend analytics for capacity or cost efficiency |
So AIOps uses analytics to convert noisy operational data into meaningful signalsĀ ā enabling automation and proactive reliability.
š§© 4ļøā£ Enterprise Architecture Perspective
As an EA, you ensure analytics is not siloed ā but part of an enterprise data and AI strategy:
EA Layer | Analytics Role |
Business Layer | Enable data-driven KPIs and OKRs |
Information Layer | Curate enterprise data models and lineage |
Application Layer | Integrate BI tools and AI services |
Technology Layer | Leverage scalable data platforms (Azure Synapse, Databricks, Power BI) |
Governance Layer | Define data quality, lineage, access, and ethics standards |
š§
āWe use analytics to derive insights from data that improve business decisions and automate operations. In AI/ML initiatives, analytics forms the backbone for data understanding, feature creation, and model validation. In AIOps, analytics enables proactive IT management through anomaly detection, trend analysis, and predictive maintenance. As an Enterprise Architect, I ensure analytics is governed, integrated, and aligned with business outcomes ā not treated as an isolated reporting activity.ā
š§© Why We Use Data Lakes for Analytics
A data lakeĀ is the central platform that stores all types of enterprise dataĀ ā structured, semi-structured, and unstructured ā at scale.
We use it because analytics and AI/ML need large volumes of clean, contextual data, and a data lake provides that foundation.
Think of it as the āsystem of intelligenceāĀ sitting on top of your system of record (core apps).
š§ Typical Data Lake Zone Architecture
[Source Systems]
āāā Core Banking, CRM, Loan Systems, Web Logs, APIs
āāā ExternalĀ Sources (Credit Bureaus, Social, Market Data)
ā (via ETL / Streaming / Batch Ingestion)
+----------------------------------------------------------+
| DATA LAKE |
+----------------------------------------------------------+
| RAW ZONEĀ Ā Ā Ā Ā Ā Ā Ā | CURATED ZONEĀ Ā Ā Ā Ā | ANALYTICS ZONEĀ Ā |
|------------------|-------------------|-------------------|
| - Unprocessed | - Cleaned, Joined | - Aggregated, |
| raw data | - Feature-rich | modeled data |
| - Landing area | - Conformed model | - Ready forĀ BI, |
| - Audit trail | - Business rules | ML, GenAI |
+----------------------------------------------------------+
āļø 1ļøā£ Raw Zone
Purpose:Ā Data as-is, directly from sourceCharacteristics:
No transformation
Retains original schema for traceability
Acts as āsource of truthā for audits or reprocessing
Example:
Loan application CSV files, transaction logs, or API JSON payloads from partner systems.
āļø 2ļøā£ Curated Zone
Purpose:Ā Data cleansing, standardization, enrichmentCharacteristics:
Cleaned, validated, schema-aligned data
Derived features or metrics added
Often partitioned by business domains (Customer, Account, Loan, etc.)
Example:
Creating income-to-debt ratio, credit utilization score, repayment behavior index
Joining customer data with bureau reports
This is where Feature Engineering TeamsĀ and Data ScientistsĀ work for ML model training.
āļø 3ļøā£ Analytics Zone
Purpose:Ā Data ready for business consumptionCharacteristics:
Optimized for queries and dashboards
Feeds ML, BI, and GenAI layers
May be structured as dimensional models (star/snowflake)
Example:
Loan default trends by region
Customer risk segmentation
Feeds Power BI, Tableau, or model training pipelines
š§© Enterprise Integration Example
[Data Sources]
ā
[Ingestion Layer]
(ADF / Kafka / Stream)
ā
[Data Lake - Raw Zone]
ā
[Data Prep / Enrichment - Curated Zone]
ā
[Analytics Zone]
ā
[BI / AI / ML / LLM]
ā
[Insight Delivery via Dashboards, APIs, Chatbots]
š Link with Analytics & AI
Zone | Used By | Purpose |
Raw | Data Engineers | Data ingestion, lineage, and auditing |
Curated | Data Scientists / ML Engineers | Model training, feature creation |
Analytics | Business Analysts / BI Teams | Dashboards, KPI monitoring, AI insights |
So yes ā analytics (and even AI/ML/GenAI) pipelines always depend on this multi-zone architectureĀ in enterprise-grade data platforms (Azure Data Lake, AWS S3 Lakehouse, GCP BigLake, etc.).
š£ļø
āYes, we use a multi-zone data lake ā Raw, Curated, and Analytics ā as the foundation for all analytics and AI/ML initiatives. Raw zone captures data as-is, curated zone enriches and standardizes it for model training, and analytics zone exposes it for business intelligence and AI use cases. This layered approach ensures data lineage, quality, and governance while enabling predictive and generative AI capabilities downstream.ā
Letās break it down practically ā starting from raw data in the data lake, and showing how analytics transforms it into business decisionsĀ step by step š
š§© 1ļøā£ Raw Data ā The Foundation
Raw data is the unprocessed feedĀ directly coming from multiple systems:
Core Banking (accounts, transactions, loans)
CRM (customer interactions)
Channels (mobile, web, call center)
External (credit bureau, KYC, market data)
At this stage, itās not directly usableĀ for decision-making because:
Itās incomplete, noisy, inconsistent, and unstructured.
Business teams canāt interpret it meaningfully.
š So we use analytics to make this data usable, insightful, and actionable.
š§ 2ļøā£ From Raw Data ā Business Decision Flow
[Raw Data]Ā
Ā Ā Ā ā (ETL / DataOps)
[Curated Data]Ā
Ā Ā Ā ā (Analytics Models)
[Business Insights]Ā
Ā Ā Ā ā (Visualization / Alerts / AI)
[Business Decision & Action]
Letās see this step-by-step š
Step 1: Data Ingestion (Raw Zone)
Collects data from all sources in data lake (Raw Zone).
Stores original records for audit, compliance, lineage.
Example:
Customer_ID, Monthly_Income, Loan_Amount, EMI_Payment_History
Step 2: Data Cleaning & Enrichment (Curated Zone)
Handle missing values, remove duplicates, standardize formats.
Enrich with derived features ā e.g.,
Debt-to-Income Ratio
Credit Utilization
Customer Lifetime Value (CLV)
Curated datasets are now analytics-ready.
Step 3: Analytics Processing (Analytics Zone)
Apply descriptive, diagnostic, predictive, or prescriptive analyticsĀ to extract meaning.
Examples:
Descriptive ā āWhich products are most used?ā
Predictive ā āWhich customers are likely to default?ā
Prescriptive ā āWhat should we offer to reduce churn?ā
Analytics models (BI dashboards, ML models, or GenAI insights) now create business intelligence.
Step 4: Visualization & Decision Support
Dashboards (Power BI, Tableau) show trends, KPIs, and anomalies.
Alerts and recommendations go to business teams or systems.
Example:
Risk team gets āTop 10 customers with rising default probabilityā.
Marketing gets āCustomer segments for upsell opportunityā.
Step 5: Business Action / Automation
Insights are operationalizedĀ into decisions:
Department | Data-driven Action |
Credit Risk | Adjust credit limit, approve/reject loans |
Marketing | Run personalized campaigns |
Operations | Automate manual workflows |
Fraud | Block suspicious transactions |
CX/Support | Route queries using AI-based assistants |
š§© Example: End-to-End Banking Scenario
š¹ Step 1: Raw Data
Data ingested from loan system, customer KYC, and bureau.
Customer_ID, Age, Income, Loan_Amount, Repayment_History
š¹ Step 2: Curated Data
Feature Engineering team derives:
debt_to_income_ratio
payment_delay_score
credit_utilization_ratio
š¹ Step 3: Analytics Layer
Predictive Analytics:Ā ML model predicts default risk.Descriptive Analytics:Ā Dashboard shows loan approval trends.Prescriptive Analytics:Ā Suggests adjusting interest rates for low-risk borrowers.
š¹ Step 4: Decision & Action
Credit committee uses these insights to automatically approve low-risk loans.
Risk team tightens policyĀ for high-risk segments.
š§© Architecture Summary (Text Diagram)
[Data Sources]
ā
[Data Lake - Raw Zone]
ā ā Cleansing, Validation
[Curated Zone]
ā ā Feature Engineering
[Analytics Zone]
ā ā BI, ML, GenAI Models
[Decision Layer]
ā
[Action: Business Strategy, Automation, CX Optimization]
š£ļø
āRaw data by itself doesnāt deliver business value ā analytics transforms it into insight.In our setup, data flows from the raw to curated to analytics zones in the lake. The curated zone creates high-quality, feature-rich datasets; the analytics zone applies descriptive, predictive, and prescriptive models. This enables business units to make data-driven decisions ā like approving loans, targeting the right customers, and proactively managing risk ā with full traceability and compliance.ā
š§ Enterprise āData-to-Decisionā Framework (for AI/ML & GenAI Enablement)
šÆ 1ļøā£ Objective
To establish an enterprise-wide framework that transforms raw operational data into actionable business insights and automated decisionsĀ using Analytics, AI/ML, and GenAI, while ensuring governance, compliance, and scalability.
š§© 2ļøā£ High-Level Flow
[Data Sources]
ā
[Data Lakehouse: Raw ā Curated ā Analytics Zones]
ā
[Analytics & AI/ML Layer]
ā
[Decision Intelligence Layer (BI, GenAI, Automation)]
ā
[Business Outcomes & Continuous Feedback Loop]
āļø 3ļøā£ Layer-by-Layer Architecture
š¹ Layer 1: Data Sources
Internal Systems:Ā Core Banking, CRM, ERP, Digital Channels
External Sources:Ā Credit Bureau, Market Feeds, Social, IoT, Regulatory APIs
Streaming Sources:Ā Kafka / Event Hub for real-time data
EA Governance:
Define Data Owners & Stewards
Metadata Catalog & Lineage (e.g., Azure Purview, Collibra)
Data Quality Rules and Policies
š¹ Layer 2: Data Lakehouse (Raw ā Curated ā Analytics Zones)
Zone | Purpose | Examples |
Raw Zone | Store all unprocessed data from various sources | Original logs, transactions, images |
Curated Zone | Clean, standardized, enriched data | De-duplicated, validated datasets |
Analytics Zone | Feature-engineered, analytics-ready datasets | Risk models, segmentation inputs |
EA Governance:
Define Data Retention & Classification Policies
Enforce Access Controls (RBAC/ABAC)
Implement DataOps Pipelines (Azure Data Factory / Databricks)
š¹ Layer 3: Analytics & AI/ML Layer
Type | Objective | Example |
Descriptive | What happened? | Loan default trend, churn rates |
Diagnostic | Why did it happen? | Feature correlation, cohort analysis |
Predictive | What will happen? | Default risk prediction, fraud likelihood |
Prescriptive | What should we do? | Adjust loan limits, cross-sell recommendation |
MLOps Governance:
Model Registry (MLflow, Azure ML)
Bias & Drift Monitoring
Explainability and Model Lifecycle Management
š¹ Layer 4: Decision Intelligence & GenAI Layer
This is where AI meets human decision-making.
Component | Role |
BI & Dashboards | Power BI / Tableau for descriptive insights |
GenAI Agents | Conversational copilots for business teams (e.g., āSummarize customer risk profileā) |
Decision Engines | Automate rule-based or model-based decisions |
Feedback Loops | Capture human feedback to retrain AI/ML models |
EA Governance:
Responsible AI Principles
GenAI Usage Policies (PII handling, prompt logging)
AI Ethics Board under Steering Committee
š¹ Layer 5: Business Outcomes Layer
Business Function | Data-Driven Decision | Outcome |
Credit Risk | Loan approval & limit adjustment | Lower NPA, faster TAT |
Fraud | Detect anomalous transactions | Reduced financial losses |
Marketing | Customer segmentation & recommendation | Higher conversion rate |
Operations | Process optimization | Reduced turnaround time |
Compliance | Regulatory reporting automation | Lower compliance risk |
š§ 4ļøā£ Continuous Learning & Feedback Loop
Insights from BI dashboards and AI/ML predictions are monitored for effectiveness.
Business feedback (approvals, rejections, overrides) flows back to data pipelinesĀ ā model retrainingĀ ā improved decisions.
This creates a closed-loop intelligence system.
[Business Action]Ā ā [Feedback Capture]Ā ā [Model Retraining]Ā ā [Improved Decision Accuracy]
šļø 5ļøā£ Governance Structure
Layer | Governance Body | Responsibilities |
Strategic | EA Steering Committee | Define AI strategy, KPIs, and ethics |
Tactical | Enterprise Architecture Review Board (EARB) | Approve AI/ML standards, reference models |
Operational | Solution Architecture Review Board (SARB) | Review AI/ML implementations & compliance |
Federated | BU AI Committees | Business-aligned adoption and local governance |
š 6ļøā£ Key Enablers
Capability | Description |
DataOps | Automate ingestion, transformation, validation |
MLOps | Standardize ML model lifecycle management |
AIOps | AI-driven monitoring & anomaly detection in operations |
FinOps | Optimize cost across cloud analytics workloads |
AI Governance Portal | One-stop view of data assets, model lineage, and risk scores |
š 7ļøā£ Example: AI-Driven Credit Decisioning
Step | Process | Tech |
Data Ingestion | Loan + KYC + Bureau data | Azure Data Factory / Kafka |
Curation | Feature engineering | Databricks |
Analytics | Predictive scoring | Azure ML |
Decision | Automated approval/rejection | GenAI + Decision Engine |
Feedback | Model tuning | MLOps pipeline |
Business Benefit:
Loan approval TAT reduced from 2 days ā 30 mins
95% accuracy in default prediction
Improved compliance and explainability
š§ 8ļøā£ EA Value Summary
Dimension | Value |
Strategic | Aligns AI/ML adoption with business KPIs |
Architectural | Standardized architecture and governance |
Operational | Automates data ā insight ā action flow |
Compliance | Ensures explainability, traceability, and ethics |
Innovation | Enables AI/GenAI copilots for decision-making |
š³ End-to-End Journey: AI/GenAI-Driven Credit Decisioning in Banking
š§ 1ļøā£ Business Objective
Enable faster and more accurate loan approval decisionsĀ while reducing credit riskĀ and ensuring fairness and compliance.
Goal:Ā Reduce loan approval turnaround from 2 days ā 30 mins
KPI:Ā 95% model accuracy, <2% false positives, 100% explainability compliance
Outcome:Ā Improved customer experience and reduced NPAs
š§© 2ļøā£ Data Ingestion & Lakehouse Setup
š¹ Sources
Core Banking System (loan applications, customer info)
Credit Bureau (CIBIL, Experian scores)
CRM (customer behavior, spending pattern)
Regulatory & Social (income tax, address validation)
š¹ Process
Ingestion:
Data is pulled in real-time using Azure Data FactoryĀ / Kafka topics.
Raw data stored in Data Lake - Raw ZoneĀ (immutable).
Curation:
Data Cleansing (remove duplicates, fix nulls, standardize formats).
Enrichment (joining with demographics, geolocation).
Derived fields like:
Debt-to-Income Ratio
Credit Utilization %
Repayment History Score
Curated data is stored in the Curated Zone.
Analytics Zone Preparation:
Feature engineering team generates ML-ready datasetsĀ (e.g., feature vectors).
Stored in Analytics ZoneĀ for model training.
Tools:Ā Azure Data Lake, Databricks, Delta Lake, Purview (for metadata & lineage).Governance:Ā Data Stewardship + DataOps Pipelines validated by Data Quality rules.
š§ 3ļøā£ Model Development & Training (AI/ML)
š¹ Feature Engineering
Feature store built for reusable engineered features (e.g., income brackets, defaults).
Feature selection using statistical correlation and SHAP importance.
š¹ Model Training
Train supervised ML models (e.g., XGBoost, LightGBM) on historical labeled data.
Split into Train/Test/Validation datasets.
š¹ Evaluation & Validation
Metrics: Accuracy, Precision, Recall, F1-score, ROC-AUC.
Fairness tests across gender, geography, and income.
Explainability generated using SHAP/LIME.
š¹ Model Registry (MLOps)
Model version, metadata, and performance logged in Azure ML Model Registry.
Approved by AI Model Review Board (SARB)Ā before deployment.
Governance:
MLOps pipeline automated using Azure DevOps.
Bias & drift tests integrated before production release.
š 4ļøā£ Model Deployment & Inference (MLOps in Action)
Containerization:
Model packaged as Docker image with FastAPI endpoint.
Deployment:
Deployed to AKS (Azure Kubernetes Service)Ā for scalable serving.
API Gateway Integration:
Exposed to the Loan Origination Microservice via Azure API Management.
Real-Time Inference:
When a loan request comes, the microservice calls ML API ā returns a risk score.
Decision Engine:
Rules + Model output combined for final decision (approve / reject / manual review).
Example:
IF risk_score < 0.3Ā ā APPROVE
ELSEĀ IF 0.3 ⤠score ⤠0.6Ā ā MANUAL REVIEW
ELSEĀ ā REJECT
š¬ 5ļøā£ GenAI Copilot for Credit Analyst
Now enters the GenAI layerĀ ā to make insights human-readable and assist decision-makers.
š¹ GenAI Components
RAG pipeline built using LangChain + Azure OpenAI + Vector Store (pgvector).
Knowledge base includes:
Risk policies
RBI credit guidelines
Historical decision explanations
Customer feedback
š¹ Use Case
Credit Officer logs into UI ā enters Loan ID ā asks:
āWhy was this loan rejected?ā
GenAI Copilot retrieves:
Model features influencing decision
Policy explanation (from knowledge base)
Similar past cases
Confidence level
Response Example:
āLoan #3421 was rejected because the applicantās debt-to-income ratio (85%) exceeds the risk policy threshold (60%). The credit score of 590 is below the bankās approval criteria. Similar cases in the last 6 months also had a default rate of 42%.ā
Governance:
Prompt templates reviewed by AI Council.
Guardrails prevent exposure of PII.
Feedback captured for improvement.
š 6ļøā£ Monitoring & Model Drift (Operational Phase)
š¹ Monitoring Dimensions
Area | Metric | Tool |
Model Performance | Accuracy, Precision, Drift | MLflow / Azure Monitor |
Data Drift | Feature distribution change | Evidently AI |
Fairness | Gender / Region bias | Responsible AI dashboard |
System Health | API latency, uptime | AIOps monitoring |
User Feedback | Analyst approval feedback | Feedback DB / retraining pipeline |
š¹ Actions
If drift > threshold, trigger retraining job via MLOps pipeline.
Governance review via SARB for production model refresh.
š 7ļøā£ Continuous Learning Loop
LoanĀ Decision ā Business Feedback ā Retrain Model ā Improved Predictions
Feedback from approved/rejected loans and analyst overrides feed into retraining.
Updated features and models go through the same MLOps cycle.
EA Governance ensures traceability via lineage and audit logs.
š§± 8ļøā£ EA Governance Alignment
Layer | Governance Body | Key Role in AI Journey |
Strategic | Steering Committee | Approve AI adoption roadmap, KPIs |
Tactical | EARB | Approve credit scoring architecture, MLOps standards |
Operational | SARB | Review deployment, drift reports, compliance |
Technology Council | Define AI principles, reference architectures, toolset (Azure ML, LangChain) | |
Data Council | Manage DataOps, lineage, and access controls |
šÆ 9ļøā£ Business Outcome Summary
Area | Before | After AI/GenAI Enablement |
Loan Approval TAT | 2 days | 30 mins |
Decision Accuracy | 80% | 95% |
Manual Review Load | 60% | 15% |
Compliance Reporting | Manual | Automated |
Explainability | Limited | GenAI Copilot-driven |
š 10ļøā£ Key Takeaways
ā Seamless integration of DataOps + MLOps + LLMOps
ā AI/GenAI made explainable, auditable, governed
ā Human + AI collaboration through GenAI Copilot
ā Federated governance ensures compliance, fairness, and transparency
š§ The Analytics Maturity Spectrum
Type | Purpose | Techniques | Example in Credit Decisioning |
Descriptive Analytics | What happened? | BI dashboards, SQL, aggregates | Loan default % by region, average approval time |
Diagnostic Analytics | Why did it happen? | Drill-downs, correlation, root cause | Defaults higher in region X due to income instability |
Predictive Analytics | What will happen? | ML models (XGBoost, Random Forest) | Predict customer default probability |
Prescriptive Analytics | What should we do? | Optimization, simulation, decision rules | Adjust interest rates or credit limits |
Cognitive / GenAI Analytics | How to explain or augment? | LLMs, RAG, Prompt Chaining | Explain loan rejections; summarize portfolio risk |
š³ End-to-End Walkthrough: Analytics in a Digital Lending Journey
Imagine a customer applies for a personal loan through your bankās mobile app.The decision to approve, reject, or send it for manual review will pass through several analytics layers.Weāll trace that journey end-to-end š
š§© Step 1: Loan Application ā Data Ingestion
Data Collected
Applicant: name, age, income, employment type, address
Loan details: requested amount, tenure, purpose
External: credit score, bureau history, bank statement features
What Happens Here
Data flows into Raw Zone ā Curated ZoneĀ of Data Lake.
Basic validation, enrichment (like deriving ādebt-to-income ratioā).
š No decision yet ā weāre only collecting and preparing data.
š§ Step 2: Descriptive Analytics ā āWhat Happened Before?ā
Purpose:Ā Understand historical loan data to create context.
Tools:Ā Power BI, SQL, Databricks Notebook
Examples:
āWhat % of loans were approved in the last quarter?ā
āWhich customer segments had the highest default rates?ā
āAverage turnaround time by region?ā
Outcome:Patterns are discovered ā e.g.
80% of defaulters had income < ā¹30K. Average DTI (Debt-to-Income) ratio for defaulters = 70%.
These insights feed business rules and model features.
š§© Where in Flow:Before model training ā this is part of historical portfolio analytics.
š Step 3: Diagnostic Analytics ā āWhy Did It Happen?ā
Purpose:Ā Identify root causesĀ of default or loan rejection patterns.
Techniques:Ā Statistical correlation, feature importance, drill-downs.
Examples:
āWhy did default rate spike in Q3?āā Root cause: layoffs in IT sector; more self-employed applicants.
āWhy do Tier-3 cities have higher rejection?āā Root cause: missing KYC and limited credit history.
Outcome:Bank updates its credit policy thresholds and ML model features accordingly.E.g., āInclude employment stability indexā as a new feature.
š§© Where in Flow:Still offline analysis ā helps refine model training and credit policy rules.
š¤ Step 4: Predictive Analytics ā āWhat Will Happen?ā
Now comes the real-time decisionĀ layer during loan processing.
Purpose:Ā Predict the likelihood of defaultĀ or loan repayment capability.
Technique:Ā ML model (e.g., XGBoost, LightGBM).
Example Flow:
Loan application hits Loan Origination Service.
Microservice sends customer + derived features ā Model API (hosted via AKS).
Model returns:
risk_scoreĀ = 0.68Ā (medium risk) probability_of_defaultĀ = 0.32
Result stored in Decision Engine DB.
š§© Where in Flow:At loan evaluation step, inside your MLOps pipeline or scoring API.
Outcome:
Low score ā auto-approve
Medium ā manual review
High ā reject
ā Predictive analytics directly drives operational decisioning.
āļø Step 5: Prescriptive Analytics ā āWhat Should We Do?ā
Purpose:Ā Recommend best possible actionĀ based on predictive insights.
Techniques:Ā Rule optimization, what-if simulation, decision matrix.
Example:
Predictive model says: risk_score = 0.6 (borderline case).
Prescriptive layer simulates outcomes:
Option 1: Approve with higher interest rate.
Option 2: Approve with guarantor.
Option 3: Reject outright.
Prescriptive engine (policy rules + optimization logic) recommends:
āApprove loan with 2% higher interest rate to offset risk.ā
š§© Where in Flow:This logic sits inside the Decision Engine microserviceĀ (post-prediction).
Outcome:Business rules combine with ML prediction ā final action (Approve / Reject / Review).
š¬ Step 6: Cognitive Analytics / GenAI ā āHow to Explain & Enhance?ā
Purpose:Ā Make AI decisions explainable and conversationalĀ to humans.
Tools:Ā LangChain + Azure OpenAI + Vector DB + RAG pattern.
Example:Credit Officer or Auditor asks:
āWhy was loan ID 42356 rejected?ā
GenAI Copilot responds:
āLoan was rejected because the customerās debt-to-income ratio (82%) exceeds the risk policy limit (60%). Credit score (585) indicates moderate risk, and past EMI delay was 3 times in last 6 months. Similar profiles had a 38% default probability last quarter.ā
Additional GenAI Tasks:
Summarize model insights in plain English.
Retrieve relevant policies from knowledge base (RAG).
Provide fairness or bias explanation (Responsible AI layer).
š§© Where in Flow:Post-decision ā in analyst dashboard, audit reports, or customer chatbot.
š Step 7: Continuous Learning (Feedback Loop)
Purpose:Ā Close the loop ā use new outcomes to improve analytics and models.
Approved loans ā actual repayment tracked ā feedback into data lake.
Defaults ā labeled for retraining model.
GenAI feedback (āexplanation not clearā) ā improve prompts.
š§© Where in Flow:End-to-end MLOps + LLMOps feedback cycle.
š Putting It All Together ā Text Diagram
[Loan Application Received]
ā
[Raw & Curated Data]
ā
[Descriptive Analytics]Ā ā Historical patterns (default %, approval trends)
ā
[Diagnostic Analytics]Ā ā Root causes (income instability, region risk)
ā
[Predictive Analytics]Ā ā ML model predicts default probability
ā
[Prescriptive Analytics]Ā ā Decision engine simulates best action
āā Auto-Approve (Low risk)
āā Manual Review (Medium risk)
āā Reject (High risk)
ā
[Cognitive / GenAI Analytics]Ā ā Explain decisions to officers & regulators
ā
[Feedback Loop]Ā ā Retrain model, refine policy thresholds
š§ EA Perspective
Layer | Analytics Type | EA Governance Focus |
Data & Platform | Descriptive, Diagnostic | Data Quality, Lineage, Metadata, Curation |
Model & Decision | Predictive, Prescriptive | MLOps, Policy Integration, Explainability |
User Experience | Cognitive / GenAI | LLMOps, Prompt Governance, Responsible AI |
Governance Bodies | ||
EARB | Approves architecture for analytics stack | |
SARB | Validates model fairness and performance | |
Technology Council | Defines tools (Power BI, Databricks, Azure ML, LangChain) |
šÆ Summary: How Analytics Enables the Loan Decision Flow
Stage | Analytics Type | Decision Influence |
Loan Trend Analysis | Descriptive | Identify approval trends |
Root Cause of Default | Diagnostic | Improve credit policy |
Risk Scoring | Predictive | Predict default probability |
Loan Action Simulation | Prescriptive | Decide approve/reject/review |
Explanation to User | Cognitive (GenAI) | Explain & justify decisions |
š¹ 1ļøā£ Data Lake Pipeline Overview
Flow:š Raw Zone ā Curated Zone ā Analytics Zone ā Model Serving / BI Dashboards
Zone | Purpose | Example Data |
Raw Zone | Ingest raw, unprocessed data from multiple systems. | Loan applications, KYC docs, income proofs, transaction logs, bureau data |
Curated Zone | Clean, standardize, and enrich data (feature engineering). | Customer profile, credit score, income-to-debt ratio, bureau risk rating |
Analytics Zone | Use curated data for analytics, AI/ML, and decision intelligence. | Derived KPIs, risk models, dashboards, alerts, trend reports |
š¹ 2ļøā£ Types of Analytics and How They Are Used
Letās take a āLoan Approval DecisionāĀ use case as an example:
šø (a) Descriptive Analytics ā āWhat happened?ā
Goal:Ā Understand the past loan trends and customer behavior.Where:Ā Performed in Analytics ZoneĀ (BI/Dashboards, SQL/PowerBI/Tableau).Example:
Average loan approval rate last quarter.
Default rate by region or income group.
Number of rejected applications due to poor credit history.
š” Output:Ā Loan summary reports, dashboards for management insights.
šø (b) Diagnostic Analytics ā āWhy did it happen?ā
Goal:Ā Investigate the reason behind past outcomes.Where:Ā Analytics Zone ā Diagnostic ML scripts or SQL analytics.Example:
Why defaults increased in the last 6 months?ā High exposure to low-income borrowers in rural areas.
Why manual reviews increased?ā Missing income proofs in 40% of applications.
š” Output:Ā Root-cause analysis ā informs lending policy adjustments.
šø (c) Predictive Analytics ā āWhat will happen next?ā
Goal:Ā Predict future outcomes based on patterns.Where:Ā Analytics Zone ā ML Models (Credit Scoring, Risk Forecasting).Example:
Predict probability of default for each applicant.
Predict which loans are likely to need manual review.
Forecast monthly loan disbursement volume.
š” Output:Ā Risk scores ā integrated into loan evaluation microserviceĀ or decision engine.
šø (d) Prescriptive Analytics ā āWhat should we do about it?ā
Goal:Ā Recommend the best action based on predictive insights.Where:Ā Analytics Zone ā AI Decision Layer / Business Rules Engine.Example:
If predicted default > 0.7 ā route to manual review.
If income/debt ratio < threshold ā auto-reject with reason.
If predicted credit score > 800 ā fast-track approval.
š” Output:Ā Automated decision rules ā integrated into loan approval workflowĀ (through APIs or decision engine).
š¹ 3ļøā£ Integration with AI/ML and GenAI
Once analytics models are validated:
Predictive & Prescriptive modelsĀ are deployed via MLOpsĀ pipelines.
Descriptive & Diagnostic insightsĀ are fed to executive dashboards.
GenAI/AI AssistantsĀ (via RAG) can summarize or explain insights in natural language to business users.
Example:
āThe increase in manual loan reviews last quarter was mainly due to missing KYC income documents in 38% of low-income applications.ā
š¹ 4ļøā£ Summary View
Layer | Analytics Type | Tools | Example Output |
Raw Zone | ā | Kafka, Data Factory | Raw ingestion logs |
Curated Zone | ā | Databricks, PySpark | Cleaned + feature engineered data |
Analytics Zone | Descriptive, Diagnostic, Predictive, Prescriptive | Power BI, MLFlow, Azure ML, LangChain | Dashboards, Risk Models, Recommendations |
Serving Layer | AI/ML Integration | MLOps, APIs | Automated loan decisions=== |
====
Dashboards are the unified visualization layerĀ on top of the Analytics Zone.
They bring together:
Descriptive analyticsĀ ā direct from curated or aggregated data
Diagnostic analyticsĀ ā from correlation and trend analysis
Predictive / Prescriptive analyticsĀ ā outputs from ML models
Cognitive analyticsĀ ā summaries or insights from GenAI
Letās break it down with examples (Loan Use Case š)
Analytics Type | Where Itās Computed | How It Appears in the Dashboard | Example |
Descriptive | BI Engine / SQL | Tables, charts, KPIs | Total loans approved, rejection rate by region |
Diagnostic | BI Engine / Advanced SQL / Python Script | Drill-down / Correlation charts | āDefaults increased due to low credit score segmentā |
Predictive | ML Model (via MLOps) ā Output stored in analytics zone | Risk score column, risk trend chart | āPredicted default risk = 0.72ā |
Prescriptive | Decision Engine / Rule Layer | Recommendation widgets | āAction: Send for manual reviewā |
Cognitive | GenAI layer / LLMOps ā API integrated with BI or Chatbot | Natural-language summary panel | āTop 3 factors driving rejections this quarterā¦ā |
š¹ How It Works in Architecture Terms
[Data Lake / Analytics Zone]
ā
[BI Semantic Layer]
ā
[Dashboard View]
āāā Descriptive KPIs (SQL / OLAP)
āāā Diagnostic Analysis (Drilldowns)
āāā Predictive Results (from ML APIs)
āāā Prescriptive Actions (from Decision Engine)
āāā Cognitive SummaryĀ (from GenAI API / LLMOps)
š¹ Practical Example (Banking Executive Dashboard)
Dashboard Sections:
Overview Tab (Descriptive)ā Total loans, NPA%, rejection trends, approval turnaround time
Root Cause Tab (Diagnostic)ā āWhyā analysis using correlation heatmaps and segment comparisons
Forecast Tab (Predictive)ā Risk forecast, default probability, disbursement projection
Recommendation Tab (Prescriptive)ā Action suggestions (e.g., adjust credit policy, tighten eligibility)
AI Insights Tab (Cognitive)ā āAsk AIā chat box powered by GenAI for narrative summariesā Example: āExplain top 3 causes for rising loan rejections in Q3ā
š§ Key Takeaway
āIn our setup, the dashboard becomes a single window for all analytics ā descriptive and diagnostic views are generated directly within BI tools, while predictive, prescriptive, and cognitive insights are integrated via APIs from the AI/ML and GenAI pipelines.This allows executives to move from data ā insight ā decisionĀ seamlessly within one analytics experience.ā
š¹ Scenario: Digital Lending Analytics ā āLoan Approval & Risk Managementā
The data pipeline runs through:Raw ā Curated ā Analytics ā Dashboard + AI Layer
1ļøā£ Descriptive Analytics ā āWhat happened?ā
Objective:Ā Give the business a factual picture of lending activity.Data Source:Ā Curated Zone (cleansed loan, customer, and repayment tables).Dashboard View:Ā Direct SQL/OLAP connection from Tableau or Power BI.
Examples
š Total loans applied, approved, rejected (this month, quarter, YTD).
š Average turn-around time from application ā disbursement.
š Regional breakdown of loan volumes.
š° Top 5 products by disbursed amount.
Who uses it:Ā CXOs, Risk and Business Heads.Purpose:Ā Baseline metrics and performance tracking.
2ļøā£ Diagnostic Analytics ā āWhy did it happen?ā
Objective:Ā Identify root causes for patterns or anomalies.Data Source:Ā Curated Zone + Feature Tables (income/debt ratio, credit utilization).Dashboard View:Ā Tableau/Power BI drill-downs or Python statistical analysis.
Examples
š āWhy did loan approvals drop 10% in Q2?āā Higher rejections in low-income segments.
š¦ āWhy did defaults increase?āā Exposure to unsecured loans in Tier-3 regions.
š Correlation analysis between loan size and default probability.
Who uses it:Ā Data Scientists, Risk Analysts.Purpose:Ā Discover underlying drivers and policy gaps.
3ļøā£ Predictive Analytics ā āWhat will happen?ā
Objective:Ā Forecast risk and future loan behaviour.Data Source:Ā Analytics Zone (model inputs from Curated Zone).Pipeline:Ā MLOpsĀ (train ā validate ā deploy credit scoring model).Dashboard View:Ā Tableau calls model API or reads model scores from Analytics DB.
Examples
š® Predicted probability of default for each applicant.
š§® Forecast of monthly disbursement volumes.
ā ļø Early-warning alerts for loans likely to turn delinquent.
Who uses it:Ā Credit Risk Teams, Operations Heads.Purpose:Ā Anticipate risk and optimize loan pipeline.
4ļøā£ Prescriptive Analytics ā āWhat should we do?ā
Objective:Ā Recommend actions based on predictive outcomes.Data Source:Ā Outputs of Predictive Models + Business Rules Engine.Pipeline:Ā Decision Engine integrated via API.Dashboard View:Ā āNext-best actionā or āRecommendationā tab.
Examples
ā If default risk < 0.3 ā Auto-approve.
šµļøāāļø If risk between 0.3ā0.7 ā Manual review.
ā If risk > 0.7 ā Reject with reason.
š” Portfolio-level actions ā āReduce exposure in Tier-3 cities.ā
Who uses it:Ā Credit Policy and Underwriting Teams.Purpose:Ā Operational decision support and automation.
5ļøā£ Cognitive Analytics (GenAI) ā āExplain and Reasonā
Objective:Ā Deliver natural-language insights and explainability.Data Source:Ā Combines Analytics Zone outputs + model metadata + business context.Pipeline:Ā LLMOpsĀ (RAG + Vector DB + Prompt Templates + Guardrails).Dashboard View:Ā Embedded GenAI chat pane or API call to LLM.
Examples
š¬ āExplain top 3 reasons for loan rejections last month.ā
š§ āSummarize credit risk trend for Q3.ā
š āGenerate executive summary of approval vs default trends.ā
š āSuggest data segments to target for new personal loan campaign.ā
Who uses it:Ā CXOs, Operations Managers, Analysts.Purpose:Ā Cognitive insight + explainability without needing SQL skills.
š¹ Text Diagram: Unified Flow to Dashboard
[Data Sources]
Loan System ⢠Bureau Data ⢠CRM ⢠Payments
ā
[Raw Zone]Ā ā [Curated Zone]
ā
Descriptive + Diagnostic Analytics
ā
[Analytics Zone]
āā ML Models ā Predictive Analytics
āā Decision Engine ā Prescriptive Analytics
āā LLMOps Layer ā Cognitive Analytics
ā
[Unified Dashboard (Tableau / Power BI)]
āā Descriptive KPIs (SQL)
āā Diagnostic Drilldowns (SQL + Python)
āā Predictive Scores (API)
āā Prescriptive Actions (API)
āā Cognitive Summaries (GenAI API)
š”
āIn our digital-lending analytics stack, curated data powers descriptive and diagnostic dashboards that show trends, volumes, and reasons for rejections.Predictive and prescriptive analytics come through our MLOps pipeline, feeding risk scores and recommended actions via APIs into Tableau.On top, a GenAI cognitive layer connected through LLMOps allows executives to ask natural-language questions like āWhy did approval rates dip last quarter?āThis unified view helps leadership move seamlessly from data ā insight ā action in one dashboard.ā
.png)

Comments