sap basis; interview
[recent]

Excel

[excel][bleft]

Database

[database][bsummary]

SAP BASIS

[basis][twocolumns]

Senior Data Engineer Interview Questions and Answers 2026

Senior Data Engineer Interview Questions and Answers 2026


Senior Data Engineer Interview Questions and Answers 2026 - Want to Crack Visa Inc. Interview? Here are latest questions asked for Senior Data Engineer role. Interview Rounds: 4 For one of my students who was interviewed recently, the clarity of thought process in his project round made him crack the offer. Round 1 - HackerRank Test (SQL, Basic Coding) Round 2 - Technical (Spark Focused) Round 3 - Technical (Project Specific) Round 4 - Mix (System Design and Managerial)


Senior Data Engineer Interview Questions and Answers


 

1) How do you calculate transaction volume for a merchant?

Concept:
Transaction volume usually means the total number of transactions or total monetary value processed by a merchant over a time period.

Case 1: Count of transactions

SELECT

    merchant_id,

    COUNT(*) AS transaction_count

FROM transactions

WHERE merchant_id = 123

  AND transaction_date BETWEEN '2025-01-01' AND '2025-01-31'

GROUP BY merchant_id;

Case 2: Total transaction value (more common in fintech)

SELECT

    merchant_id,

    SUM(amount) AS total_volume

FROM transactions

WHERE merchant_id = 123

  AND transaction_date BETWEEN '2025-01-01' AND '2025-01-31'

GROUP BY merchant_id;

Key Interview Points:

  • Clarify whether volume = count or amount
  • Handle:
    • Failed transactions (status = 'SUCCESS')
    • Currency normalization if multi-currency
  • You can also segment:
    • By day, region, payment type

2) Calculate Daily Cumulative Balance for merchant

Concept:
Running total of transactions per day.

Example:

SELECT

    merchant_id,

    transaction_date,

    SUM(amount) AS daily_amount,

    SUM(SUM(amount)) OVER (

        PARTITION BY merchant_id

        ORDER BY transaction_date

    ) AS cumulative_balance

FROM transactions

WHERE merchant_id = 123

GROUP BY merchant_id, transaction_date

ORDER BY transaction_date;

Explanation:

  • First SUM(amount) → daily total
  • Window function → cumulative sum over time

Output Example:

date

daily_amount

cumulative_balance

Jan 1

100

100

Jan 2

200

300

Jan 3

-50

250

Key Points:

  • Use window functions
  • Partition by merchant
  • Order by date
  • Handle refunds (negative values)

3) Fetch Top 3 geographies with highest customer base

Concept:
Find regions with the most unique customers.

Query:

SELECT

    geography,

    COUNT(DISTINCT customer_id) AS customer_count

FROM customers

GROUP BY geography

ORDER BY customer_count DESC

LIMIT 3;

Alternative (with ranking):

SELECT *

FROM (

    SELECT

        geography,

        COUNT(DISTINCT customer_id) AS customer_count,

        RANK() OVER (ORDER BY COUNT(DISTINCT customer_id) DESC) AS rnk

    FROM customers

    GROUP BY geography

) t

WHERE rnk <= 3;

Key Points:

  • Use COUNT(DISTINCT)
  • Ranking functions:
    • RANK() (handles ties)
    • DENSE_RANK()
  • Be careful with:
    • Duplicate users
    • Data skew

4) How do you optimize a slow SQL query?

This is a very important Visa interview question. Don’t just list points—explain how you think.


Step-by-step approach:

1. Check Execution Plan

EXPLAIN ANALYZE <query>;

Look for:

  • Full table scans
  • Expensive joins
  • High cost operations

2. Indexing

  • Add indexes on:
    • WHERE columns
    • JOIN keys
    • ORDER BY columns

Example:

CREATE INDEX idx_txn_merchant_date

ON transactions(merchant_id, transaction_date);


3. Avoid SELECT *

Instead:

SELECT merchant_id, amount


4. Filter Early (Predicate Pushdown)

Bad:

SELECT *

FROM transactions

WHERE YEAR(transaction_date) = 2025;

Good:

WHERE transaction_date BETWEEN '2025-01-01' AND '2025-12-31'


5. Optimize Joins

  • Ensure join keys are indexed
  • Use correct join type

Bad:

FROM A, B WHERE A.id = B.id

Good:

FROM A

JOIN B ON A.id = B.id


6. Use Partitioning

For large datasets:

  • Partition by date or merchant_id

7. Avoid Nested Subqueries

Convert to joins or CTEs


8. Use Aggregation Smartly

  • Pre-aggregate if needed
  • Reduce data before joins

9. Caching / Materialized Views

  • Useful for repeated queries

10. Data Volume Awareness

  • Always ask:
    • Table size?
    • Cardinality?
    • Distribution?

 

Strong Interview Closing Line:

“I first analyze the execution plan, identify bottlenecks like full scans or skewed joins, then optimize using indexing, partitioning, query rewrites, and reducing data scanned.”

Here are clear, interview-level answers with depth + examples for each Spark question—exactly the kind of explanation expected in a Senior Data Engineer interview.

 



Visit for latest Job Vacancies and News indianinQ8.com

Visit for More Forever Living Products - Forever Living Kuwait at https://foreverlivingkuwait.blogspot.com/

1) How do you handle Fault Tolerance using Spark?

Core Idea:
Spark achieves fault tolerance primarily using RDD lineage and lazy evaluation, instead of heavy replication like traditional systems.


Key Mechanisms

1. Lineage (DAG-based recovery)

  • Every dataset (RDD/DataFrame) keeps track of how it was built
  • If a partition is lost, Spark recomputes it from the lineage

Example:

rdd1 = sc.textFile("transactions.txt")

rdd2 = rdd1.map(parse)

rdd3 = rdd2.filter(lambda x: x.amount > 100)

If rdd3 partition is lost → Spark recomputes using:
rdd1 → rdd2 → rdd3


2. Lazy Evaluation

  • Transformations are not executed immediately
  • Execution happens only when an action is triggered
  • Helps rebuild lost data efficiently

3. Checkpointing (for long lineage)

  • Stores intermediate data to stable storage (HDFS/S3)
  • Prevents expensive recomputation chains

rdd.checkpoint()


4. Data Replication (for cached data)

  • When using persist() / cache(), Spark can replicate partitions

rdd.persist(StorageLevel.MEMORY_AND_DISK_2)


5. Task Retry Mechanism

  • Failed tasks are retried automatically (default: 4 times)

Interview Insight:

“Spark trades storage overhead for computation by using lineage-based recovery, making it highly efficient for large-scale distributed processing.”


2) Difference between Job, Stage, and Task + Task Count


Definitions

Job

  • Triggered by an action

Examples:

df.count()

df.collect()

Each action = 1 Job


Stage

  • A job is divided into stages based on shuffle boundaries

Two types:

  • Shuffle Map Stage
  • Result Stage

Task

  • Smallest unit of work
  • Runs on a single partition

How Spark decides number of Tasks?

👉 Number of tasks = number of partitions in that stage


Example:

rdd = sc.textFile("data.txt", 100)  # 100 partitions

rdd.map(...).filter(...).count()

  • No shuffle → 1 stage
  • Tasks = 100

With Shuffle:

rdd.groupByKey()

  • Causes shuffle → new stage created
  • Tasks in next stage = number of shuffle partitions (default: 200)

Visual Flow:

Job

 ├── Stage 1 (no shuffle)

 │     ├── Task 1

 │     ├── Task 2

 │     └── Task N

 └── Stage 2 (after shuffle)

       ├── Task 1

       └── Task M


Key Interview Points:

  • Job = Action
  • Stage = Shuffle boundary
  • Task = Partition-level execution
  • Tasks scale with partitions → tuning partitions is critical

3) Broadcast Join vs Sort Merge Join


Broadcast Join

When used:

  • One dataset is small enough to fit in memory

How it works:

  • Small table is broadcast to all worker nodes
  • Large table is scanned locally → no shuffle

from pyspark.sql.functions import broadcast

 

df_large.join(broadcast(df_small), "key")


Pros:

  • No shuffle → very fast
  • Ideal for dimension tables

Cons:

  • Memory limitation

Sort Merge Join

When used:

  • Both datasets are large

How it works:

  1. Shuffle both datasets on join key
  2. Sort them
  3. Merge join

Pros:

  • Scalable for big data

Cons:

  • Expensive (shuffle + sort)

Comparison Table:

Feature

Broadcast Join

Sort Merge Join

Data Size

Small + Large

Large + Large

Shuffle

❌ No

✅ Yes

Speed

Fast

Slower

Memory

Needs small table in memory

Distributed


AWS Data Engineering Roadmap👇


Senior Data Engineer Interview Questions and Answers 2026


Interview Tip:

“I prefer broadcast join when one dataset is small (< ~10MB–100MB depending on config), otherwise Spark defaults to sort-merge join.”


4) Detect Data Skew in Payment Transactions

Problem:
Data skew happens when some keys (e.g., merchant_id) have disproportionately large data → causing slow tasks.


Example Scenario:

  • One merchant processes millions of transactions
  • Others only thousands

Detection Techniques

1. Check key distribution

df.groupBy("merchant_id").count().orderBy("count", ascending=False).show()

👉 Look for:

  • Huge difference between top keys vs rest

2. Task Time Analysis (Spark UI)

  • Stages where:
    • Some tasks take much longer
    • Uneven execution time

3. Partition Size Check

df.rdd.glom().map(len).collect()


4. Approx Quantiles

df.stat.approxQuantile("transaction_count", [0.5, 0.9, 0.99], 0.01)


Example Insight:

merchant A → 5M records

merchant B → 10K records

👉 Clear skew


Bonus: Handling Skew (important in interview)

1. Salting Technique

from pyspark.sql.functions import rand

 

df = df.withColumn("salt", (rand() * 10).cast("int"))


2. Broadcast Join (if possible)



Kuwait bus routes and numbers, bus route kuwait CityBus, KPTC, KGL Mowsalat. find Kuwait’s public transport Muscat خط الحافلات الكويت.


Senior Data Engineer Interview Questions and Answers 2026


What is SAP Landscape?



3. Skew Join Optimization (Spark 3+)

spark.conf.set("spark.sql.adaptive.skewJoin.enabled", "true")


4. Repartitioning

df.repartition("merchant_id")


Strong Interview Closing Line:

“I detect skew by analyzing key distribution and task execution in Spark UI, then mitigate it using salting, broadcast joins, or adaptive query execution.”


Final Tip for Visa-Level Answers

Always tie your answer to:

  • Scale (millions of transactions)
  • Performance impact
  • Real-world payment systems

  

 

Below are strong, structured, interview-ready answers you can adapt to your own experience. These are written in a way that works well for a Visa-level Senior Data Engineer interview—clear, scalable, and grounded in real-world payment/data systems.


1) Give a brief about your last project. Explain architecture?

Sample Answer (customize with your domain)

“In my last project, I worked on building a real-time payment analytics pipeline that processed transaction data from multiple sources like POS systems, mobile apps, and partner APIs.”

Architecture Overview:

1. Data Ingestion

  • Streaming: Kafka (real-time transactions)
  • Batch: S3 / data lake ingestion

2. Processing Layer

  • Spark Structured Streaming for real-time
  • Batch jobs for reconciliation and reporting

3. Storage

  • Raw layer (Bronze): immutable raw data
  • Cleaned layer (Silver): validated + enriched
  • Aggregated layer (Gold): business-level metrics

4. Serving Layer

  • Data warehouse (Snowflake / Delta tables)
  • BI dashboards (Tableau/Power BI)

Simple Flow:

Sources → Kafka → Spark → Delta Lake → BI / APIs


Key Highlights:

  • Near real-time processing (latency < few minutes)
  • Handled schema evolution
  • Built data quality checks

Strong Interview Line:

“The architecture was designed for scalability, fault tolerance, and low-latency insights, which is critical in payment ecosystems.”


2) What was the data volume you were handling? How did you optimize it?

Sample Answer:

“We were processing around 200–300 million transactions per day, with peak loads during weekends and sales events.”


Optimization Techniques

1. Partitioning

  • Partitioned by transaction_date
  • Sometimes by region or merchant_id

PARTITIONED BY (transaction_date)


2. File Size Optimization

  • Maintained optimal file size (~128MB–1GB)
  • Avoided small file problem

3. Caching

df.cache()


4. Broadcast Joins

  • Used for small dimension tables (merchant, country)

5. Incremental Processing

  • Processed only new data using watermarking

6. Z-Ordering (Databricks)

  • Improved query performance on filters

7. Adaptive Query Execution (AQE)

  • Automatically handled skew and join strategies

Strong Interview Line:

“I focused on minimizing shuffle, optimizing partitioning strategy, and reducing I/O overhead, which significantly improved performance.”


3) Which part of pipeline were you handling solely?

Sample Answer:

“I was primarily responsible for the data processing and transformation layer, including both batch and streaming pipelines.”


Responsibilities:

  • Built Spark jobs for:
    • Transaction enrichment
    • Fraud feature generation
  • Designed Delta tables
  • Implemented data quality checks
  • Optimized slow queries
  • Handled job failures and monitoring

Ownership Example:

“I independently designed and implemented the merchant-level aggregation pipeline used by downstream reporting and risk systems.”


Strong Tip:

Always show:

  • Ownership
  • Impact
  • Decision-making

4) Can you redesign your project using latest Databricks tools?

This is a very high-value question—they test modernization thinking.


Modern Databricks Architecture

1. Delta Lake (Core Storage)

  • ACID transactions
  • Schema enforcement

2. Medallion Architecture

  • Bronze → Silver → Gold layers

3. Auto Loader (for ingestion)

spark.readStream.format("cloudFiles")


4. Delta Live Tables (DLT)

  • Declarative pipelines
  • Built-in quality checks

5. Unity Catalog

  • Centralized governance

6. Photon Engine

  • Faster SQL execution

7. Workflows (Job Orchestration)

  • Replace Airflow in many cases

Redesigned Flow:

Auto Loader → Bronze (Delta)

        ↓

DLT → Silver (cleaned)

        ↓

DLT → Gold (aggregated)

        ↓

Databricks SQL / BI


Improvements Over Legacy:

  • Less manual pipeline management
  • Built-in monitoring
  • Better performance (Photon)
  • Data governance (Unity Catalog)

Strong Interview Line:

“I would replace custom Spark jobs with Delta Live Tables and leverage Auto Loader for scalable ingestion, reducing operational overhead significantly.”



SAP Client Creation (SCC4) & Logical system (BD54) in SAP

5) What was the design layout of your tables?


Medallion Architecture Design

Bronze Layer (Raw)

  • Schema: minimal transformation
  • Columns:
    • raw payload
    • ingestion timestamp

Silver Layer (Cleaned)

  • Cleaned + validated data
  • Deduplication applied

PRIMARY KEY (transaction_id)


Gold Layer (Business-level)

Example: Merchant Aggregation Table

CREATE TABLE merchant_daily_summary (

    merchant_id STRING,

    transaction_date DATE,

    total_transactions BIGINT,

    total_amount DOUBLE,

    avg_transaction_value DOUBLE,

    region STRING

)


Design Best Practices

1. Partitioning

PARTITIONED BY (transaction_date)


2. Indexing / Z-Ordering

  • On:
    • merchant_id
    • customer_id

3. Normalization vs Denormalization

  • Silver → normalized
  • Gold → denormalized (for fast queries)

4. Slowly Changing Dimensions (SCD)

Example:

  • Merchant details → SCD Type 2

5. Data Quality

  • Null checks
  • Duplicate handling
  • Schema validation

Strong Interview Line:

“I designed tables based on access patterns—normalized in Silver for correctness and denormalized in Gold for performance.”


Final Advice for Your Answer Delivery

When answering:

  • Keep it structured (Architecture → Volume → Optimization → Ownership)
  • Use numbers (scale matters a lot at Visa)
  • Show trade-offs and reasoning

 

These are core Visa-level system design questions. Your answer should show scale, trade-offs, and real-world payment constraints (latency, consistency, fraud risk)—not just tools.

 

 

1) Design a system capable of handling millions of transactions per second

Step 1: Clarify Requirements

Functional:

  • Process payment transactions
  • Validate, enrich, and store transactions
  • Provide real-time response (approval/decline)

Non-Functional:

  • Throughput: Millions TPS
  • Latency: < 100–300 ms
  • Availability: 99.999%
  • Consistency: Strong for financial correctness

Step 2: High-Level Architecture

Client → API Gateway → Load Balancer → Stateless Services → Messaging → Processing → Storage


Step 3: Detailed Components

1. API Gateway

  • Authentication, rate limiting
  • Routes requests to services

2. Load Balancer

  • Distributes traffic across instances
  • Ensures horizontal scalability

3. Stateless Transaction Service

  • Validates request
  • Performs basic checks:
    • Card validity
    • Balance availability
  • Stateless → easy scaling

4. Messaging Layer (Critical for Scale)

  • Kafka / Pulsar
  • Decouples ingestion from processing

Transactions → Kafka Topic (partitioned by card_id / merchant_id)

👉 Partitioning ensures parallelism


5. Stream Processing Layer

  • Spark Streaming / Flink

Responsibilities:

  • Enrichment (merchant, geo)
  • Risk checks
  • Aggregations

6. Database Layer

OLTP (Critical path)

  • Distributed DB (Cassandra / CockroachDB)
  • Stores:
    • Transactions
    • Account balance

OLAP (Analytics)

  • Data warehouse / Delta Lake

7. Caching Layer

  • Redis
  • Stores:
    • User/session data
    • Frequent lookups

8. Downstream Systems

  • Fraud detection
  • Notifications
  • Reporting

Step 4: Key Design Decisions

1. Partitioning Strategy

  • By card_id or account_id
  • Ensures:
    • Ordering
    • Load distribution

2. Idempotency (VERY IMPORTANT)

  • Prevent duplicate transactions

transaction_id → unique key


3. Exactly-Once / At-Least-Once

  • Payments → prefer exactly-once semantics

4. Horizontal Scaling

  • Auto-scale services + Kafka partitions

5. Fault Tolerance

  • Retry mechanisms
  • Multi-region deployment

Step 5: Bottlenecks & Solutions

Problem

Solution

DB contention

Sharding

Hot keys

Better partitioning

Latency

In-memory caching

Spikes

Backpressure via Kafka


Strong Interview Closing Line:

“The system relies on partitioned streaming, stateless services, and distributed storage to achieve high throughput while maintaining low latency and strong consistency for financial correctness.”


2) Design a real-time fraud detection pipeline


Step 1: Requirements

Functional:

  • Detect fraud in real-time
  • Score transactions before approval

Non-Functional:

  • Latency: < 50–100 ms
  • High accuracy
  • Scalable

Step 2: Architecture Overview

Transaction → Stream → Feature Engineering → Model Scoring → Decision Engine


Step 3: Components

1. Data Ingestion

  • Kafka topic for transactions

2. Stream Processing (Core Layer)

  • Spark Streaming / Flink

Responsibilities:

  • Feature extraction:
    • Transaction frequency
    • Amount deviation
    • Geo-location mismatch

3. Feature Store

Stores:

  • Historical features
  • Real-time aggregates

Examples:

  • Avg transaction amount (last 1 hr)
  • Transactions per minute

4. Real-Time Feature Engineering

Example:

User makes 5 transactions in 1 minute → suspicious


5. Model Scoring Layer

  • ML model (e.g., XGBoost, Logistic Regression)

Output:

fraud_score = 0.92


6. Decision Engine

Rules:

IF fraud_score > 0.9 → Decline

IF 0.7–0.9 → Send for manual review

ELSE → Approve


7. Storage

  • Store:
    • Predictions
    • Transactions
    • Audit logs

8. Feedback Loop

  • Confirmed fraud cases → retrain model

Step 4: Handling Real-Time Constraints

1. Low Latency

  • In-memory processing
  • Precomputed features

2. Stateful Processing

  • Maintain sliding windows

Example:

Transactions in last 5 minutes


3. Exactly-Once Processing

  • Avoid duplicate scoring

4. Data Skew Handling

  • Popular merchants → skew mitigation

Step 5: Example Fraud Signals

  • Sudden location change (Kuwait → USA in 5 min)
  • High-value transaction spike
  • Multiple failed attempts
  • New device

Step 6: Scaling Strategy

  • Partition by user_id or card_id
  • Horizontal scaling of stream processors

Step 7: Trade-offs

Choice

Trade-off

Real-time vs accuracy

Faster but less context

Complex models

Higher latency

Strict rules

False positives

 

Strong Interview Closing Line:

“The key is combining real-time feature computation with low-latency model scoring, ensuring fraud decisions are made within milliseconds without impacting transaction approval time.”


Final Tip for Visa Interviews

When answering system design:

  • Always mention:
    • Latency
    • Scalability
    • Consistency
    • Fault tolerance
  • Tie everything back to:
    • Financial correctness
    • User experience

 

 

Here are strong, senior-level managerial answers tailored for a Visa / fintech environment. The goal is to sound clear, grounded, and ownership-driven—not rehearsed or overly generic.


1) What motivates you to join a fintech giant?

Strong Answer:

“What excites me about joining a fintech company like Visa is the scale, reliability, and real-world impact of the systems. Every transaction is time-sensitive and directly tied to user trust, which makes data engineering much more meaningful compared to typical analytics systems.”


Key Points to Emphasize:

1. Scale + Complexity

  • Millions of transactions per second
  • Real-time decisioning systems
  • Distributed architecture challenges

2. Criticality

  • Systems must be highly reliable (financial correctness)
  • Small errors have large impact

3. Real-Time Data Engineering

  • Streaming pipelines (Kafka, Spark, Flink)
  • Fraud detection, risk scoring

4. Learning Opportunity

  • Exposure to advanced distributed systems
  • Strong engineering culture

Strong Closing Line:

“I am motivated by building systems where data is not just stored or analyzed, but directly drives real-time financial decisions at global scale.”


2) If you are asked to handle a real-time end-to-end pipeline, which part will you need external help for?

Weak Answer (avoid):

“I can do everything.”


Strong Answer (balanced + honest):

“I am comfortable handling the end-to-end pipeline from ingestion to processing and serving, but if I were to call out areas where I might seek external help, it would be around highly specialized or cross-functional components.”


Break it down:

1. Areas I am confident in:

  • Data ingestion (Kafka, APIs)
  • Stream processing (Spark/Flink)
  • Data modeling (Delta tables, warehouse design)
  • Optimization & troubleshooting
  • Building batch + streaming pipelines

2. Where I may need external collaboration:

a) Infrastructure / Platform engineering

  • Cluster-level tuning
  • Network / cloud infra optimization
  • Kubernetes / low-level scaling issues

b) Machine Learning Model Design (if applicable)

  • Advanced fraud models
  • Feature engineering at research level

c) Security / Compliance

  • PCI-DSS compliance
  • Encryption standards in payments

Strong framing:

“I prefer collaborating with platform and security teams for infrastructure and compliance aspects, while taking full ownership of data pipeline design and performance optimization.”


Strong Closing Line:

“I believe strong systems are built through cross-functional collaboration rather than siloed ownership.”


3) Which role do you see yourself in after 3 years if selected?

Strong Answer (Senior-level growth mindset):

“In the next three years, I see myself growing into a Senior Data Engineer / Tech Lead role, owning critical data pipelines and contributing to architecture decisions for large-scale systems.”


Break it into 3 dimensions:

1. Technical Growth

  • Deep expertise in:
    • Streaming architectures
    • Real-time analytics
    • Distributed systems optimization

2. Ownership

  • Owning end-to-end pipelines
  • Designing scalable frameworks
  • Improving system reliability and cost efficiency

3. Leadership

  • Mentoring junior engineers
  • Leading design discussions
  • Driving best practices in data engineering

Optional Advanced Add-on (very strong for Visa):

“I also see myself contributing to platform-level improvements, such as reusable data frameworks or standardized streaming patterns that improve productivity across teams.”


Strong Closing Line:

“My goal is to evolve from executing pipelines to designing systems that define how data flows across the organization.”


Final Interview Strategy Tip

For managerial questions, Visa interviewers look for:

  • Stability (not job-hopping mindset)
  • Ownership mindset
  • Real-world awareness (compliance + scale)
  • Collaboration attitude

 

SQL Questions: * How do you calculate transaction volume for a merchant? * Calculate Daily Cumulative Balance for merchant? * Fetch out Top 3 geographies with highest customer base? * How do you optimize a slow SQL query? Spark Questions: * How do you handle Fault Tolerance using Spark? * What is the difference between Job, Stage and Task? How Spark decides number of tasks per stage? * What is difference between Broadcast Join vs Sort Merge Join? * Give example of how do you detect, Data Skew in a set of payment transactions? Project Related Questions: * Give a brief about your last project. Explain architecture? * What was the data volume you were handling? How did you optimize it? * Which part of pipeline were you handling solely? * Can you redesign your project using latest Databricks tools? * What was the design layout of your tables? System Design Questions: * Design a system which is capable of handling millions of transactions every second? * Design a fraud detection pipeline which functions in real time? Managerial Questions: * What motivates you to join a fintech giant? * If you are asked to handle a real time end to end pipeline, for which part will you need an external help? * Which role do you see yourself in, after 3 years, if you get selected? Tip: 

Questions during Visa Interviews are focus around real time data handling. Build strong conceptual clarity around it.


 

No comments: