Interview questions

COA Viva Interview Questions: 20 Questions with Likely Follow-Ups

April 29, 2026Updated May 5, 202621 min read
group people working out business plan office

A high-yield COA viva interview questions matrix for MBBS students and interns — with the most likely prompts, the follow-up questions examiners ask next, and

COA viva interview questions are easy to find. The problem is that every list stops at the first answer, and examiners do not.

Most students who struggle in a COA viva are not struggling because they do not know what pipelining is. They struggle because they answered "overlapping instruction execution" cleanly, felt confident, and then the examiner said "so where does it stall?" — and that is where the answer fell apart. The definition was there. The follow-up chain was not.

This is a COA viva prep matrix built for exactly that moment. Every question below comes with the likely examiner follow-up and the safest short answer path, because surviving a viva means surviving the second question, not just the first.

How Examiners Chain COA Viva Questions

Why a Good First Answer Still Gets Pushed Further

A clean opening answer does not close a topic in a viva — it opens it. When an examiner hears a confident definition, the natural next move is to find out whether that confidence has depth behind it. This is not adversarial; it is the structure of oral assessment. Research on oral examination methodology consistently shows that examiners use probing follow-ups to distinguish surface recall from functional understanding.

The classic COA viva chain looks like this: an examiner asks "what is a cache?" The student answers with something about fast memory between the CPU and RAM. The examiner immediately follows with "how does it differ from the memory hierarchy?" Now the student needs to know where cache sits relative to registers, main memory, and secondary storage — and most students who memorized the first answer did not think about the second question at all.

The Pattern Every Examiner Follows When They Smell a Weak Spot

Examiners move in a predictable sequence: definition, then comparison, then application. If the definition sounds rehearsed but thin, they will compress the comparison step and go straight to application to test whether the student actually understands the concept or just memorized a sentence.

Take "what is pipelining?" A student answers "executing multiple instructions simultaneously by dividing them into stages." Solid. The examiner then asks "why does throughput improve but latency for a single instruction does not?" That is the comparison-to-application pivot. A student who only memorized the definition has nothing left to say. A student who understood the pipeline stages — fetch, decode, execute, memory, write-back — can explain that each instruction still takes the same number of cycles, but the processor is now working on five instructions at once. That second answer is what separates a pass from a distinction.

What a 20-Second Viva Answer Should Actually Contain

The target structure is: definition, one key feature, one use case. Three elements, one breath, then stop. Use interrupts as the template.

"An interrupt is a signal sent to the CPU by a hardware or software event, indicating that it needs immediate attention. The key feature is that it suspends the current process, saves its state, and transfers control to an interrupt service routine. The use case is I/O devices — a keyboard does not make the CPU poll constantly; it interrupts only when a key is pressed."

That answer takes roughly 20 seconds. It gives the examiner a definition, a mechanism, and a real-world anchor. It also leaves no obvious gap for a destructive follow-up because it has already answered the most likely next question — why use interrupts at all — before it was asked.

The Most Asked COA Viva Interview Questions

What Is Computer Architecture?

Definition to say: Computer architecture is the set of rules and methods that describe the functionality, organization, and implementation of computer systems — essentially the interface between hardware and software.

Follow-up the examiner will ask: "Why does architecture matter in system design?"

Safe answer: Architecture determines what instructions the processor can execute and how efficiently it can execute them. A laptop CPU and a smartphone chip may run the same application, but their architectural choices — instruction sets, word length, addressing modes — determine power consumption, speed, and compatibility. The ISA is the contract between the software and the hardware.

What Is the Difference Between Computer Architecture and Computer Organization?

This is one of the highest-yield distinctions in any COA oral exam, and it is also one of the most commonly blurred.

Definition to say: Architecture is what the system does — the logical design visible to the programmer, including the instruction set, data types, and addressing modes. Organization is how the system does it — the physical implementation, including control signals, memory technology, and interconnects.

Follow-up the examiner will ask: "Give me a concrete example."

Safe answer: The ISA is architecture. Two processors can share the same ISA — say, x86 — but implement it with completely different hardware. One might use a superscalar design; another might use a simpler pipeline. Same architecture, different organization. As Patterson and Hennessy frame it: architecture is the programmer's view, organization is the implementer's view.

What Are the Main Components of a Microprocessor?

Definition to say: A microprocessor contains three core blocks — the ALU (arithmetic logic unit), the control unit, and registers. Supporting these are internal buses and interfaces to memory and I/O.

Follow-up the examiner will ask: "What does each block actually do in a basic CPU cycle?"

Safe answer: The control unit fetches the instruction from memory and decodes it. The ALU executes arithmetic or logical operations on the data. Registers hold the operands and intermediate results so the CPU does not have to reach out to main memory for every operation. In a basic fetch-decode-execute cycle, these three blocks hand off to each other in sequence.

What Is an Interrupt in a Microprocessor System?

Definition to say: An interrupt is a signal that causes the CPU to pause its current execution, save its state, and transfer control to a special routine called the interrupt service routine (ISR).

Follow-up the examiner will ask: "Why use interrupts instead of polling?"

Safe answer: Polling makes the CPU check an I/O device repeatedly in a loop — wasteful when the device responds infrequently. An interrupt lets the CPU do useful work until the device actually needs attention. A keyboard, for example, might generate a keypress once every few seconds. Polling that keyboard thousands of times per second burns CPU cycles for nothing. An interrupt fires only when the key is pressed.

What Is Pipelining?

Definition to say: Pipelining is a technique where multiple instructions are overlapped in execution by dividing the instruction cycle into discrete stages — typically fetch, decode, execute, memory access, and write-back.

Follow-up the examiner will ask: "Why does performance improve, and where can it still stall?"

Safe answer: Performance improves because the processor works on several instructions simultaneously — while one instruction is being executed, the next is being decoded, and the one after that is being fetched. Throughput increases even though each instruction takes the same number of cycles. Stalls occur when there is a dependency between instructions — for example, if instruction two needs the result of instruction one before instruction one has finished executing. That is a data hazard, and it forces the pipeline to wait.

What Is Cache Coherence?

Definition to say: Cache coherence is the problem of ensuring that multiple caches in a multiprocessor system all reflect a consistent view of shared memory at any given time.

Follow-up the examiner will ask: "Why does it matter, and what happens without it?"

Safe answer: In a dual-core system, Core 1 and Core 2 each have their own cache. If Core 1 writes a new value to a memory location and Core 2 still has the old value in its cache, the two cores are now working with different data. Without a coherence protocol — such as MESI — the system produces incorrect results silently. The examiner is looking for you to name the problem (stale data), the cause (independent caches), and the solution mechanism (invalidation or update protocols).

What Is Virtual Memory?

Definition to say: Virtual memory is a memory management technique that gives each process the illusion of having more memory than physically exists, by using secondary storage as an extension of RAM.

Follow-up the examiner will ask: "How is it different from physical memory, and what role does address translation play?"

Safe answer: Physical memory is the actual RAM installed in the machine. Virtual memory is the address space a process sees — it can be much larger. The Memory Management Unit (MMU) translates virtual addresses to physical addresses using a page table. When a process accesses a virtual address that is not currently in RAM, a page fault occurs and the OS loads the required page from disk. This is slower, but it allows the system to run programs larger than available RAM.

What Is Branch Prediction?

Definition to say: Branch prediction is a technique where the CPU guesses the outcome of a conditional branch instruction before it is resolved, so the pipeline can continue fetching and executing instructions without stalling.

Follow-up the examiner will ask: "Why does it help the pipeline, and what happens when the prediction is wrong?"

Safe answer: Without branch prediction, the pipeline would have to freeze every time it encountered a conditional jump and wait for the branch condition to be evaluated — potentially several cycles of wasted throughput. With prediction, the CPU speculatively executes the predicted path. If the prediction is correct, no time is lost. If it is wrong, the speculatively executed instructions are flushed and the pipeline restarts from the correct path — a penalty, but one that is still cheaper than always stalling.

What Is a Pipeline Hazard?

Definition to say: A pipeline hazard is any condition that prevents the next instruction in the pipeline from executing in its designated clock cycle.

Follow-up the examiner will ask: "What are the three types, and can you give an example of one?"

Safe answer: The three types are data hazards, control hazards, and structural hazards. A data hazard occurs when an instruction depends on the result of a previous instruction that has not yet completed — a read-after-write conflict is the classic example: instruction two tries to read a register that instruction one is still writing. A control hazard arises from branch instructions. A structural hazard occurs when two instructions need the same hardware resource at the same time.

Why Do We Use Cache in a Computer System?

Definition to say: Cache exists because main memory is too slow to keep pace with the CPU. By storing frequently accessed data closer to the processor, cache reduces the number of times the CPU has to wait for a memory fetch.

Follow-up the examiner will ask: "What happens when cache misses pile up?"

Safe answer: A cache miss forces the CPU to retrieve data from main memory, which takes orders of magnitude longer than a cache hit. When misses are frequent — because the working set of data is larger than the cache or because access patterns are unpredictable — the CPU spends most of its time waiting rather than computing. This is called thrashing in some contexts, and it erases the performance advantage that cache was supposed to provide.

What Is the Difference Between RAM and Cache?

Definition to say: RAM is the main memory of the system — larger, slower, and directly accessible by the CPU through the memory bus. Cache is a smaller, faster memory layer built into or very close to the CPU chip.

Follow-up the examiner will ask: "How do they differ in speed, size, and position in the memory hierarchy?"

Safe answer: Cache operates at speeds close to the CPU clock — nanosecond-level access. RAM is slower, typically tens of nanoseconds. Cache is measured in kilobytes to megabytes; RAM is measured in gigabytes. In the hierarchy, cache sits between the CPU registers and main memory. L1 cache is fastest and smallest, L2 is larger and slightly slower, L3 is larger still. RAM sits below all cache levels.

What Is the Memory Hierarchy?

Definition to say: The memory hierarchy is a structured arrangement of storage types, ordered from fastest and most expensive at the top to slowest and cheapest at the bottom, designed to balance speed and cost.

Follow-up the examiner will ask: "Where do registers and cache fit, and why is the hierarchy ordered this way?"

Safe answer: From top to bottom: CPU registers, L1 cache, L2 cache, L3 cache, main memory (RAM), secondary storage (SSD/HDD), and tertiary storage (tape, cloud). Registers are the fastest but hold only a handful of values. Each level down is slower but larger. The principle behind the hierarchy is locality of reference — programs tend to reuse the same data and instructions repeatedly, so keeping the most-used data in the fastest layer dramatically improves average access time. As Hennessy and Patterson describe it, the hierarchy exploits both temporal and spatial locality to approximate the speed of the fastest memory at the cost of the cheapest.

How to Answer Without Sounding Like You Are Reading

Start With the Line the Examiner Wants to Hear

For computer architecture viva questions, the examiner is listening for a clean definitional anchor in the first sentence. Everything after that is supporting evidence. If you lead with context or hedging — "well, it depends on how you define it" or "there are several ways to look at this" — you have already signaled uncertainty.

Take the architecture versus organization question. The answer that works is: "Architecture is what the system does; organization is how it does it." That is the line. Say it first. Then add the ISA-versus-hardware-implementation example. The examiner now knows you have the concept, and the example shows you can apply it. That sequence — anchor, then evidence — is what confident viva answers look like.

Say the Follow-Up Before They Have to Ask It

The fastest way to show control of a topic is to answer the follow-up inside your first response, before the examiner has to prompt it. This works because it demonstrates that you understand the concept well enough to know what the next question would be.

With virtual memory, for example: after defining it as the illusion of extended memory using secondary storage, add one sentence about address translation. "The MMU handles the mapping from virtual to physical addresses using a page table." That sentence answers the most likely follow-up — how does the system know where things actually are? — before it is asked. The examiner can still push further, but you have shown you are not just holding a definition at arm's length.

Stop Talking the Moment You Have Said Enough

Over-explaining is the most common viva mistake, and it is almost always self-inflicted. A student who has answered correctly keeps talking because silence feels like failure. In a viva, it is the opposite — silence after a complete answer signals confidence.

Interrupts are the clearest example. The complete answer is: interrupt, why it is used instead of polling, and one I/O device example. That is two sentences. A student who then adds a third sentence about interrupt priority levels, a fourth about nested interrupts, and a fifth about software versus hardware interrupts has now introduced three new topics the examiner can probe. Every extra sentence is a potential follow-up invitation. Say the definition, add the key feature, give the example, and stop.

Topics That Trigger Deeper Questioning

Pipelining Is Never Just Pipelining

In a computer organization and architecture viva, pipelining is almost never a single-question topic. Examiners use it as an entry point. The moment you define it, the follow-up chain is predictable: hazards, throughput versus latency, stall cycles, and forwarding or bypassing as mitigation strategies.

A concrete instruction-cycle example helps here. If you say "the five stages are fetch, decode, execute, memory, and write-back," the examiner can immediately ask what happens when stage three needs data that stage two has not finished producing. That is a data hazard. If you have already named hazards in your answer — even briefly — you have shortened the follow-up chain rather than extending it.

Cache Coherence Opens the Door to Multiprocessor Memory Talk

Cache coherence is one of those topics where the definition sounds manageable until the examiner asks the next question. Defining it as "keeping multiple caches consistent" is correct but thin. The follow-up almost always moves toward how coherence is maintained — invalidation versus update protocols, the MESI protocol, and what happens when two cores simultaneously try to write to the same address.

A two-core example makes the follow-up obvious and shows you have thought it through: Core 1 writes value 5 to address X. Core 2 still has value 3 in its cache for address X. Without a coherence protocol, Core 2 will use the wrong value. With an invalidation-based protocol, Core 2's cache line is marked invalid the moment Core 1 writes, forcing a fresh fetch. That example answers the "why does it matter" follow-up before it arrives.

Virtual Memory Is Where Address Translation Starts Multiplying Questions

Virtual memory almost always leads to a chain of related questions: page faults, paging versus segmentation, the role of the page table, and the difference between logical and physical address space. Operating systems textbooks treat these as a single interconnected cluster, and examiners do too.

The page-table example is the safest anchor. When the CPU generates a virtual address, the MMU looks it up in the page table to find the corresponding physical frame. If the page is not in RAM, a page fault triggers the OS to load it from disk. Knowing this chain — virtual address, page table lookup, physical frame or page fault — lets you follow the examiner wherever they go next in this topic cluster.

Definitions Students Blur in Viva

Architecture vs Organization Is Not a Trick Question

Students blur this because both words describe "how a computer works" in everyday language. In COA oral exam questions, the distinction is technical and specific. Architecture is the programmer-visible specification — the instruction set, data types, addressing modes, and register count. Organization is the physical realization — how the control unit is implemented, what memory technology is used, how the datapath is wired.

The clean divider the examiner is looking for is ISA versus implementation. Two processors can implement the same ISA with completely different internal organizations. That is the answer. Say it with that framing and the question is closed.

Physical Memory and Virtual Memory Are Not Interchangeable Words

Students who use these terms loosely get caught immediately in viva. Physical memory is the actual RAM chips — fixed, finite, directly addressable by hardware. Virtual memory is an abstraction created by the OS and MMU — it exists as an address space, not as physical chips.

The follow-up that exposes confusion is: "What does the OS do when a process accesses a virtual address that is not in RAM?" The correct answer involves page faults and the OS loading the required page from disk. A student who thinks virtual memory and RAM are the same thing cannot answer this at all.

Cache Is Not Just "Fast Memory"

"Fast memory" is the answer that gets students into trouble. It is not wrong, but it invites the follow-up: "What makes it useful, specifically?" The correct answer is locality of reference — the observation that programs tend to access the same data and instructions repeatedly over short time periods (temporal locality) and tend to access data near recently accessed data (spatial locality). Cache exploits both properties. A student who says "it is just faster than RAM" cannot explain why cache hit rates are high enough to make the design worthwhile.

Rapid Recall for Last-Minute Revision

The 1-Line Answer You Should Memorize for Each High-Yield Topic

Before the viva, compress each topic to a single sentence you can say without thinking. These are not full answers — they are the anchor line you say first, then build from.

  • Computer architecture: The logical design of a computer system, visible to the programmer through the ISA.
  • Computer organization: The physical implementation of that design in hardware.
  • Pipelining: Overlapping instruction stages so multiple instructions execute simultaneously.
  • Cache: Fast, small memory between the CPU and RAM that exploits locality of reference.
  • Virtual memory: An abstraction that gives processes more address space than physical RAM by using disk as overflow.
  • Interrupt: A signal that suspends the current process and transfers control to an ISR.
  • Branch prediction: The CPU guessing the next execution path before the branch condition is resolved.
  • Cache coherence: The guarantee that all caches in a multiprocessor system reflect the same memory state.
  • Pipeline hazard: Any condition that prevents the next instruction from executing on schedule.
  • Memory hierarchy: A ranked arrangement of storage types from fastest/smallest to slowest/largest.

The Questions Examiners Love Because They Expose Confusion Fast

Three topics consistently reveal shaky understanding: branch prediction, pipeline hazards, and cache coherence. They are high-yield not because they are obscure but because students half-know them — they can define the term but cannot explain the mechanism or the consequence of failure.

Branch prediction exposes confusion when the examiner asks what happens on a misprediction. Pipeline hazards expose confusion when the examiner asks which type causes which kind of stall. Cache coherence exposes confusion when the examiner asks how a coherence protocol decides whether to invalidate or update. Knowing the mechanism — not just the label — is what separates a complete answer from a half-answer on these three.

The Fastest Way to Revise the Night Before the Viva

Work in three passes. First pass: say the one-line definition for each topic out loud, from memory, without looking. If you hesitate, that is your weak spot — spend extra time there. Second pass: for each topic, say the definition and then add the most likely follow-up question and your answer to it. Do this out loud; silent revision does not replicate viva conditions. Third pass: focus only on the three confusion-prone topics — branch prediction, pipeline hazards, and cache coherence — and practice the mechanism, not just the label.

The goal is not to have memorized every possible answer. The goal is to know the next question too. If you can say the definition, add one key feature, and answer the most likely follow-up for each topic in this matrix, you are already ahead of most students walking into that room.

How Verve AI Can Help You Prepare for Your Interview With COA Viva Topics

The hardest part of COA viva prep is not finding the questions — it is practicing the follow-up chain under something that feels like real pressure. Reading answers off a page does not replicate the moment when an examiner says "and why?" and you have to produce a coherent second sentence without a script in front of you.

Verve AI Interview Copilot is built for exactly that gap. It listens in real-time to what you actually say during a practice session and responds to your specific answer — not to a generic prompt. If you define pipelining correctly but skip the hazards, Verve AI Interview Copilot follows up on the gap, the way an examiner would. If you over-explain interrupts and introduce topics you cannot support, it surfaces that pattern too. The session adapts to what you said, not to what the ideal answer was supposed to be. That is the difference between drilling a script and actually rehearsing a viva. Verve AI Interview Copilot runs mock interviews that mirror the definition-to-comparison-to-application chain described throughout this matrix, so by the time you sit in front of an actual examiner, the follow-up is not a surprise — it is a prompt you have already answered before.

Conclusion

The viva is not a test of how much you know. It is a test of how far your answer can go before it runs out. Every question in this matrix has a follow-up, and every follow-up has a mechanism behind it. You do not need to memorize every edge case in COA — you need to know the next question too.

Go through this matrix once, out loud, before your exam. Say the definition, add the key feature, answer the follow-up. If you can do that for each topic without hesitating, you are not just prepared for the first question — you are prepared for the one that comes after it.

VA

Verve AI

Interview Guidance

Ace your live interviews with AI support!

Get Started For Free

Available on Mac, Windows and iPhone