Remote procedure call Client/server architecture

18 Slides35.50 KB

Remote procedure call Client/server architecture

Client-server architecture Client sends a request, server replies w. a response – Interaction fits many applications – Naturally extends to distributed computing Why do people like client/server architecture? – Provides fault isolation between modules – Scalable performance (multiple servers) – Central server: Easy to manage Easy to program

Remote procedure call A remote procedure call makes a call to a remote service look like a local call – RPC makes transparent whether server is local or remote – RPC allows applications to become distributed transparently – RPC makes architecture of remote machine transparent

Developing with RPC 1. Define APIs between modules Split application based on function, ease of development, and ease of maintenance Don’t worry whether modules run locally or remotely 2. Decide what runs locally and remotely Decision may even be at run-time 3. Make APIs bullet proof Deal with partial failures

Stubs: obtaining transparency Compiler generates from API stubs for a procedure on the client and server Client stub – – – – Marshals arguments into machine-independent format Sends request to server Waits for response Unmarshals result and returns to caller Server stub – Unmarshals arguments and builds stack frame – Calls procedure – Server stub marshalls results and sends reply

RPC vs. LPC 4 properties of distributed computing that make achieving transparency difficult: – – – – Partial failures Concurrency Latency Memory access

Partial failures In local computing: – if machine fails, application fails In distributed computing: – if a machine fails, part of application fails – one cannot tell the difference between a machine failure and network failure How to make partial failures transparent to client?

Strawman solution Make remote behavior identical to local behavior: – Every partial failure results in complete failure You abort and reboot the whole system – You wait patiently until system is repaired Problems with this solution: – Many catastrophic failures – Clients block for long periods System might not be able to recover

Real solution: break transparency Possible semantics for RPC: – Exactly-once Impossible in practice – More than once: Only for idempotent operations – At most once Zero, don’t know, or once – Zero or once Transactional semantics At-most-once most practical – But different from LPC

Where RPC transparency breaks True concurrency – Clients run truely concurrently client() { if (exists(file)) if (!remove(file)) abort(“remove failed?”); } RPC latency is high – Orders of magnitude larger than LPC’s Memory access – Pointers are local to an address space

Summary: expose remoteness to client Expose RPC properties to client, since you cannot hide them – Consider: E-commerce application Application writers have to decide how to deal with partial failures

RPC implementation Stub compiler – Generates stubs for client and server – Language dependent – Compile into machine-independent format E.g., XDR – Format describes types and values RPC protocol RPC transport

RPC protocol Guarantee at-most-once semantics by tagging requests and response with a nonce RPC request header: – Request nonce – Service Identifier – Call identifier Protocol: – Client resends after time out – Server maintains table of nonces and replies

RPC transport Use reliable transport layer – Flow control – Congestion control – Reliable message transfer Combine RPC and transport protocol – Reduce number of messages RPC response can also function as acknowledgement for message transport protocol

Example: NFSv2 Old file system API used in distributed computing – – – – No current directory No file descriptors No streams (unnamed pipes) No close Inodes are named by file handles: – Opaque to client – Servers stores inode and device number Open() RPC translates names to file handles

NFS implementation Uses RPC – Marshals data structures in XDR format Pointers for directories – Generates stubs at client and server State-less protocol – Client maintains no state (No close() RPC!) – Server has no crucial runtime state File handles contain info to recover from server reboots NFS semantics: – Soft mounting: zero or more – Hard mounting: one or more

Implementation of write Semantics of writes: – Write-through (no close() RPC) What are possible outcomes after server failure: – Failed (stale file handle, I/O error) – Succeeded (NFS OK) – Don’t know (Client cannot tell!) Rename() example – Might happen multiple times

NFS RPC summary Remote service is not transparent – API is different – Failure semantics are different Not specified File systems semantics are unspecified! – Concurrency semantics awkard

Back to top button