/*
* Copyright 1995, 1996 Perforce Software. All rights reserved.
*
* This file is part of Perforce - the FAST SCM System.
*/
/*
* Rpc.cc - remote procedure service
*
* The general flow of the client/server protocol:
*
* The client sends a message containing the user's request to the
* server and then dispatches, i.e. reads and executes requests from
* the server until the server sends the "release" message.
*
* The server initially dispatches, i.e. reads and executes a
* request from the user. The final act of the request should
* either be to release the client (send the "release" message),
* or send a message to the client instructing it to invoke
* (eventually) another server request. In the end, the final
* message sent to the client should be "release".
*/
# include <stdhdrs.h>
# include <signaler.h>
# include <debug.h>
# include <tunable.h>
# include <strbuf.h>
# include <strdict.h>
# include <strarray.h>
# include <strtable.h>
# include <error.h>
# include <errorlog.h>
# include <tracker.h>
# include <timer.h>
# include <md5.h>
# include <ticket.h>
# include <keepalive.h>
# include <netportparser.h>
# include <netconnect.h>
# include <netbuffer.h>
# include <rpcdebug.h>
# include <msgrpc.h>
# include <p4tags.h>
# ifdef USE_SSL
extern "C" { // OpenSSL
# include "openssl/bio.h"
# include "openssl/ssl.h"
# include "openssl/err.h"
}
# endif //USE_SSL
# include "enviro.h"
# include "filesys.h"
# include "pathsys.h"
# include "rpc.h"
# include "rpcbuffer.h"
# include "rpctrans.h"
# include "rpcdispatch.h"
# include "rpcdebug.h"
# include "rpcservice.h"
# include "netsslcredentials.h"
const char *RpcTypeNames[] = {
"", // RPC_CLIENT = 0,
"rmt: ", // RPC_REMOTE,
"px: ", // RPC_PX,
"", // RPC_RH,
"bgd: ", // RPC_BACKGROUND,
"pxc: ", // RPC_PXCLIENT,
"rpl: ", // RPC_RPL,
"bs: ", // RPC_BROKER,
"bc: ", // RPC_BROKER_TO_SERVER,
"rauth: ", // RPC_RMTAUTH,
"sbx: ", // RPC_SANDBOX,
"ukn: ", // RPC_UNKNOWN
"dmrpc: ", // RPC_DMRPC
0
};
RpcService::RpcService()
{
// Initialize our members.
dispatcher = new RpcDispatcher;
protoSendBuffer = new RpcSendBuffer;
endPoint = 0;
openFlag = RPC_CONNECT;
// Load up core dispatch routines
Dispatcher( rpcServices );
}
RpcService::~RpcService()
{
delete dispatcher;
delete endPoint;
delete protoSendBuffer;
}
/**
* RpcService::SetEndpoint
*
* @brief Creates an communication's endpoint for either listening or connecting.
*
* @param address for socket
* @param error structure pointer
* @see NetEndPoint::Create
*/
void
RpcService::SetEndpoint( const char *addr, Error *e )
{
delete endPoint;
endPoint = NetEndPoint::Create( addr, e );
}
/*
* RpcService::Dispatcher() - add another list of function to call on received RPC
*/
void
RpcService::Dispatcher( const RpcDispatch *dispatch )
{
dispatcher->Add( dispatch );
}
void
RpcService::Listen( Error *e )
{
// initialize listen
openFlag = RPC_LISTEN;
endPoint->Listen( e );
if( e->Test() )
{
// Failed open -- clear open flag so connect bails, too.
e->Set( MsgRpc::Listen ) << endPoint->GetAddress();
openFlag = RPC_NOOPEN;
return;
}
}
void
RpcService::GetHost( StrPtr *peerAddress, StrBuf &hostBuf, Error *e )
{
hostBuf.Clear();
NetEndPoint *ep = NetEndPoint::Create( peerAddress->Text(), e );
if( !e->Test() )
hostBuf.Set( ep->GetPrintableHost() );
delete ep;
return;
}
void
RpcService::ListenCheck( Error *e )
{
endPoint->ListenCheck( e );
}
int
RpcService::CheaterCheck( const char *port )
{
return endPoint->CheaterCheck( port );
}
void
RpcService::Unlisten()
{
if( endPoint )
endPoint->Unlisten();
}
void
RpcService::GetExpiration( StrBuf &buf )
{
if( endPoint )
endPoint->GetExpiration( buf );
}
void
RpcService::SetProtocol( const char *var, const StrRef &value )
{
protoSendBuffer->SetVar( StrRef( (char*)var ), value );
}
void
RpcService::SetProtocolV( const char *arg )
{
StrBuf s;
const char *p;
if( p = strchr( arg, '=' ) )
{
s.Set( arg, p - arg );
protoSendBuffer->SetVar( s, StrRef( (char *)p + 1 ) );
}
else
{
protoSendBuffer->SetVar( StrRef( (char *)arg ), StrRef::Null() );
}
}
void
RpcService::GetMyFingerprint( StrBuf &value )
{
endPoint->GetMyFingerprint( value );
}
int
RpcService::IsSingle()
{
return endPoint ? endPoint->IsSingle() : 0;
}
static void RpcCleanup( Rpc *r )
{
r->FlushTransport();
}
/*
* Rpc
*/
Rpc::Rpc( RpcService *s )
{
recvBuffering = 0;
service = s;
transport = 0;
forward = 0;
sendBuffer = new RpcSendBuffer;
recvBuffer = new RpcRecvBuffer;
protoDynamic = new StrBufDict;
duplexFsend = 0;
duplexFrecv = 0;
duplexRsend = 0;
duplexRrecv = 0;
dispatchDepth = 0;
endDispatch = 0;
protocolSent = 0;
protocolServer = 0;
rpc_hi_mark_rev =
rpc_hi_mark_fwd = p4tunable.Get( P4TUNE_RPC_HIMARK );
rpc_lo_mark = p4tunable.Get( P4TUNE_RPC_LOWMARK );
TrackStart();
timer = new Timer;
}
Rpc::~Rpc( void )
{
signaler.DeleteOnIntr( this );
Disconnect();
delete sendBuffer;
delete recvBuffer;
delete protoDynamic;
delete timer;
}
void
Rpc::DoHandshake( Error *e )
{
if( transport )
transport->DoHandshake(e);
}
void
Rpc::ClientMismatch( Error *e )
{
if( transport )
transport->ClientMismatch( e );
}
void
Rpc::SetProtocolDynamic( const char *var, const StrRef &value )
{
protoDynamic->ReplaceVar( var, value.Text() );
}
void
Rpc::ClearProtocolDynamic( const char *var )
{
protoDynamic->RemoveVar( var );
}
void
Rpc::Connect( Error *e )
{
// Don't allow double connects.
if( transport )
{
e->Set( MsgRpc::Reconn );
return;
}
// If an operation fails, cruft can be left in the sendBuffer.
// Since we may use this buffer over for the next connection, we
// need to clear it now.
sendBuffer->Clear();
duplexFsend = 0;
duplexFrecv = 0;
duplexRsend = 0;
duplexRrecv = 0;
dispatchDepth = 0;
endDispatch = 0;
protocolSent = 0;
re.Clear();
se.Clear();
// Create the transport layer,
// Cleanup code is in Disconnect().
NetTransport *t = 0;
switch( service->openFlag )
{
case RPC_LISTEN:
t = service->endPoint->Accept( e );
break;
case RPC_CONNECT:
t = service->endPoint->Connect( e );
break;
default:
e->Set( MsgRpc::Unconn );
}
if( e->Test() )
{
if( t )
delete t;
// connect errors leave send & receive errors
// so as to intercept Invoke() and Dispatch()
re = *e;
se = *e;
return;
}
// Pass transport onto the buffering routines.
transport = new RpcTransport( t );
// If rpc.himark tuned beyond net.bufsize, must tell NetBuffer.
transport->SetBufferSizes( rpc_hi_mark_fwd, rpc_hi_mark_rev );
if( service->openFlag == RPC_CONNECT )
signaler.OnIntr( (SignalFunc)RpcCleanup, this );
}
void
Rpc::GetEncryptionType( StrBuf &value )
{
transport->GetEncryptionType( value );
}
void
Rpc::GetPeerFingerprint( StrBuf &value )
{
transport->GetPeerFingerprint( value );
}
void
Rpc::GetExpiration(StrBuf &value)
{
if( service )
service->GetExpiration( value );
}
/*
* Rpc::SetHiMark() - move rpc_hi_mark beyond 2000 default
*
* If we get reliable send/receive TCP buffer sizes from the client
* (they often differ from each other and from the server's), we
* can feel confident about moving our himark off the default 2000.
*
* The thinking as of Feb 2009 is that one direction of a TCP
* connection can have data outstanding that amounts to the receiver's
* SO_RCVBUF. Many systems can take more -- we tried the sum of
* the sender's SO_SNDBUF and the receiver's SO_RCVBUF -- but some
* systems can't. Solaris, for example, starts to close its window
* when it receives more than SO_RCVBUF, keeps accepting data until
* the window is 0, and won't reopen it until the application reads
* all outstanding data. Weird and unfortunate, because we generally
* read in small chunks.
*
* Sometimes we're metering client->server data, because the
* server->client side is clogged with file data (sync uses
* InvokeDuplex()). Other times we're metering server->client data,
* because the client->server side is clogged (resolve, submit use
* InvokeDuplexRev()). Since the server and client may have different
* SO_RCVBUF sizes, and we want to make the most of the connection,
* we keep separate rpc_hi_mark_fwd and rpc_hi_mark_rev limits. We
* miss out when we meter data but neither side is clogged (diff,
* open, revert). Our high mark could be somewhere around the sum
* of the SO_RCVBUFs for the client and server in this case, since
* the metered data could wait in either buffer. But we're not
* yet that sophisticated.
*
* Note that TCP limits its window size to the minimum of the
* sender's SO_SNDBUF and receiver's SO_RCVBUF, and that limits the
* amount of data travelling on the network in that direction. If
* our himark were lower than the window size, we would artificially
* limit that amount of data further. Thus we'd like the himark
* high. Fortunately, we seem safe in making it SO_RCVBUF.
*
* A SO_SNDBUF larger than the SO_RCVBUF (and thus the TCP window)
* should have no impact. A SO_RCVBUF larger than the SO_SNDBUF
* (and thus a himark larger than the TCP window) is beneficial in
* the case where we're metering data on one side but the other
* side is not clogged. It allows more data to be in transit on
* the other side of the connection.
*/
void
Rpc::SetHiMark( int sndbuf, int rcvbuf )
{
// For testing and emergencies setting rpc.himark overrides all.
if( p4tunable.IsSet( P4TUNE_RPC_HIMARK ) )
return;
const int minHiMark = p4tunable.Get( P4TUNE_RPC_HIMARK );
rpc_hi_mark_fwd = transport->GetRecvBuffering();
rpc_hi_mark_rev = rcvbuf;
// Since Invoke() needs to Dispatch() _before_ we hit
// the himark, we reduce the himark by lomark, and hope that
// no single duplex message is larger than lomark.
rpc_hi_mark_fwd -= rpc_lo_mark;
rpc_hi_mark_rev -= rpc_lo_mark;
if( rpc_hi_mark_fwd < minHiMark ) rpc_hi_mark_fwd = minHiMark;
if( rpc_hi_mark_rev < minHiMark ) rpc_hi_mark_rev = minHiMark;
// Put a stake through the heart of himark hanging: set our
// NetBuffer buffer sizes to handling any outstanding data,
// in case the lying cheating OS (linux, solaris) renegs
// on its reported TCP send or receive space.
transport->SetBufferSizes( rpc_hi_mark_fwd, rpc_hi_mark_rev );
RPC_DBG_PRINTF( DEBUG_CONNECT,
"Rpc himark: snd+rcv server %d+%d client %d+%d = %d/%d",
transport->GetSendBuffering(),
transport->GetRecvBuffering(),
sndbuf, rcvbuf,
rpc_hi_mark_fwd,
rpc_hi_mark_rev );
}
void
Rpc::Disconnect()
{
if( !transport )
return;
transport->Flush( &se );
transport->Close();
delete transport;
transport = 0;
}
StrPtr *
Rpc::GetAddress( int raf_flags )
{
return transport ? transport->GetAddress( raf_flags ) : 0;
}
StrPtr *
Rpc::GetPeerAddress( int raf_flags )
{
return transport ? transport->GetPeerAddress( raf_flags ) : 0;
}
int
Rpc::GetPortNum()
{
return transport ? transport->GetPortNum() : -1;
}
bool
Rpc::IsSockIPv6()
{
return transport ? transport->IsSockIPv6() : false;
}
StrBuf *
Rpc::MakeVar( const char *var )
{
return sendBuffer->MakeVar( StrRef( (char *)var ) );
}
void
Rpc::VSetVar( const StrPtr &var, const StrPtr &value )
{
sendBuffer->SetVar( var, value );
}
void
Rpc::Release()
{
this->Invoke( P4Tag::p_release );
}
void
Rpc::ReleaseFinal()
{
this->Invoke( P4Tag::p_release2 );
}
StrPtr *
Rpc::VGetVar( const StrPtr &var )
{
return recvBuffer->GetVar( var );
}
int
Rpc::VGetVarX( int x, StrRef &var, StrRef &val )
{
return recvBuffer->GetVar( x, var, val );
}
void
Rpc::VRemoveVar( const StrPtr &var )
{
recvBuffer->RemoveVar( var );
}
void
Rpc::VClear()
{
sendBuffer->Clear();
}
StrPtr *
Rpc::GetArgi( int i )
{
return i < recvBuffer->GetArgc() ? &recvBuffer->GetArgv()[i] : 0;
}
StrPtr *
Rpc::GetArgi( int i, Error *e )
{
StrPtr *s = GetArgi( i );
if( !s )
e->Set( MsgRpc::NoPoss );
return s;
}
int
Rpc::GetArgc()
{
return recvBuffer->GetArgc();
}
StrPtr *
Rpc::GetArgv()
{
return recvBuffer->GetArgv();
}
void
Rpc::CopyVars()
{
sendBuffer->CopyVars( recvBuffer );
}
KeepAlive *
Rpc::GetKeepAlive()
{
return transport;
}
void
Rpc::SetBreak( KeepAlive *breakCallback )
{
if( transport )
transport->SetBreak( breakCallback );
}
/*
* Invoke()
* InvokeDuplex()
* InvokeDuplexRev()
* FlushDuplex()
* InvokeFlush()
* InvokeOne()
* GotFlushed()
*/
void
Rpc::Invoke( const char *opName )
{
// If there are any outstanding requests sent via InvokeDuplexRev,
// we have to assume the return pipe is clogged, and thus meter all
// data sent.
if( duplexRrecv )
InvokeDuplex( opName );
else
InvokeOne( opName );
}
void
Rpc::InvokeDuplex( const char *opName )
{
// This data makes a loop: meter it, so we know to dispatch for it.
int sz = InvokeOne( opName );
duplexFrecv += sz;
duplexFsend += sz;
Dispatch( DfDuplex, service->dispatcher );
}
void
Rpc::InvokeDuplexRev( const char *opName )
{
// This data only goes forward, but causes the return pipe to fill.
// Meter it, but also turn on duplexRrecv so that all Invokes are
// metered until this message is flushed..
++duplexRrecv;
++duplexRsend;
InvokeDuplex( opName );
}
void
Rpc::FlushDuplex()
{
// Flush any outstanding duplex data in the pipe.
// Do nothing if there is no duplex data expected back
if( duplexFrecv > 0 )
{
// Proxies and brokers expect a flush1 message
// (with himark = 0) when flushing the channel,
// so we have to tell Dispatch() to send one.
// DfFlush moves the lomark to 0, and ensuring
// duplexFsend > 0 forces a flush1 message.
++duplexFrecv;
++duplexFsend;
Dispatch( DfFlush, service->dispatcher );
}
}
void
Rpc::InvokeOver( const char *opName )
{
// This data makes a loop, but we specifically don't flush yet:
// we return so that the outer Dispatch() can take over.
// Otherwise, we'd needlessly nest Dispatch() calls, and in
// one case ('submit') we'd try to nest too deeply.
int sz = InvokeOne( opName );
duplexFrecv += sz;
duplexFsend += sz;
Dispatch( DfOver, service->dispatcher );
}
int
Rpc::InvokeOne( const char *opName )
{
// Don't pile errors
if( se.Test() || re.Test() || !transport )
{
sendBuffer->Clear();
return 0;
}
// Send off protocol if not yet sent.
// If we're using RpcForward we want to allow it
// to send its own protocol message.
if( !protocolSent && strcmp( opName, P4Tag::p_protocol ) )
{
RpcSendBuffer buf;
int sz = transport->GetSendBuffering();
int rz = transport->GetRecvBuffering();
buf.CopyBuffer( service->protoSendBuffer );
int i = 0;
StrRef var, val;
while( protoDynamic->GetVar( i++, var, val ) )
buf.SetVar( var, val );
buf.SetVar( StrRef( P4Tag::v_sndbuf ), StrNum( sz ) );
buf.SetVar( StrRef( P4Tag::v_rcvbuf ), StrNum( rz ) );
buf.SetVar( StrRef( P4Tag::v_func ), StrRef( P4Tag::p_protocol ) );
RPC_DBG_PRINT( DEBUG_FUNCTION,
"Rpc invoking protocol" );
timer->Start();
transport->Send( buf.GetBuffer(), &re, &se );
sendTime += timer->Time();
}
protocolSent = 1;
// Set func=opName variable.
SetVar( P4Tag::v_func, opName );
// Tracking
RPC_DBG_PRINTF( DEBUG_FUNCTION,
"Rpc invoking %s", opName );
// Send the buffer to peer
timer->Start();
transport->Send( sendBuffer->GetBuffer(), &re, &se );
// time tracking
sendTime += timer->Time();
if( se.Test() )
return 0;
// Clear for next call.
// Get size of buffer so we know how full the pipe is.
// We must include RpcTransport's overhead.
int sz = sendBuffer->GetBufferSize() + transport->SendOverhead();
sendBuffer->Clear();
// tracking
sendCount++;
sendBytes += sz;
return sz;
}
void
Rpc::GotFlushed()
{
StrPtr *fseq = GetVar( P4Tag::v_fseq );
StrPtr *rseq = GetVar( P4Tag::v_rseq );
if( fseq )
duplexFrecv -= fseq->Atoi();
if( rseq )
duplexRrecv -= rseq->Atoi();
}
/*
* Rpc::Dispatch() - dispatch incoming RPC's until 'release' received
* Rpc::DispatchOne() - just dispatch from the current buffer
*/
void
Rpc::Dispatch( DispatchFlag flag, RpcDispatcher *dispatcher )
{
// Don't nest more than once: we only allow a
// Dispatch()/InvokeDuplex()/FlushDuplex() combo.
if( dispatchDepth > 1 )
return;
++dispatchDepth;
RPC_DBG_PRINTF( DEBUG_FLOW,
">>> Dispatch(%d) %d/%d %d/%d %d",
dispatchDepth, duplexFsend, duplexFrecv,
duplexRsend, duplexRrecv, flag );
// Use server's recv buffer size as himark for InvokeDuplex()
// Use client's recv buffer size as himark for InvokeDuplexRev()
int loMark = rpc_lo_mark;
int hiMark = duplexRrecv ? rpc_hi_mark_rev : rpc_hi_mark_fwd;
// Flushing means eveything goes.
// Complete: lo = N/A, hi = N/A
// Duplex: lo = LO, hi = HI
// Flush: lo = 0, hi = 0
// Over: lo = 0, hi = HI
if( flag != DfDuplex ) loMark = 0;
if( flag == DfFlush ) hiMark = 0;
// Push the recvBuffer, in case we are nesting Dispatch().
RpcRecvBuffer *savRecvBuffer = recvBuffer;
recvBuffer = 0;
// Receive (dispatching) until told to stop.
while( !endDispatch )
{
if( re.Test() && ( !transport || !transport->RecvReady() ) )
break;
// If more than loMark in pipe, send a marker.
// We flush always if Dispatch() called from FlushDuplex()
// or if Invoking from such a Dispatch().
if( duplexFsend > loMark && !se.Test() )
{
// Send a flush1 through for metering flow control.
// Note: hardcode the size of the flush1 message
// As of 2008.1 its 46 bytes (round up to 60).
const int flushMessageSize = 60;
RPC_DBG_PRINTF( DEBUG_FLOW,
"Rpc flush %d bytes", duplexFsend );
SetVar( P4Tag::v_himark, loMark ? hiMark : 0 );
duplexFrecv += flushMessageSize;
duplexFsend += flushMessageSize;
if( duplexFsend ) SetVar( P4Tag::v_fseq, duplexFsend );
if( duplexRsend ) SetVar( P4Tag::v_rseq, duplexRsend );
duplexFsend = 0;
duplexRsend = 0;
InvokeOne( P4Tag::p_flush1 );
}
// If top level Dispatch(), go forever.
// If 2nd level Dispatch(), get below hiMark.
// If flushing, go until all flush2's received.
// If error sending, go until receive error.
else if( flag == DfComplete ||
flag == DfDuplex && duplexFrecv > hiMark ||
flag == DfFlush && duplexFrecv ||
se.Test() )
{
if( !recvBuffer )
recvBuffer = new RpcRecvBuffer;
DispatchOne( dispatcher );
}
else break;
}
// Pop recvBuffer.
delete recvBuffer;
recvBuffer = savRecvBuffer;
RPC_DBG_PRINTF( DEBUG_FLOW,
"<<< Dispatch(%d) %d/%d %d/%d %d",
dispatchDepth, duplexFsend, duplexFrecv,
duplexRsend, duplexRrecv, flag );
if( !--dispatchDepth )
endDispatch = 0;
}
void
Rpc::DispatchOne( RpcDispatcher *dispatcher )
{
StrPtr *func;
// Flush any data due the other end: we can't expect
// anything if the other end is awaiting us!
// Note that a broken send pipe doesn't stop dispatching:
// we want what's in the receive pipe (they may be important
// acks), so we read until the receive pipe is broken too.
// Receive sender's buffer and then parse the variables out.
timer->Start();
int sz = transport->Receive( recvBuffer->GetBuffer(), &re, &se );
// time tracking
recvTime += timer->Time();
if( sz <= 0 )
{
// EOF doesn't set re, so we will.
if( !re.Test() )
re.Set( MsgRpc::Closed );
return;
}
// tracking
recvCount++;
recvBytes += recvBuffer->GetBufferSize();
Error e;
recvBuffer->Parse( &e );
if( e.Test() )
{
re = e;
return;
}
// Find the function to dispatch. The protocol mandates that
// this must be set: how else do we know what to do with the
// buffer?
func = GetVar( P4Tag::v_func, &e );
if( e.Test() )
{
re = e;
return;
}
RPC_DBG_PRINTF( DEBUG_FUNCTION,
"Rpc dispatch %s", func->Text() );
// Find the registered function as given with 'func'.
// If no such function is found, call 'funcHandler'.
// If any error occurs, call 'errorHandler'.
// If no 'errorHandler' just report the error and return.
ue.Clear();
const RpcDispatch *disp = dispatcher->Find( func->Text() );
if( !disp && !( disp = dispatcher->Find( P4Tag::p_funcHandler ) ) )
{
ue.Set( MsgRpc::UnReg ) << *func;
goto error;
}
// Invoke requested function.
(*disp->function)( this, &ue );
// If an error occurred, we'll call the caller's errorHandle
// function. If the error was fatal, we'll tack on some tracing
// information.
if( !ue.Test() )
return;
if( ue.IsFatal() )
ue.Set( MsgRpc::Operat ) << disp->opName;
error:
// If a user error occurred, invoke errorHandler to deal with it.
// In this case, the Error is acutally passed in to the dispatched
// function, rather than being just an output parameter.
if( disp = dispatcher->Find( P4Tag::p_errorHandler ) )
{
(*disp->function)( this, &ue );
return;
}
AssertLog.Report( &ue );
}
void
Rpc::Loopback( Error *e )
{
recvBuffer->CopyBuffer( sendBuffer->GetBuffer() );
recvBuffer->Parse( e );
sendBuffer->Clear();
}
/*
* Rpc::StartCompression() -- initiate full link compression
* Rpc::GotRecvCompress() -- turn on recv half compression
* Rpc::GotSendCompress() -- turn on send half compression
*
* Notes on compression.
*
* Both ends must support the compression protocol. Use "client" >= 6
* or "server2" >= "6" ? protocol level to be sure.
*
* Compression can be turned on any time; in practice, it happens
* after protocol exchange has happened (otherwise, how do you know?).
* To orchestrate this, the follow sequence is obeyed:
*
* 1. StartCompression() send "compress1" to the other end, and then turns
* on send compression. All data sent afterwards are compressed.
*
* 2. When the other end receives "compress1", it sends back "compress2"
* and then turns on both send and receieve compression.
*
* 3. When this end receives "compress2", it turns on receive compression.
*/
void
Rpc::StartCompression( Error *e )
{
// send the "compress1" flag, then compress the send link.
// When we get "compress2", we'll compress the recv link.
Invoke( P4Tag::p_compress1 );
transport->SendCompression( e );
}
void
Rpc::GotSendCompressed( Error *e )
{
transport->SendCompression( e );
}
void
Rpc::GotRecvCompressed( Error *e )
{
transport->RecvCompression( e );
}
void
Rpc::FlushTransport()
{
if( transport )
transport->Flush( &se );
}
int
Rpc::GetRecvBuffering()
{
if( transport )
return transport->GetRecvBuffering();
return 0;
}
/*
* Performance tracking
*
* Rpc::TrackStart() - reset tracking counters
* Rpc::Trackable() - is any track data interesting?
* Rpc::TrackReport() - report interesting track data
*/
void
Rpc::TrackStart()
{
sendCount = 0;
sendBytes = 0;
recvCount = 0;
recvBytes = 0;
sendTime = 0;
recvTime = 0;
}
int
Rpc::Trackable( int level )
{
Tracker t( level );
return t.Over( TT_RPC_ERRORS, se.Test() || re.Test() ) ||
t.Over( TT_RPC_MSGS, sendCount + recvCount ) ||
t.Over( TT_RPC_MBYTES, ( sendBytes + recvBytes ) / 1024 / 1024 );
}
void
Rpc::TrackReport( int level, StrBuf &out )
{
if( !Trackable( level ) )
return;
out
<< "--- rpc msgs/size in+out "
<< recvCount << "+"
<< sendCount << "/"
<< recvBytes / 1024 / 1024 << "mb+"
<< sendBytes / 1024 / 1024 << "mb "
<< "himarks "
<< rpc_hi_mark_fwd << "/"
<< rpc_hi_mark_rev << " snd/rcv "
<< StrMs( sendTime ) << "s/"
<< StrMs( recvTime ) << "s\n";
if( !se.Test() && !re.Test() )
return;
out << "--- rpc ";
if( se.Test() ) out << "send ";
if( re.Test() ) out << "receive ";
out << "errors, duplexing F/R "
<< duplexFrecv << "/"
<< duplexRrecv << "\n";
}
void
Rpc::GetTrack( int level, RpcTrack *track )
{
track->trackable = Trackable( level );
if( !track->trackable )
return;
ForceGetTrack( track );
}
void
Rpc::ForceGetTrack( RpcTrack *track )
{
if( !track )
return;
track->recvCount = recvCount;
track->sendCount = sendCount;
track->recvBytes = recvBytes;
track->sendBytes = sendBytes;
track->rpc_hi_mark_fwd = rpc_hi_mark_fwd;
track->rpc_hi_mark_rev = rpc_hi_mark_rev;
track->recvTime = recvTime;
track->sendTime = sendTime;
}
void
Rpc::CheckKnownHost( Error *e, const StrRef & trustfile )
{
StrBuf pubkey;
StrPtr *peer;
GetPeerFingerprint( pubkey );
// if not ssl we are done
if( !pubkey.Length() )
return;
peer = GetPeerAddress( RAF_PORT );
RPC_DBG_PRINTF( DEBUG_CONNECT,
"Checking host %s pubkey %s",
peer->Text(), pubkey.Text() );
StrRef dummyuser( "**++**" );
StrRef altuser( "++++++" );
StrBuf trustkey;
int doreplace = 0;
char *keystr;
{
Ticket hostfile( &trustfile );
keystr = hostfile.GetTicket( *peer, dummyuser );
if( keystr )
{
if( pubkey == keystr )
return;
trustkey.Set( keystr );
}
}
{
Ticket hostfile( &trustfile );
keystr = hostfile.GetTicket( *peer, altuser );
if( keystr && pubkey == keystr )
doreplace = 1;
}
if( doreplace )
{
{
Ticket hostfile( &trustfile );
hostfile.ReplaceTicket( *peer, dummyuser, pubkey, e );
}
if( !e->Test() )
{
Ticket hostfile( &trustfile );
hostfile.DeleteTicket( *peer, altuser, e );
}
return;
}
e->Set( trustkey.Length() ?
MsgRpc::HostKeyMismatch : MsgRpc::HostKeyUnknown );
*e << *peer;
*e << pubkey;
}
void
RpcUtility::Generate(RpcUtilityType type,
Error *e )
{
# ifdef USE_SSL
NetSslCredentials credentials;
switch( type )
{
case Generate_Credentials:
credentials.GenerateCredentials( e );
break;
case Generate_Fingerprint:
credentials.ReadCredentials( e );
if ( !e->Test() )
{
const StrPtr *fingerprint = credentials.GetFingerprint();
if( fingerprint )
printf("Fingerprint: %s\n", fingerprint->Text());
}
break;
}
return;
# else
e->Set( MsgRpc::SslNoSsl );
# endif //USE_SSL
}