/* * Copyright (C) 2020 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #define LOG_TAG "RpcState" #include "RpcState.h" #include #include #include #include #include #include #include "Debug.h" #include "RpcWireFormat.h" #include #include namespace android { using base::ScopeGuard; #if RPC_FLAKE_PRONE void rpcMaybeWaitToFlake() { [[clang::no_destroy]] static std::random_device r; [[clang::no_destroy]] static std::mutex m; unsigned num; { std::lock_guard lock(m); num = r(); } if (num % 10 == 0) usleep(num % 1000); } #endif RpcState::RpcState() {} RpcState::~RpcState() {} status_t RpcState::onBinderLeaving(const sp& session, const sp& binder, uint64_t* outAddress) { bool isRemote = binder->remoteBinder(); bool isRpc = isRemote && binder->remoteBinder()->isRpcBinder(); if (isRpc && binder->remoteBinder()->getPrivateAccessor().rpcSession() != session) { // We need to be able to send instructions over the socket for how to // connect to a different server, and we also need to let the host // process know that this is happening. ALOGE("Cannot send binder from unrelated binder RPC session."); return INVALID_OPERATION; } if (isRemote && !isRpc) { // Without additional work, this would have the effect of using this // process to proxy calls from the socket over to the other process, and // it would make those calls look like they come from us (not over the // sockets). In order to make this work transparently like binder, we // would instead need to send instructions over the socket for how to // connect to the host process, and we also need to let the host process // know this was happening. ALOGE("Cannot send binder proxy %p over sockets", binder.get()); return INVALID_OPERATION; } std::lock_guard _l(mNodeMutex); if (mTerminated) return DEAD_OBJECT; // TODO(b/182939933): maybe move address out of BpBinder, and keep binder->address map // in RpcState for (auto& [addr, node] : mNodeForAddress) { if (binder == node.binder) { if (isRpc) { // check integrity of data structure uint64_t actualAddr = binder->remoteBinder()->getPrivateAccessor().rpcAddress(); LOG_ALWAYS_FATAL_IF(addr != actualAddr, "Address mismatch %" PRIu64 " vs %" PRIu64, addr, actualAddr); } node.timesSent++; node.sentRef = binder; // might already be set *outAddress = addr; return OK; } } LOG_ALWAYS_FATAL_IF(isRpc, "RPC binder must have known address at this point"); bool forServer = session->server() != nullptr; // arbitrary limit for maximum number of nodes in a process (otherwise we // might run out of addresses) if (mNodeForAddress.size() > 100000) { return NO_MEMORY; } while (true) { RpcWireAddress address{ .options = RPC_WIRE_ADDRESS_OPTION_CREATED, .address = mNextId, }; if (forServer) { address.options |= RPC_WIRE_ADDRESS_OPTION_FOR_SERVER; } // avoid ubsan abort if (mNextId >= std::numeric_limits::max()) { mNextId = 0; } else { mNextId++; } auto&& [it, inserted] = mNodeForAddress.insert({RpcWireAddress::toRaw(address), BinderNode{ .binder = binder, .sentRef = binder, .timesSent = 1, }}); if (inserted) { *outAddress = it->first; return OK; } } } status_t RpcState::onBinderEntering(const sp& session, uint64_t address, sp* out) { // ensure that: if we want to use addresses for something else in the future (for // instance, allowing transitive binder sends), that we don't accidentally // send those addresses to old server. Accidentally ignoring this in that // case and considering the binder to be recognized could cause this // process to accidentally proxy transactions for that binder. Of course, // if we communicate with a binder, it could always be proxying // information. However, we want to make sure that isn't done on accident // by a client. RpcWireAddress addr = RpcWireAddress::fromRaw(address); constexpr uint32_t kKnownOptions = RPC_WIRE_ADDRESS_OPTION_CREATED | RPC_WIRE_ADDRESS_OPTION_FOR_SERVER; if (addr.options & ~kKnownOptions) { ALOGE("Address is of an unknown type, rejecting: %" PRIu64, address); return BAD_VALUE; } std::lock_guard _l(mNodeMutex); if (mTerminated) return DEAD_OBJECT; if (auto it = mNodeForAddress.find(address); it != mNodeForAddress.end()) { *out = it->second.binder.promote(); // implicitly have strong RPC refcount, since we received this binder it->second.timesRecd++; return OK; } // we don't know about this binder, so the other side of the connection // should have created it. if ((addr.options & RPC_WIRE_ADDRESS_OPTION_FOR_SERVER) == !!session->server()) { ALOGE("Server received unrecognized address which we should own the creation of %" PRIu64, address); return BAD_VALUE; } auto&& [it, inserted] = mNodeForAddress.insert({address, BinderNode{}}); LOG_ALWAYS_FATAL_IF(!inserted, "Failed to insert binder when creating proxy"); // Currently, all binders are assumed to be part of the same session (no // device global binders in the RPC world). it->second.binder = *out = BpBinder::PrivateAccessor::create(session, it->first); it->second.timesRecd = 1; return OK; } status_t RpcState::flushExcessBinderRefs(const sp& session, uint64_t address, const sp& binder) { // We can flush all references when the binder is destroyed. No need to send // extra reference counting packets now. if (binder->remoteBinder()) return OK; std::unique_lock _l(mNodeMutex); if (mTerminated) return DEAD_OBJECT; auto it = mNodeForAddress.find(address); LOG_ALWAYS_FATAL_IF(it == mNodeForAddress.end(), "Can't be deleted while we hold sp<>"); LOG_ALWAYS_FATAL_IF(it->second.binder != binder, "Caller of flushExcessBinderRefs using inconsistent arguments"); LOG_ALWAYS_FATAL_IF(it->second.timesSent <= 0, "Local binder must have been sent %p", binder.get()); // For a local binder, we only need to know that we sent it. Now that we // have an sp<> for this call, we don't need anything more. If the other // process is done with this binder, it needs to know we received the // refcount associated with this call, so we can acknowledge that we // received it. Once (or if) it has no other refcounts, it would reply with // its own decStrong so that it could be removed from this session. if (it->second.timesRecd != 0) { _l.unlock(); return session->sendDecStrongToTarget(address, 0); } return OK; } size_t RpcState::countBinders() { std::lock_guard _l(mNodeMutex); return mNodeForAddress.size(); } void RpcState::dump() { std::lock_guard _l(mNodeMutex); dumpLocked(); } void RpcState::clear() { std::unique_lock _l(mNodeMutex); if (mTerminated) { LOG_ALWAYS_FATAL_IF(!mNodeForAddress.empty(), "New state should be impossible after terminating!"); return; } if (SHOULD_LOG_RPC_DETAIL) { ALOGE("RpcState::clear()"); dumpLocked(); } // if the destructor of a binder object makes another RPC call, then calling // decStrong could deadlock. So, we must hold onto these binders until // mNodeMutex is no longer taken. std::vector> tempHoldBinder; mTerminated = true; for (auto& [address, node] : mNodeForAddress) { sp binder = node.binder.promote(); LOG_ALWAYS_FATAL_IF(binder == nullptr, "Binder %p expected to be owned.", binder.get()); if (node.sentRef != nullptr) { tempHoldBinder.push_back(node.sentRef); } } mNodeForAddress.clear(); _l.unlock(); tempHoldBinder.clear(); // explicit } void RpcState::dumpLocked() { ALOGE("DUMP OF RpcState %p", this); ALOGE("DUMP OF RpcState (%zu nodes)", mNodeForAddress.size()); for (const auto& [address, node] : mNodeForAddress) { sp binder = node.binder.promote(); const char* desc; if (binder) { if (binder->remoteBinder()) { if (binder->remoteBinder()->isRpcBinder()) { desc = "(rpc binder proxy)"; } else { desc = "(binder proxy)"; } } else { desc = "(local binder)"; } } else { desc = "(null)"; } ALOGE("- BINDER NODE: %p times sent:%zu times recd: %zu a: %" PRIu64 " type: %s", node.binder.unsafe_get(), node.timesSent, node.timesRecd, address, desc); } ALOGE("END DUMP OF RpcState"); } RpcState::CommandData::CommandData(size_t size) : mSize(size) { // The maximum size for regular binder is 1MB for all concurrent // transactions. A very small proportion of transactions are even // larger than a page, but we need to avoid allocating too much // data on behalf of an arbitrary client, or we could risk being in // a position where a single additional allocation could run out of // memory. // // Note, this limit may not reflect the total amount of data allocated for a // transaction (in some cases, additional fixed size amounts are added), // though for rough consistency, we should avoid cases where this data type // is used for multiple dynamic allocations for a single transaction. constexpr size_t kMaxTransactionAllocation = 100 * 1000; if (size == 0) return; if (size > kMaxTransactionAllocation) { ALOGW("Transaction requested too much data allocation %zu", size); return; } mData.reset(new (std::nothrow) uint8_t[size]); } status_t RpcState::rpcSend(const sp& connection, const sp& session, const char* what, iovec* iovs, int niovs, const std::function& altPoll) { for (int i = 0; i < niovs; i++) { LOG_RPC_DETAIL("Sending %s (part %d of %d) on RpcTransport %p: %s", what, i + 1, niovs, connection->rpcTransport.get(), android::base::HexString(iovs[i].iov_base, iovs[i].iov_len).c_str()); } if (status_t status = connection->rpcTransport->interruptableWriteFully(session->mShutdownTrigger.get(), iovs, niovs, altPoll); status != OK) { LOG_RPC_DETAIL("Failed to write %s (%d iovs) on RpcTransport %p, error: %s", what, niovs, connection->rpcTransport.get(), statusToString(status).c_str()); (void)session->shutdownAndWait(false); return status; } return OK; } status_t RpcState::rpcRec(const sp& connection, const sp& session, const char* what, iovec* iovs, int niovs) { if (status_t status = connection->rpcTransport->interruptableReadFully(session->mShutdownTrigger.get(), iovs, niovs, {}); status != OK) { LOG_RPC_DETAIL("Failed to read %s (%d iovs) on RpcTransport %p, error: %s", what, niovs, connection->rpcTransport.get(), statusToString(status).c_str()); (void)session->shutdownAndWait(false); return status; } for (int i = 0; i < niovs; i++) { LOG_RPC_DETAIL("Received %s (part %d of %d) on RpcTransport %p: %s", what, i + 1, niovs, connection->rpcTransport.get(), android::base::HexString(iovs[i].iov_base, iovs[i].iov_len).c_str()); } return OK; } status_t RpcState::readNewSessionResponse(const sp& connection, const sp& session, uint32_t* version) { RpcNewSessionResponse response; iovec iov{&response, sizeof(response)}; if (status_t status = rpcRec(connection, session, "new session response", &iov, 1); status != OK) { return status; } *version = response.version; return OK; } status_t RpcState::sendConnectionInit(const sp& connection, const sp& session) { RpcOutgoingConnectionInit init{ .msg = RPC_CONNECTION_INIT_OKAY, }; iovec iov{&init, sizeof(init)}; return rpcSend(connection, session, "connection init", &iov, 1); } status_t RpcState::readConnectionInit(const sp& connection, const sp& session) { RpcOutgoingConnectionInit init; iovec iov{&init, sizeof(init)}; if (status_t status = rpcRec(connection, session, "connection init", &iov, 1); status != OK) return status; static_assert(sizeof(init.msg) == sizeof(RPC_CONNECTION_INIT_OKAY)); if (0 != strncmp(init.msg, RPC_CONNECTION_INIT_OKAY, sizeof(init.msg))) { ALOGE("Connection init message unrecognized %.*s", static_cast(sizeof(init.msg)), init.msg); return BAD_VALUE; } return OK; } sp RpcState::getRootObject(const sp& connection, const sp& session) { Parcel data; data.markForRpc(session); Parcel reply; status_t status = transactAddress(connection, 0, RPC_SPECIAL_TRANSACT_GET_ROOT, data, session, &reply, 0); if (status != OK) { ALOGE("Error getting root object: %s", statusToString(status).c_str()); return nullptr; } return reply.readStrongBinder(); } status_t RpcState::getMaxThreads(const sp& connection, const sp& session, size_t* maxThreadsOut) { Parcel data; data.markForRpc(session); Parcel reply; status_t status = transactAddress(connection, 0, RPC_SPECIAL_TRANSACT_GET_MAX_THREADS, data, session, &reply, 0); if (status != OK) { ALOGE("Error getting max threads: %s", statusToString(status).c_str()); return status; } int32_t maxThreads; status = reply.readInt32(&maxThreads); if (status != OK) return status; if (maxThreads <= 0) { ALOGE("Error invalid max maxThreads: %d", maxThreads); return BAD_VALUE; } *maxThreadsOut = maxThreads; return OK; } status_t RpcState::getSessionId(const sp& connection, const sp& session, std::vector* sessionIdOut) { Parcel data; data.markForRpc(session); Parcel reply; status_t status = transactAddress(connection, 0, RPC_SPECIAL_TRANSACT_GET_SESSION_ID, data, session, &reply, 0); if (status != OK) { ALOGE("Error getting session ID: %s", statusToString(status).c_str()); return status; } return reply.readByteVector(sessionIdOut); } status_t RpcState::transact(const sp& connection, const sp& binder, uint32_t code, const Parcel& data, const sp& session, Parcel* reply, uint32_t flags) { if (!data.isForRpc()) { ALOGE("Refusing to send RPC with parcel not crafted for RPC call on binder %p code " "%" PRIu32, binder.get(), code); return BAD_TYPE; } if (data.objectsCount() != 0) { ALOGE("Parcel at %p has attached objects but is being used in an RPC call on binder %p " "code %" PRIu32, &data, binder.get(), code); return BAD_TYPE; } uint64_t address; if (status_t status = onBinderLeaving(session, binder, &address); status != OK) return status; return transactAddress(connection, address, code, data, session, reply, flags); } status_t RpcState::transactAddress(const sp& connection, uint64_t address, uint32_t code, const Parcel& data, const sp& session, Parcel* reply, uint32_t flags) { LOG_ALWAYS_FATAL_IF(!data.isForRpc()); LOG_ALWAYS_FATAL_IF(data.objectsCount() != 0); uint64_t asyncNumber = 0; if (address != 0) { std::unique_lock _l(mNodeMutex); if (mTerminated) return DEAD_OBJECT; // avoid fatal only, otherwise races auto it = mNodeForAddress.find(address); LOG_ALWAYS_FATAL_IF(it == mNodeForAddress.end(), "Sending transact on unknown address %" PRIu64, address); if (flags & IBinder::FLAG_ONEWAY) { asyncNumber = it->second.asyncNumber; if (!nodeProgressAsyncNumber(&it->second)) { _l.unlock(); (void)session->shutdownAndWait(false); return DEAD_OBJECT; } } } LOG_ALWAYS_FATAL_IF(std::numeric_limits::max() - sizeof(RpcWireHeader) - sizeof(RpcWireTransaction) < data.dataSize(), "Too much data %zu", data.dataSize()); RpcWireHeader command{ .command = RPC_COMMAND_TRANSACT, .bodySize = static_cast(sizeof(RpcWireTransaction) + data.dataSize()), }; RpcWireTransaction transaction{ .address = RpcWireAddress::fromRaw(address), .code = code, .flags = flags, .asyncNumber = asyncNumber, }; constexpr size_t kWaitMaxUs = 1000000; constexpr size_t kWaitLogUs = 10000; size_t waitUs = 0; // Oneway calls have no sync point, so if many are sent before, whether this // is a twoway or oneway transaction, they may have filled up the socket. // So, make sure we drain them before polling. std::function drainRefs = [&] { if (waitUs > kWaitLogUs) { ALOGE("Cannot send command, trying to process pending refcounts. Waiting %zuus. Too " "many oneway calls?", waitUs); } if (waitUs > 0) { usleep(waitUs); waitUs = std::min(kWaitMaxUs, waitUs * 2); } else { waitUs = 1; } return drainCommands(connection, session, CommandType::CONTROL_ONLY); }; iovec iovs[]{ {&command, sizeof(RpcWireHeader)}, {&transaction, sizeof(RpcWireTransaction)}, {const_cast(data.data()), data.dataSize()}, }; if (status_t status = rpcSend(connection, session, "transaction", iovs, arraysize(iovs), drainRefs); status != OK) { // TODO(b/167966510): need to undo onBinderLeaving - we know the // refcount isn't successfully transferred. return status; } if (flags & IBinder::FLAG_ONEWAY) { LOG_RPC_DETAIL("Oneway command, so no longer waiting on RpcTransport %p", connection->rpcTransport.get()); // Do not wait on result. return OK; } LOG_ALWAYS_FATAL_IF(reply == nullptr, "Reply parcel must be used for synchronous transaction."); return waitForReply(connection, session, reply); } static void cleanup_reply_data(Parcel* p, const uint8_t* data, size_t dataSize, const binder_size_t* objects, size_t objectsCount) { (void)p; delete[] const_cast(data - offsetof(RpcWireReply, data)); (void)dataSize; LOG_ALWAYS_FATAL_IF(objects != nullptr); LOG_ALWAYS_FATAL_IF(objectsCount != 0, "%zu objects remaining", objectsCount); } status_t RpcState::waitForReply(const sp& connection, const sp& session, Parcel* reply) { RpcWireHeader command; while (true) { iovec iov{&command, sizeof(command)}; if (status_t status = rpcRec(connection, session, "command header (for reply)", &iov, 1); status != OK) return status; if (command.command == RPC_COMMAND_REPLY) break; if (status_t status = processCommand(connection, session, command, CommandType::ANY); status != OK) return status; } CommandData data(command.bodySize); if (!data.valid()) return NO_MEMORY; iovec iov{data.data(), command.bodySize}; if (status_t status = rpcRec(connection, session, "reply body", &iov, 1); status != OK) return status; if (command.bodySize < sizeof(RpcWireReply)) { ALOGE("Expecting %zu but got %" PRId32 " bytes for RpcWireReply. Terminating!", sizeof(RpcWireReply), command.bodySize); (void)session->shutdownAndWait(false); return BAD_VALUE; } RpcWireReply* rpcReply = reinterpret_cast(data.data()); if (rpcReply->status != OK) return rpcReply->status; data.release(); reply->ipcSetDataReference(rpcReply->data, command.bodySize - offsetof(RpcWireReply, data), nullptr, 0, cleanup_reply_data); reply->markForRpc(session); return OK; } status_t RpcState::sendDecStrongToTarget(const sp& connection, const sp& session, uint64_t addr, size_t target) { RpcDecStrong body = { .address = RpcWireAddress::fromRaw(addr), }; { std::lock_guard _l(mNodeMutex); if (mTerminated) return DEAD_OBJECT; // avoid fatal only, otherwise races auto it = mNodeForAddress.find(addr); LOG_ALWAYS_FATAL_IF(it == mNodeForAddress.end(), "Sending dec strong on unknown address %" PRIu64, addr); LOG_ALWAYS_FATAL_IF(it->second.timesRecd < target, "Can't dec count of %zu to %zu.", it->second.timesRecd, target); // typically this happens when multiple threads send dec refs at the // same time - the transactions will get combined automatically if (it->second.timesRecd == target) return OK; body.amount = it->second.timesRecd - target; it->second.timesRecd = target; LOG_ALWAYS_FATAL_IF(nullptr != tryEraseNode(it), "Bad state. RpcState shouldn't own received binder"); } RpcWireHeader cmd = { .command = RPC_COMMAND_DEC_STRONG, .bodySize = sizeof(RpcDecStrong), }; iovec iovs[]{{&cmd, sizeof(cmd)}, {&body, sizeof(body)}}; return rpcSend(connection, session, "dec ref", iovs, arraysize(iovs)); } status_t RpcState::getAndExecuteCommand(const sp& connection, const sp& session, CommandType type) { LOG_RPC_DETAIL("getAndExecuteCommand on RpcTransport %p", connection->rpcTransport.get()); RpcWireHeader command; iovec iov{&command, sizeof(command)}; if (status_t status = rpcRec(connection, session, "command header (for server)", &iov, 1); status != OK) return status; return processCommand(connection, session, command, type); } status_t RpcState::drainCommands(const sp& connection, const sp& session, CommandType type) { uint8_t buf; while (true) { size_t num_bytes; status_t status = connection->rpcTransport->peek(&buf, sizeof(buf), &num_bytes); if (status == WOULD_BLOCK) break; if (status != OK) return status; if (!num_bytes) break; status = getAndExecuteCommand(connection, session, type); if (status != OK) return status; } return OK; } status_t RpcState::processCommand(const sp& connection, const sp& session, const RpcWireHeader& command, CommandType type) { IPCThreadState* kernelBinderState = IPCThreadState::selfOrNull(); IPCThreadState::SpGuard spGuard{ .address = __builtin_frame_address(0), .context = "processing binder RPC command", }; const IPCThreadState::SpGuard* origGuard; if (kernelBinderState != nullptr) { origGuard = kernelBinderState->pushGetCallingSpGuard(&spGuard); } ScopeGuard guardUnguard = [&]() { if (kernelBinderState != nullptr) { kernelBinderState->restoreGetCallingSpGuard(origGuard); } }; switch (command.command) { case RPC_COMMAND_TRANSACT: if (type != CommandType::ANY) return BAD_TYPE; return processTransact(connection, session, command); case RPC_COMMAND_DEC_STRONG: return processDecStrong(connection, session, command); } // We should always know the version of the opposing side, and since the // RPC-binder-level wire protocol is not self synchronizing, we have no way // to understand where the current command ends and the next one begins. We // also can't consider it a fatal error because this would allow any client // to kill us, so ending the session for misbehaving client. ALOGE("Unknown RPC command %d - terminating session", command.command); (void)session->shutdownAndWait(false); return DEAD_OBJECT; } status_t RpcState::processTransact(const sp& connection, const sp& session, const RpcWireHeader& command) { LOG_ALWAYS_FATAL_IF(command.command != RPC_COMMAND_TRANSACT, "command: %d", command.command); CommandData transactionData(command.bodySize); if (!transactionData.valid()) { return NO_MEMORY; } iovec iov{transactionData.data(), transactionData.size()}; if (status_t status = rpcRec(connection, session, "transaction body", &iov, 1); status != OK) return status; return processTransactInternal(connection, session, std::move(transactionData)); } static void do_nothing_to_transact_data(Parcel* p, const uint8_t* data, size_t dataSize, const binder_size_t* objects, size_t objectsCount) { (void)p; (void)data; (void)dataSize; (void)objects; (void)objectsCount; } status_t RpcState::processTransactInternal(const sp& connection, const sp& session, CommandData transactionData) { // for 'recursive' calls to this, we have already read and processed the // binder from the transaction data and taken reference counts into account, // so it is cached here. sp target; processTransactInternalTailCall: if (transactionData.size() < sizeof(RpcWireTransaction)) { ALOGE("Expecting %zu but got %zu bytes for RpcWireTransaction. Terminating!", sizeof(RpcWireTransaction), transactionData.size()); (void)session->shutdownAndWait(false); return BAD_VALUE; } RpcWireTransaction* transaction = reinterpret_cast(transactionData.data()); uint64_t addr = RpcWireAddress::toRaw(transaction->address); bool oneway = transaction->flags & IBinder::FLAG_ONEWAY; status_t replyStatus = OK; if (addr != 0) { if (!target) { replyStatus = onBinderEntering(session, addr, &target); } if (replyStatus != OK) { // do nothing } else if (target == nullptr) { // This can happen if the binder is remote in this process, and // another thread has called the last decStrong on this binder. // However, for local binders, it indicates a misbehaving client // (any binder which is being transacted on should be holding a // strong ref count), so in either case, terminating the // session. ALOGE("While transacting, binder has been deleted at address %" PRIu64 ". Terminating!", addr); (void)session->shutdownAndWait(false); replyStatus = BAD_VALUE; } else if (target->localBinder() == nullptr) { ALOGE("Unknown binder address or non-local binder, not address %" PRIu64 ". Terminating!", addr); (void)session->shutdownAndWait(false); replyStatus = BAD_VALUE; } else if (oneway) { std::unique_lock _l(mNodeMutex); auto it = mNodeForAddress.find(addr); if (it->second.binder.promote() != target) { ALOGE("Binder became invalid during transaction. Bad client? %" PRIu64, addr); replyStatus = BAD_VALUE; } else if (transaction->asyncNumber != it->second.asyncNumber) { // we need to process some other asynchronous transaction // first it->second.asyncTodo.push(BinderNode::AsyncTodo{ .ref = target, .data = std::move(transactionData), .asyncNumber = transaction->asyncNumber, }); size_t numPending = it->second.asyncTodo.size(); LOG_RPC_DETAIL("Enqueuing %" PRIu64 " on %" PRIu64 " (%zu pending)", transaction->asyncNumber, addr, numPending); constexpr size_t kArbitraryOnewayCallTerminateLevel = 10000; constexpr size_t kArbitraryOnewayCallWarnLevel = 1000; constexpr size_t kArbitraryOnewayCallWarnPer = 1000; if (numPending >= kArbitraryOnewayCallWarnLevel) { if (numPending >= kArbitraryOnewayCallTerminateLevel) { ALOGE("WARNING: %zu pending oneway transactions. Terminating!", numPending); _l.unlock(); (void)session->shutdownAndWait(false); return FAILED_TRANSACTION; } if (numPending % kArbitraryOnewayCallWarnPer == 0) { ALOGW("Warning: many oneway transactions built up on %p (%zu)", target.get(), numPending); } } return OK; } } } Parcel reply; reply.markForRpc(session); if (replyStatus == OK) { Parcel data; // transaction->data is owned by this function. Parcel borrows this data and // only holds onto it for the duration of this function call. Parcel will be // deleted before the 'transactionData' object. data.ipcSetDataReference(transaction->data, transactionData.size() - offsetof(RpcWireTransaction, data), nullptr /*object*/, 0 /*objectCount*/, do_nothing_to_transact_data); data.markForRpc(session); if (target) { bool origAllowNested = connection->allowNested; connection->allowNested = !oneway; replyStatus = target->transact(transaction->code, data, &reply, transaction->flags); connection->allowNested = origAllowNested; } else { LOG_RPC_DETAIL("Got special transaction %u", transaction->code); switch (transaction->code) { case RPC_SPECIAL_TRANSACT_GET_MAX_THREADS: { replyStatus = reply.writeInt32(session->getMaxIncomingThreads()); break; } case RPC_SPECIAL_TRANSACT_GET_SESSION_ID: { // for client connections, this should always report the value // originally returned from the server, so this is asserting // that it exists replyStatus = reply.writeByteVector(session->mId); break; } default: { sp server = session->server(); if (server) { switch (transaction->code) { case RPC_SPECIAL_TRANSACT_GET_ROOT: { sp root = session->mSessionSpecificRootObject ?: server->getRootObject(); replyStatus = reply.writeStrongBinder(root); break; } default: { replyStatus = UNKNOWN_TRANSACTION; } } } else { ALOGE("Special command sent, but no server object attached."); } } } } } if (oneway) { if (replyStatus != OK) { ALOGW("Oneway call failed with error: %d", replyStatus); } LOG_RPC_DETAIL("Processed async transaction %" PRIu64 " on %" PRIu64, transaction->asyncNumber, addr); // Check to see if there is another asynchronous transaction to process. // This behavior differs from binder behavior, since in the binder // driver, asynchronous transactions will be processed after existing // pending binder transactions on the queue. The downside of this is // that asynchronous transactions can be drowned out by synchronous // transactions. However, we have no easy way to queue these // transactions after the synchronous transactions we may want to read // from the wire. So, in socket binder here, we have the opposite // downside: asynchronous transactions may drown out synchronous // transactions. { std::unique_lock _l(mNodeMutex); auto it = mNodeForAddress.find(addr); // last refcount dropped after this transaction happened if (it == mNodeForAddress.end()) return OK; if (!nodeProgressAsyncNumber(&it->second)) { _l.unlock(); (void)session->shutdownAndWait(false); return DEAD_OBJECT; } if (it->second.asyncTodo.size() == 0) return OK; if (it->second.asyncTodo.top().asyncNumber == it->second.asyncNumber) { LOG_RPC_DETAIL("Found next async transaction %" PRIu64 " on %" PRIu64, it->second.asyncNumber, addr); // justification for const_cast (consider avoiding priority_queue): // - AsyncTodo operator< doesn't depend on 'data' or 'ref' objects // - gotta go fast auto& todo = const_cast(it->second.asyncTodo.top()); // reset up arguments transactionData = std::move(todo.data); LOG_ALWAYS_FATAL_IF(target != todo.ref, "async list should be associated with a binder"); it->second.asyncTodo.pop(); goto processTransactInternalTailCall; } } // done processing all the async commands on this binder that we can, so // write decstrongs on the binder if (addr != 0 && replyStatus == OK) { return flushExcessBinderRefs(session, addr, target); } return OK; } // Binder refs are flushed for oneway calls only after all calls which are // built up are executed. Otherwise, they fill up the binder buffer. if (addr != 0 && replyStatus == OK) { replyStatus = flushExcessBinderRefs(session, addr, target); } LOG_ALWAYS_FATAL_IF(std::numeric_limits::max() - sizeof(RpcWireHeader) - sizeof(RpcWireReply) < reply.dataSize(), "Too much data for reply %zu", reply.dataSize()); RpcWireHeader cmdReply{ .command = RPC_COMMAND_REPLY, .bodySize = static_cast(sizeof(RpcWireReply) + reply.dataSize()), }; RpcWireReply rpcReply{ .status = replyStatus, }; iovec iovs[]{ {&cmdReply, sizeof(RpcWireHeader)}, {&rpcReply, sizeof(RpcWireReply)}, {const_cast(reply.data()), reply.dataSize()}, }; return rpcSend(connection, session, "reply", iovs, arraysize(iovs)); } status_t RpcState::processDecStrong(const sp& connection, const sp& session, const RpcWireHeader& command) { LOG_ALWAYS_FATAL_IF(command.command != RPC_COMMAND_DEC_STRONG, "command: %d", command.command); CommandData commandData(command.bodySize); if (!commandData.valid()) { return NO_MEMORY; } iovec iov{commandData.data(), commandData.size()}; if (status_t status = rpcRec(connection, session, "dec ref body", &iov, 1); status != OK) return status; if (command.bodySize != sizeof(RpcDecStrong)) { ALOGE("Expecting %zu but got %" PRId32 " bytes for RpcDecStrong. Terminating!", sizeof(RpcDecStrong), command.bodySize); (void)session->shutdownAndWait(false); return BAD_VALUE; } RpcDecStrong* body = reinterpret_cast(commandData.data()); uint64_t addr = RpcWireAddress::toRaw(body->address); std::unique_lock _l(mNodeMutex); auto it = mNodeForAddress.find(addr); if (it == mNodeForAddress.end()) { ALOGE("Unknown binder address %" PRIu64 " for dec strong.", addr); return OK; } sp target = it->second.binder.promote(); if (target == nullptr) { ALOGE("While requesting dec strong, binder has been deleted at address %" PRIu64 ". Terminating!", addr); _l.unlock(); (void)session->shutdownAndWait(false); return BAD_VALUE; } if (it->second.timesSent < body->amount) { ALOGE("Record of sending binder %zu times, but requested decStrong for %" PRIu64 " of %u", it->second.timesSent, addr, body->amount); return OK; } LOG_ALWAYS_FATAL_IF(it->second.sentRef == nullptr, "Inconsistent state, lost ref for %" PRIu64, addr); LOG_RPC_DETAIL("Processing dec strong of %" PRIu64 " by %u from %zu", addr, body->amount, it->second.timesSent); it->second.timesSent -= body->amount; sp tempHold = tryEraseNode(it); _l.unlock(); tempHold = nullptr; // destructor may make binder calls on this session return OK; } sp RpcState::tryEraseNode(std::map::iterator& it) { sp ref; if (it->second.timesSent == 0) { ref = std::move(it->second.sentRef); if (it->second.timesRecd == 0) { LOG_ALWAYS_FATAL_IF(!it->second.asyncTodo.empty(), "Can't delete binder w/ pending async transactions"); mNodeForAddress.erase(it); } } return ref; } bool RpcState::nodeProgressAsyncNumber(BinderNode* node) { // 2**64 =~ 10**19 =~ 1000 transactions per second for 585 million years to // a single binder if (node->asyncNumber >= std::numeric_limitsasyncNumber)>::max()) { ALOGE("Out of async transaction IDs. Terminating"); return false; } node->asyncNumber++; return true; } } // namespace android