| /* |
| * Copyright (C) 2020 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #define LOG_TAG "RpcSession" |
| |
| #include <binder/RpcSession.h> |
| |
| #include <dlfcn.h> |
| #include <inttypes.h> |
| #include <poll.h> |
| #include <pthread.h> |
| #include <unistd.h> |
| |
| #include <string_view> |
| |
| #include <android-base/hex.h> |
| #include <android-base/macros.h> |
| #include <android-base/scopeguard.h> |
| #include <binder/BpBinder.h> |
| #include <binder/Parcel.h> |
| #include <binder/RpcServer.h> |
| #include <binder/RpcTransportRaw.h> |
| #include <binder/Stability.h> |
| #include <utils/String8.h> |
| |
| #include "FdTrigger.h" |
| #include "RpcSocketAddress.h" |
| #include "RpcState.h" |
| #include "RpcWireFormat.h" |
| #include "Utils.h" |
| |
| #ifdef __GLIBC__ |
| extern "C" pid_t gettid(); |
| #endif |
| |
| #ifndef __ANDROID_RECOVERY__ |
| #include <android_runtime/vm.h> |
| #include <jni.h> |
| #endif |
| |
| namespace android { |
| |
| using base::unique_fd; |
| |
| RpcSession::RpcSession(std::unique_ptr<RpcTransportCtx> ctx) : mCtx(std::move(ctx)) { |
| LOG_RPC_DETAIL("RpcSession created %p", this); |
| |
| mRpcBinderState = std::make_unique<RpcState>(); |
| } |
| RpcSession::~RpcSession() { |
| LOG_RPC_DETAIL("RpcSession destroyed %p", this); |
| |
| std::lock_guard<std::mutex> _l(mMutex); |
| LOG_ALWAYS_FATAL_IF(mConnections.mIncoming.size() != 0, |
| "Should not be able to destroy a session with servers in use."); |
| } |
| |
| sp<RpcSession> RpcSession::make() { |
| // Default is without TLS. |
| return make(RpcTransportCtxFactoryRaw::make()); |
| } |
| |
| sp<RpcSession> RpcSession::make(std::unique_ptr<RpcTransportCtxFactory> rpcTransportCtxFactory) { |
| auto ctx = rpcTransportCtxFactory->newClientCtx(); |
| if (ctx == nullptr) return nullptr; |
| return sp<RpcSession>::make(std::move(ctx)); |
| } |
| |
| void RpcSession::setMaxIncomingThreads(size_t threads) { |
| std::lock_guard<std::mutex> _l(mMutex); |
| LOG_ALWAYS_FATAL_IF(!mConnections.mOutgoing.empty() || !mConnections.mIncoming.empty(), |
| "Must set max incoming threads before setting up connections, but has %zu " |
| "client(s) and %zu server(s)", |
| mConnections.mOutgoing.size(), mConnections.mIncoming.size()); |
| mMaxIncomingThreads = threads; |
| } |
| |
| size_t RpcSession::getMaxIncomingThreads() { |
| std::lock_guard<std::mutex> _l(mMutex); |
| return mMaxIncomingThreads; |
| } |
| |
| void RpcSession::setMaxOutgoingThreads(size_t threads) { |
| std::lock_guard<std::mutex> _l(mMutex); |
| LOG_ALWAYS_FATAL_IF(!mConnections.mOutgoing.empty() || !mConnections.mIncoming.empty(), |
| "Must set max outgoing threads before setting up connections, but has %zu " |
| "client(s) and %zu server(s)", |
| mConnections.mOutgoing.size(), mConnections.mIncoming.size()); |
| mMaxOutgoingThreads = threads; |
| } |
| |
| size_t RpcSession::getMaxOutgoingThreads() { |
| std::lock_guard<std::mutex> _l(mMutex); |
| return mMaxOutgoingThreads; |
| } |
| |
| bool RpcSession::setProtocolVersion(uint32_t version) { |
| if (version >= RPC_WIRE_PROTOCOL_VERSION_NEXT && |
| version != RPC_WIRE_PROTOCOL_VERSION_EXPERIMENTAL) { |
| ALOGE("Cannot start RPC session with version %u which is unknown (current protocol version " |
| "is %u).", |
| version, RPC_WIRE_PROTOCOL_VERSION); |
| return false; |
| } |
| |
| std::lock_guard<std::mutex> _l(mMutex); |
| if (mProtocolVersion && version > *mProtocolVersion) { |
| ALOGE("Cannot upgrade explicitly capped protocol version %u to newer version %u", |
| *mProtocolVersion, version); |
| return false; |
| } |
| |
| mProtocolVersion = version; |
| return true; |
| } |
| |
| std::optional<uint32_t> RpcSession::getProtocolVersion() { |
| std::lock_guard<std::mutex> _l(mMutex); |
| return mProtocolVersion; |
| } |
| |
| status_t RpcSession::setupUnixDomainClient(const char* path) { |
| return setupSocketClient(UnixSocketAddress(path)); |
| } |
| |
| status_t RpcSession::setupVsockClient(unsigned int cid, unsigned int port) { |
| return setupSocketClient(VsockSocketAddress(cid, port)); |
| } |
| |
| status_t RpcSession::setupInetClient(const char* addr, unsigned int port) { |
| auto aiStart = InetSocketAddress::getAddrInfo(addr, port); |
| if (aiStart == nullptr) return UNKNOWN_ERROR; |
| for (auto ai = aiStart.get(); ai != nullptr; ai = ai->ai_next) { |
| InetSocketAddress socketAddress(ai->ai_addr, ai->ai_addrlen, addr, port); |
| if (status_t status = setupSocketClient(socketAddress); status == OK) return OK; |
| } |
| ALOGE("None of the socket address resolved for %s:%u can be added as inet client.", addr, port); |
| return NAME_NOT_FOUND; |
| } |
| |
| status_t RpcSession::setupPreconnectedClient(unique_fd fd, std::function<unique_fd()>&& request) { |
| return setupClient([&](const std::vector<uint8_t>& sessionId, bool incoming) -> status_t { |
| // std::move'd from fd becomes -1 (!ok()) |
| if (!fd.ok()) { |
| fd = request(); |
| if (!fd.ok()) return BAD_VALUE; |
| } |
| if (auto res = setNonBlocking(fd); !res.ok()) { |
| ALOGE("setupPreconnectedClient: %s", res.error().message().c_str()); |
| return res.error().code() == 0 ? UNKNOWN_ERROR : -res.error().code(); |
| } |
| return initAndAddConnection(std::move(fd), sessionId, incoming); |
| }); |
| } |
| |
| status_t RpcSession::addNullDebuggingClient() { |
| // Note: only works on raw sockets. |
| if (auto status = initShutdownTrigger(); status != OK) return status; |
| |
| unique_fd serverFd(TEMP_FAILURE_RETRY(open("/dev/null", O_WRONLY | O_CLOEXEC))); |
| |
| if (serverFd == -1) { |
| int savedErrno = errno; |
| ALOGE("Could not connect to /dev/null: %s", strerror(savedErrno)); |
| return -savedErrno; |
| } |
| |
| auto server = mCtx->newTransport(std::move(serverFd), mShutdownTrigger.get()); |
| if (server == nullptr) { |
| ALOGE("Unable to set up RpcTransport"); |
| return UNKNOWN_ERROR; |
| } |
| return addOutgoingConnection(std::move(server), false); |
| } |
| |
| sp<IBinder> RpcSession::getRootObject() { |
| ExclusiveConnection connection; |
| status_t status = ExclusiveConnection::find(sp<RpcSession>::fromExisting(this), |
| ConnectionUse::CLIENT, &connection); |
| if (status != OK) return nullptr; |
| return state()->getRootObject(connection.get(), sp<RpcSession>::fromExisting(this)); |
| } |
| |
| status_t RpcSession::getRemoteMaxThreads(size_t* maxThreads) { |
| ExclusiveConnection connection; |
| status_t status = ExclusiveConnection::find(sp<RpcSession>::fromExisting(this), |
| ConnectionUse::CLIENT, &connection); |
| if (status != OK) return status; |
| return state()->getMaxThreads(connection.get(), sp<RpcSession>::fromExisting(this), maxThreads); |
| } |
| |
| bool RpcSession::shutdownAndWait(bool wait) { |
| std::unique_lock<std::mutex> _l(mMutex); |
| LOG_ALWAYS_FATAL_IF(mShutdownTrigger == nullptr, "Shutdown trigger not installed"); |
| |
| mShutdownTrigger->trigger(); |
| |
| if (wait) { |
| LOG_ALWAYS_FATAL_IF(mShutdownListener == nullptr, "Shutdown listener not installed"); |
| mShutdownListener->waitForShutdown(_l, sp<RpcSession>::fromExisting(this)); |
| |
| LOG_ALWAYS_FATAL_IF(!mConnections.mThreads.empty(), "Shutdown failed"); |
| } |
| |
| _l.unlock(); |
| mRpcBinderState->clear(); |
| |
| return true; |
| } |
| |
| status_t RpcSession::transact(const sp<IBinder>& binder, uint32_t code, const Parcel& data, |
| Parcel* reply, uint32_t flags) { |
| ExclusiveConnection connection; |
| status_t status = |
| ExclusiveConnection::find(sp<RpcSession>::fromExisting(this), |
| (flags & IBinder::FLAG_ONEWAY) ? ConnectionUse::CLIENT_ASYNC |
| : ConnectionUse::CLIENT, |
| &connection); |
| if (status != OK) return status; |
| return state()->transact(connection.get(), binder, code, data, |
| sp<RpcSession>::fromExisting(this), reply, flags); |
| } |
| |
| status_t RpcSession::sendDecStrong(const BpBinder* binder) { |
| // target is 0 because this is used to free BpBinder objects |
| return sendDecStrongToTarget(binder->getPrivateAccessor().rpcAddress(), 0 /*target*/); |
| } |
| |
| status_t RpcSession::sendDecStrongToTarget(uint64_t address, size_t target) { |
| ExclusiveConnection connection; |
| status_t status = ExclusiveConnection::find(sp<RpcSession>::fromExisting(this), |
| ConnectionUse::CLIENT_REFCOUNT, &connection); |
| if (status != OK) return status; |
| return state()->sendDecStrongToTarget(connection.get(), sp<RpcSession>::fromExisting(this), |
| address, target); |
| } |
| |
| status_t RpcSession::readId() { |
| { |
| std::lock_guard<std::mutex> _l(mMutex); |
| LOG_ALWAYS_FATAL_IF(mForServer != nullptr, "Can only update ID for client."); |
| } |
| |
| ExclusiveConnection connection; |
| status_t status = ExclusiveConnection::find(sp<RpcSession>::fromExisting(this), |
| ConnectionUse::CLIENT, &connection); |
| if (status != OK) return status; |
| |
| status = state()->getSessionId(connection.get(), sp<RpcSession>::fromExisting(this), &mId); |
| if (status != OK) return status; |
| |
| LOG_RPC_DETAIL("RpcSession %p has id %s", this, |
| base::HexString(mId.data(), mId.size()).c_str()); |
| return OK; |
| } |
| |
| void RpcSession::WaitForShutdownListener::onSessionAllIncomingThreadsEnded( |
| const sp<RpcSession>& session) { |
| (void)session; |
| } |
| |
| void RpcSession::WaitForShutdownListener::onSessionIncomingThreadEnded() { |
| mCv.notify_all(); |
| } |
| |
| void RpcSession::WaitForShutdownListener::waitForShutdown(std::unique_lock<std::mutex>& lock, |
| const sp<RpcSession>& session) { |
| while (session->mConnections.mIncoming.size() > 0) { |
| if (std::cv_status::timeout == mCv.wait_for(lock, std::chrono::seconds(1))) { |
| ALOGE("Waiting for RpcSession to shut down (1s w/o progress): %zu incoming connections " |
| "still.", |
| session->mConnections.mIncoming.size()); |
| } |
| } |
| } |
| |
| void RpcSession::preJoinThreadOwnership(std::thread thread) { |
| LOG_ALWAYS_FATAL_IF(thread.get_id() != std::this_thread::get_id(), "Must own this thread"); |
| |
| { |
| std::lock_guard<std::mutex> _l(mMutex); |
| mConnections.mThreads[thread.get_id()] = std::move(thread); |
| } |
| } |
| |
| RpcSession::PreJoinSetupResult RpcSession::preJoinSetup( |
| std::unique_ptr<RpcTransport> rpcTransport) { |
| // must be registered to allow arbitrary client code executing commands to |
| // be able to do nested calls (we can't only read from it) |
| sp<RpcConnection> connection = assignIncomingConnectionToThisThread(std::move(rpcTransport)); |
| |
| status_t status; |
| |
| if (connection == nullptr) { |
| status = DEAD_OBJECT; |
| } else { |
| status = |
| mRpcBinderState->readConnectionInit(connection, sp<RpcSession>::fromExisting(this)); |
| } |
| |
| return PreJoinSetupResult{ |
| .connection = std::move(connection), |
| .status = status, |
| }; |
| } |
| |
| namespace { |
| #ifdef __ANDROID_RECOVERY__ |
| class JavaThreadAttacher {}; |
| #else |
| // RAII object for attaching / detaching current thread to JVM if Android Runtime exists. If |
| // Android Runtime doesn't exist, no-op. |
| class JavaThreadAttacher { |
| public: |
| JavaThreadAttacher() { |
| // Use dlsym to find androidJavaAttachThread because libandroid_runtime is loaded after |
| // libbinder. |
| auto vm = getJavaVM(); |
| if (vm == nullptr) return; |
| |
| char threadName[16]; |
| if (0 != pthread_getname_np(pthread_self(), threadName, sizeof(threadName))) { |
| constexpr const char* defaultThreadName = "UnknownRpcSessionThread"; |
| memcpy(threadName, defaultThreadName, |
| std::min<size_t>(sizeof(threadName), strlen(defaultThreadName) + 1)); |
| } |
| LOG_RPC_DETAIL("Attaching current thread %s to JVM", threadName); |
| JavaVMAttachArgs args; |
| args.version = JNI_VERSION_1_2; |
| args.name = threadName; |
| args.group = nullptr; |
| JNIEnv* env; |
| |
| LOG_ALWAYS_FATAL_IF(vm->AttachCurrentThread(&env, &args) != JNI_OK, |
| "Cannot attach thread %s to JVM", threadName); |
| mAttached = true; |
| } |
| ~JavaThreadAttacher() { |
| if (!mAttached) return; |
| auto vm = getJavaVM(); |
| LOG_ALWAYS_FATAL_IF(vm == nullptr, |
| "Unable to detach thread. No JavaVM, but it was present before!"); |
| |
| LOG_RPC_DETAIL("Detaching current thread from JVM"); |
| if (vm->DetachCurrentThread() != JNI_OK) { |
| mAttached = false; |
| } else { |
| ALOGW("Unable to detach current thread from JVM"); |
| } |
| } |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(JavaThreadAttacher); |
| bool mAttached = false; |
| |
| static JavaVM* getJavaVM() { |
| static auto fn = reinterpret_cast<decltype(&AndroidRuntimeGetJavaVM)>( |
| dlsym(RTLD_DEFAULT, "AndroidRuntimeGetJavaVM")); |
| if (fn == nullptr) return nullptr; |
| return fn(); |
| } |
| }; |
| #endif |
| } // namespace |
| |
| void RpcSession::join(sp<RpcSession>&& session, PreJoinSetupResult&& setupResult) { |
| sp<RpcConnection>& connection = setupResult.connection; |
| |
| if (setupResult.status == OK) { |
| LOG_ALWAYS_FATAL_IF(!connection, "must have connection if setup succeeded"); |
| [[maybe_unused]] JavaThreadAttacher javaThreadAttacher; |
| while (true) { |
| status_t status = session->state()->getAndExecuteCommand(connection, session, |
| RpcState::CommandType::ANY); |
| if (status != OK) { |
| LOG_RPC_DETAIL("Binder connection thread closing w/ status %s", |
| statusToString(status).c_str()); |
| break; |
| } |
| } |
| } else { |
| ALOGE("Connection failed to init, closing with status %s", |
| statusToString(setupResult.status).c_str()); |
| } |
| |
| sp<RpcSession::EventListener> listener; |
| { |
| std::lock_guard<std::mutex> _l(session->mMutex); |
| auto it = session->mConnections.mThreads.find(std::this_thread::get_id()); |
| LOG_ALWAYS_FATAL_IF(it == session->mConnections.mThreads.end()); |
| it->second.detach(); |
| session->mConnections.mThreads.erase(it); |
| |
| listener = session->mEventListener.promote(); |
| } |
| |
| // done after all cleanup, since session shutdown progresses via callbacks here |
| if (connection != nullptr) { |
| LOG_ALWAYS_FATAL_IF(!session->removeIncomingConnection(connection), |
| "bad state: connection object guaranteed to be in list"); |
| } |
| |
| session = nullptr; |
| |
| if (listener != nullptr) { |
| listener->onSessionIncomingThreadEnded(); |
| } |
| } |
| |
| sp<RpcServer> RpcSession::server() { |
| RpcServer* unsafeServer = mForServer.unsafe_get(); |
| sp<RpcServer> server = mForServer.promote(); |
| |
| LOG_ALWAYS_FATAL_IF((unsafeServer == nullptr) != (server == nullptr), |
| "wp<> is to avoid strong cycle only"); |
| return server; |
| } |
| |
| status_t RpcSession::setupClient(const std::function<status_t(const std::vector<uint8_t>& sessionId, |
| bool incoming)>& connectAndInit) { |
| { |
| std::lock_guard<std::mutex> _l(mMutex); |
| LOG_ALWAYS_FATAL_IF(mConnections.mOutgoing.size() != 0, |
| "Must only setup session once, but already has %zu clients", |
| mConnections.mOutgoing.size()); |
| } |
| |
| if (auto status = initShutdownTrigger(); status != OK) return status; |
| |
| auto oldProtocolVersion = mProtocolVersion; |
| auto cleanup = base::ScopeGuard([&] { |
| // if any threads are started, shut them down |
| (void)shutdownAndWait(true); |
| |
| mShutdownListener = nullptr; |
| mEventListener.clear(); |
| |
| mId.clear(); |
| |
| mShutdownTrigger = nullptr; |
| mRpcBinderState = std::make_unique<RpcState>(); |
| |
| // protocol version may have been downgraded - if we reuse this object |
| // to connect to another server, force that server to request a |
| // downgrade again |
| mProtocolVersion = oldProtocolVersion; |
| |
| mConnections = {}; |
| }); |
| |
| if (status_t status = connectAndInit({}, false /*incoming*/); status != OK) return status; |
| |
| { |
| ExclusiveConnection connection; |
| if (status_t status = ExclusiveConnection::find(sp<RpcSession>::fromExisting(this), |
| ConnectionUse::CLIENT, &connection); |
| status != OK) |
| return status; |
| |
| uint32_t version; |
| if (status_t status = |
| state()->readNewSessionResponse(connection.get(), |
| sp<RpcSession>::fromExisting(this), &version); |
| status != OK) |
| return status; |
| if (!setProtocolVersion(version)) return BAD_VALUE; |
| } |
| |
| // TODO(b/189955605): we should add additional sessions dynamically |
| // instead of all at once. |
| size_t numThreadsAvailable; |
| if (status_t status = getRemoteMaxThreads(&numThreadsAvailable); status != OK) { |
| ALOGE("Could not get max threads after initial session setup: %s", |
| statusToString(status).c_str()); |
| return status; |
| } |
| |
| if (status_t status = readId(); status != OK) { |
| ALOGE("Could not get session id after initial session setup: %s", |
| statusToString(status).c_str()); |
| return status; |
| } |
| |
| size_t outgoingThreads = std::min(numThreadsAvailable, mMaxOutgoingThreads); |
| ALOGI_IF(outgoingThreads != numThreadsAvailable, |
| "Server hints client to start %zu outgoing threads, but client will only start %zu " |
| "because it is preconfigured to start at most %zu outgoing threads.", |
| numThreadsAvailable, outgoingThreads, mMaxOutgoingThreads); |
| |
| // TODO(b/189955605): we should add additional sessions dynamically |
| // instead of all at once - the other side should be responsible for setting |
| // up additional connections. We need to create at least one (unless 0 are |
| // requested to be set) in order to allow the other side to reliably make |
| // any requests at all. |
| |
| // we've already setup one client |
| LOG_RPC_DETAIL("RpcSession::setupClient() instantiating %zu outgoing (server max: %zu) and %zu " |
| "incoming threads", |
| outgoingThreads, numThreadsAvailable, mMaxIncomingThreads); |
| for (size_t i = 0; i + 1 < outgoingThreads; i++) { |
| if (status_t status = connectAndInit(mId, false /*incoming*/); status != OK) return status; |
| } |
| |
| for (size_t i = 0; i < mMaxIncomingThreads; i++) { |
| if (status_t status = connectAndInit(mId, true /*incoming*/); status != OK) return status; |
| } |
| |
| cleanup.Disable(); |
| |
| return OK; |
| } |
| |
| status_t RpcSession::setupSocketClient(const RpcSocketAddress& addr) { |
| return setupClient([&](const std::vector<uint8_t>& sessionId, bool incoming) { |
| return setupOneSocketConnection(addr, sessionId, incoming); |
| }); |
| } |
| |
| status_t RpcSession::setupOneSocketConnection(const RpcSocketAddress& addr, |
| const std::vector<uint8_t>& sessionId, |
| bool incoming) { |
| for (size_t tries = 0; tries < 5; tries++) { |
| if (tries > 0) usleep(10000); |
| |
| unique_fd serverFd(TEMP_FAILURE_RETRY( |
| socket(addr.addr()->sa_family, SOCK_STREAM | SOCK_CLOEXEC | SOCK_NONBLOCK, 0))); |
| if (serverFd == -1) { |
| int savedErrno = errno; |
| ALOGE("Could not create socket at %s: %s", addr.toString().c_str(), |
| strerror(savedErrno)); |
| return -savedErrno; |
| } |
| |
| if (0 != TEMP_FAILURE_RETRY(connect(serverFd.get(), addr.addr(), addr.addrSize()))) { |
| int connErrno = errno; |
| if (connErrno == EAGAIN || connErrno == EINPROGRESS) { |
| // For non-blocking sockets, connect() may return EAGAIN (for unix domain socket) or |
| // EINPROGRESS (for others). Call poll() and getsockopt() to get the error. |
| status_t pollStatus = mShutdownTrigger->triggerablePoll(serverFd, POLLOUT); |
| if (pollStatus != OK) { |
| ALOGE("Could not POLLOUT after connect() on non-blocking socket: %s", |
| statusToString(pollStatus).c_str()); |
| return pollStatus; |
| } |
| // Set connErrno to the errno that connect() would have set if the fd were blocking. |
| socklen_t connErrnoLen = sizeof(connErrno); |
| int ret = |
| getsockopt(serverFd.get(), SOL_SOCKET, SO_ERROR, &connErrno, &connErrnoLen); |
| if (ret == -1) { |
| int savedErrno = errno; |
| ALOGE("Could not getsockopt() after connect() on non-blocking socket: %s. " |
| "(Original error from connect() is: %s)", |
| strerror(savedErrno), strerror(connErrno)); |
| return -savedErrno; |
| } |
| // Retrieved the real connErrno as if connect() was called with a blocking socket |
| // fd. Continue checking connErrno. |
| } |
| if (connErrno == ECONNRESET) { |
| ALOGW("Connection reset on %s", addr.toString().c_str()); |
| continue; |
| } |
| // connErrno could be zero if getsockopt determines so. Hence zero-check again. |
| if (connErrno != 0) { |
| ALOGE("Could not connect socket at %s: %s", addr.toString().c_str(), |
| strerror(connErrno)); |
| return -connErrno; |
| } |
| } |
| LOG_RPC_DETAIL("Socket at %s client with fd %d", addr.toString().c_str(), serverFd.get()); |
| |
| return initAndAddConnection(std::move(serverFd), sessionId, incoming); |
| } |
| |
| ALOGE("Ran out of retries to connect to %s", addr.toString().c_str()); |
| return UNKNOWN_ERROR; |
| } |
| |
| status_t RpcSession::initAndAddConnection(unique_fd fd, const std::vector<uint8_t>& sessionId, |
| bool incoming) { |
| LOG_ALWAYS_FATAL_IF(mShutdownTrigger == nullptr); |
| auto server = mCtx->newTransport(std::move(fd), mShutdownTrigger.get()); |
| if (server == nullptr) { |
| ALOGE("%s: Unable to set up RpcTransport", __PRETTY_FUNCTION__); |
| return UNKNOWN_ERROR; |
| } |
| |
| LOG_RPC_DETAIL("Socket at client with RpcTransport %p", server.get()); |
| |
| if (sessionId.size() > std::numeric_limits<uint16_t>::max()) { |
| ALOGE("Session ID too big %zu", sessionId.size()); |
| return BAD_VALUE; |
| } |
| |
| RpcConnectionHeader header{ |
| .version = mProtocolVersion.value_or(RPC_WIRE_PROTOCOL_VERSION), |
| .options = 0, |
| .sessionIdSize = static_cast<uint16_t>(sessionId.size()), |
| }; |
| |
| if (incoming) { |
| header.options |= RPC_CONNECTION_OPTION_INCOMING; |
| } |
| |
| auto sendHeaderStatus = |
| server->interruptableWriteFully(mShutdownTrigger.get(), &header, sizeof(header), {}); |
| if (sendHeaderStatus != OK) { |
| ALOGE("Could not write connection header to socket: %s", |
| statusToString(sendHeaderStatus).c_str()); |
| return sendHeaderStatus; |
| } |
| |
| if (sessionId.size() > 0) { |
| auto sendSessionIdStatus = |
| server->interruptableWriteFully(mShutdownTrigger.get(), sessionId.data(), |
| sessionId.size(), {}); |
| if (sendSessionIdStatus != OK) { |
| ALOGE("Could not write session ID ('%s') to socket: %s", |
| base::HexString(sessionId.data(), sessionId.size()).c_str(), |
| statusToString(sendSessionIdStatus).c_str()); |
| return sendSessionIdStatus; |
| } |
| } |
| |
| LOG_RPC_DETAIL("Socket at client: header sent"); |
| |
| if (incoming) { |
| return addIncomingConnection(std::move(server)); |
| } else { |
| return addOutgoingConnection(std::move(server), true /*init*/); |
| } |
| } |
| |
| status_t RpcSession::addIncomingConnection(std::unique_ptr<RpcTransport> rpcTransport) { |
| std::mutex mutex; |
| std::condition_variable joinCv; |
| std::unique_lock<std::mutex> lock(mutex); |
| std::thread thread; |
| sp<RpcSession> thiz = sp<RpcSession>::fromExisting(this); |
| bool ownershipTransferred = false; |
| thread = std::thread([&]() { |
| std::unique_lock<std::mutex> threadLock(mutex); |
| std::unique_ptr<RpcTransport> movedRpcTransport = std::move(rpcTransport); |
| // NOLINTNEXTLINE(performance-unnecessary-copy-initialization) |
| sp<RpcSession> session = thiz; |
| session->preJoinThreadOwnership(std::move(thread)); |
| |
| // only continue once we have a response or the connection fails |
| auto setupResult = session->preJoinSetup(std::move(movedRpcTransport)); |
| |
| ownershipTransferred = true; |
| threadLock.unlock(); |
| joinCv.notify_one(); |
| // do not use & vars below |
| |
| RpcSession::join(std::move(session), std::move(setupResult)); |
| }); |
| joinCv.wait(lock, [&] { return ownershipTransferred; }); |
| LOG_ALWAYS_FATAL_IF(!ownershipTransferred); |
| return OK; |
| } |
| |
| status_t RpcSession::initShutdownTrigger() { |
| // first client connection added, but setForServer not called, so |
| // initializaing for a client. |
| if (mShutdownTrigger == nullptr) { |
| mShutdownTrigger = FdTrigger::make(); |
| mEventListener = mShutdownListener = sp<WaitForShutdownListener>::make(); |
| if (mShutdownTrigger == nullptr) return INVALID_OPERATION; |
| } |
| return OK; |
| } |
| |
| status_t RpcSession::addOutgoingConnection(std::unique_ptr<RpcTransport> rpcTransport, bool init) { |
| sp<RpcConnection> connection = sp<RpcConnection>::make(); |
| { |
| std::lock_guard<std::mutex> _l(mMutex); |
| connection->rpcTransport = std::move(rpcTransport); |
| connection->exclusiveTid = gettid(); |
| mConnections.mOutgoing.push_back(connection); |
| } |
| |
| status_t status = OK; |
| if (init) { |
| status = |
| mRpcBinderState->sendConnectionInit(connection, sp<RpcSession>::fromExisting(this)); |
| } |
| |
| { |
| std::lock_guard<std::mutex> _l(mMutex); |
| connection->exclusiveTid = std::nullopt; |
| } |
| |
| return status; |
| } |
| |
| bool RpcSession::setForServer(const wp<RpcServer>& server, const wp<EventListener>& eventListener, |
| const std::vector<uint8_t>& sessionId, |
| const sp<IBinder>& sessionSpecificRoot) { |
| LOG_ALWAYS_FATAL_IF(mForServer != nullptr); |
| LOG_ALWAYS_FATAL_IF(server == nullptr); |
| LOG_ALWAYS_FATAL_IF(mEventListener != nullptr); |
| LOG_ALWAYS_FATAL_IF(eventListener == nullptr); |
| LOG_ALWAYS_FATAL_IF(mShutdownTrigger != nullptr); |
| |
| mShutdownTrigger = FdTrigger::make(); |
| if (mShutdownTrigger == nullptr) return false; |
| |
| mId = sessionId; |
| mForServer = server; |
| mEventListener = eventListener; |
| mSessionSpecificRootObject = sessionSpecificRoot; |
| return true; |
| } |
| |
| sp<RpcSession::RpcConnection> RpcSession::assignIncomingConnectionToThisThread( |
| std::unique_ptr<RpcTransport> rpcTransport) { |
| std::lock_guard<std::mutex> _l(mMutex); |
| |
| if (mConnections.mIncoming.size() >= mMaxIncomingThreads) { |
| ALOGE("Cannot add thread to session with %zu threads (max is set to %zu)", |
| mConnections.mIncoming.size(), mMaxIncomingThreads); |
| return nullptr; |
| } |
| |
| // Don't accept any more connections, some have shutdown. Usually this |
| // happens when new connections are still being established as part of a |
| // very short-lived session which shuts down after it already started |
| // accepting new connections. |
| if (mConnections.mIncoming.size() < mConnections.mMaxIncoming) { |
| return nullptr; |
| } |
| |
| sp<RpcConnection> session = sp<RpcConnection>::make(); |
| session->rpcTransport = std::move(rpcTransport); |
| session->exclusiveTid = gettid(); |
| |
| mConnections.mIncoming.push_back(session); |
| mConnections.mMaxIncoming = mConnections.mIncoming.size(); |
| |
| return session; |
| } |
| |
| bool RpcSession::removeIncomingConnection(const sp<RpcConnection>& connection) { |
| std::unique_lock<std::mutex> _l(mMutex); |
| if (auto it = |
| std::find(mConnections.mIncoming.begin(), mConnections.mIncoming.end(), connection); |
| it != mConnections.mIncoming.end()) { |
| mConnections.mIncoming.erase(it); |
| if (mConnections.mIncoming.size() == 0) { |
| sp<EventListener> listener = mEventListener.promote(); |
| if (listener) { |
| _l.unlock(); |
| listener->onSessionAllIncomingThreadsEnded(sp<RpcSession>::fromExisting(this)); |
| } |
| } |
| return true; |
| } |
| return false; |
| } |
| |
| std::vector<uint8_t> RpcSession::getCertificate(RpcCertificateFormat format) { |
| return mCtx->getCertificate(format); |
| } |
| |
| status_t RpcSession::ExclusiveConnection::find(const sp<RpcSession>& session, ConnectionUse use, |
| ExclusiveConnection* connection) { |
| connection->mSession = session; |
| connection->mConnection = nullptr; |
| connection->mReentrant = false; |
| |
| pid_t tid = gettid(); |
| std::unique_lock<std::mutex> _l(session->mMutex); |
| |
| session->mConnections.mWaitingThreads++; |
| while (true) { |
| sp<RpcConnection> exclusive; |
| sp<RpcConnection> available; |
| |
| // CHECK FOR DEDICATED CLIENT SOCKET |
| // |
| // A server/looper should always use a dedicated connection if available |
| findConnection(tid, &exclusive, &available, session->mConnections.mOutgoing, |
| session->mConnections.mOutgoingOffset); |
| |
| // WARNING: this assumes a server cannot request its client to send |
| // a transaction, as mIncoming is excluded below. |
| // |
| // Imagine we have more than one thread in play, and a single thread |
| // sends a synchronous, then an asynchronous command. Imagine the |
| // asynchronous command is sent on the first client connection. Then, if |
| // we naively send a synchronous command to that same connection, the |
| // thread on the far side might be busy processing the asynchronous |
| // command. So, we move to considering the second available thread |
| // for subsequent calls. |
| if (use == ConnectionUse::CLIENT_ASYNC && (exclusive != nullptr || available != nullptr)) { |
| session->mConnections.mOutgoingOffset = (session->mConnections.mOutgoingOffset + 1) % |
| session->mConnections.mOutgoing.size(); |
| } |
| |
| // USE SERVING SOCKET (e.g. nested transaction) |
| if (use != ConnectionUse::CLIENT_ASYNC) { |
| sp<RpcConnection> exclusiveIncoming; |
| // server connections are always assigned to a thread |
| findConnection(tid, &exclusiveIncoming, nullptr /*available*/, |
| session->mConnections.mIncoming, 0 /* index hint */); |
| |
| // asynchronous calls cannot be nested, we currently allow ref count |
| // calls to be nested (so that you can use this without having extra |
| // threads). Note 'drainCommands' is used so that these ref counts can't |
| // build up. |
| if (exclusiveIncoming != nullptr) { |
| if (exclusiveIncoming->allowNested) { |
| // guaranteed to be processed as nested command |
| exclusive = exclusiveIncoming; |
| } else if (use == ConnectionUse::CLIENT_REFCOUNT && available == nullptr) { |
| // prefer available socket, but if we don't have one, don't |
| // wait for one |
| exclusive = exclusiveIncoming; |
| } |
| } |
| } |
| |
| // if our thread is already using a connection, prioritize using that |
| if (exclusive != nullptr) { |
| connection->mConnection = exclusive; |
| connection->mReentrant = true; |
| break; |
| } else if (available != nullptr) { |
| connection->mConnection = available; |
| connection->mConnection->exclusiveTid = tid; |
| break; |
| } |
| |
| if (session->mConnections.mOutgoing.size() == 0) { |
| ALOGE("Session has no client connections. This is required for an RPC server to make " |
| "any non-nested (e.g. oneway or on another thread) calls. Use: %d. Server " |
| "connections: %zu", |
| static_cast<int>(use), session->mConnections.mIncoming.size()); |
| return WOULD_BLOCK; |
| } |
| |
| LOG_RPC_DETAIL("No available connections (have %zu clients and %zu servers). Waiting...", |
| session->mConnections.mOutgoing.size(), |
| session->mConnections.mIncoming.size()); |
| session->mAvailableConnectionCv.wait(_l); |
| } |
| session->mConnections.mWaitingThreads--; |
| |
| return OK; |
| } |
| |
| void RpcSession::ExclusiveConnection::findConnection(pid_t tid, sp<RpcConnection>* exclusive, |
| sp<RpcConnection>* available, |
| std::vector<sp<RpcConnection>>& sockets, |
| size_t socketsIndexHint) { |
| LOG_ALWAYS_FATAL_IF(sockets.size() > 0 && socketsIndexHint >= sockets.size(), |
| "Bad index %zu >= %zu", socketsIndexHint, sockets.size()); |
| |
| if (*exclusive != nullptr) return; // consistent with break below |
| |
| for (size_t i = 0; i < sockets.size(); i++) { |
| sp<RpcConnection>& socket = sockets[(i + socketsIndexHint) % sockets.size()]; |
| |
| // take first available connection (intuition = caching) |
| if (available && *available == nullptr && socket->exclusiveTid == std::nullopt) { |
| *available = socket; |
| continue; |
| } |
| |
| // though, prefer to take connection which is already inuse by this thread |
| // (nested transactions) |
| if (exclusive && socket->exclusiveTid == tid) { |
| *exclusive = socket; |
| break; // consistent with return above |
| } |
| } |
| } |
| |
| RpcSession::ExclusiveConnection::~ExclusiveConnection() { |
| // reentrant use of a connection means something less deep in the call stack |
| // is using this fd, and it retains the right to it. So, we don't give up |
| // exclusive ownership, and no thread is freed. |
| if (!mReentrant && mConnection != nullptr) { |
| std::unique_lock<std::mutex> _l(mSession->mMutex); |
| mConnection->exclusiveTid = std::nullopt; |
| if (mSession->mConnections.mWaitingThreads > 0) { |
| _l.unlock(); |
| mSession->mAvailableConnectionCv.notify_one(); |
| } |
| } |
| } |
| |
| } // namespace android |