2 * Copyright (C) 2005 The Android Open Source Project
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
8 * http://www.apache.org/licenses/LICENSE-2.0
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
17 #define LOG_TAG "Parcel"
18 //#define LOG_NDEBUG 0
29 #include <sys/types.h>
30 #include <sys/resource.h>
33 #include <binder/Binder.h>
34 #include <binder/BpBinder.h>
35 #include <binder/IPCThreadState.h>
36 #include <binder/Parcel.h>
37 #include <binder/ProcessState.h>
38 #include <binder/Status.h>
39 #include <binder/TextOutput.h>
40 #include <binder/Value.h>
42 #include <cutils/ashmem.h>
43 #include <utils/Debug.h>
44 #include <utils/Flattenable.h>
45 #include <utils/Log.h>
46 #include <utils/misc.h>
47 #include <utils/String8.h>
48 #include <utils/String16.h>
50 #include <private/binder/binder_module.h>
51 #include <private/binder/Static.h>
54 #define INT32_MAX ((int32_t)(2147483647))
58 //#define LOG_REFS(...) ALOG(LOG_DEBUG, LOG_TAG, __VA_ARGS__)
59 #define LOG_ALLOC(...)
60 //#define LOG_ALLOC(...) ALOG(LOG_DEBUG, LOG_TAG, __VA_ARGS__)
62 // ---------------------------------------------------------------------------
64 // This macro should never be used at runtime, as a too large value
65 // of s could cause an integer overflow. Instead, you should always
66 // use the wrapper function pad_size()
67 #define PAD_SIZE_UNSAFE(s) (((s)+3)&~3)
69 static size_t pad_size(size_t s) {
70 if (s > (SIZE_T_MAX - 3)) {
73 return PAD_SIZE_UNSAFE(s);
76 // Note: must be kept in sync with android/os/StrictMode.java's PENALTY_GATHER
77 #define STRICT_MODE_PENALTY_GATHER (0x40 << 16)
79 // XXX This can be made public if we want to provide
80 // support for typed data.
81 struct small_flat_data
89 static pthread_mutex_t gParcelGlobalAllocSizeLock = PTHREAD_MUTEX_INITIALIZER;
90 static size_t gParcelGlobalAllocSize = 0;
91 static size_t gParcelGlobalAllocCount = 0;
93 static size_t gMaxFds = 0;
95 // Maximum size of a blob to transfer in-place.
96 static const size_t BLOB_INPLACE_LIMIT = 16 * 1024;
100 BLOB_ASHMEM_IMMUTABLE = 1,
101 BLOB_ASHMEM_MUTABLE = 2,
104 void acquire_object(const sp<ProcessState>& proc,
105 const flat_binder_object& obj, const void* who, size_t* outAshmemSize)
108 case BINDER_TYPE_BINDER:
110 LOG_REFS("Parcel %p acquiring reference on local %p", who, obj.cookie);
111 reinterpret_cast<IBinder*>(obj.cookie)->incStrong(who);
114 case BINDER_TYPE_WEAK_BINDER:
116 reinterpret_cast<RefBase::weakref_type*>(obj.binder)->incWeak(who);
118 case BINDER_TYPE_HANDLE: {
119 const sp<IBinder> b = proc->getStrongProxyForHandle(obj.handle);
121 LOG_REFS("Parcel %p acquiring reference on remote %p", who, b.get());
126 case BINDER_TYPE_WEAK_HANDLE: {
127 const wp<IBinder> b = proc->getWeakProxyForHandle(obj.handle);
128 if (b != NULL) b.get_refs()->incWeak(who);
131 case BINDER_TYPE_FD: {
132 if ((obj.cookie != 0) && (outAshmemSize != NULL) && ashmem_valid(obj.handle)) {
133 // If we own an ashmem fd, keep track of how much memory it refers to.
134 int size = ashmem_get_size_region(obj.handle);
136 *outAshmemSize += size;
143 ALOGD("Invalid object type 0x%08x", obj.type);
146 void acquire_object(const sp<ProcessState>& proc,
147 const flat_binder_object& obj, const void* who)
149 acquire_object(proc, obj, who, NULL);
152 static void release_object(const sp<ProcessState>& proc,
153 const flat_binder_object& obj, const void* who, size_t* outAshmemSize)
156 case BINDER_TYPE_BINDER:
158 LOG_REFS("Parcel %p releasing reference on local %p", who, obj.cookie);
159 reinterpret_cast<IBinder*>(obj.cookie)->decStrong(who);
162 case BINDER_TYPE_WEAK_BINDER:
164 reinterpret_cast<RefBase::weakref_type*>(obj.binder)->decWeak(who);
166 case BINDER_TYPE_HANDLE: {
167 const sp<IBinder> b = proc->getStrongProxyForHandle(obj.handle);
169 LOG_REFS("Parcel %p releasing reference on remote %p", who, b.get());
174 case BINDER_TYPE_WEAK_HANDLE: {
175 const wp<IBinder> b = proc->getWeakProxyForHandle(obj.handle);
176 if (b != NULL) b.get_refs()->decWeak(who);
179 case BINDER_TYPE_FD: {
180 if (obj.cookie != 0) { // owned
181 if ((outAshmemSize != NULL) && ashmem_valid(obj.handle)) {
182 int size = ashmem_get_size_region(obj.handle);
184 *outAshmemSize -= size;
194 ALOGE("Invalid object type 0x%08x", obj.type);
197 void release_object(const sp<ProcessState>& proc,
198 const flat_binder_object& obj, const void* who)
200 release_object(proc, obj, who, NULL);
203 inline static status_t finish_flatten_binder(
204 const sp<IBinder>& /*binder*/, const flat_binder_object& flat, Parcel* out)
206 return out->writeObject(flat, false);
209 status_t flatten_binder(const sp<ProcessState>& /*proc*/,
210 const sp<IBinder>& binder, Parcel* out)
212 flat_binder_object obj;
214 if (IPCThreadState::self()->backgroundSchedulingDisabled()) {
215 /* minimum priority for all nodes is nice 0 */
216 obj.flags = FLAT_BINDER_FLAG_ACCEPTS_FDS;
218 /* minimum priority for all nodes is MAX_NICE(19) */
219 obj.flags = 0x13 | FLAT_BINDER_FLAG_ACCEPTS_FDS;
222 if (binder != NULL) {
223 IBinder *local = binder->localBinder();
225 BpBinder *proxy = binder->remoteBinder();
229 const int32_t handle = proxy ? proxy->handle() : 0;
230 obj.type = BINDER_TYPE_HANDLE;
231 obj.binder = 0; /* Don't pass uninitialized stack data to a remote process */
235 obj.type = BINDER_TYPE_BINDER;
236 obj.binder = reinterpret_cast<uintptr_t>(local->getWeakRefs());
237 obj.cookie = reinterpret_cast<uintptr_t>(local);
240 obj.type = BINDER_TYPE_BINDER;
245 return finish_flatten_binder(binder, obj, out);
248 status_t flatten_binder(const sp<ProcessState>& /*proc*/,
249 const wp<IBinder>& binder, Parcel* out)
251 flat_binder_object obj;
253 obj.flags = 0x7f | FLAT_BINDER_FLAG_ACCEPTS_FDS;
254 if (binder != NULL) {
255 sp<IBinder> real = binder.promote();
257 IBinder *local = real->localBinder();
259 BpBinder *proxy = real->remoteBinder();
263 const int32_t handle = proxy ? proxy->handle() : 0;
264 obj.type = BINDER_TYPE_WEAK_HANDLE;
265 obj.binder = 0; /* Don't pass uninitialized stack data to a remote process */
269 obj.type = BINDER_TYPE_WEAK_BINDER;
270 obj.binder = reinterpret_cast<uintptr_t>(binder.get_refs());
271 obj.cookie = reinterpret_cast<uintptr_t>(binder.unsafe_get());
273 return finish_flatten_binder(real, obj, out);
276 // XXX How to deal? In order to flatten the given binder,
277 // we need to probe it for information, which requires a primary
278 // reference... but we don't have one.
280 // The OpenBinder implementation uses a dynamic_cast<> here,
281 // but we can't do that with the different reference counting
282 // implementation we are using.
283 ALOGE("Unable to unflatten Binder weak reference!");
284 obj.type = BINDER_TYPE_BINDER;
287 return finish_flatten_binder(NULL, obj, out);
290 obj.type = BINDER_TYPE_BINDER;
293 return finish_flatten_binder(NULL, obj, out);
297 inline static status_t finish_unflatten_binder(
298 BpBinder* /*proxy*/, const flat_binder_object& /*flat*/,
299 const Parcel& /*in*/)
304 status_t unflatten_binder(const sp<ProcessState>& proc,
305 const Parcel& in, sp<IBinder>* out)
307 const flat_binder_object* flat = in.readObject(false);
310 switch (flat->type) {
311 case BINDER_TYPE_BINDER:
312 *out = reinterpret_cast<IBinder*>(flat->cookie);
313 return finish_unflatten_binder(NULL, *flat, in);
314 case BINDER_TYPE_HANDLE:
315 *out = proc->getStrongProxyForHandle(flat->handle);
316 return finish_unflatten_binder(
317 static_cast<BpBinder*>(out->get()), *flat, in);
323 status_t unflatten_binder(const sp<ProcessState>& proc,
324 const Parcel& in, wp<IBinder>* out)
326 const flat_binder_object* flat = in.readObject(false);
329 switch (flat->type) {
330 case BINDER_TYPE_BINDER:
331 *out = reinterpret_cast<IBinder*>(flat->cookie);
332 return finish_unflatten_binder(NULL, *flat, in);
333 case BINDER_TYPE_WEAK_BINDER:
334 if (flat->binder != 0) {
335 out->set_object_and_refs(
336 reinterpret_cast<IBinder*>(flat->cookie),
337 reinterpret_cast<RefBase::weakref_type*>(flat->binder));
341 return finish_unflatten_binder(NULL, *flat, in);
342 case BINDER_TYPE_HANDLE:
343 case BINDER_TYPE_WEAK_HANDLE:
344 *out = proc->getWeakProxyForHandle(flat->handle);
345 return finish_unflatten_binder(
346 static_cast<BpBinder*>(out->unsafe_get()), *flat, in);
352 // ---------------------------------------------------------------------------
356 LOG_ALLOC("Parcel %p: constructing", this);
363 LOG_ALLOC("Parcel %p: destroyed", this);
366 size_t Parcel::getGlobalAllocSize() {
367 pthread_mutex_lock(&gParcelGlobalAllocSizeLock);
368 size_t size = gParcelGlobalAllocSize;
369 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock);
373 size_t Parcel::getGlobalAllocCount() {
374 pthread_mutex_lock(&gParcelGlobalAllocSizeLock);
375 size_t count = gParcelGlobalAllocCount;
376 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock);
380 const uint8_t* Parcel::data() const
385 size_t Parcel::dataSize() const
387 return (mDataSize > mDataPos ? mDataSize : mDataPos);
390 size_t Parcel::dataAvail() const
392 size_t result = dataSize() - dataPosition();
393 if (result > INT32_MAX) {
399 size_t Parcel::dataPosition() const
404 size_t Parcel::dataCapacity() const
406 return mDataCapacity;
409 status_t Parcel::setDataSize(size_t size)
411 if (size > INT32_MAX) {
412 // don't accept size_t values which may have come from an
413 // inadvertent conversion from a negative int.
418 err = continueWrite(size);
419 if (err == NO_ERROR) {
421 ALOGV("setDataSize Setting data size of %p to %zu", this, mDataSize);
426 void Parcel::setDataPosition(size_t pos) const
428 if (pos > INT32_MAX) {
429 // don't accept size_t values which may have come from an
430 // inadvertent conversion from a negative int.
436 mObjectsSorted = false;
439 status_t Parcel::setDataCapacity(size_t size)
441 if (size > INT32_MAX) {
442 // don't accept size_t values which may have come from an
443 // inadvertent conversion from a negative int.
447 if (size > mDataCapacity) return continueWrite(size);
451 status_t Parcel::setData(const uint8_t* buffer, size_t len)
453 if (len > INT32_MAX) {
454 // don't accept size_t values which may have come from an
455 // inadvertent conversion from a negative int.
459 status_t err = restartWrite(len);
460 if (err == NO_ERROR) {
461 memcpy(const_cast<uint8_t*>(data()), buffer, len);
468 status_t Parcel::appendFrom(const Parcel *parcel, size_t offset, size_t len)
470 const sp<ProcessState> proc(ProcessState::self());
472 const uint8_t *data = parcel->mData;
473 const binder_size_t *objects = parcel->mObjects;
474 size_t size = parcel->mObjectsSize;
475 int startPos = mDataPos;
476 int firstIndex = -1, lastIndex = -2;
482 if (len > INT32_MAX) {
483 // don't accept size_t values which may have come from an
484 // inadvertent conversion from a negative int.
488 // range checks against the source parcel size
489 if ((offset > parcel->mDataSize)
490 || (len > parcel->mDataSize)
491 || (offset + len > parcel->mDataSize)) {
495 // Count objects in range
496 for (int i = 0; i < (int) size; i++) {
497 size_t off = objects[i];
498 if ((off >= offset) && (off + sizeof(flat_binder_object) <= offset + len)) {
499 if (firstIndex == -1) {
505 int numObjects = lastIndex - firstIndex + 1;
507 if ((mDataSize+len) > mDataCapacity) {
510 if (err != NO_ERROR) {
516 memcpy(mData + mDataPos, data + offset, len);
522 if (numObjects > 0) {
524 if (mObjectsCapacity < mObjectsSize + numObjects) {
525 size_t newSize = ((mObjectsSize + numObjects)*3)/2;
526 if (newSize*sizeof(binder_size_t) < mObjectsSize) return NO_MEMORY; // overflow
527 binder_size_t *objects =
528 (binder_size_t*)realloc(mObjects, newSize*sizeof(binder_size_t));
529 if (objects == (binder_size_t*)0) {
533 mObjectsCapacity = newSize;
536 // append and acquire objects
537 int idx = mObjectsSize;
538 for (int i = firstIndex; i <= lastIndex; i++) {
539 size_t off = objects[i] - offset + startPos;
540 mObjects[idx++] = off;
543 flat_binder_object* flat
544 = reinterpret_cast<flat_binder_object*>(mData + off);
545 acquire_object(proc, *flat, this, &mOpenAshmemSize);
547 if (flat->type == BINDER_TYPE_FD) {
548 // If this is a file descriptor, we need to dup it so the
549 // new Parcel now owns its own fd, and can declare that we
550 // officially know we have fds.
551 flat->handle = fcntl(flat->handle, F_DUPFD_CLOEXEC, 0);
553 mHasFds = mFdsKnown = true;
555 err = FDS_NOT_ALLOWED;
564 int Parcel::compareData(const Parcel& other) {
565 size_t size = dataSize();
566 if (size != other.dataSize()) {
567 return size < other.dataSize() ? -1 : 1;
569 return memcmp(data(), other.data(), size);
572 bool Parcel::allowFds() const
577 bool Parcel::pushAllowFds(bool allowFds)
579 const bool origValue = mAllowFds;
586 void Parcel::restoreAllowFds(bool lastValue)
588 mAllowFds = lastValue;
591 bool Parcel::hasFileDescriptors() const
599 // Write RPC headers. (previously just the interface token)
600 status_t Parcel::writeInterfaceToken(const String16& interface)
602 writeInt32(IPCThreadState::self()->getStrictModePolicy() |
603 STRICT_MODE_PENALTY_GATHER);
604 // currently the interface identification token is just its name as a string
605 return writeString16(interface);
608 bool Parcel::checkInterface(IBinder* binder) const
610 return enforceInterface(binder->getInterfaceDescriptor());
613 bool Parcel::enforceInterface(const String16& interface,
614 IPCThreadState* threadState) const
616 int32_t strictPolicy = readInt32();
617 if (threadState == NULL) {
618 threadState = IPCThreadState::self();
620 if ((threadState->getLastTransactionBinderFlags() &
621 IBinder::FLAG_ONEWAY) != 0) {
622 // For one-way calls, the callee is running entirely
623 // disconnected from the caller, so disable StrictMode entirely.
624 // Not only does disk/network usage not impact the caller, but
625 // there's no way to commuicate back any violations anyway.
626 threadState->setStrictModePolicy(0);
628 threadState->setStrictModePolicy(strictPolicy);
630 const String16 str(readString16());
631 if (str == interface) {
634 ALOGW("**** enforceInterface() expected '%s' but read '%s'",
635 String8(interface).string(), String8(str).string());
640 const binder_size_t* Parcel::objects() const
645 size_t Parcel::objectsCount() const
650 status_t Parcel::errorCheck() const
655 void Parcel::setError(status_t err)
660 status_t Parcel::finishWrite(size_t len)
662 if (len > INT32_MAX) {
663 // don't accept size_t values which may have come from an
664 // inadvertent conversion from a negative int.
668 //printf("Finish write of %d\n", len);
670 ALOGV("finishWrite Setting data pos of %p to %zu", this, mDataPos);
671 if (mDataPos > mDataSize) {
672 mDataSize = mDataPos;
673 ALOGV("finishWrite Setting data size of %p to %zu", this, mDataSize);
675 //printf("New pos=%d, size=%d\n", mDataPos, mDataSize);
679 status_t Parcel::writeUnpadded(const void* data, size_t len)
681 if (len > INT32_MAX) {
682 // don't accept size_t values which may have come from an
683 // inadvertent conversion from a negative int.
687 size_t end = mDataPos + len;
688 if (end < mDataPos) {
693 if (end <= mDataCapacity) {
695 memcpy(mData+mDataPos, data, len);
696 return finishWrite(len);
699 status_t err = growData(len);
700 if (err == NO_ERROR) goto restart_write;
704 status_t Parcel::write(const void* data, size_t len)
706 if (len > INT32_MAX) {
707 // don't accept size_t values which may have come from an
708 // inadvertent conversion from a negative int.
712 void* const d = writeInplace(len);
714 memcpy(d, data, len);
720 void* Parcel::writeInplace(size_t len)
722 if (len > INT32_MAX) {
723 // don't accept size_t values which may have come from an
724 // inadvertent conversion from a negative int.
728 const size_t padded = pad_size(len);
730 // sanity check for integer overflow
731 if (mDataPos+padded < mDataPos) {
735 if ((mDataPos+padded) <= mDataCapacity) {
737 //printf("Writing %ld bytes, padded to %ld\n", len, padded);
738 uint8_t* const data = mData+mDataPos;
740 // Need to pad at end?
742 #if BYTE_ORDER == BIG_ENDIAN
743 static const uint32_t mask[4] = {
744 0x00000000, 0xffffff00, 0xffff0000, 0xff000000
747 #if BYTE_ORDER == LITTLE_ENDIAN
748 static const uint32_t mask[4] = {
749 0x00000000, 0x00ffffff, 0x0000ffff, 0x000000ff
752 //printf("Applying pad mask: %p to %p\n", (void*)mask[padded-len],
753 // *reinterpret_cast<void**>(data+padded-4));
754 *reinterpret_cast<uint32_t*>(data+padded-4) &= mask[padded-len];
761 status_t err = growData(padded);
762 if (err == NO_ERROR) goto restart_write;
766 status_t Parcel::writeUtf8AsUtf16(const std::string& str) {
767 const uint8_t* strData = (uint8_t*)str.data();
768 const size_t strLen= str.length();
769 const ssize_t utf16Len = utf8_to_utf16_length(strData, strLen);
770 if (utf16Len < 0 || utf16Len > std::numeric_limits<int32_t>::max()) {
774 status_t err = writeInt32(utf16Len);
779 // Allocate enough bytes to hold our converted string and its terminating NULL.
780 void* dst = writeInplace((utf16Len + 1) * sizeof(char16_t));
785 utf8_to_utf16(strData, strLen, (char16_t*)dst, (size_t) utf16Len + 1);
790 status_t Parcel::writeUtf8AsUtf16(const std::unique_ptr<std::string>& str) {
792 return writeInt32(-1);
794 return writeUtf8AsUtf16(*str);
800 status_t writeByteVectorInternal(Parcel* parcel, const std::vector<T>& val)
803 if (val.size() > std::numeric_limits<int32_t>::max()) {
808 status = parcel->writeInt32(val.size());
813 void* data = parcel->writeInplace(val.size());
819 memcpy(data, val.data(), val.size());
824 status_t writeByteVectorInternalPtr(Parcel* parcel,
825 const std::unique_ptr<std::vector<T>>& val)
828 return parcel->writeInt32(-1);
831 return writeByteVectorInternal(parcel, *val);
836 status_t Parcel::writeByteVector(const std::vector<int8_t>& val) {
837 return writeByteVectorInternal(this, val);
840 status_t Parcel::writeByteVector(const std::unique_ptr<std::vector<int8_t>>& val)
842 return writeByteVectorInternalPtr(this, val);
845 status_t Parcel::writeByteVector(const std::vector<uint8_t>& val) {
846 return writeByteVectorInternal(this, val);
849 status_t Parcel::writeByteVector(const std::unique_ptr<std::vector<uint8_t>>& val)
851 return writeByteVectorInternalPtr(this, val);
854 status_t Parcel::writeInt32Vector(const std::vector<int32_t>& val)
856 return writeTypedVector(val, &Parcel::writeInt32);
859 status_t Parcel::writeInt32Vector(const std::unique_ptr<std::vector<int32_t>>& val)
861 return writeNullableTypedVector(val, &Parcel::writeInt32);
864 status_t Parcel::writeInt64Vector(const std::vector<int64_t>& val)
866 return writeTypedVector(val, &Parcel::writeInt64);
869 status_t Parcel::writeInt64Vector(const std::unique_ptr<std::vector<int64_t>>& val)
871 return writeNullableTypedVector(val, &Parcel::writeInt64);
874 status_t Parcel::writeFloatVector(const std::vector<float>& val)
876 return writeTypedVector(val, &Parcel::writeFloat);
879 status_t Parcel::writeFloatVector(const std::unique_ptr<std::vector<float>>& val)
881 return writeNullableTypedVector(val, &Parcel::writeFloat);
884 status_t Parcel::writeDoubleVector(const std::vector<double>& val)
886 return writeTypedVector(val, &Parcel::writeDouble);
889 status_t Parcel::writeDoubleVector(const std::unique_ptr<std::vector<double>>& val)
891 return writeNullableTypedVector(val, &Parcel::writeDouble);
894 status_t Parcel::writeBoolVector(const std::vector<bool>& val)
896 return writeTypedVector(val, &Parcel::writeBool);
899 status_t Parcel::writeBoolVector(const std::unique_ptr<std::vector<bool>>& val)
901 return writeNullableTypedVector(val, &Parcel::writeBool);
904 status_t Parcel::writeCharVector(const std::vector<char16_t>& val)
906 return writeTypedVector(val, &Parcel::writeChar);
909 status_t Parcel::writeCharVector(const std::unique_ptr<std::vector<char16_t>>& val)
911 return writeNullableTypedVector(val, &Parcel::writeChar);
914 status_t Parcel::writeString16Vector(const std::vector<String16>& val)
916 return writeTypedVector(val, &Parcel::writeString16);
919 status_t Parcel::writeString16Vector(
920 const std::unique_ptr<std::vector<std::unique_ptr<String16>>>& val)
922 return writeNullableTypedVector(val, &Parcel::writeString16);
925 status_t Parcel::writeUtf8VectorAsUtf16Vector(
926 const std::unique_ptr<std::vector<std::unique_ptr<std::string>>>& val) {
927 return writeNullableTypedVector(val, &Parcel::writeUtf8AsUtf16);
930 status_t Parcel::writeUtf8VectorAsUtf16Vector(const std::vector<std::string>& val) {
931 return writeTypedVector(val, &Parcel::writeUtf8AsUtf16);
934 status_t Parcel::writeInt32(int32_t val)
936 return writeAligned(val);
939 status_t Parcel::writeUint32(uint32_t val)
941 return writeAligned(val);
944 status_t Parcel::writeInt32Array(size_t len, const int32_t *val) {
945 if (len > INT32_MAX) {
946 // don't accept size_t values which may have come from an
947 // inadvertent conversion from a negative int.
952 return writeInt32(-1);
954 status_t ret = writeInt32(static_cast<uint32_t>(len));
955 if (ret == NO_ERROR) {
956 ret = write(val, len * sizeof(*val));
960 status_t Parcel::writeByteArray(size_t len, const uint8_t *val) {
961 if (len > INT32_MAX) {
962 // don't accept size_t values which may have come from an
963 // inadvertent conversion from a negative int.
968 return writeInt32(-1);
970 status_t ret = writeInt32(static_cast<uint32_t>(len));
971 if (ret == NO_ERROR) {
972 ret = write(val, len * sizeof(*val));
977 status_t Parcel::writeBool(bool val)
979 return writeInt32(int32_t(val));
982 status_t Parcel::writeChar(char16_t val)
984 return writeInt32(int32_t(val));
987 status_t Parcel::writeByte(int8_t val)
989 return writeInt32(int32_t(val));
992 status_t Parcel::writeInt64(int64_t val)
994 return writeAligned(val);
997 status_t Parcel::writeUint64(uint64_t val)
999 return writeAligned(val);
1002 status_t Parcel::writePointer(uintptr_t val)
1004 return writeAligned<binder_uintptr_t>(val);
1007 status_t Parcel::writeFloat(float val)
1009 return writeAligned(val);
1012 #if defined(__mips__) && defined(__mips_hard_float)
1014 status_t Parcel::writeDouble(double val)
1018 unsigned long long ll;
1021 return writeAligned(u.ll);
1026 status_t Parcel::writeDouble(double val)
1028 return writeAligned(val);
1033 status_t Parcel::writeCString(const char* str)
1035 return write(str, strlen(str)+1);
1038 status_t Parcel::writeString8(const String8& str)
1040 status_t err = writeInt32(str.bytes());
1041 // only write string if its length is more than zero characters,
1042 // as readString8 will only read if the length field is non-zero.
1043 // this is slightly different from how writeString16 works.
1044 if (str.bytes() > 0 && err == NO_ERROR) {
1045 err = write(str.string(), str.bytes()+1);
1050 status_t Parcel::writeString16(const std::unique_ptr<String16>& str)
1053 return writeInt32(-1);
1056 return writeString16(*str);
1059 status_t Parcel::writeString16(const String16& str)
1061 return writeString16(str.string(), str.size());
1064 status_t Parcel::writeString16(const char16_t* str, size_t len)
1066 if (str == NULL) return writeInt32(-1);
1068 status_t err = writeInt32(len);
1069 if (err == NO_ERROR) {
1070 len *= sizeof(char16_t);
1071 uint8_t* data = (uint8_t*)writeInplace(len+sizeof(char16_t));
1073 memcpy(data, str, len);
1074 *reinterpret_cast<char16_t*>(data+len) = 0;
1082 status_t Parcel::writeStrongBinder(const sp<IBinder>& val)
1084 return flatten_binder(ProcessState::self(), val, this);
1087 status_t Parcel::writeStrongBinderVector(const std::vector<sp<IBinder>>& val)
1089 return writeTypedVector(val, &Parcel::writeStrongBinder);
1092 status_t Parcel::writeStrongBinderVector(const std::unique_ptr<std::vector<sp<IBinder>>>& val)
1094 return writeNullableTypedVector(val, &Parcel::writeStrongBinder);
1097 status_t Parcel::readStrongBinderVector(std::unique_ptr<std::vector<sp<IBinder>>>* val) const {
1098 return readNullableTypedVector(val, &Parcel::readNullableStrongBinder);
1101 status_t Parcel::readStrongBinderVector(std::vector<sp<IBinder>>* val) const {
1102 return readTypedVector(val, &Parcel::readStrongBinder);
1105 status_t Parcel::writeWeakBinder(const wp<IBinder>& val)
1107 return flatten_binder(ProcessState::self(), val, this);
1110 status_t Parcel::writeRawNullableParcelable(const Parcelable* parcelable) {
1112 return writeInt32(0);
1115 return writeParcelable(*parcelable);
1118 status_t Parcel::writeParcelable(const Parcelable& parcelable) {
1119 status_t status = writeInt32(1); // parcelable is not null.
1123 return parcelable.writeToParcel(this);
1126 status_t Parcel::writeValue(const binder::Value& value) {
1127 return value.writeToParcel(this);
1130 status_t Parcel::writeNativeHandle(const native_handle* handle)
1132 if (!handle || handle->version != sizeof(native_handle))
1136 err = writeInt32(handle->numFds);
1137 if (err != NO_ERROR) return err;
1139 err = writeInt32(handle->numInts);
1140 if (err != NO_ERROR) return err;
1142 for (int i=0 ; err==NO_ERROR && i<handle->numFds ; i++)
1143 err = writeDupFileDescriptor(handle->data[i]);
1145 if (err != NO_ERROR) {
1146 ALOGD("write native handle, write dup fd failed");
1149 err = write(handle->data + handle->numFds, sizeof(int)*handle->numInts);
1153 status_t Parcel::writeFileDescriptor(int fd, bool takeOwnership)
1155 flat_binder_object obj;
1156 obj.type = BINDER_TYPE_FD;
1157 obj.flags = 0x7f | FLAT_BINDER_FLAG_ACCEPTS_FDS;
1158 obj.binder = 0; /* Don't pass uninitialized stack data to a remote process */
1160 obj.cookie = takeOwnership ? 1 : 0;
1161 return writeObject(obj, true);
1164 status_t Parcel::writeDupFileDescriptor(int fd)
1166 int dupFd = fcntl(fd, F_DUPFD_CLOEXEC, 0);
1170 status_t err = writeFileDescriptor(dupFd, true /*takeOwnership*/);
1177 status_t Parcel::writeParcelFileDescriptor(int fd, bool takeOwnership)
1180 return writeFileDescriptor(fd, takeOwnership);
1183 status_t Parcel::writeUniqueFileDescriptor(const base::unique_fd& fd) {
1184 return writeDupFileDescriptor(fd.get());
1187 status_t Parcel::writeUniqueFileDescriptorVector(const std::vector<base::unique_fd>& val) {
1188 return writeTypedVector(val, &Parcel::writeUniqueFileDescriptor);
1191 status_t Parcel::writeUniqueFileDescriptorVector(const std::unique_ptr<std::vector<base::unique_fd>>& val) {
1192 return writeNullableTypedVector(val, &Parcel::writeUniqueFileDescriptor);
1195 status_t Parcel::writeBlob(size_t len, bool mutableCopy, WritableBlob* outBlob)
1197 if (len > INT32_MAX) {
1198 // don't accept size_t values which may have come from an
1199 // inadvertent conversion from a negative int.
1204 if (!mAllowFds || len <= BLOB_INPLACE_LIMIT) {
1205 ALOGV("writeBlob: write in place");
1206 status = writeInt32(BLOB_INPLACE);
1207 if (status) return status;
1209 void* ptr = writeInplace(len);
1210 if (!ptr) return NO_MEMORY;
1212 outBlob->init(-1, ptr, len, false);
1216 ALOGV("writeBlob: write to ashmem");
1217 int fd = ashmem_create_region("Parcel Blob", len);
1218 if (fd < 0) return NO_MEMORY;
1220 int result = ashmem_set_prot_region(fd, PROT_READ | PROT_WRITE);
1224 void* ptr = ::mmap(NULL, len, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
1225 if (ptr == MAP_FAILED) {
1229 result = ashmem_set_prot_region(fd, PROT_READ);
1234 status = writeInt32(mutableCopy ? BLOB_ASHMEM_MUTABLE : BLOB_ASHMEM_IMMUTABLE);
1236 status = writeFileDescriptor(fd, true /*takeOwnership*/);
1238 outBlob->init(fd, ptr, len, mutableCopy);
1250 status_t Parcel::writeDupImmutableBlobFileDescriptor(int fd)
1252 // Must match up with what's done in writeBlob.
1253 if (!mAllowFds) return FDS_NOT_ALLOWED;
1254 status_t status = writeInt32(BLOB_ASHMEM_IMMUTABLE);
1255 if (status) return status;
1256 return writeDupFileDescriptor(fd);
1259 status_t Parcel::write(const FlattenableHelperInterface& val)
1264 const size_t len = val.getFlattenedSize();
1265 const size_t fd_count = val.getFdCount();
1267 if ((len > INT32_MAX) || (fd_count >= gMaxFds)) {
1268 // don't accept size_t values which may have come from an
1269 // inadvertent conversion from a negative int.
1273 err = this->writeInt32(len);
1274 if (err) return err;
1276 err = this->writeInt32(fd_count);
1277 if (err) return err;
1280 void* const buf = this->writeInplace(pad_size(len));
1286 fds = new (std::nothrow) int[fd_count];
1287 if (fds == nullptr) {
1288 ALOGE("write: failed to allocate requested %zu fds", fd_count);
1293 err = val.flatten(buf, len, fds, fd_count);
1294 for (size_t i=0 ; i<fd_count && err==NO_ERROR ; i++) {
1295 err = this->writeDupFileDescriptor( fds[i] );
1305 status_t Parcel::writeObject(const flat_binder_object& val, bool nullMetaData)
1307 const bool enoughData = (mDataPos+sizeof(val)) <= mDataCapacity;
1308 const bool enoughObjects = mObjectsSize < mObjectsCapacity;
1309 if (enoughData && enoughObjects) {
1311 *reinterpret_cast<flat_binder_object*>(mData+mDataPos) = val;
1313 // remember if it's a file descriptor
1314 if (val.type == BINDER_TYPE_FD) {
1316 // fail before modifying our object index
1317 return FDS_NOT_ALLOWED;
1319 mHasFds = mFdsKnown = true;
1322 // Need to write meta-data?
1323 if (nullMetaData || val.binder != 0) {
1324 mObjects[mObjectsSize] = mDataPos;
1325 acquire_object(ProcessState::self(), val, this, &mOpenAshmemSize);
1329 return finishWrite(sizeof(flat_binder_object));
1333 const status_t err = growData(sizeof(val));
1334 if (err != NO_ERROR) return err;
1336 if (!enoughObjects) {
1337 size_t newSize = ((mObjectsSize+2)*3)/2;
1338 if (newSize*sizeof(binder_size_t) < mObjectsSize) return NO_MEMORY; // overflow
1339 binder_size_t* objects = (binder_size_t*)realloc(mObjects, newSize*sizeof(binder_size_t));
1340 if (objects == NULL) return NO_MEMORY;
1342 mObjectsCapacity = newSize;
1348 status_t Parcel::writeNoException()
1350 binder::Status status;
1351 return status.writeToParcel(this);
1354 status_t Parcel::writeMap(const ::android::binder::Map& map_in)
1357 using ::android::binder::Value;
1358 using ::android::binder::Map;
1360 Map::const_iterator iter;
1363 ret = writeInt32(map_in.size());
1365 if (ret != NO_ERROR) {
1369 for (iter = map_in.begin(); iter != map_in.end(); ++iter) {
1370 ret = writeValue(Value(iter->first));
1371 if (ret != NO_ERROR) {
1375 ret = writeValue(iter->second);
1376 if (ret != NO_ERROR) {
1384 status_t Parcel::writeNullableMap(const std::unique_ptr<binder::Map>& map)
1387 return writeInt32(-1);
1390 return writeMap(*map.get());
1393 status_t Parcel::readMap(::android::binder::Map* map_out)const
1396 using ::android::String16;
1397 using ::android::String8;
1398 using ::android::binder::Value;
1399 using ::android::binder::Map;
1401 status_t ret = NO_ERROR;
1404 ret = readInt32(&count);
1405 if (ret != NO_ERROR) {
1410 ALOGE("readMap: Unexpected count: %d", count);
1411 return (count == -1)
1422 ret = readValue(&value);
1423 if (ret != NO_ERROR) {
1427 if (!value.getString(&key)) {
1428 ALOGE("readMap: Key type not a string (parcelType = %d)", value.parcelType());
1432 ret = readValue(&value);
1433 if (ret != NO_ERROR) {
1437 (*map_out)[key] = value;
1443 status_t Parcel::readNullableMap(std::unique_ptr<binder::Map>* map) const
1445 const size_t start = dataPosition();
1447 status_t status = readInt32(&count);
1450 if (status != OK || count == -1) {
1454 setDataPosition(start);
1455 map->reset(new binder::Map());
1457 status = readMap(map->get());
1468 void Parcel::remove(size_t /*start*/, size_t /*amt*/)
1470 LOG_ALWAYS_FATAL("Parcel::remove() not yet implemented!");
1473 status_t Parcel::validateReadData(size_t upperBound) const
1475 // Don't allow non-object reads on object data
1476 if (mObjectsSorted || mObjectsSize <= 1) {
1478 // Expect to check only against the next object
1479 if (mNextObjectHint < mObjectsSize && upperBound > mObjects[mNextObjectHint]) {
1480 // For some reason the current read position is greater than the next object
1481 // hint. Iterate until we find the right object
1482 size_t nextObject = mNextObjectHint;
1484 if (mDataPos < mObjects[nextObject] + sizeof(flat_binder_object)) {
1485 // Requested info overlaps with an object
1486 ALOGE("Attempt to read from protected data in Parcel %p", this);
1487 return PERMISSION_DENIED;
1490 } while (nextObject < mObjectsSize && upperBound > mObjects[nextObject]);
1491 mNextObjectHint = nextObject;
1495 // Quickly determine if mObjects is sorted.
1496 binder_size_t* currObj = mObjects + mObjectsSize - 1;
1497 binder_size_t* prevObj = currObj;
1498 while (currObj > mObjects) {
1500 if(*prevObj > *currObj) {
1505 mObjectsSorted = true;
1509 // Insertion Sort mObjects
1510 // Great for mostly sorted lists. If randomly sorted or reverse ordered mObjects become common,
1511 // switch to std::sort(mObjects, mObjects + mObjectsSize);
1512 for (binder_size_t* iter0 = mObjects + 1; iter0 < mObjects + mObjectsSize; iter0++) {
1513 binder_size_t temp = *iter0;
1514 binder_size_t* iter1 = iter0 - 1;
1515 while (iter1 >= mObjects && *iter1 > temp) {
1516 *(iter1 + 1) = *iter1;
1519 *(iter1 + 1) = temp;
1521 mNextObjectHint = 0;
1522 mObjectsSorted = true;
1526 status_t Parcel::read(void* outData, size_t len) const
1528 if (len > INT32_MAX) {
1529 // don't accept size_t values which may have come from an
1530 // inadvertent conversion from a negative int.
1534 if ((mDataPos+pad_size(len)) >= mDataPos && (mDataPos+pad_size(len)) <= mDataSize
1535 && len <= pad_size(len)) {
1536 if (mObjectsSize > 0) {
1537 status_t err = validateReadData(mDataPos + pad_size(len));
1538 if(err != NO_ERROR) return err;
1540 memcpy(outData, mData+mDataPos, len);
1541 mDataPos += pad_size(len);
1542 ALOGV("read Setting data pos of %p to %zu", this, mDataPos);
1545 return NOT_ENOUGH_DATA;
1548 const void* Parcel::readInplace(size_t len) const
1550 if (len > INT32_MAX) {
1551 // don't accept size_t values which may have come from an
1552 // inadvertent conversion from a negative int.
1556 if ((mDataPos+pad_size(len)) >= mDataPos && (mDataPos+pad_size(len)) <= mDataSize
1557 && len <= pad_size(len)) {
1558 if (mObjectsSize > 0) {
1559 status_t err = validateReadData(mDataPos + pad_size(len));
1560 if(err != NO_ERROR) return NULL;
1563 const void* data = mData+mDataPos;
1564 mDataPos += pad_size(len);
1565 ALOGV("readInplace Setting data pos of %p to %zu", this, mDataPos);
1572 status_t Parcel::readAligned(T *pArg) const {
1573 COMPILE_TIME_ASSERT_FUNCTION_SCOPE(PAD_SIZE_UNSAFE(sizeof(T)) == sizeof(T));
1575 if ((mDataPos+sizeof(T)) <= mDataSize) {
1576 if (mObjectsSize > 0) {
1577 status_t err = validateReadData(mDataPos + sizeof(T));
1578 if(err != NO_ERROR) return err;
1581 const void* data = mData+mDataPos;
1582 mDataPos += sizeof(T);
1583 *pArg = *reinterpret_cast<const T*>(data);
1586 return NOT_ENOUGH_DATA;
1591 T Parcel::readAligned() const {
1593 if (readAligned(&result) != NO_ERROR) {
1601 status_t Parcel::writeAligned(T val) {
1602 COMPILE_TIME_ASSERT_FUNCTION_SCOPE(PAD_SIZE_UNSAFE(sizeof(T)) == sizeof(T));
1604 if ((mDataPos+sizeof(val)) <= mDataCapacity) {
1606 *reinterpret_cast<T*>(mData+mDataPos) = val;
1607 return finishWrite(sizeof(val));
1610 status_t err = growData(sizeof(val));
1611 if (err == NO_ERROR) goto restart_write;
1617 template<typename T>
1618 status_t readByteVectorInternal(const Parcel* parcel,
1619 std::vector<T>* val) {
1623 status_t status = parcel->readInt32(&size);
1630 status = UNEXPECTED_NULL;
1633 if (size_t(size) > parcel->dataAvail()) {
1638 T* data = const_cast<T*>(reinterpret_cast<const T*>(parcel->readInplace(size)));
1644 val->insert(val->end(), data, data + size);
1649 template<typename T>
1650 status_t readByteVectorInternalPtr(
1651 const Parcel* parcel,
1652 std::unique_ptr<std::vector<T>>* val) {
1653 const int32_t start = parcel->dataPosition();
1655 status_t status = parcel->readInt32(&size);
1658 if (status != OK || size < 0) {
1662 parcel->setDataPosition(start);
1663 val->reset(new (std::nothrow) std::vector<T>());
1665 status = readByteVectorInternal(parcel, val->get());
1676 status_t Parcel::readByteVector(std::vector<int8_t>* val) const {
1677 return readByteVectorInternal(this, val);
1680 status_t Parcel::readByteVector(std::vector<uint8_t>* val) const {
1681 return readByteVectorInternal(this, val);
1684 status_t Parcel::readByteVector(std::unique_ptr<std::vector<int8_t>>* val) const {
1685 return readByteVectorInternalPtr(this, val);
1688 status_t Parcel::readByteVector(std::unique_ptr<std::vector<uint8_t>>* val) const {
1689 return readByteVectorInternalPtr(this, val);
1692 status_t Parcel::readInt32Vector(std::unique_ptr<std::vector<int32_t>>* val) const {
1693 return readNullableTypedVector(val, &Parcel::readInt32);
1696 status_t Parcel::readInt32Vector(std::vector<int32_t>* val) const {
1697 return readTypedVector(val, &Parcel::readInt32);
1700 status_t Parcel::readInt64Vector(std::unique_ptr<std::vector<int64_t>>* val) const {
1701 return readNullableTypedVector(val, &Parcel::readInt64);
1704 status_t Parcel::readInt64Vector(std::vector<int64_t>* val) const {
1705 return readTypedVector(val, &Parcel::readInt64);
1708 status_t Parcel::readFloatVector(std::unique_ptr<std::vector<float>>* val) const {
1709 return readNullableTypedVector(val, &Parcel::readFloat);
1712 status_t Parcel::readFloatVector(std::vector<float>* val) const {
1713 return readTypedVector(val, &Parcel::readFloat);
1716 status_t Parcel::readDoubleVector(std::unique_ptr<std::vector<double>>* val) const {
1717 return readNullableTypedVector(val, &Parcel::readDouble);
1720 status_t Parcel::readDoubleVector(std::vector<double>* val) const {
1721 return readTypedVector(val, &Parcel::readDouble);
1724 status_t Parcel::readBoolVector(std::unique_ptr<std::vector<bool>>* val) const {
1725 const int32_t start = dataPosition();
1727 status_t status = readInt32(&size);
1730 if (status != OK || size < 0) {
1734 setDataPosition(start);
1735 val->reset(new (std::nothrow) std::vector<bool>());
1737 status = readBoolVector(val->get());
1746 status_t Parcel::readBoolVector(std::vector<bool>* val) const {
1748 status_t status = readInt32(&size);
1755 return UNEXPECTED_NULL;
1760 /* C++ bool handling means a vector of bools isn't necessarily addressable
1761 * (we might use individual bits)
1764 for (int32_t i = 0; i < size; ++i) {
1765 status = readBool(&data);
1776 status_t Parcel::readCharVector(std::unique_ptr<std::vector<char16_t>>* val) const {
1777 return readNullableTypedVector(val, &Parcel::readChar);
1780 status_t Parcel::readCharVector(std::vector<char16_t>* val) const {
1781 return readTypedVector(val, &Parcel::readChar);
1784 status_t Parcel::readString16Vector(
1785 std::unique_ptr<std::vector<std::unique_ptr<String16>>>* val) const {
1786 return readNullableTypedVector(val, &Parcel::readString16);
1789 status_t Parcel::readString16Vector(std::vector<String16>* val) const {
1790 return readTypedVector(val, &Parcel::readString16);
1793 status_t Parcel::readUtf8VectorFromUtf16Vector(
1794 std::unique_ptr<std::vector<std::unique_ptr<std::string>>>* val) const {
1795 return readNullableTypedVector(val, &Parcel::readUtf8FromUtf16);
1798 status_t Parcel::readUtf8VectorFromUtf16Vector(std::vector<std::string>* val) const {
1799 return readTypedVector(val, &Parcel::readUtf8FromUtf16);
1802 status_t Parcel::readInt32(int32_t *pArg) const
1804 return readAligned(pArg);
1807 int32_t Parcel::readInt32() const
1809 return readAligned<int32_t>();
1812 status_t Parcel::readUint32(uint32_t *pArg) const
1814 return readAligned(pArg);
1817 uint32_t Parcel::readUint32() const
1819 return readAligned<uint32_t>();
1822 status_t Parcel::readInt64(int64_t *pArg) const
1824 return readAligned(pArg);
1828 int64_t Parcel::readInt64() const
1830 return readAligned<int64_t>();
1833 status_t Parcel::readUint64(uint64_t *pArg) const
1835 return readAligned(pArg);
1838 uint64_t Parcel::readUint64() const
1840 return readAligned<uint64_t>();
1843 status_t Parcel::readPointer(uintptr_t *pArg) const
1846 binder_uintptr_t ptr;
1847 ret = readAligned(&ptr);
1853 uintptr_t Parcel::readPointer() const
1855 return readAligned<binder_uintptr_t>();
1859 status_t Parcel::readFloat(float *pArg) const
1861 return readAligned(pArg);
1865 float Parcel::readFloat() const
1867 return readAligned<float>();
1870 #if defined(__mips__) && defined(__mips_hard_float)
1872 status_t Parcel::readDouble(double *pArg) const
1876 unsigned long long ll;
1880 status = readAligned(&u.ll);
1885 double Parcel::readDouble() const
1889 unsigned long long ll;
1891 u.ll = readAligned<unsigned long long>();
1897 status_t Parcel::readDouble(double *pArg) const
1899 return readAligned(pArg);
1902 double Parcel::readDouble() const
1904 return readAligned<double>();
1909 status_t Parcel::readIntPtr(intptr_t *pArg) const
1911 return readAligned(pArg);
1915 intptr_t Parcel::readIntPtr() const
1917 return readAligned<intptr_t>();
1920 status_t Parcel::readBool(bool *pArg) const
1923 status_t ret = readInt32(&tmp);
1928 bool Parcel::readBool() const
1930 return readInt32() != 0;
1933 status_t Parcel::readChar(char16_t *pArg) const
1936 status_t ret = readInt32(&tmp);
1937 *pArg = char16_t(tmp);
1941 char16_t Parcel::readChar() const
1943 return char16_t(readInt32());
1946 status_t Parcel::readByte(int8_t *pArg) const
1949 status_t ret = readInt32(&tmp);
1950 *pArg = int8_t(tmp);
1954 int8_t Parcel::readByte() const
1956 return int8_t(readInt32());
1959 status_t Parcel::readUtf8FromUtf16(std::string* str) const {
1960 size_t utf16Size = 0;
1961 const char16_t* src = readString16Inplace(&utf16Size);
1963 return UNEXPECTED_NULL;
1966 // Save ourselves the trouble, we're done.
1967 if (utf16Size == 0u) {
1972 // Allow for closing '\0'
1973 ssize_t utf8Size = utf16_to_utf8_length(src, utf16Size) + 1;
1977 // Note that while it is probably safe to assume string::resize keeps a
1978 // spare byte around for the trailing null, we still pass the size including the trailing null
1979 str->resize(utf8Size);
1980 utf16_to_utf8(src, utf16Size, &((*str)[0]), utf8Size);
1981 str->resize(utf8Size - 1);
1985 status_t Parcel::readUtf8FromUtf16(std::unique_ptr<std::string>* str) const {
1986 const int32_t start = dataPosition();
1988 status_t status = readInt32(&size);
1991 if (status != OK || size < 0) {
1995 setDataPosition(start);
1996 str->reset(new (std::nothrow) std::string());
1997 return readUtf8FromUtf16(str->get());
2000 const char* Parcel::readCString() const
2002 const size_t avail = mDataSize-mDataPos;
2004 const char* str = reinterpret_cast<const char*>(mData+mDataPos);
2005 // is the string's trailing NUL within the parcel's valid bounds?
2006 const char* eos = reinterpret_cast<const char*>(memchr(str, 0, avail));
2008 const size_t len = eos - str;
2009 mDataPos += pad_size(len+1);
2010 ALOGV("readCString Setting data pos of %p to %zu", this, mDataPos);
2017 String8 Parcel::readString8() const
2020 status_t status = readString8(&retString);
2022 // We don't care about errors here, so just return an empty string.
2028 status_t Parcel::readString8(String8* pArg) const
2031 status_t status = readInt32(&size);
2035 // watch for potential int overflow from size+1
2036 if (size < 0 || size >= INT32_MAX) {
2039 // |writeString8| writes nothing for empty string.
2044 const char* str = (const char*)readInplace(size + 1);
2048 pArg->setTo(str, size);
2052 String16 Parcel::readString16() const
2055 const char16_t* str = readString16Inplace(&len);
2056 if (str) return String16(str, len);
2057 ALOGE("Reading a NULL string not supported here.");
2061 status_t Parcel::readString16(std::unique_ptr<String16>* pArg) const
2063 const int32_t start = dataPosition();
2065 status_t status = readInt32(&size);
2068 if (status != OK || size < 0) {
2072 setDataPosition(start);
2073 pArg->reset(new (std::nothrow) String16());
2075 status = readString16(pArg->get());
2084 status_t Parcel::readString16(String16* pArg) const
2087 const char16_t* str = readString16Inplace(&len);
2089 pArg->setTo(str, len);
2093 return UNEXPECTED_NULL;
2097 const char16_t* Parcel::readString16Inplace(size_t* outLen) const
2099 int32_t size = readInt32();
2100 // watch for potential int overflow from size+1
2101 if (size >= 0 && size < INT32_MAX) {
2103 const char16_t* str = (const char16_t*)readInplace((size+1)*sizeof(char16_t));
2112 status_t Parcel::readStrongBinder(sp<IBinder>* val) const
2114 status_t status = readNullableStrongBinder(val);
2115 if (status == OK && !val->get()) {
2116 status = UNEXPECTED_NULL;
2121 status_t Parcel::readNullableStrongBinder(sp<IBinder>* val) const
2123 return unflatten_binder(ProcessState::self(), *this, val);
2126 sp<IBinder> Parcel::readStrongBinder() const
2129 // Note that a lot of code in Android reads binders by hand with this
2130 // method, and that code has historically been ok with getting nullptr
2131 // back (while ignoring error codes).
2132 readNullableStrongBinder(&val);
2136 wp<IBinder> Parcel::readWeakBinder() const
2139 unflatten_binder(ProcessState::self(), *this, &val);
2143 status_t Parcel::readParcelable(Parcelable* parcelable) const {
2144 int32_t have_parcelable = 0;
2145 status_t status = readInt32(&have_parcelable);
2149 if (!have_parcelable) {
2150 return UNEXPECTED_NULL;
2152 return parcelable->readFromParcel(this);
2155 status_t Parcel::readValue(binder::Value* value) const {
2156 return value->readFromParcel(this);
2159 int32_t Parcel::readExceptionCode() const
2161 binder::Status status;
2162 status.readFromParcel(*this);
2163 return status.exceptionCode();
2166 native_handle* Parcel::readNativeHandle() const
2168 int numFds, numInts;
2170 err = readInt32(&numFds);
2171 if (err != NO_ERROR) return 0;
2172 err = readInt32(&numInts);
2173 if (err != NO_ERROR) return 0;
2175 native_handle* h = native_handle_create(numFds, numInts);
2180 for (int i=0 ; err==NO_ERROR && i<numFds ; i++) {
2181 h->data[i] = fcntl(readFileDescriptor(), F_DUPFD_CLOEXEC, 0);
2182 if (h->data[i] < 0) {
2183 for (int j = 0; j < i; j++) {
2186 native_handle_delete(h);
2190 err = read(h->data + numFds, sizeof(int)*numInts);
2191 if (err != NO_ERROR) {
2192 native_handle_close(h);
2193 native_handle_delete(h);
2199 int Parcel::readFileDescriptor() const
2201 const flat_binder_object* flat = readObject(true);
2203 if (flat && flat->type == BINDER_TYPE_FD) {
2204 return flat->handle;
2210 int Parcel::readParcelFileDescriptor() const
2212 int32_t hasComm = readInt32();
2213 int fd = readFileDescriptor();
2216 readFileDescriptor();
2221 status_t Parcel::readUniqueFileDescriptor(base::unique_fd* val) const
2223 int got = readFileDescriptor();
2225 if (got == BAD_TYPE) {
2229 val->reset(fcntl(got, F_DUPFD_CLOEXEC, 0));
2231 if (val->get() < 0) {
2239 status_t Parcel::readUniqueFileDescriptorVector(std::unique_ptr<std::vector<base::unique_fd>>* val) const {
2240 return readNullableTypedVector(val, &Parcel::readUniqueFileDescriptor);
2243 status_t Parcel::readUniqueFileDescriptorVector(std::vector<base::unique_fd>* val) const {
2244 return readTypedVector(val, &Parcel::readUniqueFileDescriptor);
2247 status_t Parcel::readBlob(size_t len, ReadableBlob* outBlob) const
2250 status_t status = readInt32(&blobType);
2251 if (status) return status;
2253 if (blobType == BLOB_INPLACE) {
2254 ALOGV("readBlob: read in place");
2255 const void* ptr = readInplace(len);
2256 if (!ptr) return BAD_VALUE;
2258 outBlob->init(-1, const_cast<void*>(ptr), len, false);
2262 ALOGV("readBlob: read from ashmem");
2263 bool isMutable = (blobType == BLOB_ASHMEM_MUTABLE);
2264 int fd = readFileDescriptor();
2265 if (fd == int(BAD_TYPE)) return BAD_VALUE;
2267 void* ptr = ::mmap(NULL, len, isMutable ? PROT_READ | PROT_WRITE : PROT_READ,
2269 if (ptr == MAP_FAILED) return NO_MEMORY;
2271 outBlob->init(fd, ptr, len, isMutable);
2275 status_t Parcel::read(FlattenableHelperInterface& val) const
2278 const size_t len = this->readInt32();
2279 const size_t fd_count = this->readInt32();
2281 if ((len > INT32_MAX) || (fd_count >= gMaxFds)) {
2282 // don't accept size_t values which may have come from an
2283 // inadvertent conversion from a negative int.
2288 void const* const buf = this->readInplace(pad_size(len));
2294 fds = new (std::nothrow) int[fd_count];
2295 if (fds == nullptr) {
2296 ALOGE("read: failed to allocate requested %zu fds", fd_count);
2301 status_t err = NO_ERROR;
2302 for (size_t i=0 ; i<fd_count && err==NO_ERROR ; i++) {
2303 int fd = this->readFileDescriptor();
2304 if (fd < 0 || ((fds[i] = fcntl(fd, F_DUPFD_CLOEXEC, 0)) < 0)) {
2306 ALOGE("fcntl(F_DUPFD_CLOEXEC) failed in Parcel::read, i is %zu, fds[i] is %d, fd_count is %zu, error: %s",
2307 i, fds[i], fd_count, strerror(fd < 0 ? -fd : errno));
2308 // Close all the file descriptors that were dup-ed.
2309 for (size_t j=0; j<i ;j++) {
2315 if (err == NO_ERROR) {
2316 err = val.unflatten(buf, len, fds, fd_count);
2325 const flat_binder_object* Parcel::readObject(bool nullMetaData) const
2327 const size_t DPOS = mDataPos;
2328 if ((DPOS+sizeof(flat_binder_object)) <= mDataSize) {
2329 const flat_binder_object* obj
2330 = reinterpret_cast<const flat_binder_object*>(mData+DPOS);
2331 mDataPos = DPOS + sizeof(flat_binder_object);
2332 if (!nullMetaData && (obj->cookie == 0 && obj->binder == 0)) {
2333 // When transferring a NULL object, we don't write it into
2334 // the object list, so we don't want to check for it when
2336 ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos);
2340 // Ensure that this object is valid...
2341 binder_size_t* const OBJS = mObjects;
2342 const size_t N = mObjectsSize;
2343 size_t opos = mNextObjectHint;
2346 ALOGV("Parcel %p looking for obj at %zu, hint=%zu",
2349 // Start at the current hint position, looking for an object at
2350 // the current data position.
2352 while (opos < (N-1) && OBJS[opos] < DPOS) {
2358 if (OBJS[opos] == DPOS) {
2360 ALOGV("Parcel %p found obj %zu at index %zu with forward search",
2362 mNextObjectHint = opos+1;
2363 ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos);
2367 // Look backwards for it...
2368 while (opos > 0 && OBJS[opos] > DPOS) {
2371 if (OBJS[opos] == DPOS) {
2373 ALOGV("Parcel %p found obj %zu at index %zu with backward search",
2375 mNextObjectHint = opos+1;
2376 ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos);
2380 ALOGW("Attempt to read object from Parcel %p at offset %zu that is not in the object list",
2386 void Parcel::closeFileDescriptors()
2388 size_t i = mObjectsSize;
2390 //ALOGI("Closing file descriptors for %zu objects...", i);
2394 const flat_binder_object* flat
2395 = reinterpret_cast<flat_binder_object*>(mData+mObjects[i]);
2396 if (flat->type == BINDER_TYPE_FD) {
2397 //ALOGI("Closing fd: %ld", flat->handle);
2398 close(flat->handle);
2403 uintptr_t Parcel::ipcData() const
2405 return reinterpret_cast<uintptr_t>(mData);
2408 size_t Parcel::ipcDataSize() const
2410 return (mDataSize > mDataPos ? mDataSize : mDataPos);
2413 uintptr_t Parcel::ipcObjects() const
2415 return reinterpret_cast<uintptr_t>(mObjects);
2418 size_t Parcel::ipcObjectsCount() const
2420 return mObjectsSize;
2423 void Parcel::ipcSetDataReference(const uint8_t* data, size_t dataSize,
2424 const binder_size_t* objects, size_t objectsCount, release_func relFunc, void* relCookie)
2426 binder_size_t minOffset = 0;
2429 mData = const_cast<uint8_t*>(data);
2430 mDataSize = mDataCapacity = dataSize;
2431 //ALOGI("setDataReference Setting data size of %p to %lu (pid=%d)", this, mDataSize, getpid());
2433 ALOGV("setDataReference Setting data pos of %p to %zu", this, mDataPos);
2434 mObjects = const_cast<binder_size_t*>(objects);
2435 mObjectsSize = mObjectsCapacity = objectsCount;
2436 mNextObjectHint = 0;
2437 mObjectsSorted = false;
2439 mOwnerCookie = relCookie;
2440 for (size_t i = 0; i < mObjectsSize; i++) {
2441 binder_size_t offset = mObjects[i];
2442 if (offset < minOffset) {
2443 ALOGE("%s: bad object offset %" PRIu64 " < %" PRIu64 "\n",
2444 __func__, (uint64_t)offset, (uint64_t)minOffset);
2448 minOffset = offset + sizeof(flat_binder_object);
2453 void Parcel::print(TextOutput& to, uint32_t /*flags*/) const
2457 if (errorCheck() != NO_ERROR) {
2458 const status_t err = errorCheck();
2459 to << "Error: " << (void*)(intptr_t)err << " \"" << strerror(-err) << "\"";
2460 } else if (dataSize() > 0) {
2461 const uint8_t* DATA = data();
2462 to << indent << HexDump(DATA, dataSize()) << dedent;
2463 const binder_size_t* OBJS = objects();
2464 const size_t N = objectsCount();
2465 for (size_t i=0; i<N; i++) {
2466 const flat_binder_object* flat
2467 = reinterpret_cast<const flat_binder_object*>(DATA+OBJS[i]);
2468 to << endl << "Object #" << i << " @ " << (void*)OBJS[i] << ": "
2469 << TypeCode(flat->type & 0x7f7f7f00)
2470 << " = " << flat->binder;
2479 void Parcel::releaseObjects()
2481 const sp<ProcessState> proc(ProcessState::self());
2482 size_t i = mObjectsSize;
2483 uint8_t* const data = mData;
2484 binder_size_t* const objects = mObjects;
2487 const flat_binder_object* flat
2488 = reinterpret_cast<flat_binder_object*>(data+objects[i]);
2489 release_object(proc, *flat, this, &mOpenAshmemSize);
2493 void Parcel::acquireObjects()
2495 const sp<ProcessState> proc(ProcessState::self());
2496 size_t i = mObjectsSize;
2497 uint8_t* const data = mData;
2498 binder_size_t* const objects = mObjects;
2501 const flat_binder_object* flat
2502 = reinterpret_cast<flat_binder_object*>(data+objects[i]);
2503 acquire_object(proc, *flat, this, &mOpenAshmemSize);
2507 void Parcel::freeData()
2513 void Parcel::freeDataNoInit()
2516 LOG_ALLOC("Parcel %p: freeing other owner data", this);
2517 //ALOGI("Freeing data ref of %p (pid=%d)", this, getpid());
2518 mOwner(this, mData, mDataSize, mObjects, mObjectsSize, mOwnerCookie);
2520 LOG_ALLOC("Parcel %p: freeing allocated data", this);
2523 LOG_ALLOC("Parcel %p: freeing with %zu capacity", this, mDataCapacity);
2524 pthread_mutex_lock(&gParcelGlobalAllocSizeLock);
2525 if (mDataCapacity <= gParcelGlobalAllocSize) {
2526 gParcelGlobalAllocSize = gParcelGlobalAllocSize - mDataCapacity;
2528 gParcelGlobalAllocSize = 0;
2530 if (gParcelGlobalAllocCount > 0) {
2531 gParcelGlobalAllocCount--;
2533 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock);
2536 if (mObjects) free(mObjects);
2540 status_t Parcel::growData(size_t len)
2542 if (len > INT32_MAX) {
2543 // don't accept size_t values which may have come from an
2544 // inadvertent conversion from a negative int.
2548 size_t newSize = ((mDataSize+len)*3)/2;
2549 return (newSize <= mDataSize)
2550 ? (status_t) NO_MEMORY
2551 : continueWrite(newSize);
2554 status_t Parcel::restartWrite(size_t desired)
2556 if (desired > INT32_MAX) {
2557 // don't accept size_t values which may have come from an
2558 // inadvertent conversion from a negative int.
2564 return continueWrite(desired);
2567 uint8_t* data = (uint8_t*)realloc(mData, desired);
2568 if (!data && desired > mDataCapacity) {
2576 LOG_ALLOC("Parcel %p: restart from %zu to %zu capacity", this, mDataCapacity, desired);
2577 pthread_mutex_lock(&gParcelGlobalAllocSizeLock);
2578 gParcelGlobalAllocSize += desired;
2579 gParcelGlobalAllocSize -= mDataCapacity;
2581 gParcelGlobalAllocCount++;
2583 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock);
2585 mDataCapacity = desired;
2588 mDataSize = mDataPos = 0;
2589 ALOGV("restartWrite Setting data size of %p to %zu", this, mDataSize);
2590 ALOGV("restartWrite Setting data pos of %p to %zu", this, mDataPos);
2594 mObjectsSize = mObjectsCapacity = 0;
2595 mNextObjectHint = 0;
2596 mObjectsSorted = false;
2604 status_t Parcel::continueWrite(size_t desired)
2606 if (desired > INT32_MAX) {
2607 // don't accept size_t values which may have come from an
2608 // inadvertent conversion from a negative int.
2612 // If shrinking, first adjust for any objects that appear
2613 // after the new data size.
2614 size_t objectsSize = mObjectsSize;
2615 if (desired < mDataSize) {
2619 while (objectsSize > 0) {
2620 if (mObjects[objectsSize-1] < desired)
2628 // If the size is going to zero, just release the owner's data.
2634 // If there is a different owner, we need to take
2636 uint8_t* data = (uint8_t*)malloc(desired);
2641 binder_size_t* objects = NULL;
2644 objects = (binder_size_t*)calloc(objectsSize, sizeof(binder_size_t));
2652 // Little hack to only acquire references on objects
2653 // we will be keeping.
2654 size_t oldObjectsSize = mObjectsSize;
2655 mObjectsSize = objectsSize;
2657 mObjectsSize = oldObjectsSize;
2661 memcpy(data, mData, mDataSize < desired ? mDataSize : desired);
2663 if (objects && mObjects) {
2664 memcpy(objects, mObjects, objectsSize*sizeof(binder_size_t));
2666 //ALOGI("Freeing data ref of %p (pid=%d)", this, getpid());
2667 mOwner(this, mData, mDataSize, mObjects, mObjectsSize, mOwnerCookie);
2670 LOG_ALLOC("Parcel %p: taking ownership of %zu capacity", this, desired);
2671 pthread_mutex_lock(&gParcelGlobalAllocSizeLock);
2672 gParcelGlobalAllocSize += desired;
2673 gParcelGlobalAllocCount++;
2674 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock);
2678 mDataSize = (mDataSize < desired) ? mDataSize : desired;
2679 ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize);
2680 mDataCapacity = desired;
2681 mObjectsSize = mObjectsCapacity = objectsSize;
2682 mNextObjectHint = 0;
2683 mObjectsSorted = false;
2686 if (objectsSize < mObjectsSize) {
2687 // Need to release refs on any objects we are dropping.
2688 const sp<ProcessState> proc(ProcessState::self());
2689 for (size_t i=objectsSize; i<mObjectsSize; i++) {
2690 const flat_binder_object* flat
2691 = reinterpret_cast<flat_binder_object*>(mData+mObjects[i]);
2692 if (flat->type == BINDER_TYPE_FD) {
2693 // will need to rescan because we may have lopped off the only FDs
2696 release_object(proc, *flat, this, &mOpenAshmemSize);
2698 binder_size_t* objects =
2699 (binder_size_t*)realloc(mObjects, objectsSize*sizeof(binder_size_t));
2703 mObjectsSize = objectsSize;
2704 mNextObjectHint = 0;
2705 mObjectsSorted = false;
2708 // We own the data, so we can just do a realloc().
2709 if (desired > mDataCapacity) {
2710 uint8_t* data = (uint8_t*)realloc(mData, desired);
2712 LOG_ALLOC("Parcel %p: continue from %zu to %zu capacity", this, mDataCapacity,
2714 pthread_mutex_lock(&gParcelGlobalAllocSizeLock);
2715 gParcelGlobalAllocSize += desired;
2716 gParcelGlobalAllocSize -= mDataCapacity;
2717 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock);
2719 mDataCapacity = desired;
2720 } else if (desired > mDataCapacity) {
2725 if (mDataSize > desired) {
2726 mDataSize = desired;
2727 ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize);
2729 if (mDataPos > desired) {
2731 ALOGV("continueWrite Setting data pos of %p to %zu", this, mDataPos);
2736 // This is the first data. Easy!
2737 uint8_t* data = (uint8_t*)malloc(desired);
2743 if(!(mDataCapacity == 0 && mObjects == NULL
2744 && mObjectsCapacity == 0)) {
2745 ALOGE("continueWrite: %zu/%p/%zu/%zu", mDataCapacity, mObjects, mObjectsCapacity, desired);
2748 LOG_ALLOC("Parcel %p: allocating with %zu capacity", this, desired);
2749 pthread_mutex_lock(&gParcelGlobalAllocSizeLock);
2750 gParcelGlobalAllocSize += desired;
2751 gParcelGlobalAllocCount++;
2752 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock);
2755 mDataSize = mDataPos = 0;
2756 ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize);
2757 ALOGV("continueWrite Setting data pos of %p to %zu", this, mDataPos);
2758 mDataCapacity = desired;
2764 void Parcel::initState()
2766 LOG_ALLOC("Parcel %p: initState", this);
2772 ALOGV("initState Setting data size of %p to %zu", this, mDataSize);
2773 ALOGV("initState Setting data pos of %p to %zu", this, mDataPos);
2776 mObjectsCapacity = 0;
2777 mNextObjectHint = 0;
2778 mObjectsSorted = false;
2783 mOpenAshmemSize = 0;
2785 // racing multiple init leads only to multiple identical write
2787 struct rlimit result;
2788 if (!getrlimit(RLIMIT_NOFILE, &result)) {
2789 gMaxFds = (size_t)result.rlim_cur;
2790 //ALOGI("parcel fd limit set to %zu", gMaxFds);
2792 ALOGW("Unable to getrlimit: %s", strerror(errno));
2798 void Parcel::scanForFds() const
2800 bool hasFds = false;
2801 for (size_t i=0; i<mObjectsSize; i++) {
2802 const flat_binder_object* flat
2803 = reinterpret_cast<const flat_binder_object*>(mData + mObjects[i]);
2804 if (flat->type == BINDER_TYPE_FD) {
2813 size_t Parcel::getBlobAshmemSize() const
2815 // This used to return the size of all blobs that were written to ashmem, now we're returning
2816 // the ashmem currently referenced by this Parcel, which should be equivalent.
2817 // TODO: Remove method once ABI can be changed.
2818 return mOpenAshmemSize;
2821 size_t Parcel::getOpenAshmemSize() const
2823 return mOpenAshmemSize;
2826 // --- Parcel::Blob ---
2828 Parcel::Blob::Blob() :
2829 mFd(-1), mData(NULL), mSize(0), mMutable(false) {
2832 Parcel::Blob::~Blob() {
2836 void Parcel::Blob::release() {
2837 if (mFd != -1 && mData) {
2838 ::munmap(mData, mSize);
2843 void Parcel::Blob::init(int fd, void* data, size_t size, bool isMutable) {
2847 mMutable = isMutable;
2850 void Parcel::Blob::clear() {
2857 }; // namespace android