571 lines
19 KiB
C++
571 lines
19 KiB
C++
//===-- MachVMMemory.cpp ----------------------------------------*- C++ -*-===//
|
|
//
|
|
// The LLVM Compiler Infrastructure
|
|
//
|
|
// This file is distributed under the University of Illinois Open Source
|
|
// License. See LICENSE.TXT for details.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// Created by Greg Clayton on 6/26/07.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "MachVMMemory.h"
|
|
#include "DNBLog.h"
|
|
#include "MachVMRegion.h"
|
|
#include <dlfcn.h>
|
|
#include <mach/mach_vm.h>
|
|
#include <mach/shared_region.h>
|
|
#include <sys/sysctl.h>
|
|
|
|
static const vm_size_t kInvalidPageSize = ~0;
|
|
|
|
MachVMMemory::MachVMMemory() : m_page_size(kInvalidPageSize), m_err(0) {}
|
|
|
|
MachVMMemory::~MachVMMemory() {}
|
|
|
|
nub_size_t MachVMMemory::PageSize(task_t task) {
|
|
if (m_page_size == kInvalidPageSize) {
|
|
#if defined(TASK_VM_INFO) && TASK_VM_INFO >= 22
|
|
if (task != TASK_NULL) {
|
|
kern_return_t kr;
|
|
mach_msg_type_number_t info_count = TASK_VM_INFO_COUNT;
|
|
task_vm_info_data_t vm_info;
|
|
kr = task_info(task, TASK_VM_INFO, (task_info_t)&vm_info, &info_count);
|
|
if (kr == KERN_SUCCESS) {
|
|
DNBLogThreadedIf(
|
|
LOG_TASK,
|
|
"MachVMMemory::PageSize task_info returned page size of 0x%x",
|
|
(int)vm_info.page_size);
|
|
m_page_size = vm_info.page_size;
|
|
return m_page_size;
|
|
} else {
|
|
DNBLogThreadedIf(LOG_TASK, "MachVMMemory::PageSize task_info call "
|
|
"failed to get page size, TASK_VM_INFO %d, "
|
|
"TASK_VM_INFO_COUNT %d, kern return %d",
|
|
TASK_VM_INFO, TASK_VM_INFO_COUNT, kr);
|
|
}
|
|
}
|
|
#endif
|
|
m_err = ::host_page_size(::mach_host_self(), &m_page_size);
|
|
if (m_err.Fail())
|
|
m_page_size = 0;
|
|
}
|
|
return m_page_size;
|
|
}
|
|
|
|
nub_size_t MachVMMemory::MaxBytesLeftInPage(task_t task, nub_addr_t addr,
|
|
nub_size_t count) {
|
|
const nub_size_t page_size = PageSize(task);
|
|
if (page_size > 0) {
|
|
nub_size_t page_offset = (addr % page_size);
|
|
nub_size_t bytes_left_in_page = page_size - page_offset;
|
|
if (count > bytes_left_in_page)
|
|
count = bytes_left_in_page;
|
|
}
|
|
return count;
|
|
}
|
|
|
|
nub_bool_t MachVMMemory::GetMemoryRegionInfo(task_t task, nub_addr_t address,
|
|
DNBRegionInfo *region_info) {
|
|
MachVMRegion vmRegion(task);
|
|
|
|
if (vmRegion.GetRegionForAddress(address)) {
|
|
region_info->addr = vmRegion.StartAddress();
|
|
region_info->size = vmRegion.GetByteSize();
|
|
region_info->permissions = vmRegion.GetDNBPermissions();
|
|
} else {
|
|
region_info->addr = address;
|
|
region_info->size = 0;
|
|
if (vmRegion.GetError().Success()) {
|
|
// vmRegion.GetRegionForAddress() return false, indicating that "address"
|
|
// wasn't in a valid region, but the "vmRegion" info was successfully
|
|
// read from the task which means the info describes the next valid
|
|
// region from which we can infer the size of this invalid region
|
|
mach_vm_address_t start_addr = vmRegion.StartAddress();
|
|
if (address < start_addr)
|
|
region_info->size = start_addr - address;
|
|
}
|
|
// If we can't get any info about the size from the next region it means
|
|
// we asked about an address that was past all mappings, so the size
|
|
// of this region will take up all remaining address space.
|
|
if (region_info->size == 0)
|
|
region_info->size = INVALID_NUB_ADDRESS - region_info->addr;
|
|
|
|
// Not readable, writeable or executable
|
|
region_info->permissions = 0;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
// For integrated graphics chip, this makes the accounting info for 'wired'
|
|
// memory more like top.
|
|
uint64_t MachVMMemory::GetStolenPages(task_t task) {
|
|
static uint64_t stolenPages = 0;
|
|
static bool calculated = false;
|
|
if (calculated)
|
|
return stolenPages;
|
|
|
|
static int mib_reserved[CTL_MAXNAME];
|
|
static int mib_unusable[CTL_MAXNAME];
|
|
static int mib_other[CTL_MAXNAME];
|
|
static size_t mib_reserved_len = 0;
|
|
static size_t mib_unusable_len = 0;
|
|
static size_t mib_other_len = 0;
|
|
int r;
|
|
|
|
/* This can be used for testing: */
|
|
// tsamp->pages_stolen = (256 * 1024 * 1024ULL) / tsamp->pagesize;
|
|
|
|
if (0 == mib_reserved_len) {
|
|
mib_reserved_len = CTL_MAXNAME;
|
|
|
|
r = sysctlnametomib("machdep.memmap.Reserved", mib_reserved,
|
|
&mib_reserved_len);
|
|
|
|
if (-1 == r) {
|
|
mib_reserved_len = 0;
|
|
return 0;
|
|
}
|
|
|
|
mib_unusable_len = CTL_MAXNAME;
|
|
|
|
r = sysctlnametomib("machdep.memmap.Unusable", mib_unusable,
|
|
&mib_unusable_len);
|
|
|
|
if (-1 == r) {
|
|
mib_reserved_len = 0;
|
|
return 0;
|
|
}
|
|
|
|
mib_other_len = CTL_MAXNAME;
|
|
|
|
r = sysctlnametomib("machdep.memmap.Other", mib_other, &mib_other_len);
|
|
|
|
if (-1 == r) {
|
|
mib_reserved_len = 0;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
if (mib_reserved_len > 0 && mib_unusable_len > 0 && mib_other_len > 0) {
|
|
uint64_t reserved = 0, unusable = 0, other = 0;
|
|
size_t reserved_len;
|
|
size_t unusable_len;
|
|
size_t other_len;
|
|
|
|
reserved_len = sizeof(reserved);
|
|
unusable_len = sizeof(unusable);
|
|
other_len = sizeof(other);
|
|
|
|
/* These are all declared as QUAD/uint64_t sysctls in the kernel. */
|
|
|
|
if (sysctl(mib_reserved, static_cast<u_int>(mib_reserved_len), &reserved,
|
|
&reserved_len, NULL, 0)) {
|
|
return 0;
|
|
}
|
|
|
|
if (sysctl(mib_unusable, static_cast<u_int>(mib_unusable_len), &unusable,
|
|
&unusable_len, NULL, 0)) {
|
|
return 0;
|
|
}
|
|
|
|
if (sysctl(mib_other, static_cast<u_int>(mib_other_len), &other, &other_len,
|
|
NULL, 0)) {
|
|
return 0;
|
|
}
|
|
|
|
if (reserved_len == sizeof(reserved) && unusable_len == sizeof(unusable) &&
|
|
other_len == sizeof(other)) {
|
|
uint64_t stolen = reserved + unusable + other;
|
|
uint64_t mb128 = 128 * 1024 * 1024ULL;
|
|
|
|
if (stolen >= mb128) {
|
|
stolen = (stolen & ~((128 * 1024 * 1024ULL) - 1)); // rounding down
|
|
stolenPages = stolen / PageSize(task);
|
|
}
|
|
}
|
|
}
|
|
|
|
calculated = true;
|
|
return stolenPages;
|
|
}
|
|
|
|
static uint64_t GetPhysicalMemory() {
|
|
// This doesn't change often at all. No need to poll each time.
|
|
static uint64_t physical_memory = 0;
|
|
static bool calculated = false;
|
|
if (calculated)
|
|
return physical_memory;
|
|
|
|
size_t len = sizeof(physical_memory);
|
|
sysctlbyname("hw.memsize", &physical_memory, &len, NULL, 0);
|
|
|
|
calculated = true;
|
|
return physical_memory;
|
|
}
|
|
|
|
// rsize and dirty_size is not adjusted for dyld shared cache and multiple
|
|
// __LINKEDIT segment, as in vmmap. In practice, dirty_size doesn't differ much
|
|
// but rsize may. There is performance penalty for the adjustment. Right now,
|
|
// only use the dirty_size.
|
|
void MachVMMemory::GetRegionSizes(task_t task, mach_vm_size_t &rsize,
|
|
mach_vm_size_t &dirty_size) {
|
|
#if defined(TASK_VM_INFO) && TASK_VM_INFO >= 22
|
|
|
|
task_vm_info_data_t vm_info;
|
|
mach_msg_type_number_t info_count;
|
|
kern_return_t kr;
|
|
|
|
info_count = TASK_VM_INFO_COUNT;
|
|
kr = task_info(task, TASK_VM_INFO_PURGEABLE, (task_info_t)&vm_info,
|
|
&info_count);
|
|
if (kr == KERN_SUCCESS)
|
|
dirty_size = vm_info.internal;
|
|
#endif
|
|
}
|
|
|
|
// Test whether the virtual address is within the architecture's shared region.
|
|
static bool InSharedRegion(mach_vm_address_t addr, cpu_type_t type) {
|
|
mach_vm_address_t base = 0, size = 0;
|
|
|
|
switch (type) {
|
|
#if defined(CPU_TYPE_ARM64) && defined(SHARED_REGION_BASE_ARM64)
|
|
case CPU_TYPE_ARM64:
|
|
base = SHARED_REGION_BASE_ARM64;
|
|
size = SHARED_REGION_SIZE_ARM64;
|
|
break;
|
|
#endif
|
|
|
|
case CPU_TYPE_ARM:
|
|
base = SHARED_REGION_BASE_ARM;
|
|
size = SHARED_REGION_SIZE_ARM;
|
|
break;
|
|
|
|
case CPU_TYPE_X86_64:
|
|
base = SHARED_REGION_BASE_X86_64;
|
|
size = SHARED_REGION_SIZE_X86_64;
|
|
break;
|
|
|
|
case CPU_TYPE_I386:
|
|
base = SHARED_REGION_BASE_I386;
|
|
size = SHARED_REGION_SIZE_I386;
|
|
break;
|
|
|
|
default: {
|
|
// Log error abut unknown CPU type
|
|
break;
|
|
}
|
|
}
|
|
|
|
return (addr >= base && addr < (base + size));
|
|
}
|
|
|
|
void MachVMMemory::GetMemorySizes(task_t task, cpu_type_t cputype,
|
|
nub_process_t pid, mach_vm_size_t &rprvt,
|
|
mach_vm_size_t &vprvt) {
|
|
// Collecting some other info cheaply but not reporting for now.
|
|
mach_vm_size_t empty = 0;
|
|
mach_vm_size_t fw_private = 0;
|
|
|
|
mach_vm_size_t aliased = 0;
|
|
bool global_shared_text_data_mapped = false;
|
|
vm_size_t pagesize = PageSize(task);
|
|
|
|
for (mach_vm_address_t addr = 0, size = 0;; addr += size) {
|
|
vm_region_top_info_data_t info;
|
|
mach_msg_type_number_t count = VM_REGION_TOP_INFO_COUNT;
|
|
mach_port_t object_name;
|
|
|
|
kern_return_t kr =
|
|
mach_vm_region(task, &addr, &size, VM_REGION_TOP_INFO,
|
|
(vm_region_info_t)&info, &count, &object_name);
|
|
if (kr != KERN_SUCCESS)
|
|
break;
|
|
|
|
if (InSharedRegion(addr, cputype)) {
|
|
// Private Shared
|
|
fw_private += info.private_pages_resident * pagesize;
|
|
|
|
// Check if this process has the globally shared text and data regions
|
|
// mapped in. If so, set global_shared_text_data_mapped to TRUE and avoid
|
|
// checking again.
|
|
if (global_shared_text_data_mapped == FALSE &&
|
|
info.share_mode == SM_EMPTY) {
|
|
vm_region_basic_info_data_64_t b_info;
|
|
mach_vm_address_t b_addr = addr;
|
|
mach_vm_size_t b_size = size;
|
|
count = VM_REGION_BASIC_INFO_COUNT_64;
|
|
|
|
kr = mach_vm_region(task, &b_addr, &b_size, VM_REGION_BASIC_INFO,
|
|
(vm_region_info_t)&b_info, &count, &object_name);
|
|
if (kr != KERN_SUCCESS)
|
|
break;
|
|
|
|
if (b_info.reserved) {
|
|
global_shared_text_data_mapped = TRUE;
|
|
}
|
|
}
|
|
|
|
// Short circuit the loop if this isn't a shared private region, since
|
|
// that's the only region type we care about within the current address
|
|
// range.
|
|
if (info.share_mode != SM_PRIVATE) {
|
|
continue;
|
|
}
|
|
}
|
|
|
|
// Update counters according to the region type.
|
|
if (info.share_mode == SM_COW && info.ref_count == 1) {
|
|
// Treat single reference SM_COW as SM_PRIVATE
|
|
info.share_mode = SM_PRIVATE;
|
|
}
|
|
|
|
switch (info.share_mode) {
|
|
case SM_LARGE_PAGE:
|
|
// Treat SM_LARGE_PAGE the same as SM_PRIVATE
|
|
// since they are not shareable and are wired.
|
|
case SM_PRIVATE:
|
|
rprvt += info.private_pages_resident * pagesize;
|
|
rprvt += info.shared_pages_resident * pagesize;
|
|
vprvt += size;
|
|
break;
|
|
|
|
case SM_EMPTY:
|
|
empty += size;
|
|
break;
|
|
|
|
case SM_COW:
|
|
case SM_SHARED: {
|
|
if (pid == 0) {
|
|
// Treat kernel_task specially
|
|
if (info.share_mode == SM_COW) {
|
|
rprvt += info.private_pages_resident * pagesize;
|
|
vprvt += size;
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (info.share_mode == SM_COW) {
|
|
rprvt += info.private_pages_resident * pagesize;
|
|
vprvt += info.private_pages_resident * pagesize;
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
// log that something is really bad.
|
|
break;
|
|
}
|
|
}
|
|
|
|
rprvt += aliased;
|
|
}
|
|
|
|
static void GetPurgeableAndAnonymous(task_t task, uint64_t &purgeable,
|
|
uint64_t &anonymous) {
|
|
#if defined(TASK_VM_INFO) && TASK_VM_INFO >= 22
|
|
|
|
kern_return_t kr;
|
|
mach_msg_type_number_t info_count;
|
|
task_vm_info_data_t vm_info;
|
|
|
|
info_count = TASK_VM_INFO_COUNT;
|
|
kr = task_info(task, TASK_VM_INFO_PURGEABLE, (task_info_t)&vm_info,
|
|
&info_count);
|
|
if (kr == KERN_SUCCESS) {
|
|
purgeable = vm_info.purgeable_volatile_resident;
|
|
anonymous =
|
|
vm_info.internal + vm_info.compressed - vm_info.purgeable_volatile_pmap;
|
|
}
|
|
|
|
#endif
|
|
}
|
|
|
|
#if defined(HOST_VM_INFO64_COUNT)
|
|
nub_bool_t MachVMMemory::GetMemoryProfile(
|
|
DNBProfileDataScanType scanType, task_t task, struct task_basic_info ti,
|
|
cpu_type_t cputype, nub_process_t pid, vm_statistics64_data_t &vminfo,
|
|
uint64_t &physical_memory, mach_vm_size_t &rprvt, mach_vm_size_t &rsize,
|
|
mach_vm_size_t &vprvt, mach_vm_size_t &vsize, mach_vm_size_t &dirty_size,
|
|
mach_vm_size_t &purgeable, mach_vm_size_t &anonymous)
|
|
#else
|
|
nub_bool_t MachVMMemory::GetMemoryProfile(
|
|
DNBProfileDataScanType scanType, task_t task, struct task_basic_info ti,
|
|
cpu_type_t cputype, nub_process_t pid, vm_statistics_data_t &vminfo,
|
|
uint64_t &physical_memory, mach_vm_size_t &rprvt, mach_vm_size_t &rsize,
|
|
mach_vm_size_t &vprvt, mach_vm_size_t &vsize, mach_vm_size_t &dirty_size,
|
|
mach_vm_size_t &purgeable, mach_vm_size_t &anonymous)
|
|
#endif
|
|
{
|
|
if (scanType & eProfileHostMemory)
|
|
physical_memory = GetPhysicalMemory();
|
|
|
|
if (scanType & eProfileMemory) {
|
|
static mach_port_t localHost = mach_host_self();
|
|
#if defined(HOST_VM_INFO64_COUNT)
|
|
mach_msg_type_number_t count = HOST_VM_INFO64_COUNT;
|
|
host_statistics64(localHost, HOST_VM_INFO64, (host_info64_t)&vminfo,
|
|
&count);
|
|
#else
|
|
mach_msg_type_number_t count = HOST_VM_INFO_COUNT;
|
|
host_statistics(localHost, HOST_VM_INFO, (host_info_t)&vminfo, &count);
|
|
vminfo.wire_count += GetStolenPages(task);
|
|
#endif
|
|
|
|
/* We are no longer reporting these. Let's not waste time.
|
|
GetMemorySizes(task, cputype, pid, rprvt, vprvt);
|
|
rsize = ti.resident_size;
|
|
vsize = ti.virtual_size;
|
|
|
|
if (scanType & eProfileMemoryDirtyPage)
|
|
{
|
|
// This uses vmmap strategy. We don't use the returned rsize for now. We
|
|
prefer to match top's version since that's what we do for the rest of the
|
|
metrics.
|
|
GetRegionSizes(task, rsize, dirty_size);
|
|
}
|
|
*/
|
|
|
|
if (scanType & eProfileMemoryAnonymous) {
|
|
GetPurgeableAndAnonymous(task, purgeable, anonymous);
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
nub_size_t MachVMMemory::Read(task_t task, nub_addr_t address, void *data,
|
|
nub_size_t data_count) {
|
|
if (data == NULL || data_count == 0)
|
|
return 0;
|
|
|
|
nub_size_t total_bytes_read = 0;
|
|
nub_addr_t curr_addr = address;
|
|
uint8_t *curr_data = (uint8_t *)data;
|
|
while (total_bytes_read < data_count) {
|
|
mach_vm_size_t curr_size =
|
|
MaxBytesLeftInPage(task, curr_addr, data_count - total_bytes_read);
|
|
mach_msg_type_number_t curr_bytes_read = 0;
|
|
vm_offset_t vm_memory = 0;
|
|
m_err = ::mach_vm_read(task, curr_addr, curr_size, &vm_memory,
|
|
&curr_bytes_read);
|
|
|
|
if (DNBLogCheckLogBit(LOG_MEMORY))
|
|
m_err.LogThreaded("::mach_vm_read ( task = 0x%4.4x, addr = 0x%8.8llx, "
|
|
"size = %llu, data => %8.8p, dataCnt => %i )",
|
|
task, (uint64_t)curr_addr, (uint64_t)curr_size,
|
|
vm_memory, curr_bytes_read);
|
|
|
|
if (m_err.Success()) {
|
|
if (curr_bytes_read != curr_size) {
|
|
if (DNBLogCheckLogBit(LOG_MEMORY))
|
|
m_err.LogThreaded(
|
|
"::mach_vm_read ( task = 0x%4.4x, addr = 0x%8.8llx, size = %llu, "
|
|
"data => %8.8p, dataCnt=>%i ) only read %u of %llu bytes",
|
|
task, (uint64_t)curr_addr, (uint64_t)curr_size, vm_memory,
|
|
curr_bytes_read, curr_bytes_read, (uint64_t)curr_size);
|
|
}
|
|
::memcpy(curr_data, (void *)vm_memory, curr_bytes_read);
|
|
::vm_deallocate(mach_task_self(), vm_memory, curr_bytes_read);
|
|
total_bytes_read += curr_bytes_read;
|
|
curr_addr += curr_bytes_read;
|
|
curr_data += curr_bytes_read;
|
|
} else {
|
|
break;
|
|
}
|
|
}
|
|
return total_bytes_read;
|
|
}
|
|
|
|
nub_size_t MachVMMemory::Write(task_t task, nub_addr_t address,
|
|
const void *data, nub_size_t data_count) {
|
|
MachVMRegion vmRegion(task);
|
|
|
|
nub_size_t total_bytes_written = 0;
|
|
nub_addr_t curr_addr = address;
|
|
const uint8_t *curr_data = (const uint8_t *)data;
|
|
|
|
while (total_bytes_written < data_count) {
|
|
if (vmRegion.GetRegionForAddress(curr_addr)) {
|
|
mach_vm_size_t curr_data_count = data_count - total_bytes_written;
|
|
mach_vm_size_t region_bytes_left = vmRegion.BytesRemaining(curr_addr);
|
|
if (region_bytes_left == 0) {
|
|
break;
|
|
}
|
|
if (curr_data_count > region_bytes_left)
|
|
curr_data_count = region_bytes_left;
|
|
|
|
if (vmRegion.SetProtections(curr_addr, curr_data_count,
|
|
VM_PROT_READ | VM_PROT_WRITE)) {
|
|
nub_size_t bytes_written =
|
|
WriteRegion(task, curr_addr, curr_data, curr_data_count);
|
|
if (bytes_written <= 0) {
|
|
// Status should have already be posted by WriteRegion...
|
|
break;
|
|
} else {
|
|
total_bytes_written += bytes_written;
|
|
curr_addr += bytes_written;
|
|
curr_data += bytes_written;
|
|
}
|
|
} else {
|
|
DNBLogThreadedIf(
|
|
LOG_MEMORY_PROTECTIONS, "Failed to set read/write protections on "
|
|
"region for address: [0x%8.8llx-0x%8.8llx)",
|
|
(uint64_t)curr_addr, (uint64_t)(curr_addr + curr_data_count));
|
|
break;
|
|
}
|
|
} else {
|
|
DNBLogThreadedIf(LOG_MEMORY_PROTECTIONS,
|
|
"Failed to get region for address: 0x%8.8llx",
|
|
(uint64_t)address);
|
|
break;
|
|
}
|
|
}
|
|
|
|
return total_bytes_written;
|
|
}
|
|
|
|
nub_size_t MachVMMemory::WriteRegion(task_t task, const nub_addr_t address,
|
|
const void *data,
|
|
const nub_size_t data_count) {
|
|
if (data == NULL || data_count == 0)
|
|
return 0;
|
|
|
|
nub_size_t total_bytes_written = 0;
|
|
nub_addr_t curr_addr = address;
|
|
const uint8_t *curr_data = (const uint8_t *)data;
|
|
while (total_bytes_written < data_count) {
|
|
mach_msg_type_number_t curr_data_count =
|
|
static_cast<mach_msg_type_number_t>(MaxBytesLeftInPage(
|
|
task, curr_addr, data_count - total_bytes_written));
|
|
m_err =
|
|
::mach_vm_write(task, curr_addr, (pointer_t)curr_data, curr_data_count);
|
|
if (DNBLogCheckLogBit(LOG_MEMORY) || m_err.Fail())
|
|
m_err.LogThreaded("::mach_vm_write ( task = 0x%4.4x, addr = 0x%8.8llx, "
|
|
"data = %8.8p, dataCnt = %u )",
|
|
task, (uint64_t)curr_addr, curr_data, curr_data_count);
|
|
|
|
#if !defined(__i386__) && !defined(__x86_64__)
|
|
vm_machine_attribute_val_t mattr_value = MATTR_VAL_CACHE_FLUSH;
|
|
|
|
m_err = ::vm_machine_attribute(task, curr_addr, curr_data_count,
|
|
MATTR_CACHE, &mattr_value);
|
|
if (DNBLogCheckLogBit(LOG_MEMORY) || m_err.Fail())
|
|
m_err.LogThreaded("::vm_machine_attribute ( task = 0x%4.4x, addr = "
|
|
"0x%8.8llx, size = %u, attr = MATTR_CACHE, mattr_value "
|
|
"=> MATTR_VAL_CACHE_FLUSH )",
|
|
task, (uint64_t)curr_addr, curr_data_count);
|
|
#endif
|
|
|
|
if (m_err.Success()) {
|
|
total_bytes_written += curr_data_count;
|
|
curr_addr += curr_data_count;
|
|
curr_data += curr_data_count;
|
|
} else {
|
|
break;
|
|
}
|
|
}
|
|
return total_bytes_written;
|
|
}
|