summaryrefslogtreecommitdiff
path: root/core/pool_allocator.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'core/pool_allocator.cpp')
-rw-r--r--core/pool_allocator.cpp370
1 files changed, 170 insertions, 200 deletions
diff --git a/core/pool_allocator.cpp b/core/pool_allocator.cpp
index 3260225ac3..68de05a765 100644
--- a/core/pool_allocator.cpp
+++ b/core/pool_allocator.cpp
@@ -28,43 +28,39 @@
/*************************************************************************/
#include "pool_allocator.h"
-#include "error_macros.h"
#include "core/os/os.h"
-#include "os/memory.h"
+#include "error_macros.h"
#include "os/copymem.h"
+#include "os/memory.h"
#include "print_string.h"
#include <assert.h>
-#define COMPACT_CHUNK( m_entry , m_to_pos ) \
-do { \
- void *_dst=&((unsigned char*)pool)[m_to_pos]; \
- void *_src=&((unsigned char*)pool)[(m_entry).pos]; \
- movemem(_dst,_src,aligned((m_entry).len)); \
- (m_entry).pos=m_to_pos; \
-} while (0);
+#define COMPACT_CHUNK(m_entry, m_to_pos) \
+ do { \
+ void *_dst = &((unsigned char *)pool)[m_to_pos]; \
+ void *_src = &((unsigned char *)pool)[(m_entry).pos]; \
+ movemem(_dst, _src, aligned((m_entry).len)); \
+ (m_entry).pos = m_to_pos; \
+ } while (0);
void PoolAllocator::mt_lock() const {
-
}
void PoolAllocator::mt_unlock() const {
-
}
+bool PoolAllocator::get_free_entry(EntryArrayPos *p_pos) {
-bool PoolAllocator::get_free_entry(EntryArrayPos* p_pos) {
-
- if (entry_count==entry_max)
+ if (entry_count == entry_max)
return false;
- for (int i=0;i<entry_max;i++) {
+ for (int i = 0; i < entry_max; i++) {
- if (entry_array[i].len==0) {
- *p_pos=i;
+ if (entry_array[i].len == 0) {
+ *p_pos = i;
return true;
}
-
}
ERR_PRINT("Out of memory Chunks!");
@@ -82,134 +78,121 @@ bool PoolAllocator::find_hole(EntryArrayPos *p_pos, int p_for_size) {
/* position where previous entry ends. Defaults to zero (begin of pool) */
- int prev_entry_end_pos=0;
+ int prev_entry_end_pos = 0;
- for (int i=0;i<entry_count;i++) {
+ for (int i = 0; i < entry_count; i++) {
-
- Entry &entry=entry_array[ entry_indices[ i ] ];
+ Entry &entry = entry_array[entry_indices[i]];
/* determine hole size to previous entry */
- int hole_size=entry.pos-prev_entry_end_pos;
+ int hole_size = entry.pos - prev_entry_end_pos;
/* detemine if what we want fits in that hole */
- if (hole_size>=p_for_size) {
- *p_pos=i;
+ if (hole_size >= p_for_size) {
+ *p_pos = i;
return true;
}
/* prepare for next one */
- prev_entry_end_pos=entry_end(entry);
+ prev_entry_end_pos = entry_end(entry);
}
/* No holes between entrys, check at the end..*/
- if ( (pool_size-prev_entry_end_pos)>=p_for_size ) {
- *p_pos=entry_count;
+ if ((pool_size - prev_entry_end_pos) >= p_for_size) {
+ *p_pos = entry_count;
return true;
}
return false;
-
}
-
void PoolAllocator::compact(int p_up_to) {
- uint32_t prev_entry_end_pos=0;
-
- if (p_up_to<0)
- p_up_to=entry_count;
- for (int i=0;i<p_up_to;i++) {
+ uint32_t prev_entry_end_pos = 0;
+ if (p_up_to < 0)
+ p_up_to = entry_count;
+ for (int i = 0; i < p_up_to; i++) {
- Entry &entry=entry_array[ entry_indices[ i ] ];
+ Entry &entry = entry_array[entry_indices[i]];
/* determine hole size to previous entry */
- int hole_size=entry.pos-prev_entry_end_pos;
+ int hole_size = entry.pos - prev_entry_end_pos;
/* if we can compact, do it */
- if (hole_size>0 && !entry.lock) {
-
- COMPACT_CHUNK(entry,prev_entry_end_pos);
+ if (hole_size > 0 && !entry.lock) {
+ COMPACT_CHUNK(entry, prev_entry_end_pos);
}
/* prepare for next one */
- prev_entry_end_pos=entry_end(entry);
+ prev_entry_end_pos = entry_end(entry);
}
-
-
}
void PoolAllocator::compact_up(int p_from) {
- uint32_t next_entry_end_pos=pool_size; // - static_area_size;
+ uint32_t next_entry_end_pos = pool_size; // - static_area_size;
- for (int i=entry_count-1;i>=p_from;i--) {
+ for (int i = entry_count - 1; i >= p_from; i--) {
-
- Entry &entry=entry_array[ entry_indices[ i ] ];
+ Entry &entry = entry_array[entry_indices[i]];
/* determine hole size to nextious entry */
- int hole_size=next_entry_end_pos-(entry.pos+aligned(entry.len));
+ int hole_size = next_entry_end_pos - (entry.pos + aligned(entry.len));
/* if we can compact, do it */
- if (hole_size>0 && !entry.lock) {
-
- COMPACT_CHUNK(entry,(next_entry_end_pos-aligned(entry.len)));
+ if (hole_size > 0 && !entry.lock) {
+ COMPACT_CHUNK(entry, (next_entry_end_pos - aligned(entry.len)));
}
/* prepare for next one */
- next_entry_end_pos=entry.pos;
+ next_entry_end_pos = entry.pos;
}
-
}
+bool PoolAllocator::find_entry_index(EntryIndicesPos *p_map_pos, Entry *p_entry) {
-bool PoolAllocator::find_entry_index(EntryIndicesPos *p_map_pos,Entry *p_entry) {
+ EntryArrayPos entry_pos = entry_max;
- EntryArrayPos entry_pos=entry_max;
+ for (int i = 0; i < entry_count; i++) {
- for (int i=0;i<entry_count;i++) {
+ if (&entry_array[entry_indices[i]] == p_entry) {
- if (&entry_array[ entry_indices[ i ] ]==p_entry) {
-
- entry_pos=i;
+ entry_pos = i;
break;
}
}
- if (entry_pos==entry_max)
+ if (entry_pos == entry_max)
return false;
- *p_map_pos=entry_pos;
+ *p_map_pos = entry_pos;
return true;
-
}
PoolAllocator::ID PoolAllocator::alloc(int p_size) {
- ERR_FAIL_COND_V(p_size<1,POOL_ALLOCATOR_INVALID_ID);
+ ERR_FAIL_COND_V(p_size < 1, POOL_ALLOCATOR_INVALID_ID);
#ifdef DEBUG_ENABLED
if (p_size > free_mem) OS::get_singleton()->debug_break();
#endif
- ERR_FAIL_COND_V(p_size>free_mem,POOL_ALLOCATOR_INVALID_ID);
+ ERR_FAIL_COND_V(p_size > free_mem, POOL_ALLOCATOR_INVALID_ID);
mt_lock();
- if (entry_count==entry_max) {
+ if (entry_count == entry_max) {
mt_unlock();
ERR_PRINT("entry_count==entry_max");
return POOL_ALLOCATOR_INVALID_ID;
}
-
- int size_to_alloc=aligned(p_size);
+ int size_to_alloc = aligned(p_size);
EntryIndicesPos new_entry_indices_pos;
@@ -228,60 +211,59 @@ PoolAllocator::ID PoolAllocator::alloc(int p_size) {
EntryArrayPos new_entry_array_pos;
- bool found_free_entry=get_free_entry(&new_entry_array_pos);
+ bool found_free_entry = get_free_entry(&new_entry_array_pos);
if (!found_free_entry) {
mt_unlock();
- ERR_FAIL_COND_V( !found_free_entry , POOL_ALLOCATOR_INVALID_ID );
+ ERR_FAIL_COND_V(!found_free_entry, POOL_ALLOCATOR_INVALID_ID);
}
/* move all entry indices up, make room for this one */
- for (int i=entry_count;i>new_entry_indices_pos;i-- ) {
+ for (int i = entry_count; i > new_entry_indices_pos; i--) {
- entry_indices[i]=entry_indices[i-1];
+ entry_indices[i] = entry_indices[i - 1];
}
- entry_indices[new_entry_indices_pos]=new_entry_array_pos;
+ entry_indices[new_entry_indices_pos] = new_entry_array_pos;
entry_count++;
- Entry &entry=entry_array[ entry_indices[ new_entry_indices_pos ] ];
+ Entry &entry = entry_array[entry_indices[new_entry_indices_pos]];
- entry.len=p_size;
- entry.pos=(new_entry_indices_pos==0)?0:entry_end(entry_array[ entry_indices[ new_entry_indices_pos-1 ] ]); //alloc either at begining or end of previous
- entry.lock=0;
- entry.check=(check_count++)&CHECK_MASK;
- free_mem-=size_to_alloc;
- if (free_mem<free_mem_peak)
- free_mem_peak=free_mem;
+ entry.len = p_size;
+ entry.pos = (new_entry_indices_pos == 0) ? 0 : entry_end(entry_array[entry_indices[new_entry_indices_pos - 1]]); //alloc either at begining or end of previous
+ entry.lock = 0;
+ entry.check = (check_count++) & CHECK_MASK;
+ free_mem -= size_to_alloc;
+ if (free_mem < free_mem_peak)
+ free_mem_peak = free_mem;
- ID retval = (entry_indices[ new_entry_indices_pos ]<<CHECK_BITS)|entry.check;
+ ID retval = (entry_indices[new_entry_indices_pos] << CHECK_BITS) | entry.check;
mt_unlock();
//ERR_FAIL_COND_V( (uintptr_t)get(retval)%align != 0, retval );
return retval;
-
}
-PoolAllocator::Entry * PoolAllocator::get_entry(ID p_mem) {
+PoolAllocator::Entry *PoolAllocator::get_entry(ID p_mem) {
- unsigned int check=p_mem&CHECK_MASK;
- int entry=p_mem>>CHECK_BITS;
- ERR_FAIL_INDEX_V(entry,entry_max,NULL);
- ERR_FAIL_COND_V(entry_array[entry].check!=check,NULL);
- ERR_FAIL_COND_V(entry_array[entry].len==0,NULL);
+ unsigned int check = p_mem & CHECK_MASK;
+ int entry = p_mem >> CHECK_BITS;
+ ERR_FAIL_INDEX_V(entry, entry_max, NULL);
+ ERR_FAIL_COND_V(entry_array[entry].check != check, NULL);
+ ERR_FAIL_COND_V(entry_array[entry].len == 0, NULL);
return &entry_array[entry];
}
-const PoolAllocator::Entry * PoolAllocator::get_entry(ID p_mem) const {
+const PoolAllocator::Entry *PoolAllocator::get_entry(ID p_mem) const {
- unsigned int check=p_mem&CHECK_MASK;
- int entry=p_mem>>CHECK_BITS;
- ERR_FAIL_INDEX_V(entry,entry_max,NULL);
- ERR_FAIL_COND_V(entry_array[entry].check!=check,NULL);
- ERR_FAIL_COND_V(entry_array[entry].len==0,NULL);
+ unsigned int check = p_mem & CHECK_MASK;
+ int entry = p_mem >> CHECK_BITS;
+ ERR_FAIL_INDEX_V(entry, entry_max, NULL);
+ ERR_FAIL_COND_V(entry_array[entry].check != check, NULL);
+ ERR_FAIL_COND_V(entry_array[entry].len == 0, NULL);
return &entry_array[entry];
}
@@ -289,7 +271,7 @@ const PoolAllocator::Entry * PoolAllocator::get_entry(ID p_mem) const {
void PoolAllocator::free(ID p_mem) {
mt_lock();
- Entry *e=get_entry(p_mem);
+ Entry *e = get_entry(p_mem);
if (!e) {
mt_unlock();
ERR_PRINT("!e");
@@ -303,22 +285,20 @@ void PoolAllocator::free(ID p_mem) {
EntryIndicesPos entry_indices_pos;
- bool index_found = find_entry_index(&entry_indices_pos,e);
+ bool index_found = find_entry_index(&entry_indices_pos, e);
if (!index_found) {
mt_unlock();
ERR_FAIL_COND(!index_found);
}
+ for (int i = entry_indices_pos; i < (entry_count - 1); i++) {
-
- for (int i=entry_indices_pos;i<(entry_count-1);i++) {
-
- entry_indices[ i ] = entry_indices[ i+1 ];
+ entry_indices[i] = entry_indices[i + 1];
}
entry_count--;
- free_mem+=aligned(e->len);
+ free_mem += aligned(e->len);
e->clear();
mt_unlock();
}
@@ -328,7 +308,7 @@ int PoolAllocator::get_size(ID p_mem) const {
int size;
mt_lock();
- const Entry *e=get_entry(p_mem);
+ const Entry *e = get_entry(p_mem);
if (!e) {
mt_unlock();
@@ -336,40 +316,40 @@ int PoolAllocator::get_size(ID p_mem) const {
return 0;
}
- size=e->len;
+ size = e->len;
mt_unlock();
return size;
}
-Error PoolAllocator::resize(ID p_mem,int p_new_size) {
+Error PoolAllocator::resize(ID p_mem, int p_new_size) {
mt_lock();
- Entry *e=get_entry(p_mem);
+ Entry *e = get_entry(p_mem);
if (!e) {
mt_unlock();
- ERR_FAIL_COND_V(!e,ERR_INVALID_PARAMETER);
+ ERR_FAIL_COND_V(!e, ERR_INVALID_PARAMETER);
}
if (needs_locking && e->lock) {
mt_unlock();
- ERR_FAIL_COND_V(e->lock,ERR_ALREADY_IN_USE);
+ ERR_FAIL_COND_V(e->lock, ERR_ALREADY_IN_USE);
}
int alloc_size = aligned(p_new_size);
- if (aligned(e->len)==alloc_size) {
+ if (aligned(e->len) == alloc_size) {
- e->len=p_new_size;
+ e->len = p_new_size;
mt_unlock();
return OK;
- } else if (e->len>(uint32_t)p_new_size) {
+ } else if (e->len > (uint32_t)p_new_size) {
free_mem += aligned(e->len);
free_mem -= alloc_size;
- e->len=p_new_size;
+ e->len = p_new_size;
mt_unlock();
return OK;
}
@@ -379,77 +359,74 @@ Error PoolAllocator::resize(ID p_mem,int p_new_size) {
if ((_free + aligned(e->len)) - alloc_size < 0) {
mt_unlock();
- ERR_FAIL_V( ERR_OUT_OF_MEMORY );
+ ERR_FAIL_V(ERR_OUT_OF_MEMORY);
};
EntryIndicesPos entry_indices_pos;
- bool index_found = find_entry_index(&entry_indices_pos,e);
+ bool index_found = find_entry_index(&entry_indices_pos, e);
if (!index_found) {
mt_unlock();
- ERR_FAIL_COND_V(!index_found,ERR_BUG);
+ ERR_FAIL_COND_V(!index_found, ERR_BUG);
}
//no need to move stuff around, it fits before the next block
int next_pos;
- if (entry_indices_pos+1 == entry_count) {
+ if (entry_indices_pos + 1 == entry_count) {
next_pos = pool_size; // - static_area_size;
} else {
- next_pos = entry_array[entry_indices[entry_indices_pos+1]].pos;
+ next_pos = entry_array[entry_indices[entry_indices_pos + 1]].pos;
};
if ((next_pos - e->pos) > alloc_size) {
- free_mem+=aligned(e->len);
- e->len=p_new_size;
- free_mem-=alloc_size;
+ free_mem += aligned(e->len);
+ e->len = p_new_size;
+ free_mem -= alloc_size;
mt_unlock();
return OK;
}
//it doesn't fit, compact around BEFORE current index (make room behind)
- compact(entry_indices_pos+1);
-
+ compact(entry_indices_pos + 1);
if ((next_pos - e->pos) > alloc_size) {
//now fits! hooray!
- free_mem+=aligned(e->len);
- e->len=p_new_size;
- free_mem-=alloc_size;
+ free_mem += aligned(e->len);
+ e->len = p_new_size;
+ free_mem -= alloc_size;
mt_unlock();
- if (free_mem<free_mem_peak)
- free_mem_peak=free_mem;
+ if (free_mem < free_mem_peak)
+ free_mem_peak = free_mem;
return OK;
}
//STILL doesn't fit, compact around AFTER current index (make room after)
- compact_up(entry_indices_pos+1);
+ compact_up(entry_indices_pos + 1);
- if ((entry_array[entry_indices[entry_indices_pos+1]].pos - e->pos) > alloc_size) {
+ if ((entry_array[entry_indices[entry_indices_pos + 1]].pos - e->pos) > alloc_size) {
//now fits! hooray!
- free_mem+=aligned(e->len);
- e->len=p_new_size;
- free_mem-=alloc_size;
+ free_mem += aligned(e->len);
+ e->len = p_new_size;
+ free_mem -= alloc_size;
mt_unlock();
- if (free_mem<free_mem_peak)
- free_mem_peak=free_mem;
+ if (free_mem < free_mem_peak)
+ free_mem_peak = free_mem;
return OK;
}
mt_unlock();
ERR_FAIL_V(ERR_OUT_OF_MEMORY);
-
}
-
Error PoolAllocator::lock(ID p_mem) {
if (!needs_locking)
return OK;
mt_lock();
- Entry *e=get_entry(p_mem);
+ Entry *e = get_entry(p_mem);
if (!e) {
mt_unlock();
@@ -467,7 +444,7 @@ bool PoolAllocator::is_locked(ID p_mem) const {
return false;
mt_lock();
- const Entry *e=((PoolAllocator*)(this))->get_entry(p_mem);
+ const Entry *e = ((PoolAllocator *)(this))->get_entry(p_mem);
if (!e) {
mt_unlock();
@@ -483,91 +460,87 @@ const void *PoolAllocator::get(ID p_mem) const {
if (!needs_locking) {
- const Entry *e=get_entry(p_mem);
- ERR_FAIL_COND_V(!e,NULL);
+ const Entry *e = get_entry(p_mem);
+ ERR_FAIL_COND_V(!e, NULL);
return &pool[e->pos];
-
}
mt_lock();
- const Entry *e=get_entry(p_mem);
+ const Entry *e = get_entry(p_mem);
if (!e) {
mt_unlock();
- ERR_FAIL_COND_V(!e,NULL);
+ ERR_FAIL_COND_V(!e, NULL);
}
- if (e->lock==0) {
+ if (e->lock == 0) {
mt_unlock();
- ERR_PRINT( "e->lock == 0" );
+ ERR_PRINT("e->lock == 0");
return NULL;
}
- if ((int)e->pos>=pool_size) {
+ if ((int)e->pos >= pool_size) {
mt_unlock();
ERR_PRINT("e->pos<0 || e->pos>=pool_size");
return NULL;
}
- const void *ptr=&pool[e->pos];
+ const void *ptr = &pool[e->pos];
mt_unlock();
return ptr;
-
}
void *PoolAllocator::get(ID p_mem) {
if (!needs_locking) {
- Entry *e=get_entry(p_mem);
+ Entry *e = get_entry(p_mem);
if (!e) {
- ERR_FAIL_COND_V(!e,NULL);
+ ERR_FAIL_COND_V(!e, NULL);
};
return &pool[e->pos];
-
}
mt_lock();
- Entry *e=get_entry(p_mem);
+ Entry *e = get_entry(p_mem);
if (!e) {
mt_unlock();
- ERR_FAIL_COND_V(!e,NULL);
+ ERR_FAIL_COND_V(!e, NULL);
}
- if (e->lock==0) {
+ if (e->lock == 0) {
//assert(0);
mt_unlock();
- ERR_PRINT( "e->lock == 0" );
+ ERR_PRINT("e->lock == 0");
return NULL;
}
- if ((int)e->pos>=pool_size) {
+ if ((int)e->pos >= pool_size) {
mt_unlock();
ERR_PRINT("e->pos<0 || e->pos>=pool_size");
return NULL;
}
- void *ptr=&pool[e->pos];
+ void *ptr = &pool[e->pos];
mt_unlock();
return ptr;
-
}
void PoolAllocator::unlock(ID p_mem) {
if (!needs_locking)
return;
mt_lock();
- Entry *e=get_entry(p_mem);
- if (e->lock == 0 ) {
+ Entry *e = get_entry(p_mem);
+ if (e->lock == 0) {
mt_unlock();
- ERR_PRINT( "e->lock == 0" );
+ ERR_PRINT("e->lock == 0");
return;
}
e->lock--;
@@ -576,7 +549,7 @@ void PoolAllocator::unlock(ID p_mem) {
int PoolAllocator::get_used_mem() const {
- return pool_size-free_mem;
+ return pool_size - free_mem;
}
int PoolAllocator::get_free_peak() {
@@ -589,72 +562,69 @@ int PoolAllocator::get_free_mem() {
return free_mem;
}
-void PoolAllocator::create_pool(void * p_mem,int p_size,int p_max_entries) {
+void PoolAllocator::create_pool(void *p_mem, int p_size, int p_max_entries) {
- pool=(uint8_t*)p_mem;
- pool_size=p_size;
+ pool = (uint8_t *)p_mem;
+ pool_size = p_size;
- entry_array = memnew_arr( Entry, p_max_entries );
- entry_indices = memnew_arr( int, p_max_entries );
+ entry_array = memnew_arr(Entry, p_max_entries);
+ entry_indices = memnew_arr(int, p_max_entries);
entry_max = p_max_entries;
- entry_count=0;
+ entry_count = 0;
- free_mem=p_size;
- free_mem_peak=p_size;
+ free_mem = p_size;
+ free_mem_peak = p_size;
- check_count=0;
+ check_count = 0;
}
-PoolAllocator::PoolAllocator(int p_size,bool p_needs_locking,int p_max_entries) {
+PoolAllocator::PoolAllocator(int p_size, bool p_needs_locking, int p_max_entries) {
- mem_ptr=memalloc( p_size);
+ mem_ptr = memalloc(p_size);
ERR_FAIL_COND(!mem_ptr);
- align=1;
- create_pool(mem_ptr,p_size,p_max_entries);
- needs_locking=p_needs_locking;
-
+ align = 1;
+ create_pool(mem_ptr, p_size, p_max_entries);
+ needs_locking = p_needs_locking;
}
-PoolAllocator::PoolAllocator(void * p_mem,int p_size, int p_align ,bool p_needs_locking,int p_max_entries) {
+PoolAllocator::PoolAllocator(void *p_mem, int p_size, int p_align, bool p_needs_locking, int p_max_entries) {
if (p_align > 1) {
- uint8_t *mem8=(uint8_t*)p_mem;
+ uint8_t *mem8 = (uint8_t *)p_mem;
uint64_t ofs = (uint64_t)mem8;
- if (ofs%p_align) {
- int dif = p_align-(ofs%p_align);
- mem8+=p_align-(ofs%p_align);
+ if (ofs % p_align) {
+ int dif = p_align - (ofs % p_align);
+ mem8 += p_align - (ofs % p_align);
p_size -= dif;
- p_mem = (void*)mem8;
+ p_mem = (void *)mem8;
};
};
- create_pool( p_mem,p_size,p_max_entries);
- needs_locking=p_needs_locking;
- align=p_align;
- mem_ptr=NULL;
+ create_pool(p_mem, p_size, p_max_entries);
+ needs_locking = p_needs_locking;
+ align = p_align;
+ mem_ptr = NULL;
}
-PoolAllocator::PoolAllocator(int p_align,int p_size,bool p_needs_locking,int p_max_entries) {
+PoolAllocator::PoolAllocator(int p_align, int p_size, bool p_needs_locking, int p_max_entries) {
- ERR_FAIL_COND(p_align<1);
- mem_ptr=Memory::alloc_static( p_size+p_align,"PoolAllocator()");
- uint8_t *mem8=(uint8_t*)mem_ptr;
+ ERR_FAIL_COND(p_align < 1);
+ mem_ptr = Memory::alloc_static(p_size + p_align, "PoolAllocator()");
+ uint8_t *mem8 = (uint8_t *)mem_ptr;
uint64_t ofs = (uint64_t)mem8;
- if (ofs%p_align)
- mem8+=p_align-(ofs%p_align);
- create_pool( mem8 ,p_size,p_max_entries);
- needs_locking=p_needs_locking;
- align=p_align;
+ if (ofs % p_align)
+ mem8 += p_align - (ofs % p_align);
+ create_pool(mem8, p_size, p_max_entries);
+ needs_locking = p_needs_locking;
+ align = p_align;
}
PoolAllocator::~PoolAllocator() {
if (mem_ptr)
- memfree( mem_ptr );
-
- memdelete_arr( entry_array );
- memdelete_arr( entry_indices );
+ memfree(mem_ptr);
+ memdelete_arr(entry_array);
+ memdelete_arr(entry_indices);
}
-