@@ -40,4 +40,24 @@
#define likely(X) __builtin_expect((X) != 0, 1)
#define unlikely(X) __builtin_expect((X), 0)
+namespace GTM HIDDEN {
+
+// Locally defined protected allocation functions.
+//
+// To avoid dependency on libstdc++ new/delete, as well as to not
+// interfere with the wrapping of the global new/delete we wrap for
+// the user in alloc_cpp.cc, use class-local versions that defer
+// to malloc/free. Recall that operator new/delete does not go through
+// normal lookup and so we cannot simply inject a version into the
+// GTM namespace.
+// If separate_cl is true, the allocator will try to return memory that is on
+// cache lines that are not shared with any object used by another thread.
+extern void * xmalloc (size_t s, bool separate_cl = false)
+ __attribute__((malloc, nothrow));
+extern void * xrealloc (void *p, size_t s, bool separate_cl = false)
+ __attribute__((malloc, nothrow));
+
+} // namespace GTM
+
+
#endif // COMMON_H
new file mode 100644
@@ -0,0 +1,110 @@
+/* Copyright (C) 2011 Free Software Foundation, Inc.
+ Contributed by Torvald Riegel <triegel@redhat.com>.
+
+ This file is part of the GNU Transactional Memory Library (libitm).
+
+ Libitm is free software; you can redistribute it and/or modify it
+ under the terms of the GNU General Public License as published by
+ the Free Software Foundation; either version 3 of the License, or
+ (at your option) any later version.
+
+ Libitm is distributed in the hope that it will be useful, but WITHOUT ANY
+ WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
+ FOR A PARTICULAR PURPOSE. See the GNU General Public License for
+ more details.
+
+ Under Section 7 of GPL version 3, you are granted additional
+ permissions described in the GCC Runtime Library Exception, version
+ 3.1, as published by the Free Software Foundation.
+
+ You should have received a copy of the GNU General Public License and
+ a copy of the GCC Runtime Library Exception along with this program;
+ see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
+ <http://www.gnu.org/licenses/>. */
+
+#ifndef LIBITM_CONTAINERS_H
+#define LIBITM_CONTAINERS_H 1
+
+#include "common.h"
+#include <stdio.h>
+
+namespace GTM HIDDEN {
+
+// A simple vector-like container.
+// If alloc_seperate_cl is true, allocations will happen on separate cache
+// lines.
+template <typename T, bool alloc_separate_cl = true>
+class vector
+{
+ private:
+ size_t m_capacity;
+ size_t m_size;
+ T* entries;
+
+ // Initial capacity of the vector.
+ static const size_t default_initial_capacity = 32;
+ // Above that capacity, grow vector by that size for each call.
+ static const size_t default_resize_max = 2048;
+ // Resize vector to at least this capacity.
+ static const size_t default_resize_min = 32;
+
+ // Don't try to copy this vector.
+ vector<T, alloc_separate_cl>(const vector<T, alloc_separate_cl>& x);
+
+ public:
+ typedef T datatype;
+ typedef T* iterator;
+
+ iterator begin() const { return entries; }
+ iterator end() const { return entries + m_size; }
+ T& operator[] (size_t pos) { return entries[pos]; }
+ const T& operator[] (size_t pos) const { return entries[pos]; }
+
+ vector<T, alloc_separate_cl>(size_t initial_size = default_initial_capacity)
+ : m_capacity(initial_size),
+ m_size(0)
+ {
+ if (m_capacity > 0)
+ entries = (T*) xmalloc(sizeof(T) * m_capacity, alloc_separate_cl);
+ else
+ entries = 0;
+ }
+ ~vector<T, alloc_separate_cl>() { if (m_capacity) free(entries); }
+
+ void resize()
+ {
+ if (m_capacity >= default_resize_max)
+ m_capacity = m_capacity + default_resize_max;
+ else
+ m_capacity = m_capacity * 2;
+ if (m_capacity < default_resize_min)
+ m_capacity = default_resize_min;
+ entries = (T*) xrealloc(entries, sizeof(T) * m_capacity, alloc_separate_cl);
+ }
+ void resize_noinline() __attribute__((noinline)) { resize(); }
+
+ size_t size() const { return m_size; }
+ size_t capacity() const { return this->capacity; }
+
+ void clear() { m_size = 0; }
+
+ iterator push() {
+ // We don't want inlining here since push() is often on the fast path.
+ if (m_size == m_capacity) resize_noinline();
+ fprintf(stdout, "cap=%zu size=%zu ent=%p\n", m_capacity, m_size, entries);
+ return &entries[m_size++];
+ }
+
+ iterator pop() {
+ if (m_size > 0)
+ {
+ m_size--;
+ return entries + m_size;
+ }
+ else return 0;
+ }
+};
+
+} // namespace GTM
+
+#endif // LIBITM_CONTAINERS_H
@@ -53,18 +53,6 @@ template<> struct sized_integral<8> { typedef uint64_t type; };
typedef unsigned int gtm_word __attribute__((mode (word)));
-// Locally defined protected allocation functions.
-//
-// To avoid dependency on libstdc++ new/delete, as well as to not
-// interfere with the wrapping of the global new/delete we wrap for
-// the user in alloc_cpp.cc, use class-local versions that defer
-// to malloc/free. Recall that operator new/delete does not go through
-// normal lookup and so we cannot simply inject a version into the
-// GTM namespace.
-
-extern void * xmalloc (size_t s) __attribute__((malloc, nothrow));
-extern void * xrealloc (void *p, size_t s) __attribute__((malloc, nothrow));
-
} // namespace GTM
#include "target.h"
@@ -74,6 +62,7 @@ extern void * xrealloc (void *p, size_t s) __attribute__((malloc, nothrow));
#include "cachepage.h"
#include "stmlock.h"
#include "dispatch.h"
+#include "containers.h"
namespace GTM HIDDEN {
@@ -118,9 +107,7 @@ struct gtm_transaction
gtm_jmpbuf jb;
// Data used by local.c for the local memory undo log.
- struct gtm_local_undo **local_undo;
- size_t n_local_undo;
- size_t size_local_undo;
+ vector<gtm_local_undo*> local_undo;
// Data used by alloc.c for the malloc/free undo log.
aa_tree<uintptr_t, gtm_alloc_action> alloc_actions;
@@ -37,28 +37,20 @@ struct gtm_local_undo
void
gtm_transaction::commit_local ()
{
- gtm_local_undo **local_undo = this->local_undo;
- size_t i, n = this->n_local_undo;
+ size_t i, n = local_undo.size();
if (n > 0)
{
for (i = 0; i < n; ++i)
free (local_undo[i]);
- this->n_local_undo = 0;
- }
- if (local_undo)
- {
- free (local_undo);
- this->local_undo = NULL;
- this->size_local_undo = 0;
+ this->local_undo.clear();
}
}
void
gtm_transaction::rollback_local (void)
{
- gtm_local_undo **local_undo = this->local_undo;
- size_t i, n = this->n_local_undo;
+ size_t i, n = local_undo.size();
if (n > 0)
{
@@ -71,7 +63,7 @@ gtm_transaction::rollback_local (void)
free (u);
}
}
- this->n_local_undo = 0;
+ local_undo.clear();
}
}
@@ -80,8 +72,7 @@ gtm_transaction::rollback_local (void)
void
gtm_transaction::drop_references_local (const void *ptr, size_t len)
{
- gtm_local_undo **local_undo = this->local_undo;
- size_t i, n = this->n_local_undo;
+ size_t i, n = local_undo.size();
if (n > 0)
{
@@ -110,19 +101,7 @@ GTM_LB (const void *ptr, size_t len)
undo->addr = (void *) ptr;
undo->len = len;
- if (tx->local_undo == NULL)
- {
- tx->size_local_undo = 32;
- tx->local_undo = (gtm_local_undo **)
- xmalloc (sizeof (undo) * tx->size_local_undo);
- }
- else if (tx->n_local_undo == tx->size_local_undo)
- {
- tx->size_local_undo *= 2;
- tx->local_undo = (gtm_local_undo **)
- xrealloc (tx->local_undo, sizeof (undo) * tx->size_local_undo);
- }
- tx->local_undo[tx->n_local_undo++] = undo;
+ tx->local_undo.push()[0] = undo;
memcpy (undo->saved, ptr, len);
}
@@ -59,8 +59,11 @@ GTM_fatal (const char *fmt, ...)
}
void *
-xmalloc (size_t size)
+xmalloc (size_t size, bool separate_cl)
{
+ // TODO Use posix_memalign if separate_cl is true, or some other allocation
+ // method that will avoid sharing cache lines with data used by other
+ // threads.
void *r = malloc (size);
if (r == 0)
GTM_fatal ("Out of memory allocating %lu bytes", (unsigned long) size);
@@ -68,8 +71,11 @@ xmalloc (size_t size)
}
void *
-xrealloc (void *old, size_t size)
+xrealloc (void *old, size_t size, bool separate_cl)
{
+ // TODO Use posix_memalign if separate_cl is true, or some other allocation
+ // method that will avoid sharing cache lines with data used by other
+ // threads.
void *r = realloc (old, size);
if (r == 0)
GTM_fatal ("Out of memory allocating %lu bytes", (unsigned long) size);