stasis-aries-wal/doc/paper/LLADD-Freenix.tex

1210 lines
64 KiB
TeX
Raw Normal View History

2004-10-22 04:57:25 +00:00
%% LyX 1.3 created this file. For more info, see http://www.lyx.org/.
%% Do not edit unless you really know what you are doing.
%\documentclass[letterpaper,twocolumn,english]{article}
%\usepackage[T1]{fontenc}
%\usepackage[latin1]{inputenc}
%\usepackage{graphicx}
\documentclass[letterpaper,twocolumn,english]{article}
\usepackage[latin1]{inputenc}
\usepackage{graphicx}
\usepackage{usenix,epsfig}
2004-10-23 02:19:01 +00:00
%\makeatletter
%\documentclass{article}
%\usepackage{usenix,epsfig,twocolumn}
2004-10-22 04:57:25 +00:00
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% LyX specific LaTeX commands.
%% Bold symbol macro for standard LaTeX users
%\newcommand{\boldsymbol}[1]{\mbox{\boldmath $#1$}}
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% User specified LaTeX commands.
\usepackage[T1]{fontenc}
\usepackage{ae,aecompl}
%\usepackage{babel}
%\makeatother
\begin{document}
\date{}
2004-10-23 02:06:10 +00:00
\title{\Large \bf LLADD: An Extensible Transactional Storage Layer\\
\normalsize{(yaahd)}}
2004-10-22 04:57:25 +00:00
\author{
2004-10-22 21:09:45 +00:00
Russell Sears and Eric Brewer\\
{\em UC Berkeley}\\
2004-10-22 04:57:25 +00:00
% is there a standard format for email/URLs??
% remember that ~ doesn't do what you expect, use \~{}.
2004-10-22 21:09:45 +00:00
{\normalsize \{sears,brewer\}@cs.berkeley.edu, http://lladd.sourceforge.net} \\
2004-10-22 04:57:25 +00:00
%
% copy the following lines to add more authors
2004-10-22 21:09:45 +00:00
% \smallskip
% Name Two Here \\
%{\em Two's Institution}\\
2004-10-22 04:57:25 +00:00
%% is there a standard format for email/URLs??
2004-10-22 21:09:45 +00:00
%{\normalsize two@host.site.dom, http://host.site.dom/twourl}
2004-10-22 04:57:25 +00:00
%
} % end author
\maketitle
2004-10-22 21:09:45 +00:00
\thispagestyle{plain}
2004-10-22 04:57:25 +00:00
\subsection*{Abstract}
2004-10-22 05:07:22 +00:00
Although many systems provide transactionally consistent data management,
existing implementations are generally monolithic and tied to a higher-level DBMS, limiting the scope of their usefulness to a single application,
2004-10-22 04:57:25 +00:00
or a specific type of problem. As a result, many systems are forced
2004-10-22 05:07:22 +00:00
to ``work around'' the data models provided by a transactional storage
layer. Manifestation of this problem include ``impedence mismatch''
2004-10-22 04:57:25 +00:00
in the database world and the limited number of data models provided
2004-10-22 21:02:10 +00:00
by existing libraries such as Berkeley DB. In this paper, we describe
2004-10-22 05:07:22 +00:00
a light-weight, easily extensible library, LLADD, that allows application
2004-10-22 04:57:25 +00:00
developers to develop scalable and transactional application-specific
data structures. We demonstrate that LLADD is simpler than prior systems
2004-10-22 05:07:22 +00:00
and is very flexible, while performing favorably in a number of
2004-10-22 04:57:25 +00:00
micro-benchmarks. We also describe, in simple and concrete terms,
the issues inherent in the design and implementation of robust, scalable
transactional data structures. In addition to the source code, we
have also made a comprehensive suite of unit-tests, API documentation,
and debugging mechanisms publicly available.%
\footnote{http://lladd.sourceforge.net/%
}
\section{Introduction}
Changes in data models, consistency requirements, system scalibility,
communication models and fault models require changes to the storage
2004-10-22 05:07:22 +00:00
and recovery subsystems of modern applications.
For applications that are willing to store all of their data in a
DBMS, and access it only via SQL, existing databases are just fine and
LLADD has little to offer. However, for those applications that need
more direct management of data, LLADD offers a layered architecture
that enables simple but robust data management.\footnote{A large class
of such applications are deemed ``navigational'' in the database
vocabulary, as they directly navigate data structures rather than
2004-10-22 21:09:45 +00:00
perform set operations.}
We also believe that LLADD is applicable in
the context of new, special-purpose database systems such as XML databases,
2004-10-23 02:19:01 +00:00
streaming databases, and extensible/semantic file systems~\cite{reiser, semantic}. These form a
2004-10-22 21:09:45 +00:00
fruitful area of current research, but existing monolithic database systems tend to be a poor fit for these new areas.
2004-10-22 05:07:22 +00:00
2004-10-22 21:09:45 +00:00
The basic approach of LLADD, taken from ARIES~\cite{aries}, is to build
2004-10-22 05:07:22 +00:00
\emph{transactional pages}, which enables recovery on a page-by-page
basis, despite support for high concurrency and the minimization of
2004-10-23 02:06:10 +00:00
disk seeks during commit (by using a log). We show how to build a variety
2004-10-22 05:07:22 +00:00
of useful data managers on top of this layer, including persistent
2004-10-23 02:19:01 +00:00
hash tables, lightweight recoverable virtual memory (LRVM)~\cite{lrvm}, and simple
2004-10-22 05:07:22 +00:00
databases. We also cover the details of crash recovery,
2004-10-23 02:06:10 +00:00
application-level support for transaction abort and commit, and latching for multithreaded applications.
Finally, we discuss the shortcomings of common applications, and explain
why LLADD provides an appropriate solution to these problems.
%[more coverage of kinds of apps? imap, lrvm, cht, file system, database]
2004-10-22 04:57:25 +00:00
Many implementations of transactional pages exist in industry and
in the literature. Unfortunately, these algorithms tend either to
be straightforward and unsuitable for real-world deployment, or are
robust and scalable, but achieve these properties by relying upon
2004-10-22 21:09:45 +00:00
intricate sets of internal and often implicit interactions. The
ARIES algorithm falls into the second category, and has been extremely
2004-10-22 05:07:22 +00:00
sucessful as part of the IBM DB2 database system.
It provides performance and reliability that is comparable to that of current
2004-10-22 04:57:25 +00:00
commercial and open-source products. Unfortunately, while the algorithm
2004-10-23 02:06:10 +00:00
is conceptually simple, many subtleties arise in its implementation.
2004-10-22 04:57:25 +00:00
We chose ARIES as the basis of LLADD, and have made a significant
2004-10-22 05:07:22 +00:00
effort to document these interactions. Although a complete discussion
2004-10-22 21:09:45 +00:00
of the ARIES algorithm is beyond the scope of this paper, we will
2004-10-23 02:06:10 +00:00
provide a brief overview, and explain the details that are relevant
2004-10-22 04:57:25 +00:00
to developers that wish to extend LLADD.
2004-10-22 21:09:45 +00:00
By documenting the interface between ARIES and higher-level primitives
2004-10-22 05:07:22 +00:00
such as data structures, and by structuring LLADD to make this
interface explicit in both the library and its extensions, we hope to
make it easy to produce correct and efficient durable data
structures. In existing systems (and indeed, in earlier versions of
LLADD), the implementation of such structures is extremely
complicated, and subject to the introduction of incredibly subtle
errors that would only be evident during crash recovery or at other
inconvenient times. Thus there is great value is reusing these lower
layers once developed.
2004-10-22 04:57:25 +00:00
Finally, by approaching this problem by implementing a number of simple
2004-10-22 05:07:22 +00:00
modules that ``do one thing and do it well'', we believe that
LLADD can provide competitive performance while making future improvements
2004-10-22 04:57:25 +00:00
to its core implementation significantly easier. In order to achieve
this goal, LLADD has been split into a number of modules forming a
2004-10-22 21:09:45 +00:00
{\em core library}, and a number of extensions called {\em operations} that
2004-10-22 04:57:25 +00:00
build upon the core library. Since each of these modules exports a
stable interface, they can be independently improved.
\subsection{Prior Work\label{sub:Prior-Work}}
An extensive amount of prior work covers the algorithms presented in
this paper. Most fundamentally, systems that provide transactional
consistency to their users generally include a number of common
2004-10-22 21:09:45 +00:00
modules. Figure~\ref{cap:DB-Architecture} presents a high-level overview of a typical system.
2004-10-22 04:57:25 +00:00
\begin{figure}
\includegraphics[%
width=1.0\columnwidth]{DB-Architecture.pdf}
\caption{\label{cap:DB-Architecture}Conceptual view of a modern
2004-10-23 02:06:10 +00:00
transactional application. Current systems include high-level
2004-10-22 04:57:25 +00:00
functionality, such as indices and locking, but are not designed to
2004-10-23 02:06:10 +00:00
allow developers to replace this functionality with
application-specific modules.}
2004-10-22 04:57:25 +00:00
\end{figure}
Many applications make use of transactional storage, and each is
designed for a specific application, or set of applications. LLADD
provides a flexible substrate that allows such applications to be
developed. The complexity of existing systems varies widely, as do
the applications for which these systems are designed.
On the database side of things, relational databases excel in areas
where performance is important, but where the consistency and
2004-10-23 02:06:10 +00:00
durability of the data are crucial. Often, databases significantly
2004-10-22 04:57:25 +00:00
outlive the software that uses them, and must be able to cope with
2004-10-23 02:19:01 +00:00
changes in business practices, system architectures, etc.~\cite{relational}
2004-10-22 04:57:25 +00:00
2004-10-23 02:19:01 +00:00
Object-oriented databases are more focused on facilitating the
development of complex applications that require reliable storage, and
may take advantage of less-flexible, more efficient data models,
2004-10-22 04:57:25 +00:00
as they often only interact with a single application, or a handful of
2004-10-23 02:19:01 +00:00
variants of that application.~\cite{lamb}
2004-10-22 04:57:25 +00:00
Databases are designed for circumstances where development time may
dominate cost, many users must share access to the same data, and
where security, scalability, and a host of other concerns are
important. In many, if not most, circumstances these issues are less
2004-10-22 04:57:25 +00:00
important, or even irrelevant. Therefore, applying a database in
these situations is likely overkill, which may partially explain the
2004-10-23 02:19:01 +00:00
popularity of MySQL~\cite{mysql}, which allows some of these constraints to be
2004-10-22 04:57:25 +00:00
relaxed at the discretion of a developer or end user.
Still, there are many applications where MySQL is still too
inflexible. In order to serve these applications, a host of software
solutions have been devised. Some are extremely complex, such as
semantic file systems, where the file system understands the contents
of the files that it contains, and is able to provide services such as
rapid search, or file-type specific operations such as thumbnailing,
automatic content updates, and so on. Others are simpler, such as
2004-10-23 02:19:01 +00:00
Berkeley~DB,~\cite{berkeleyDB, bdb} which provides transactional storage of data in unindexed
2004-10-22 05:44:40 +00:00
form, in indexed form using a hash table, or a tree. LRVM is a version
2004-10-22 04:57:25 +00:00
of malloc() that provides transacational memory, and is similar to an
2004-10-22 05:44:40 +00:00
object-oriented database, but is much lighter weight, and more
2004-10-23 02:06:10 +00:00
flexible~\cite{lrvm}.
2004-10-22 04:57:25 +00:00
Finally, some applications require incredibly simple, but extremely
scalable storage mechanisms. Cluster Hash Tables are a good example
of the type of system that serves these applications well, due to
their relative simplicity, and extremely good scalability
2004-10-22 05:44:40 +00:00
characteristics. Depending on the fault model on which a cluster hash table is
2004-10-22 19:24:03 +00:00
implemented, it is quite plausible that key portions of
2004-10-22 04:57:25 +00:00
the transactional mechanism, such as forcing log entries to disk, will
be replaced with other durability schemes, such as in-memory
replication across many nodes, or multiplexing log entries across
multiple systems. This level of flexibility would be difficult to
retrofit into existing transactional applications, but is appropriate
in many environments.
We have only provided a small sampling of the many applications that
make use of transactional storage. Unfortunately, it is extremely
difficult to implement a correct, efficient and scalable transactional
2004-10-23 02:06:10 +00:00
data store, and we know of no library that provides low-level access
2004-10-22 19:24:03 +00:00
to the primitives of such a durability algorithm. These algorithms
2004-10-22 04:57:25 +00:00
have a reputation of being complex, with many intricate interactions,
which prevent them from being implemented in a modular, easily
understandable, and extensible way.
Because of this, many applications that would benefit from
2004-10-23 02:06:10 +00:00
transactional storage, such as CVS and many implementations of IMAP,
either ignore the problem, leaving the burden of recovery to system
administrators or users, or implement ad-hoc solutions that employ
2004-10-23 02:06:10 +00:00
complex, application-specific consistency protocols in order to ensure
the consistency of their data. This increases the complexity of such
applications, and often provides only a partial solution to the
transactional storage problem, resulting in erratic and unpredictable
application behavior.
In addition to describing such an
2004-10-22 19:24:03 +00:00
implementation of ARIES, a well-tested
``industrial strength'' algorithm for transactional storage, this paper
outlines the most important interactions that we discovered (that
2004-10-22 04:57:25 +00:00
is, the ones that could not be encapsulated within our
2004-10-22 19:24:03 +00:00
implementation), and gives the reader a sense of how to use the
primitives the library provides.
2004-10-22 04:57:25 +00:00
%Many plausible lock managers, can do any one you want.
%too much implemented part of DB; need more 'flexible' substrate.
\section{ARIES from an Operation's Perspective}
Instead of providing a comprehensive discussion of ARIES, we will
focus upon those features of the algorithm that are most relevant
to a developer attempting to add a new set of operations. Correctly
implementing such extensions is complicated by concerns regarding
concurrency, recovery, and the possibility that any operation may
be rolled back at runtime.
We first sketch the constraints placed upon operation implementations,
and then describe the properties of our implementation of ARIES that
2004-10-22 05:44:40 +00:00
make these constraints necessary. Because comprehensive discussions of
2004-10-23 02:19:01 +00:00
write ahead logging protocols and ARIES are available elsewhere,~\cite{haerder, aries} we
2004-10-22 05:44:40 +00:00
only discuss those details relevant to the implementation of new
operations in LLADD.
2004-10-22 04:57:25 +00:00
\subsection{Properties of an Operation\label{sub:OperationProperties}}
A LLADD operation consists of some code that performs some action on
the developer's behalf. These operations implement the high-level
actions that are composed into transactions. They are implemented at
a relatively low level, and have full access to the ARIES algorithm.
We expect the majority of an application to reason in terms of the
interface provided by custom operations, allowing the the application,
the operation, and LLADD itself to be independently improved.
Since transactions may be aborted,
2004-10-22 04:57:25 +00:00
the effects of an operation must be reversible. Furthermore, aborting
and comitting transactions may be interleaved, and LLADD does not
2004-10-22 19:24:03 +00:00
allow cascading aborts,%
2004-10-22 04:57:25 +00:00
\footnote{That is, by aborting, one transaction may not cause other transactions
to abort. To understand why operation implementors must worry about
this, imagine that transaction A split a node in a tree, transaction
B added some data to the node that A just created, and then A aborted.
When A was undone, what would become of the data that B inserted?%
} so in order to implement an operation, we must implement some sort
2004-10-22 19:24:03 +00:00
of locking, or other concurrency mechanism that isolates transactions
2004-10-23 02:19:01 +00:00
from each other. LLADD only provides physical consistency; due to the variety of locking systems available, and their interaction with application workload,~\cite{multipleGenericLocking} we leave
2004-10-22 05:44:40 +00:00
it to the application to decide what sort of transaction isolation is
appropriate. For example, it is relatively easy to
2004-10-23 02:19:01 +00:00
build a strict two-phase locking lock manager~\cite{hierarcicalLocking} on top of LLADD, as
2004-10-22 05:44:40 +00:00
needed by a DBMS, or a simpler lock-per-folder approach that would
suffice for an IMAP server. Thus, data dependencies among
transactions are allowed, but we still must ensure the physical
consistency of our data structures, such as operations on pages or locks.
2004-10-22 04:57:25 +00:00
2004-10-22 19:24:03 +00:00
Also, all actions performed by a transaction that committed must be
2004-10-22 04:57:25 +00:00
restored in the case of a crash, and all actions performed by aborting
transactions must be undone. In order for LLADD to arrange for this
to happen at recovery, operations must produce log entries that contain
all information necessary for undo and redo.
2004-10-22 05:44:40 +00:00
An important concept in ARIES is the ``log sequence number'' or LSN.
An LSN is essentially a virtual timestamp that goes on every page; it
2004-10-23 02:23:50 +00:00
marks the last log entry that is reflected on the page, which
2004-10-22 05:44:40 +00:00
implies that all previous log entries are also reflected. Given the
LSN, you can tell where to start playing back the log to bring a page
up to date. The LSN goes on the page so that it is always written to
disk atomically with the data of the page.
ARIES (and thus LLADD) allows pages to be {\em stolen}, i.e. written
back to disk while they still contain uncommitted data. It is
tempting to disallow this, but to do so has serious consequences such as
2004-10-22 05:44:40 +00:00
a increased need for buffer memory (to hold all dirty pages). Worse,
as we allow multiple transactions to run concurrently on the same page
(but not typically the same item), it may be that a given page {\em
always} contains some uncommitted data and thus could never be written
back to disk. To handle stolen pages, we log UNDO records that
we can use to undo the uncommitted changes in case we crash. LLADD
2004-10-23 02:23:50 +00:00
ensures that the UNDO record is durable in the log before the
2004-10-22 05:44:40 +00:00
page is written back to disk, and that the page LSN reflects this log entry.
Similarly, we do not force pages out to disk every time a transaction
commits, as this limits performance. Instead, we log REDO records
that we can use to redo the change in case the committed version never
makes it to disk. LLADD ensures that the REDO entry is durable in the
log before the transaction commits. REDO entries are physical changes
to a single page (``page-oriented redo''), and thus must be redone in
the exact order.
One unique aspect of LLADD, which
is not true for ARIES, is that {\em normal} operations use the REDO
function; i.e. there is no way to modify the page except via the REDO
operation. This has the great property that the REDO code is known to
work, since even the original update is a ``redo''.
2004-10-22 19:24:03 +00:00
In general, the LLADD philosophy is that you
define operations in terms of their REDO/UNDO behavior, and then build
the actual update methods around those.
2004-10-22 05:44:40 +00:00
Eventually, the page makes it to disk, but the REDO entry is still
useful: we can use it to roll forward a single page from an archived
copy. Thus one of the nice properties of LLADD, which has been
tested, is that we can handle media failures very gracefully: lost
disk blocks or even whole files can be recovered given an old version
and the log.
2004-10-22 04:57:25 +00:00
\subsection{Normal Processing}
Operation implementors follow the pattern in Figure \ref{cap:Tset},
and need only implement a wrapper function (``Tset()'' in the figure,
and a pair of redo and undo functions will be registered with LLADD.
The Tupdate function, which is built into LLADD, handles most of the
runtime complexity. LLADD also uses the undo and redo functions
during recovery, in the same way that they are used during normal
processing.
2004-10-22 04:57:25 +00:00
\subsubsection{The buffer manager}
LLADD manages memory on behalf of the application and prevents pages
2004-10-23 02:23:50 +00:00
from being stolen prematurely. Although LLADD uses the STEAL policy
and may write buffer pages to disk before transaction commit, it still
must make sure that the UNDO log entries have been forced to disk
before the page is written to disk. Therefore, operations must inform
the buffer manager when they write to a page, and update the LSN of
the page. This is handled automatically by the write methods that LLADD
provides to operation implementors (such as writeRecord()). However,
it is also possible to create your own low-level page manipulation
routines, in which case these routines must follow the protocol.
2004-10-22 04:57:25 +00:00
2004-10-23 02:23:50 +00:00
\subsubsection{Log entries and forward operation\\ (the Tupdate() function)\label{sub:Tupdate}}
2004-10-22 04:57:25 +00:00
In order to handle crashes correctly, and in order to the undo the
effects of aborted transactions, LLADD provides operation implementors
with a mechanism to log undo and redo information for their actions.
This takes the form of the log entry interface, which works as follows.
Operations consist of a wrapper function that performs some pre-calculations
and perhaps acquires latches. The wrapper function then passes a log
2004-10-23 02:23:50 +00:00
entry to LLADD. LLADD passes this entry to the logger, {\em and then processes
it as though it were redoing the action during recovery}, calling a function
2004-10-22 04:57:25 +00:00
that the operation implementor registered with
LLADD. When the function returns, control is passed back to the wrapper
function, which performs any post processing (such as generating return
values), and releases any latches that it acquired. %
\begin{figure}
2004-10-23 00:42:54 +00:00
\begin{center}
\includegraphics[%
2004-10-22 04:57:25 +00:00
width=0.70\columnwidth]{TSetCall.pdf}
2004-10-23 00:42:54 +00:00
\end{center}
2004-10-22 04:57:25 +00:00
\caption{\label{cap:Tset}Runtime behavior of a simple operation. Tset() and do\_set() are
2004-10-22 04:57:25 +00:00
implemented as extensions, while Tupdate() is built in. New operations
need not be aware of the complexities of LLADD.}
\end{figure}
This way, the operation's behavior during recovery's redo phase (an
uncommon case) will be identical to the behavior during normal processing,
making it easier to spot bugs. Similarly, undo and redo operations take
an identical set of parameters, and undo during recovery is the same
as undo during normal processing. This makes recovery bugs more obvious and allows redo
functions to be reused to implement undo.
Although any latches acquired by the wrapper function will not be
reacquired during recovery, the redo phase of the recovery process
is single threaded. Since latches acquired by the wrapper function
are held while the log entry and page are updated, the ordering of
the log entries and page updates associated with a particular latch
2004-10-22 21:09:45 +00:00
will be consistent. Because undo occurs during normal operation,
2004-10-22 19:24:03 +00:00
some care must be taken to ensure that undo operations obtain the
2004-10-22 05:44:40 +00:00
proper latches.
2004-10-22 04:57:25 +00:00
\subsection{Recovery}
2004-10-22 21:09:45 +00:00
In this section, we present the details of crach recovery, user-defined logging, and atomic actions that commit even if their enclosing transaction aborts.
2004-10-22 04:57:25 +00:00
\subsubsection{ANALYSIS / REDO / UNDO}
2004-10-22 21:09:45 +00:00
Recovery in ARIES consists of three stages, analysis, redo and undo.
2004-10-22 19:40:13 +00:00
The first, analysis, is
2004-10-22 19:24:03 +00:00
implemented by LLADD, but will not be discussed in this
2004-10-22 04:57:25 +00:00
paper. The second, redo, ensures that each redo entry in the log
will have been applied each page in the page file exactly once.
2004-10-22 19:24:03 +00:00
The third phase, undo, rolls back any transactions that were active
2004-10-22 04:57:25 +00:00
when the crash occured, as though the application manually aborted
2004-10-22 19:40:13 +00:00
them with the {}``abort'' function call.
2004-10-22 04:57:25 +00:00
After the analysis phase, the on-disk version of the page file
is in the same state it was in when LLADD crashed. This means that
some subset of the page updates performed during normal operation
have made it to disk, and that the log contains full redo and undo
information for the version of each page present in the page file.%
\footnote{Although this discussion assumes that the entire log is present, the
ARIES algorithm supports log truncation, which allows us to discard
old portions of the log, bounding its size on disk.%
} However, we make no further assumptions regarding the order in which
pages were propogated to disk. Therefore, redo must assume that any
data structures, lookup tables, etc. that span more than a single
page are in an inconsistent state. Therefore, as the redo phase re-applies
the information in the log to the page file, it must address all pages directly.
Therefore, the redo information for each operation in the log
must contain the physical address (page number) of the information
that it modifies, and the portion of the operation executed by a single
2004-10-22 19:40:13 +00:00
redo log entry must only rely upon the contents of the page that the log
2004-10-22 04:57:25 +00:00
entry refers to. Since we assume that pages are propagated to disk
2004-10-22 19:24:03 +00:00
atomically, the REDO phase may rely upon information contained within
2004-10-22 04:57:25 +00:00
a single page.
2004-10-22 19:24:03 +00:00
Once redo completes, we have applied some prefix of the run-time log that contains
complete entries for all committed transactions. Therefore, we know that the page file is in
a physically consistent state, although it contains portions of the
results of uncomitted transactions. The final stage of recovery is
2004-10-22 04:57:25 +00:00
the undo phase, which simply aborts all uncomitted transactions. Since
2004-10-22 19:40:13 +00:00
the page file is physically consistent, the transactions may be aborted
2004-10-22 04:57:25 +00:00
exactly as they would be during normal operation.
\subsubsection{Physical, Logical and Phisiological Logging.}
The above discussion avoided the use of some terminology that is common
2004-10-22 21:09:45 +00:00
in the database literature and which should be presented here. ``Physical
loggging'' is the practice of logging physical (byte-level) updates
and the physical (page number) addresses to which they are applied.
2004-10-22 04:57:25 +00:00
2004-10-22 21:09:45 +00:00
It is subtly different than ``physiological logging,'' which is
2004-10-22 04:57:25 +00:00
what LLADD recommends for its redo records. In physiological logging,
2004-10-22 21:09:45 +00:00
the physical address (page number) is stored, but the byte offset
2004-10-22 04:57:25 +00:00
and the actual difference are stored implicitly in the parameters
of some function. When the parameters are applied to the function,
it will update the page in a way that preserves application semantics.
2004-10-22 21:09:45 +00:00
The common use for this is {\em slotted pages}, which use a level of indirection to allow records to be rearranged on the page; redo operations use the index as the parameter rather than the page offset. For example, data within
2004-10-22 04:57:25 +00:00
a single page can be re-arranged at runtime to produce contiguous
2004-10-23 02:19:01 +00:00
regions of free space. LLADD generalizes this model; for example, the parameters passed to the function may be significantly smaller than the physical change made to the page.~\cite{physiological}
2004-10-22 04:57:25 +00:00
{}``Logical logging'' can only be used for undo entries in LLADD,
and is identical to physiological logging, except that it stores a
logical address (the key of a hash table, for instance) instead of
a physical address. This allows the location of data in the page file
to change, even if outstanding transactions may have to roll back
changes made to that data. Clearly, for LLADD to be able to apply
logical log entries, the page file must be physically consistent,
ruling out use of logical logging for redo operations.
LLADD supports all three types of logging, and allows developers to
register new operations, which is the key to its extensibility. After
discussing LLADD's architecture, we will revisit this topic with a
concrete example.
\subsection{Concurrency and Aborted Transactions}
2004-10-22 21:54:35 +00:00
Section~\ref{sub:OperationProperties} states that LLADD does not
allow cascading aborts, implying that operation implementors must
protect transactions from any structural changes made to data structures
by uncomitted transactions, but LLADD does not provide any mechanisms
designed for long-term locking. However, one of LLADD's goals is to
make it easy to implement custom data structures for use within safe,
multi-threaded transactions. Clearly, an additional mechanism is needed.
2004-10-22 21:54:35 +00:00
The solution is to allow portions of an operation to ``commit'' before
the operation returns.
\footnote{We considered the use of nested top actions, which LLADD could easily
support. However, we currently use the slightly simpler (and lighter-weight)
mechanism described here. If the need arises, we will add support
2004-10-22 21:54:35 +00:00
for nested top actions.}
An operation's wrapper is just a normal function, and therefore may
generate multiple log entries. First, it writes an undo-only entry
to the log. This entry will cause the \emph{logical} inverse of the
current operation to be performed at recovery or abort, must be idempotent,
and must fail gracefully if applied to a version of the database that
does not contain the results of the current operation. Also, it must
behave correctly even if an arbitrary number of intervening operations
are performed on the data structure.
2004-10-22 21:54:35 +00:00
Next, the operation writes one or more redo-only log entries that may perform structural
2004-10-22 22:21:40 +00:00
modifications to the data structure. These redo entries have the constraint that any prefix of them must leave the database in a consistent state, since only a prefix might execute before a crash. This is not as hard as it sounds, and in fact the
2004-10-23 02:19:01 +00:00
$B^{LINK}$ tree~\cite{blink} is an example of a B-Tree implementation
that behaves in this way, while the linear hash table implementation
2004-10-22 21:54:35 +00:00
discussed in Section~\ref{sub:Linear-Hash-Table} is a scalable
hash table that meets these constraints.
2004-10-22 21:54:35 +00:00
[EAB: I still think there must be a way to log all of the redoes
before any of the actions take place, thus ensuring that you can redo
the whole thing if needed. Alternatively, we could pin a page until
the set completes, in which case we know that that all of the records
are in the log before any page is stolen.]
2004-10-22 04:57:25 +00:00
\subsection{Summary}
This section presented a relatively simple set of rules and patterns
that a developer must follow in order to implement a durable, transactional
and highly-concurrent data structure using LLADD:
\begin{itemize}
\item Pages should only be updated inside of a redo or undo function.
\item An update to a page should update the LSN.
\item If the data read by the wrapper function must match the state of
the page that the redo function sees, then the wrapper should latch
the relevant data.
\item Redo operations should address pages by their physical offset,
2004-10-22 19:40:13 +00:00
while Undo operations should use a more permanent address (such as
2004-10-22 04:57:25 +00:00
index key) if the data may move between pages over time.
\item An undo operation must correctly update a data structure if any
prefix of its corresponding redo operations are applied to the
structure, and if any number of intervening operations are applied to
the structure.
\end{itemize}
Because undo and redo operations during normal operation and recovery
are similar, most bugs will be found with conventional testing
strategies. It is difficult to verify the final property, although a
number of tools could be written to simulate various crash scenarios,
2004-10-22 19:40:13 +00:00
and check the behavior of operations under these scenarios. Of course,
such a tool could easily be applied to existing LLADD operations.
2004-10-22 04:57:25 +00:00
Note that the ARIES algorithm is extremely complex, and we have left
2004-10-22 05:44:40 +00:00
out most of the details needed to understand how ARIES works, or to
2004-10-23 02:19:01 +00:00
implement it correctly.\footnote{The original ARIES paper is around 70 pages, and the ARIES/IM paper~\cite{ariesim}, which coversd index implementation is roughly the same length.} Yet, we believe we have covered everything that a programmer needs
2004-10-22 19:40:13 +00:00
to know in order to implement new data structures using the
2004-10-22 05:44:40 +00:00
functionality that ARIES provides. This was possible due to the encapsulation
2004-10-22 04:57:25 +00:00
of the ARIES algorithm inside of LLADD, which is the feature that
most strongly differentiates LLADD from other, similar libraries.
We hope that this will increase the availability of transactional
2004-10-22 19:24:03 +00:00
data primitives to application developers.
2004-10-22 04:57:25 +00:00
\section{LLADD Architecture}
%
\begin{figure}
\includegraphics[%
width=1.0\columnwidth]{LLADD-Arch2.pdf}
\caption{\label{cap:LLADD-Architecture}Simplified LLADD Architecture: The
core of the library places as few restrictions on the application's
data layout as possible. Custom {}``operations'' implement the client's
2004-10-22 19:24:03 +00:00
desired data layout. The separation of these two sets of modules makes
2004-10-22 04:57:25 +00:00
it easy to improve and customize LLADD.}
\end{figure}
2004-10-22 19:24:03 +00:00
LLADD is a toolkit for building transaction managers.
It provides user-defined redo and undo behavior, and has an extendible
2004-10-22 19:40:13 +00:00
logging system with 19 types of log entries so far (not counting those
internal to LLADD, such as ``begin'', ``abort'', and ``clr''). Most of these
2004-10-22 04:57:25 +00:00
extensions deal with data layout or modification, but some deal with
other aspects of LLADD, such as extensions to recovery semantics (Section
\ref{sub:Two-Phase-Commit}). LLADD comes with some default page layout
schemes, but allows its users to redefine this layout as is appropriate.
Currently LLADD imposes two requirements on page layouts. The first
2004-10-22 19:24:03 +00:00
32 bits must contain an LSN for recovery purposes,
and the second 32 bits must contain the page type (since we allow multple page formats).
2004-10-22 04:57:25 +00:00
2004-10-22 19:24:03 +00:00
Although it ships with basic operations that support variable length
2004-10-22 04:57:25 +00:00
records, hash tables and other common data types, our goal is to
decouple all decisions regarding data format from the implementation
of the logging and recovery systems. Therefore, the preceeding section
2004-10-22 19:40:13 +00:00
is essentially documentation for users of the library, while
2004-10-22 04:57:25 +00:00
the purpose of the performance numbers in our evaluation section are
not to validate our hash table, but to show that the underlying architecture
is able to efficiently support interesting data structures.
2004-10-22 19:40:13 +00:00
Despite the complexity of the interactions between its modules, the
basic ARIES algorithm itself is quite simple. Therefore, in order to
keep LLADD simple, we started with a set of modules, and iteratively
2004-10-22 21:09:45 +00:00
refined the boundaries between these modules. Figure~\ref{cap:LLADD-Architecture} presents the resulting architecture. The
2004-10-22 19:40:13 +00:00
core of the LLADD library is quite small at 2218 lines of code, 2155
lines of implementations of operations and other extensions, and 408
2004-10-22 21:09:45 +00:00
lines of installable header files.\footnote{These counts were generated using David
2004-10-23 02:19:01 +00:00
A. Wheeler's {\tt SLOCCount}. \cite{sloccount}} The code has been documented extensively,
2004-10-22 19:40:13 +00:00
and we hope that we have exposed most of the subtle interactions
between internal modules in the online documentation.
2004-10-22 04:57:25 +00:00
As LLADD has evolved, many of its sub-systems have been incrementally
improved, and we believe that the current set of modules is amenable
to the addition of new functionality. For instance, the logging module
interface encapsulates all of the details regarding its on disk format,
which would make it straightforward to implement more exotic logging
2004-10-22 19:40:13 +00:00
techniques such as using log shipping to maintain a ``warm replica''
2004-10-22 04:57:25 +00:00
for failover purposes, or the use of log replication to avoid physical
disk access at commit time. Similarly, the interface encodes the dependencies
between the logger and other subsystems, so, for instance, the requirements
that the buffer manager places on the logger would be obvious to someone
that attempted to alter the logging functionality.%
\footnote{The buffer manager must ensure that the logger has forced the appropriate
log entries to disk before writing a dirty page to disk. Otherwise,
it would be impossible to undo the changes that had been made to the
page.%
}
The buffer manager itself is another potential area for extension.
Because the interface between the buffer manager and LLADD is simple,
we would like to support transactional access to resources beyond
simple page files. Some examples include transactional updates of
multiple files on disk, transactional groups of program executions
or network requests, or even leveraging some of the advances being
2004-10-22 21:09:45 +00:00
made in the Linux and other modern OS kernels. For example,
ReiserFS recently added support for atomic file-system operations.
This could be used to provide variable-sized pages
2004-10-22 19:40:13 +00:00
to LLADD. Combining these ideas should make it easy to
implement some interesting applications, and to improve existing
systems such as CVS, IMAP, and a host of ``simple'' desktop applications.
2004-10-22 04:57:25 +00:00
From the testing point of view, the advantage of LLADD's division
into subsystems with simple interfaces is obvious. We are able to
use standard unit-testing techniques to test each of LLADD's subsystems
independently, and have documented both external and internal interfaces,
making it easy to add new tests and debug old ones. Furthermore, by
adding a 'simulate crash' operation to a few of the key components,
we can simulate application level crashes by clearing LLADD's internal
state, re-initializing the library and verifying that recovery was
2004-10-22 19:40:13 +00:00
successful. These tests currently cover approximately
2004-10-23 02:19:01 +00:00
90\%\footnote{generated using ``gcov'', which is part of gcc, and ``lcov,'' which interprets gcov's output.~\cite{lcov}}
2004-10-22 19:40:13 +00:00
of the code. We have not yet developed a mechanism that will allow us to
2004-10-22 04:57:25 +00:00
accurately model hardware failures, which is an area where futher
2004-10-22 05:03:16 +00:00
work is needed. However, the basis for this work will be the development
of test harnesses that verify operation behavior in exceptional circumstances.
2004-10-22 04:57:25 +00:00
LLADD's performance requirements vary wildly depending on the workload
with which it is presented. Its performance on a large number of small,
sequential transactions will always be limited by the amount time
required to flush a page to disk. To some extent, compact logical
and physiological log entries improve this situation. On the other
hand, long running transactions only rarely force-write to disk and
become CPU bound. Standard profiling techniques of the overall library's
2004-10-22 05:03:16 +00:00
performance and microbenchmarks of crucial modules handle such situations
2004-10-22 04:57:25 +00:00
nicely.
A more interesting set of performance requirements are imposed by
multithreaded workloads. Each module of LLADD is reentrant, and a
C preprocessor directive allows the entire library to be instrumented
in order to profile latching behavior, which is useful both for perfomance
tuning and for debugging purposes. A thread that is not involved in
an I/O request never needs to wait for a latch held by a thread that
is waiting for I/O.%
\footnote{Strictly speaking, this statement is only true for the LLADD's core.
However, there are variants of most popular data structures that allow
us to preserve these invariants. LLADD can correctly support operations
whether or not they have these properties.%
}
There are a number of performance optimizations that are specific
to multithreaded operations that we do not perform. The most glaring
omission is log bundling; if multiple transactions commit at once,
LLADD must force the log to disk one time per transaction. This problem
is not fundamental, but simply has not made it into the current code
base. Similarly, since page eviction requires a force-write if the
full ARIES recovery algorithm is in use, we could implement a thread
2004-10-22 05:03:16 +00:00
that asynchronously maintained a set of free buffer pages. We plan to
implement such optimizations, but they are not reflected
2004-10-22 04:57:25 +00:00
in this paper's performance figures.
\section{Sample Operations}
In order to validate LLADD's architecture, and to show that it simplifies
the creation of efficient data structures, we have have implemented
a number of simple extensions. In this section, we describe their
design, and provide some concrete examples of our experiences extending
LLADD.
\subsection{Linear Hash Table\label{sub:Linear-Hash-Table}}
Linear hash tables are hash tables that are able to extend their bucket
list incrementally at runtime. They work as follows. Imagine that
we want to double the size of a hash table of size $2^{n}$, and that
the hash table has been constructed with some hash function $h_{n}(x)=h(x)\, mod\,2^{n}$
Choose $h_{n+1}(x)=h(x)\, mod\,2^{n+1}$ as the hash function for
the new table. Conceptually we are simply prepending a random bit
to the old value of the hash function, so all lower order bits remain
the same. At this point, we could simply block all concurrent access
and iterate over the entire hash table, reinserting values according
to the new hash function.
However, because of the way we chose $h_{n+1}(x),$ we know that the
contents of each bucket, $m$, will be split betwen bucket $m$ and
bucket $m+2^{n}$. Therefore, if we keep track of the last bucket
that was split, we can split a few buckets at a time, resizing the
hash table without introducing long pauses while we reorganize the
2004-10-23 02:19:01 +00:00
hash table.~\cite{lht} We can handle overflow using standard techniques.
2004-10-22 04:57:25 +00:00
LLADD's linear hash table uses linked lists of overflow buckets.
For this scheme to work, we must be able to address a portion of the
page file as though it were an expandable array. We have implemented
2004-10-22 19:24:03 +00:00
this functionality as a separate module, but will not discuss it here.
2004-10-22 04:57:25 +00:00
For the purposes of comparison, we provide two linear hash implementations.
The first is straightforward, and is layered on top of LLADD's standard
record setting operation, Tset(), and therefore performs physical
undo. This implementation provided a stepping stone to the more sophisticated
version which employs logical undo, and uses an identical on-disk
layout. As we discussed earlier, logical undo provides more opportunities
for concurrency, while decreasing the size of log entries. In fact,
2004-10-22 05:03:16 +00:00
the physical-undo implementation of the linear hash table cannot support
concurrent transactions, while threads utilizing the physical-undo
implementation never hold locks on more than two buckets.%
\footnote{However, only one thread may expand the hashtable at once. In order to amortize the overhead of initiating an expansion, and to allow concurrent insertions, the hash table is expanded in increments of a few thousand buckets.}%
2004-10-22 04:57:25 +00:00
\begin{figure}
2004-10-23 00:42:54 +00:00
\begin{center}
\includegraphics[%
width=0.70\columnwidth]{LinkedList.pdf}
2004-10-23 00:42:54 +00:00
\end{center}
2004-10-22 04:57:25 +00:00
\caption{\label{cap:Linear-Hash-Table}Linear Hash Table Bucket operations.}
\end{figure}
Because another module provides the resizable arrays needed for the
bucket list, the complexity of the linear hash algorithm is in two
areas. The first, linked list management, is straightforward in the
physical case, but must be performed in a specific order in the logical
case. See Figure \ref{cap:Linear-Hash-Table} for a sequence of steps
that safely implement the necessary linked list operations. Note that
in the first two cases, the portion of the linked list that is visible
from LLADD's point of view is always consistent. This is important
for crash recovery; it is possible that LLADD will crash before the
entire sequence of operations has been completed. The logging protocol
guarantees that some prefix of the log will be available. Therefore,
2004-10-22 05:03:16 +00:00
as long as the run-time version of the hash table is always consistent,
2004-10-22 21:02:10 +00:00
we may be certain that the logical consistency of the linked list is maintained
2004-10-22 05:03:16 +00:00
at all steps. Here, the challenge comes from the fact that the buffer
2004-10-22 04:57:25 +00:00
manager only provides atomic updates of single pages; in practice,
a linked list may span pages.
2004-10-22 05:03:16 +00:00
The last case, where buckets are split as the bucket list is expanded,
2004-10-22 04:57:25 +00:00
is a bit more complicated. We must maintain consistency between two
linked lists, and a page at the begining of the hash table that contains
the last bucket that we successfully split. Here, we misuse the undo
entry to ensure proper crash recovery. Our bucket split algorithm
is idempotent, so it may be applied an arbitrary number of times to
a given bucket with no ill-effects. Also note that (for our purposes),
there is never a good reason to undo a bucket split, so we can safely
apply the split whether or not the current transaction commits.
2004-10-22 21:02:10 +00:00
First, we write an ``undo'' record that checks the hash table's metadata and
redoes the split if necessary (this record has no effect
2004-10-22 05:03:16 +00:00
unless we crash during this bucket split). Second, we write (and execute) a series
of redo-only records to the log. These encode the bucket split, and follow
2004-10-22 04:57:25 +00:00
the linked list protocols listed above. Finally, we write a redo-only
entry that updates the hash table's metadata.%
\footnote{Had we been using nested top actions, we would not need the special
2004-10-22 19:24:03 +00:00
undo entry, but we would need to store {\em physical} undo information for
each of the modifications made to the bucket, since any subset of the pages may have been stolen. This method does have
2004-10-22 04:57:25 +00:00
the disadvantage of producing a few redo-only entries during recovery,
2004-10-22 21:02:10 +00:00
but the number of such entries is
2004-10-22 05:03:16 +00:00
bounded by the number of entries that would be produced during normal
operation.%
2004-10-22 04:57:25 +00:00
}
We allow pointer aliasing at this step so that a given key can be
present for a short period of time in both buckets. If we crash before
the undo entry is written, no harm is done. If we crash after the
entire update makes it to log, the redo stage will set the hash's
metadata appropriately, and the 'undo' record becomes a no-op. If
we crash in the middle of the bucket split, we know that the current
transaction did not commit, and that recovery will execute the 'undo'
record. It will see that the bucket split is still pending and finish
splitting the bucket appropriately. Since the bucket split is idempotent,
and we've arranged for it to behave correctly regardless of the point
at which it was interrupted, the hastable is correctly restored.
Note that there is a point during the undo phase where the bucket
is in an inconsistent physical state, although normally the redo phase
is able to bring the database to a fully consistent physical state.
We handle this by obtaining a runtime lock on the bucket during normal
operation. This runtime lock blocks any attempt to write log entries
2004-10-22 05:03:16 +00:00
that alter a bucket that is being split, so we know that no other
2004-10-22 04:57:25 +00:00
logical operations will attempt to access an inconsistent bucket.
Since the second implementation of the linear hash table uses logical
2004-10-22 05:03:16 +00:00
undo, we are able to allow concurrent updates to different portions
2004-10-22 04:57:25 +00:00
of the table. This is not true in the case of the implementation that
uses pure physical logging, as physical undo cannot generally tolerate
concurrent structural modifications to data structures.
2004-10-22 21:54:35 +00:00
\subsection{Two-Phase Commit\label{sub:Two-Phase-Commit}}
2004-10-22 04:57:25 +00:00
2004-10-22 21:54:35 +00:00
The two-phase commit protocol is used in clustering applications where
2004-10-22 04:57:25 +00:00
multiple, well maintained, well connected computers must agree upon
a set of successful transactions. Some of the systems could crash,
or the network could fail during operation, but we assume that such
2004-10-22 21:54:35 +00:00
failures are temporary. Two-phase commit designates a single computer
2004-10-22 04:57:25 +00:00
as the coordinator of a given transaction. This computer contacts
the other systems participating in the transaction, and asks them
to prepare to commit the transaction. If a subordinate system sees
that an error has occurred, or the transaction should be aborted for
some other reason, then it informs the coordinator. Otherwise, it
enters the \emph{prepared} state, and tells the coordinator that it
2004-10-22 21:02:10 +00:00
is ready to commit. At some point in the future the coordinator will
reply, telling the subordinate to commit or abort. From LLADD's point
2004-10-22 04:57:25 +00:00
of view, the interesting portion of this algorithm is the \emph{prepared}
state, since it must be able to commit a prepared transaction if it
crashes before the coordinator responds, but cannot commit before
hearing the response, since it may be asked to abort the transaction.
Implementing the prepare state on top of the ARIES algorithm constists
of writing a special log entry that informs the undo portion of the
recovery phase that it should stop rolling back the current transaction
and instead add it to the list of active transactions.%
\footnote{Also, any locks that the transaction obtained should be restored,
which is outside of the scope of LLADD, although this functionality
could be added relatively easily if a lock manager were implemented
on top of LLADD.%
2004-10-22 05:03:16 +00:00
} Due to LLADD's extendible logging system, and the simplicity
2004-10-22 04:57:25 +00:00
of its recovery code, it took an afternoon to add a prepare operation
2004-10-22 21:02:10 +00:00
to LLADD, allowing it to support applications that require two-phase commit.
A preliminary implementation of a cluster hash table that employs two-phase
commit is included in LLADD's CVS repository, but is not ready for
real-world deployment.
\subsection{Other Applications}
Previously, we mentioned a few programs that we think would benefit
from LLADD. Here we sketch the process of implementing such
applictions. LRVM implements a transactional version of malloc(). It
employs the operating system's virtual memory system to generate page
faults if the application accesses a portion of memory that have not
been swapped in. These page faults are intercepted and processed by a
transactional storage layer which loads the corresponding pages from
disk. A few simple functions such as abort() and commit() are
provided to the application, and allow it to control the duration of
its transactions. LLADD provides such a layer and the necessary
calls, reducing the LRVM implementation to an implementation of the
page fault handling code. The performance of the transactional
storage system is crucial for this sort of application, and the
variable length, keyed access, and higher levels of abstractions
provided by existing libraries would be overkill. LLADD could easily
be extended so that it employs an appropriate on-disk structure that
provides efficient, offset based access to aligned, fixed length
blocks of data. Furthermore, LRVM requires a set\_range() operation
that efficiently updates a range of a record, saving logging overhead.
All of these features could easily added to LLADD, providing a simple,
fast version of LRVM that would benefit from the infrastructure
surrounding LLADD.
CVS provides version control over large sets of files. Multiple users
may concurrently update the repository of files, and CVS attempts to
merge conflicts, and maintain the consistency of the file tree. By
adding the ability to perform file system manipulations to LLADD, we
could easily support applications with requirements similar to those
of CVS. Furthermore, we could combine the file-system manipulation
with record-oriented storage to store application-level logs, and
other important metadata. This would allow a single mechanism to
support applications such as CVS, simplifying fault tolerance, and
improving the scalibility of such applications.
IMAP is similar to CVS, but benefits further since it uses a simple,
folder-based locking protocol, which would be extremely easy to
implement using LLADD.
These last two examples highlight some of the potential advantages of
extending LLADD to manipulate the file system, although it is possible
that LLADD's page file would provide improved performance over the
file system, at the expense of some complexity, and the transparency
of file-system based storage mechanisms.
Another area of interest is in transactional serialization mechanisms
for programming languages. Existing solutions are often complex, or
are layered on top of a relational database, or other system that uses
a data format that is different than the representation the
programming language uses. The wide variety of persistance mechanisms
available for Java provide a nice survey of the potential design
choices and tradeoffs. Since LLADD can easily be adapted to an
application's desired data format, we believe that it is a good match
for such persistance mechanisms.
2004-10-22 04:57:25 +00:00
\section{Performance}
2004-10-22 04:57:25 +00:00
We hope that the preceeding sections have given the reader an idea
of the usefulness and extensibility of the LLADD library. In this
section we focus on performance evaluation.
In order to evaluate the physical and logical hashtable implementations,
we first ran a test that inserts some tuples into the database. For
2004-10-23 02:19:01 +00:00
this test, we chose fixed-length (key, value) pairs of integers. For
2004-10-22 04:57:25 +00:00
simplicity, our hashtable implementations currently only support fixed-length
keys and values, so this this test puts us at a significant advantage.
It also provides an example of the type of workload that LLADD handles
well, since LLADD is specifically designed to support application
2004-10-22 05:44:40 +00:00
specific transactional data structures. For comparison, we ran
2004-10-22 21:02:10 +00:00
``Record Number'' trials, named after the Berkeley DB access method.
2004-10-22 05:44:40 +00:00
In this case, the two programs essentially stored the data in a large
array on disk. This test provides a measurement of the speed of the
2004-10-22 21:02:10 +00:00
lowest level primitive supported by Berkeley DB, and the corresponding LLADD extension.
2004-10-22 04:57:25 +00:00
%
\begin{figure*}
2004-10-23 00:42:54 +00:00
\begin{center}
2004-10-22 05:03:16 +00:00
\includegraphics[%
2004-10-23 00:42:54 +00:00
width=0.75\textwidth]{INSERT.pdf}
\end{center}
2004-10-22 04:57:25 +00:00
\caption{\label{cap:INSERTS}The final data points for LLADD's and Berkeley
DB's record number based storage are 7.4 and 9.5 seconds, respectively.
LLADD's hash table is significantly faster than Berkeley DB in this
test, but provides less functionality than the Berkeley DB hash. Finally,
the logical logging version of LLADD's hash table is faster than the
physical version, and handles the multi-threaded test well. The threaded
2004-10-22 19:24:03 +00:00
test spawned 200 threads and split its workload into 200 separate transactions.}
2004-10-22 04:57:25 +00:00
\end{figure*}
2004-10-23 00:42:54 +00:00
%
\begin{figure*}
\begin{center}
\includegraphics[%
width=0.75\textwidth]{THREADS-sparse.pdf}
\end{center}
\caption{\label{cap:THREADS}The time required to perform a fixed
amount of processing, split across various numbers of threads. This
test was run agains the highly concurrent Logical Logging version of
the linear hash table. No significant performance degradation was
seen within the range measured. The inserts were done in serial, and
the lookups were performed in parallel.}
\end{figure*}
2004-10-22 04:57:25 +00:00
The times included in Figure \ref{cap:INSERTS} include page file
and log creation, insertion of the tuples as a single transaction,
2004-10-22 21:02:10 +00:00
and a clean program shutdown. We used the ``transapp.cs'' program from
2004-10-22 04:57:25 +00:00
the Berkeley DB 4.2 tutorial to run the Berkeley DB tests, and hardcoded
it to use integers instead of strings. We used the Berkeley DB {}``DB\_HASH''
index type for the hashtable implementation, and {}``DB\_RECNO''
2004-10-22 05:44:40 +00:00
in order to run the {}``Record Number'' test.
2004-10-22 04:57:25 +00:00
Since LLADD addresses records as \{Page, Slot, Size\} triples, which
2004-10-22 05:44:40 +00:00
is a lower level interface than Berkeley DB exports, we used the expandable
2004-10-22 04:57:25 +00:00
array that supports the hashtable implementation to run the {}``LLADD
Record Number'' test.
One should not look at Figure \ref{cap:INSERTS}, and conclude {}``LLADD
is almost five times faster than Berkeley DB,'' since we chose a
hash table implementation that is tuned for fixed-length data. Instead,
2004-10-22 19:24:03 +00:00
the conclusions we draw from this test are that, first, LLADD's primitive
2004-10-22 04:57:25 +00:00
operations are on par, perforance wise, with Berkeley DB's, which
we find very encouraging. Second, even a highly tuned implementation
2004-10-22 21:02:10 +00:00
of a ``simple,'' general purpose data structure is not without overhead,
2004-10-22 04:57:25 +00:00
and for applications where performance is important a special purpose
structure may be appropriate.
2004-10-22 22:21:40 +00:00
%Also, the multithreaded test run shows that the library is capable of
%handling a large number of threads. The performance degradation
%associated with running 200 concurrent threads was negligible. Figure
%TODO expands upon this point by plotting the time taken for various
%numbers of threads to perform a total of 500,000 (TODO-CHECK) read operations. The
%logical logging version of LLADD's hashtable outperformed the physical
The logical logging version of LLADD's hashtable outperformed the physical
2004-10-22 06:07:44 +00:00
logging version for two reasons. First, since it writes fewer undo
records, it generates a smaller log file. Second, in order to
emphasize the performance benefits of our extension mechanism, we use
lower level primitives for the logical logging version. The logical
logging version implements locking at the bucket level, so many
mutexes that are acquired by LLADD's default mechanisms are redundant.
The physical logging version of the hashtable serves as a rough proxy
for an implementation on top of a non-extendible system. Therefore,
it uses LLADD's default mechanisms, which include the redundant
acquisition of locks.
2004-10-22 05:44:40 +00:00
2004-10-22 22:21:40 +00:00
As a final note on our first performance graph, we would like to address
2004-10-22 04:57:25 +00:00
the fact that LLADD's hashtable curve is non-linear. LLADD currently
uses a fixed-size in-memory hashtable implementation in many areas,
2004-10-22 06:07:44 +00:00
and it is possible that we exceeded the fixed-size of this hashtable
2004-10-22 04:57:25 +00:00
on the larger test sets. Also, LLADD's buffer manager is currently
fixed size. Regardless of the cause of this non-linearity, we do not
believe that it is fundamental to our implementation.
2004-10-22 22:21:40 +00:00
The multithreaded test run in the first figure shows that the library
is capable of handling a large number of threads. The performance
degradation associated with running 200 concurrent threads was
2004-10-23 00:42:54 +00:00
negligible. Figure~\ref{cap:THREADS} expands upon this point by plotting the time
2004-10-22 22:21:40 +00:00
taken for various numbers of threads to perform a total of 500,000
2004-10-23 00:42:54 +00:00
read operations. The performance of LLADD in this figure
2004-10-22 22:21:40 +00:00
is essentially flat, showing only a negligable slowdown up to 250
threads. (Our test system prevented us from spawning more than 250
simultaneous threads, and we suspect that the ``true'' limit of
LLADD's scalability is must higher than 250 threads. This test was
performed on a uni-processor machine, so we did not expect to see a
significant speedup when we moved from a single thread to multiple
threads.
Unfortuantely, when ran this test on a multi-processor machine, we saw
a further degradation in performance instead of the expected speed up.
The problem seems to be the additional overhead incurred by
multi-threaded applications running on SMP machines under Linux 2.6,
as the single thread test spent a small amount of time in the Linux
kernel, while even the two thread version of the test spent a
significant time in kernel code. We suspect that the large number of
briefly-held latches that LLADD acquires caused this problem. We plan
to investigate this problem further, adopting LLADD to a more advanced
2004-10-23 02:19:01 +00:00
threading package~\cite{capriccio}, or providing a 'SMP Mode' compile time option that
2004-10-22 22:21:40 +00:00
decreases the number of latches that LLADD acquires at the expense of
opportunities for concurrency.
2004-10-22 04:57:25 +00:00
\section{Future Work}
LLADD is an extendible implementation of the ARIES algorithm. This
allows application developers to incorporate transactional recovery
into a wide range of systems. We have a few ideas along these lines,
and also have some ideas for extensions to LLADD itself.
2004-10-22 06:21:51 +00:00
LLADD currently relies upon its buffer manager for page-oriented storage.
2004-10-22 04:57:25 +00:00
Although we did not have space to discuss it in this paper, we have
a blob implementation that stores large data outside of the page file.
2004-10-22 06:21:51 +00:00
This concept could be extended to arbitrary primitives, such as transactional
updates to file system directory trees, or integration of networking
2004-10-22 04:57:25 +00:00
or other operations directly into LLADD transactions. Doing this would
2004-10-22 06:21:51 +00:00
allow LLADD to act as a sort of ``glue code'' among various systems,
2004-10-22 04:57:25 +00:00
ensuring data integrity and adding database-style functionality, such
as continuous backup to systems that currently do not provide such
mechanisms. We believe that there is quite a bit of room for the developement
of new software systems in the space between the high-level, but sometimes
2004-10-22 21:02:10 +00:00
inappropriate interfaces exported by existing transactiona storage systems,
and the unsafe, low-level primitives provided supported by current file systems.
2004-10-22 04:57:25 +00:00
2004-10-22 06:21:51 +00:00
Currently, although we have implemented a two-phase commit algorithm,
2004-10-22 04:57:25 +00:00
LLADD really is not very network aware. If we provided a clean abstraction
that allowed LLADD extensions and operations to cross network boundaries,
then we could provide a wider range of network consistency algorithms,
and cleanly support the implementation of operations that perform
well in networked and in local environments.
2004-10-22 06:21:51 +00:00
Although LLADD is re-entrant, its latching mechanisms only provide physical
consistency. Traditionally, lock managers, which provide higher levels
of consistency, have been tightly coupled with transactional page implementations.
2004-10-22 04:57:25 +00:00
Generally, the semantics of undo and redo operations provided by the
transactional page layer and its associated data structures determine
the level of concurrency that is possible. Since prior systems provide
2004-10-22 06:21:51 +00:00
a monolithic set of primitives to their users, these systems typically had complex interactions among the lock manager, on-disk formats and the transactional
2004-10-22 04:57:25 +00:00
page layer. Finally, at recovery time it is often desirable to reacquire
locks on behalf of a transaction. Without extensible logging and without
modifying the recovery code, it is impossible to ensure that such
locks are correctly restored. By providing extensible logging, data-structures,
and undo/redo semantics, LLADD removes these reasons for coupling
the lock manager and the rest of the storage mechanisms. The flexiblity
offered by splitting the lock manager and the ARIES algorithm into
independent sub-systems, and allowing users to independently extend
either module seems to outweigh the extra complexity that will be
added to LLADD's interface. In particular, most difficulties related
to locking seem to be data-structure dependent, suggesting that, like
page layout or the semantics of various types of log entires, they
2004-10-22 06:21:51 +00:00
are largely orthogonal to the atomicity and durability algorithms
2004-10-22 04:57:25 +00:00
implemented by LLADD.
\section{Conclusion}
2004-10-22 21:20:30 +00:00
We have outlined the design and implementation of a library for the
development of transactional storage systems. By decoupling the
on-disk format from the transactional storage system, we provide
applications with customizable, high-performance, transactional
storage. By summarizing and documenting the interactions between
these customizations and the storage system, we make it easy to
implement such customizations.
Current applications generally must choose between high-level, general
purpose libraries which impose severe performance penalties, and
ad-hoc ``from-scratch'' atomicity and durability mechanisms. By
bridging this gap, we hope to make it easier to implement a class of
applications and algorithms whose implementations are generally
complex, or fail to provide reliable storage to their users.
By releasing LLADD to the community, we hope that we will be able to
provide a toolkit that aids in the development of real-world
applications, and is flexible enough for use as a research platform.
Because of the interface between operation extensions and the
underlying implementation of the ARIES algorithm, we allow operation
extensions and the implementation of the library to evolve
independently, allowing applications to adopt to advanced replication
techniques as the circumstances in which they are deployed changes.
2004-10-22 04:57:25 +00:00
\section{Acknowledgements}
2004-10-23 02:19:01 +00:00
We would like to thank Jason Bayer, Jim Blomo and Jimmy
Kittiyachavalit for their implementation work and contributions to
earlier versions of LLADD. Joe Hellerstein and Mike Franlin provided
us with invaluable advice. Rob von Behren provided us with some last
minute assistance during the benchmarking process.
2004-10-22 04:57:25 +00:00
\section{Availability}
LLADD is free software, available at:
\begin{center}
{\tt http://www.sourceforge.net/projects/lladd}\\
\end{center}
\begin{thebibliography}{99}
2004-10-23 00:42:54 +00:00
\bibitem[Lamb]{lamb} Lamb, et al., {\em The ObjectStore System.} CACM 34(10) (1991) p. 50-63
2004-10-23 02:19:01 +00:00
\bibitem[Lehman]{blink} Lehman \& Yao, {\em Efficient Locking for Concurrent Operations in B-trees.} TODS 6(4) (1981) p. 650-670
\bibitem[Mohan92]{aries} Mohan, et al., {\em ARIES: A Transaction Recovery Method Supporting Fine-Granularity Locking and Partial Rollbacks Using Write-Ahead Logging.} TODS 17(1) (1992) p. 94-162
2004-10-23 00:42:54 +00:00
\bibitem[Mohan86]{twopc} Mohan, Lindsay \& Obermarck, {\em Transaction Management in the R* Distributed Database Management System} TODS 11(4) (1986) p. 378-396
\bibitem[Stonebraker]{newTypes} Stonebraker. {\em Inclusion of New Types in Relational Data Base } ICDE (1986) p. 262-269
\bibitem[Gray]{physiological} Gray, J. and Reuter, A. {\em Transaction Processing: Concepts and Techniques}. Morgan Kaufmann (199) San Mateo, CA
\bibitem[Agrawal]{multipleGenericLocking} Agrawal, et al. {\em Concurrency Control Performance Modeling: Alternatives and Implications}. TODS 12(4): (1987) 609-654
2004-10-22 04:57:25 +00:00
2004-10-23 00:42:54 +00:00
\bibitem[Gray75]{hierarcicalLocking} Jim Gray, Raymond A. Lorie, and Gianfranco R. Putzulo. {\em Granularity of locks and degrees of consistency in a shared database}. In 1st International Conference on VLDB, pages 428--431, September 1975. Reprinted in Readings in Database Systems, 3rd edition.
2004-10-22 04:57:25 +00:00
2004-10-23 00:42:54 +00:00
\bibitem[Gifford]{semantic} David K. Gifford, P. Jouvelot, Mark A. Sheldon, and Jr. James W. O'Toole. {\em Semantic file systems}. Proceedings of the Thirteenth ACM Symposium on Operating Systems Principles, (1991) p. 16-25.
2004-10-22 04:57:25 +00:00
2004-10-23 00:42:54 +00:00
\bibitem[Seltzer]{berkeleyDB} M. Seltzer, M. Olsen. {\em LIBTP: Portable, Modular Transactions for UNIX}. Proceedings of the 1992 Winter Usenix (1992)
2004-10-22 04:57:25 +00:00
2004-10-23 00:42:54 +00:00
\bibitem[Behren]{capriccio} R. von Behren, J Condit, F. Zhou, G. Necula, and E. Brewer. {\em Capriccio: Scalable Threads for Internet Services} SOSP 19 (2003).
2004-10-22 04:57:25 +00:00
2004-10-23 00:42:54 +00:00
\bibitem[Evangelos]{lru2s} Envangelos P. Markatos. {\em On Caching Search Engine Results}. Institute of Computer Science, Foundation for Research \& Technology - Hellas (FORTH) Technical Report 241 (1999)
2004-10-22 04:57:25 +00:00
2004-10-23 02:19:01 +00:00
\bibitem[Mohan]{ariesim} Mohan, Levine. {\em ARIES/IM: an efficient and high concurrency index management method using write-ahead logging} International Converence on Management of Data, SIGMOD (1992) p. 371-380
\bibitem[BDB]{bdb} Berkeley~DB~4.2.52, {\tt http://www.sleepycat.com/}
\bibitem[SLOCCount]{sloccount} SLOCCount, {\tt http://www.dwheeler.com/sloccount/ }
%
\bibitem[lcov]{lcov} The~LTP~gcov~extension, {\tt http://ltp.sourceforge.net/coverage/lcov.php }
%
\bibitem[Reiser]{reiser} Reiser,~Hans~T. {\em ReiserFS 4} {\tt http://www.namesys.com/v4/v4.html } (2004)
%
\bibitem[Haerder]{haerder} Haerder \& Reuter {\em "Principles of Transaction-Oriented Database Recovery." } Computing Surveys 15(4) p 287-317 (1983)
\bibitem[Satyanarayanan]{lrvm} Satyanarayanan, M., Mashburn, H. H., Kumar, P., Steere, D. C., AND Kistler, J. J. {\em Lightweight Recoverable Virtual Memory}. ACM Transactions on Computer Systems 12, 1 (Februrary 1994) p. 33-57. Corrigendum: May 1994, Vol. 12, No. 2, pp. 165-172.
\bibitem[Codd]{relational} E. F. Codd, {\em A Relational Model of Data for Large Shared Data Banks.} CACM 13(6) p. 377-387 (1970)
\bibitem[MySQL]{mysql} {\em MySQL Documentation}, {\tt http://dev.mysql.com/doc }
2004-10-22 04:57:25 +00:00
2004-10-23 02:19:01 +00:00
\bibitem[Litwin]{lht} Litwin, W., {\em Linear Hashing: A New Tool for File and Table Addressing}. Proc. 6th VLDB, Montreal, Canada, (Oct. 1980) p. 212-223
2004-10-23 00:42:54 +00:00
%\bibitem[Beazley]{beazley} D.~M.~Beazley and P.~S.~Lomdahl,
%{\em Message-Passing Multi-Cell Molecular Dynamics on the Connection
%Machine 5}, Parall.~Comp.~ 20 (1994) p. 173-195.
%
%\bibitem[RealName]{CitePetName} A.~N.~Author and A.~N.~Other,
%{\em Title of Riveting Article}, JournalName VolNum (Year) p. Start-End
%
%\bibitem[ET]{embed} Embedded Tk, \\
%{\tt ftp://ftp.vnet.net/pub/users/drh/ET.html}
%
%\bibitem[Expect]{expect} Don Libes, {\em Exploring Expect}, O'Reilly \& Associates, Inc. (1995).
%
%\bibitem[Heidrich]{heidrich} Wolfgang Heidrich and Philipp Slusallek, {\em
%Automatic Generation of Tcl Bindings for C and C++ Libraries.},
%USENIX 3rd Annual Tcl/Tk Workshop (1995).
%
%\bibitem[Ousterhout]{ousterhout} John K. Ousterhout, {\em Tcl and the Tk Toolkit}, Addison-Wesley Publishers (1994).
%
%\bibitem[Perl5]{perl5} Perl5 Programmers reference,\\
%{\tt http://www.metronet.com/perlinfo/doc}, (1996).
%
%\bibitem[Wetherall]{otcl} D. Wetherall, C. J. Lindblad, ``Extending Tcl for
%Dynamic Object-Oriented Programming'', Proceedings of the USENIX 3rd Annual Tcl/Tk Workshop (1995).
2004-10-22 04:57:25 +00:00
\end{thebibliography}
\end{document}