Обсуждение: Re: [HACKERS] Mariposa

Поиск
Список
Период
Сортировка

Re: [HACKERS] Mariposa

От
Bob Devine
Дата:
Ross J. Reedstrom wrote:

> Right. As I've been able to make out so far, in Mariposa a query passes
> through the regular parser and single-site optimizer, then the selected
> plan tree is handed to a 'fragmenter' to break the work up into chunks,
> which are then handed around to a 'broker' which uses a microeconomic
> 'bid' process to parcels them out to both local and remote executors. The
> results from each site then go through a local 'coordinator' which merges
> the result sets, and hands them back to the original client.
> 
> Whew!
> 
> It's interesting to compare the theory describing the workings of Mariposa
> (such as the paper in VLDB), and the code. For the fragmenter, the paper
> describes basically a rational decomposition of the plan, while the code
> applies non-deterministic, but tuneable, methods (lots of calls to random
> and comparisions to user specified odds ratios).
> 
> It strikes me as a bit odd to optimize the plan for a single site,
> then break it all apart again. My thoughts on this are to implement
> a two new node types: one a remote table, and one which represents
> access to a remote table. Remote tables have host info in them, and
> always be added to the plan with a remote-access node directly above
> them. Remote-access nodes would be seperate from their remote-table,
> to allow the communications cost to be slid up the plan tree, and merged
> with other remote-access nodes talking to the same server. This should
> maintain the order-agnostic nature of the optimizer. The executor will
> need to build SQL statements and from the sub-plans and submit them via
> standard network db access client librarys.
> 
> First step, create a remote-table node, and teach the excutor how to get
> info from it. Later, add the seperable remote-access node.
> 
> How insane does this sound now? Am I still a mad scientist? (...always!)

Let me give a brief "what, where, and why" about Mariposa.

The impetus for Mariposa was that every distributed database
idea either died or failed to scale up beyond a handful of nodes.
There was Ingres/Star, IBM's important R* project, DEC's failed
RdbStar project, and others.  The idea of grouping together a
bunch of servers is a seductive, but very hard, one.

What exists today is a group of simple extensions (basically they
are: treating remote tables differently, use replication instead
of distributed consistency, and pushing the problem up to the
programmer's level).  Not too transparent or even seamless.

Mariposa proposed that tables can dynamically move.  And fragments
of tables.  And queries too!  This helps a lot towards meeting
the big problem of load balancing and configuring any huge system.
Compare this with web servers -- web sites can become overloaded
plus everyone knows the permanent location of the data on the
server (there is some tricky footwork behind the scenes that allow
multiple servers to share the load).

Soooo, how to optimize a query where _everything_ can move?
That is the basis for the "optimize for a single site, later
split it for distributed execution" idea.  The loss of data
mobility would have been more painful than a more complicated
optimizer (at least in theory! ;-)  So it does a bit insane but
the alternative would have been to assume all tables are remote
but this collapses to the same idea as "all local" tables that
just have higher access costs.

Fast forward to 1999, Mariposa is now being commercialized by
Cohera (www.cohera.com) as a middleware distribution layer.

--
Bob Devine  devine@cs.utah.edu

(PS: Just for background, I proposed a lot of the Mariposa ideas
way back in 1992 at Berkeley after working on DEC's RdbStar.
Now working on Impulse - www.cs.utah.edu/projects/impulse)




Re: [HACKERS] Mariposa

От
"Ross J. Reedstrom"
Дата:
On Tue, Aug 03, 1999 at 11:34:47AM -0600, Bob Devine wrote:
> Let me give a brief "what, where, and why" about Mariposa.
> 
<SNIPPED Very interesting precise>

> 
> Fast forward to 1999, Mariposa is now being commercialized by
> Cohera (www.cohera.com) as a middleware distribution layer.

Right - at $40000 a crack for a two backend db version :-(

> 
> --
> Bob Devine  devine@cs.utah.edu
> 
> (PS: Just for background, I proposed a lot of the Mariposa ideas
> way back in 1992 at Berkeley after working on DEC's RdbStar.
> Now working on Impulse - www.cs.utah.edu/projects/impulse)

Bob - 

Thanks for the background. As you say, Mariposa is coming at this problem
from a different angle, and it wasn't clear to me _why_ that was. Now it
is. I think it's still be good to have the small, add-on version of remote
tables that I proposed. Folding the rest of Mariposa in later might be an
interesting project, as well, and could co-exist (or be built on top of)
fix-location remote tables. Perhaps I should see if it'd be possible to
use the Mariposa syntax extensions with fixed remote tables. That way,
distributed dbs could be made mobile seamlessly (Ha!)

Ross

-- 
Ross J. Reedstrom, Ph.D., <reedstrm@rice.edu> 
NSBRI Research Scientist/Programmer
Computer and Information Technology Institute
Rice University, 6100 S. Main St.,  Houston, TX 77005


Re: [HACKERS] Mariposa

От
Adriaan Joubert
Дата:
I've been reading all this with interest, even though I know nothing
about distributed database design. I've used Tandem's a bit though, and
they do a rather good job of parallelising queries. A key part of
building an efficient database system on a Tandem is figuring out how
the database is distributed over the disks, which used to correspond (on
the K10000 anyway) to processors. This partitioning is explicitly
declared. I believe the query optimizer used this information to figure
out where it had to go for data. If yor partitioning was wrong,
performance would be dismal, if it was right  -- pheew, it would fly.
Bit more onus on the dba, or application developer, but, having worked
on lots of parallel applications, it is my experience that a completely
automatic solution is never terribly good. Distributing work/data
optimally is just too complex a problem to automate.

Adriaan