MPS AMC pool class
This document contains a guide to the MPS AMC pool class, followed by the historical initial design. References, History, Copyright and License are at the end.
Guide
Readership: any MPS developer. Not confidential.
Introduction
The AMC pool class is a general-purpose automatic (collecting) pool class. It is intended for most client objects. AMC is "Automatic, Mostly Copying": it preserves objects by copying, except when an ambiguous reference 'nails' the object in place. It is generational. Chain: specify capacity and mortality of generations 0..N-1. Survivors from N-1 get promoted into an arena-wide topGen (often anachronistically called the "dynamic" generation).
Design documentation for AMC is rather incomplete. There is some in the wiki.
AMC Seg States
AMC Segs are in one of three states: "mobile", "boarded", or "stuck" [This is new terminology, not used in the initial design, and not yet widespread in the source code -- RHSK 2009-09-14]. Segs are normally "mobile": all objects on the seg are un-nailed, and thus may be preserved by copying. An ambiguous reference to any address within an AMC seg makes that seg "boarded": a nailboard is allocated to record ambiguous refs ("nails"), but un-nailed objects on the seg are still preserved by copying. Stuck segments only occur in emergency tracing: a discovery fix to an object in a mobile segment is recorded in the only non-allocating way available: by making the entire segment "stuck".
Pads
(See also job001809 and job001811, and mps/branch/2009-03-31/padding)
A pad is logically a trivial client object. Pads are created by the MPS asking the client's format code to create them, to fill up a space in a seg. Thereafter, the pad appears to the MPS as a normal client object (that is: the MPS cannot distinguish a pad from a client object).
AMC creates pads for three reasons: buffer empty, LSP, or NMR. (LSP pads are new with job001811).
Buffer empty pads are made by AMCBufferEmpty whenever it detaches a non-empty buffer from an AMC segment. Buffer detachment is most often caused because the buffer is too small for the current buffer reserve request (which may be either a client requested or a forwarding allocation). Detachment may happen for other reasons, such as trace flip.
LSP pads (Large Segment Padding) are made by AMCBufferFill when the requested fill size is 'large' (see "The LSP payoff calculation" below). AMCBufferFill fills the buffer to exactly the size requested by the current buffer reserve operation; that is: it does not round up to the whole segment size. This prevents subsequent small objects being placed in the same segment as a single very large object. If the buffer fill size is less than the segment size, AMCBufferFill fills any remainder with an LSP pad.
NMR pads (Non-Mobile Reclaim) are made by amcReclaimNailed, when performing reclaim on a non-mobile (that is: either boarded or stuck) segment:
The more common NMR scenario is reclaim of a boarded seg after a non-emergency trace. Ambiguous refs into the seg are recorded as nails. Subsequent exact refs to a nailed object do nothing further, but exact refs that do not match a nail cause preserve-by-copy and leave a forwarding object. Unreachable objects are not touched during the scan+fix part of the trace. On reclaim, only nailed objects need to be preserved; others (namely forwarding pointers and unreachable objects) are replaced by an NMR pad. (Note that a BE or LSP pad appears to be an unreachable object, and is therefore overwritten by an NMR pad).
The less common NMR scenario is after emergency tracing. Boarded segs still occur; they may have nailed objects from ambiguous refs, forwarding objects from pre-emergency exact fixes, nailed objects from mid-emergency exact fixes, and unpreserved objects; reclaim is as in the non-emergency case. Stuck segs may have forwarding objects from pre-emergency exact fixes, objects from mid-emergency fixes, and unreachable objects -- but the latter two are not distinguishable because there is no nailboard. On reclaim, all objects except forwarding pointers are preserved; each forwarding object is replaced by an NMR pad.
If amcReclaimNailed finds no objects to be preserved then it calls SegFree (new with job001809).
Placement Pads are okay
Placement Pads are the BE and LSP pads created in "to-space" when placing objects into segments. This wasted space is an expected space-cost of AMC's naive (but time-efficient) approach to placement of objects into segments. This is normally not a severe problem. (The worst case is a client that always requests ArenaAlign()+1 byte objects: this has a nearly 100% overhead).
Retained Pads could be a problem
Retained Pads are the NMR pads stuck in "from-space": non-mobile segments that were condemned but have preserved-in-place objects cannot be freed by amcReclaimNailed. The space around the preserved objects is filled with NMR pads.
In the worst case, retained pads could waste an enormous amount of space! A small (1 byte) object could retain a multi-page segment for as long as the ambiguous reference persists; that is: indefinitely. Imagine a 256-page (1 MiB) segment containing a very large object followed by a handful of small objects. An ambiguous ref to one of the small objects will unfortunately cause the entire 256-page seg to be retained, mostly as an NMR pad; this is a massive overhead of wasted space.
AMC mitigates this worst-case behaviour, by treating large segments specially.
Small, Medium, and Large Segments
AMC categorises segments as "small" (1 page), "medium" (several pages), or "large" (>= AMCLargeSegPAGES pages):
pages = SegSize(seg) / ArenaAlign(arena); if(pages == 1) { /* small */ } else if(pages < AMCLargeSegPAGES) { /* medium */ } else { /* large */ }
AMCLargeSegPAGES is currently 8 -- see "The LSP payoff calculation" below.
AMC might treat "Large" segments specially, in two ways:
- .large.single-reserve: A large segment is only used for a single (large) buffer reserve request; the remainder of the segment (if any) is immediately padded with an LSP pad.
- .large.lsp-no-retain: Nails to such an LSP pad do not cause AMCReclaimNailed to retain the segment.
.large.single-reserve is implemented. See job001811.
.large.lsp-no-retain is NOT currently implemented.
The point of .large.lsp-no-retain would be to avoid retention of the (large) segment when there is a spurious ambiguous reference to the LSP pad at the end of the segment. Such an ambiguous reference might happen naturally and repeatably if the preceding large object is an array, the array is accessed by an ambiguous element pointer (eg. on the stack), and the element pointer ends up pointing just off the end of the large object (as is normal for sequential element access in C) and remains with that value for a while. (Such an ambiguous reference could also occur by chance, eg. by coincidence with an int or float, or when the stack grows to include old unerased values).
Implementing .large.lsp-no-retain is a little tricky. A pad is indistinguishable from a client object, so AMC has no direct way to detect, and safely ignore, the final LSP object in the seg. If AMC could guarantee that the single buffer reserve (.large.single-reserve) is only used for a single object, then AMCReclaimNailed could honour a nail at the start of a large seg and ignore all others; this would be extremely simple to implement. But AMC cannot guarantee this, because in the MPS Allocation Point Protocol the client is permitted to make a large buffer reserve and then fill it with many small objects. In such a case, AMC must honour all nails (if the buffer reserve request was an exact multiple of ArenaAlign), or all nails except to the last object (if there was a remainder filled with an LSP pad). Because an LSP pad cannot be distinguished from a client object, and the requested allocation size is not recorded, AMC cannot distinguish these two conditions at reclaim time. Therefore AMC must record whether or not the last object in the seg is a pad, in order to ignore nails to it. This could be done by adding a flag to AMCSegStruct. (This can be done without increasing the structure size, by making the "Bool new;" field smaller than its current 32 bits.)
The LSP payoff calculation
The job001811 LSP fix treats large segments differently. Without it, after allocating a very large object (in a new very large multi-page segment), MPS would happily place subsequent small objects in any remaining space at the end of the segment. This would risk pathological fragmentation: if these small objects were systematically preserved by ambiguous refs, enormous NMR pads would be retained along with them.
The payoff calculation is a bit like deciding whether or not to purchase insurance. For single-page and medium-sized segments, we go ahead and use the remaining space for subsequent small objects. This is equivalent to choosing not to purchase insurance. If the small objects were to be preserved by ambiguous refs, the retained NMR pads would be big, but not massive. We expect such ambiguous refs to be uncommon, so we choose to live with this slight risk of bad fragmentation. The benefit is that the remaining space is used.
For large segments, we decide that the risk of using the remainder is just too great, and the benefit too small, so we throw it away as an LSP pad. This is equivalent to purchasing insurance: we choose to pay a known small cost every time, to avoid risking an occasional disaster.
To decide what size of segment counts as "large", we must decide how much uninsured risk we can tolerate, versus how much insurance cost we can tolerate. The likelihood of ambiguous references retaining objects is entirely dependent on client behaviour. However, as a sufficient 'one size fits all' policy, I (RHSK 2009-09-14) have judged that segments smaller than 8 pages long do not need to be treated as large: the insurance cost to 'play safe' would be considerable (wasting up to 1 page of remainder per 7 pages of allocation), and the fragmentation overhead risk is not that great (at most 8 times worse than the unavoidable minimum). So AMCLargeSegPAGES is defined as 8 in config.h. As long as the assumption that most segments are not ambiguously referenced remains correct, I expect this policy will be satisfactory.
To verify that this threshold is acceptable for a given client, poolamc.c calculates metrics; see "Feedback about retained pages" below. If this one-size-fits-all approach is not satisfactory, AMCLargeSegPAGES could be made a client-tunable parameter.
Retained pages
The reasons why a segment and its pages might be retained are:
- ambiguous ref to first-obj: unavoidable page retention (only the mutator can reduce this, if they so wish, by nulling out ambig refs);
- ambiguous ref to rest-obj: tuning MPS LSP policy could mitigate this, reducing the likelihood of rest-objs being co-located with large first-objs;
- ambiguous ref to final pad: implementing .large.lsp-no-retain could mitigate this;
- ambiguous ref to other (NMR) pad: hard to mitigate, as pads are indistinguishable from client objects;
- emergency trace;
- non-object-aligned ambiguous ref: fixed by job001809;
- other reason (eg. buffered at flip): not expected to be a problem.
(This list puts the reasons that are more 'obvious' to the client programmer first, and the more obscure reasons last).
Feedback about retained pages
(New with job001811). AMC now accumulates counts of pages condemned and retained during a trace, in categories according to size and reason for retention, and emits diagnostic at trace-end via the pool->class->traceEnd method. See comments on the PageRetStruct in poolamc.c. These page-based metrics are not as precise as actually counting the size of objects, but they require much less intrusive code to implement, and should be sufficient to assess whether AMC's page retention policies and behaviour are acceptable.
Initial Design
THE DESIGN OF THE AUTOMATIC MOSTLY-COPYING MEMORY POOL CLASS design.mps.poolamc incomplete design richard 1995-08-25 INTRODUCTION .intro: This is the design of the AMC Pool Class. AMC stands for Automatic Mostly-Copying. This design is highly fragmentory and some may even be sufficiently old to be misleading. .readership: The intended readership is any MPS developer. OVERVIEW .overview: This class is intended to be the main pool class used by Harlequin Dylan. It provides garbage collection of objects (hence "automatic"). It uses generational copying algorithms, but with some facility for handling small numbers of ambiguous references. Ambiguous references prevent the pool from copying objects (hence "mostly copying"). It provides incremental collection. [ lot of this design is awesomely old -- drj 1998-02-04] DEFINITIONS .def.grain: Grain. An quantity of memory which is both aligned to the pool's alignment and equal to the pool's alignment in size. IE the smallest amount of memory worth talking about. DESIGN Segments .seg.class: AMC allocates segments of class AMCSegClass, which is a subclass of GCSegClass. Instances contain a segTypeP field, which is of type int*. .seg.gen : AMC organizes the segments it manages into generations. .seg.gen.map: Every segment is in exactly one generation. .seg.gen.ind: The segment's segTypeP field indicates which generation (that the segment is in) (an AMCGenStruct see blah below). .seg.typep: The segTypeP field actually points to either the type field of a generation or to the type field of a nail board. .seg.typep.distinguish: The type field (which can be accessed in either case) determines whether the segTypeP field is pointing to a generation or to a nail board. .seg.gen.get: The map from segment to generation is implemented by AMCSegGen which deals with all this. Fixing and Nailing [.fix.nail.* are placeholders for design rather than design really-- drj 1998-02-04] .fix.nail: .nailboard: AMC uses a nail board structure for recording ambiguous references to segments. A nail board is a bit table with one bit per grain in the segment. .nailboard.create: Nail boards are allocated dynamically whenever a segment becomes newly ambiguously referenced. .nailboard.destroy: They are deallocated during reclaim. Ambiguous fixes simply set the appropriate bit in this table. This table is used by subsequent scans and reclaims in order to work out what objects were marked. .nailboard.emergency: During emergency tracing two things relating to nail boards happen that don't normally: .nailboard.emergency.nonew: Nail boards aren't allocated when we have new ambiguous references to segments (.nailbaord.emergency.nonew.justify: We could try and allocate a nail board, but we're in emergency mode so short of memory so it's unlikely to succeed, and there would be additional code for yet another error path which complicates things); .nailboard.emergency.exact: nail boards are used to record exact references in order to avoid copying the objects. .nailboard.hyper-c onservative: Not creating new nail boards (.nailboard.emergency.nonew above) means that when we have a new reference to a segment during emergency tracing then we nail the entire segment and preserve everything in place. .fix.nail.states: Partition the segment states into 4 sets: white segment and not nailed (and has no nail board) white segment and nailed and has no nail board white segment and nailed and has nail board the rest .fix.nail.why: A segment is recorded as being nailed when either there is an ambiguous reference to it, or there is an exact reference to it and the object couldn't be copied off the segment (because there wasn't enough memory to allocate the copy). In either of these cases reclaim cannot simply destroy the segment (usually the segment will not be destroyed because it will have live objects on it, though see .nailboard.limitations.middle below). If the segment is nailed then we might be using a nail board to mark objects on the segment. However, we cannot guarantee that being nailed implies a nail board, because we might not be able to allocate the nail board. Hence all these states actually occur in practice. .fix.nail.distinguish: The nailed bits in the segment descriptor (SegStruct) are used to record whether a segment is nailed or not. The segTypeP field of the segment either points to (the "type" field of) an AMCGen or to an AMCNailBoard, the type field can be used to determine which of these is the case. (see .seg.typep above). .nailboard.limitations.single: Just having a single nail board per segment prevents traces from improving on the findings of each other: a later trace could find that a nailed object is no longer nailed or even dead. Until the nail board is discarded, that is. .nailboard.limitations.middle: An ambiguous reference into the middle of an object will cause the segment to survive, even if there are no surviving objects on it. .nailboard.limitations.reclaim: AMCReclaimNailed could cover each block of reclaimed objects between two nailed objects with a single padding object, speeding up further scans. Emergency Tracing .emergency.fix: AMCFixEmergency is at the core of AMC's emergency tracing policy (unsurprisingly). AMCFixEmergency chooses exactly one of three options: a) use the existing nail board structure to record the fix, b) preserve and nail the segment in its entirety, c) snapout an exact (or high rank) pointer to a broken heart to the broken heart's forwarding pointer. If the rank of the reference is AMBIG then it either does a) or b) depending on wether there is an existing nail board or not. Otherwise (the rank is exact or higher) if there is a broken heart it is used to snapout the pointer. Otherwise it is as for an AMBIG ref (we either do a) or b)). .emergency.scan: This is basically as before, the only complication is that when scanning a nailed segment we may need to do multiple passes, as FixEmergency may introduce new marks into the nail board. Buffers .buffer.class: AMC uses buffer of class AMCBufClass (a subclass of SegBufClass) .buffer.gen: Each buffer allocates into exactly one generation. .buffer.field.gen: AMCBuf buffer contain a gen field which points to the generation that the buffer allocates into. .buffer.fill.gen: AMCBufferFill uses the generation (obtained from the gen field) to initialise the segment's segTypeP field which is how segments get allocated in that generation. .buffer.condemn: We condemn buffered segments, but not the contents of the buffers themselves, because we can't reclaim uncommitted buffers (see design.mps.buffer for details). If the segment has a forwarding buffer on it, we detach it [why? @@@@ forwarding buffers are detached because they used to cause objects on the same segment to not get condemned, hence caused retention of garbage. Now that we condemn the non-buffered portion of buffered segments this is probably unnecessary -- drj 1998-06-01 But it's probably more efficient than keeping the buffer on the segment, because then the other stuff gets nailed -- pekka 1998-07-10]. If the segment has a mutator buffer on it, we nail the buffer. If the buffer cannot be nailed, we give up condemning, since nailing the whole segment would make it survive anyway. The scan methods skip over buffers and fix methods don't do anything to things that have already been nailed, so the buffer is effectively black. AMCStruct .struct: AMCStruct is the pool class AMC instance structure. .struct.pool: Like other pool class instances, it contains a PoolStruct containing the generic pool fields. .struct.format: The "format" field points to a Format structure describing the object format of objects allocated in the pool. The field is intialized by AMCInit from a parameter, and thereafter it is not changed until the pool is destroyed. [actually the format field is in the generic PoolStruct these days. drj 1998-09-21] [lots more fields here] Generations .gen: Generations partition the segments that a pool manages (see .seg.gen.map above). .gen.collect: Generations are more or less the units of condemnation in AMC. And also the granularity for forwarding (when copying objects during a collection): all the objects which are copied out of a generation use the same forwarding buffer for allocating the new copies, and a forwarding buffer results in allocation in exactly one generation. .gen.rep: Generations are represented using an AMCGenStruct structure. .gen.create: All the generation are create when the pool is created (during AMCInitComm). .gen.manage.ring: An AMC's generations are kept on a ring attached to the AMCStruct (the genRing field). .gen.manage.array: They are also kept in an array which is allocated when the pool is created and attached to the AMCStruct (the gens field holds the number of generations, the gen field points to an array of AMCGen). [it seems to me that we could probably get rid of the ring -- drj 1998-09-22] .gen.number: There are AMCTopGen + 2 generations in total. "normal" generations numbered from 0 to AMCTopGen inclusive and an extra "ramp" generation (see .gen.ramp below). .gen.forward: Each generation has an associated forwarding buffer (stored in the "forward" field of AMCGen). This is the buffer that is used to forward objects out of this generation. When a generation is created in AMCGenCreate, its forwarding buffer has a NULL p field, indicating that the forwarding buffer has no generation to allocate in. The collector will assert out (in AMCBufferFill where it checks that buffer->p is an AMCGen) if you try to forward an object out of such a generation. .gen.forward.setup: All the generation's forwarding buffer's are associated with generations when the pool is created (just after the generations are created in AMCInitComm). Ramps .ramp: Ramps usefully implement the begin/end mps_alloc_pattern_ramp interface. .gen.ramp: To implement ramping (request.dylan.170423), AMC uses a special "ramping mode", where promotions are redirected. One generation is designated the "ramp generation" (amc->rampGen in the code). .gen.ramp.ordinary: Ordinarily, that is whilst not ramping, objects are promoted into the ramp generation from younger generations and are promoted out to older generations. The generation that the ramp generation ordinarily promotes into is designated the "after-ramp generation" (amc->afterRampGen). .gen.ramp.particular: As of 2002-02-06 the ramp generation is the second oldest generation and the after-ramp generation is the oldest generation. .gen.ramp.possible: (Old design notes) In alternative designs it might be possible to make the ramp generation a special generation that is only promoted into during ramping (AMCRampGenFollows), however, this is not done. .gen.ramp.ramping: The ramp generation is promoted into itself during ramping mode; .gen.ramp.after: after this mode ends, the ramp generation is promoted into the after-ramp generation as usual. .gen.ramp.after.once: Care is taken to ensure that there is at least one collection where stuff is promoted from the ramp generation to the after-ramp generation even if ramping mode is immediately re-entered. .ramp.mode: This behaviour is controlled in a slightly convoluted manner by a state machine. The rampMode field of the pool forms an important part of the state of the machine. There are five states: outside, begin, ramping, finish, collecting These appear in the code as RampOUTSIDE, and so on. .ramp.state.cycle.usual: The usual progression of states is a cycle: outside -> begin -> ramping -> finish -> collecting -> outside. .ramp.count: The pool just counts the number of ap's that have begun ramp mode (and not ended). No state changes occur unless this count goes from 0 to 1 (entering ramping) or from 1 to 0 (leaving ramping). In other words, all nested ramps are ignored (see code in AMCRampBegin and AMCRampEnd). .ramp.state.invariants: In states collecting and outside the count must be 0. In states begin and ramping the count must be > 0. In state finish the count is not constrained (but it must be >= 0 anyway). .ramp.begin: When the count goes up from zero, the states moves from collecting or outside to begin; if the state is finish then no transition takes place (it cannot be in state begin or ramping). .ramp.begin.leave: We can leave the begin state to either the state outside or the state ramping. .ramp.begin.leave.outside: We go to outside if the count drops to 0 before a collection starts. This shortcuts the usual cycle of states for small enough ramps. .ramp.begin.leave.ramping: We enter the ramping state if a collection starts that condemns the ramp generation (pedantically when a new GC begins, and a segment in the ramp generation is condemned we leave the begin state, see AMCWhiten). At this point we switch the forwarding generations to forward in the ramping way (.gen.ramp); the pool enters ramping mode. (This assumes that each generation is condemned together with all lower generations.) .ramp.end: After the ramp count goes back to zero, the pool enters the finish state if it was in the ramping state, or shortcuts the usual cycle by going to the outside state if it was in the begin state. If it was already in the finish state then it stay there. It can't be in any other state. .ramp.finish: We stay in the finish state until we do a GC that condemns the ramp generation, and we ensure that such a GC will promote from the ramp generation to the after-ramp generation. In fact, the next new GC should collect the ramp generation because we jig the benefits to ensure that [is this true? drj 2002-03-06]. .ramp.finish.leave: We leave this state when the GC starts the condemn phase and we enter collecting state. At this point the forwarding generations are also switched so that the ramp generation promotes into the after-ramp generation. .ramp.collecting.leave: We leave the collecting state when the GC enters reclaim (specifically, when a segment in the ramp generation is reclaimed). Ordinarily we enter the outside state, but if the client has increased the ramp count then we go directly to the begin state. The forwarding generations are not switched until we leave this state, ensuring that we perform a whole collection with the ramp generation promoting to the after-ramp generation. .ramp.collect-all There are two flavours of ramp collections: the normal one that collects the ramp generation and the younger ones, and the collect-all flavour that does a full GC (this is a hack for producing certain Dylan statistics). The collection will be of collect-all flavour, if any of the RampBegins during the corresponding rank asked for that. Ramp beginnings and collections are asynchronous, so we need two fields to implement this behaviour: collectAll to indicate whether the ramp collection that is about to start should be collect-all, and collectAllNext to keep track of whether the current ramp has any requests for it. Headers .header: AMC supports a fixed-size header on objects, with the client pointers pointing after the header, rather than the base of the memory block. See format documentation for details of the interface. .header.client: The code mostly deals in client pointers, only computing the base and limit of a block when these are needed (such as when an object is copied). In several places, the code gets a block of some sort, a segment or a buffer, and creates a client pointer by adding the header length (pool->format->headerLength). .header.fix: There are two versions of the fix method, due to its criticality, with (AMCHeaderFix) and without (AMCFix) headers. The correct one is selected in AMCInitComm, and placed in the pool's fix field. This is the main reason why fix methods dispatch through the instance, rather than the class like all other methods. OLD AND AGING NOTES BELOW HERE: AMCFinish .finish: .finish.forward: 103 /* If the pool is being destroyed it is OK to destroy */ 104 /* the forwarding buffers, as the condemned set is about */ 105 /* to disappear. */ AMCBufferEmpty .flush: Removes the connexion between a buffer and a group, so that the group is no longer buffered, and the buffer is reset and will cause a refill when next used. .flush.pad: The group is padded out with a dummy object so that it appears full. .flush.expose: The buffer needs exposing before writing the padding object onto it. If the buffer is being used for forwarding it might already be exposed, in this case the segment attached to it must be covered when it leaves the buffer. See .fill.expose. .flush.cover: The buffer needs covering whether it was being used for forwarding or not. See .flush.expose. AMCBufferFill .fill: 185 * Reserve was called on an allocation buffer which was reset, 186 * or there wasn't enough room left in the buffer. Allocate a group 187 * for the new object and attach it to the buffer. 188 * .fill.expose: 189 * .fill.expose: If the buffer is being used for forwarding it may 190 * be exposed, in which case the group attached to it should be 191 * exposed. See .flush.cover. AMCBufferTrip .trip: 239 * A flip occurred between a reserve and commit on a buffer, and 240 * the buffer was "tripped" (limit set to zero). The object wasn't 241 * scanned, and must therefore be assumed to be invalid, so the 242 * reservation must be rolled back. This function detaches the 243 * buffer from the group completely. The next allocation in the 244 * buffer will cause a refill, and reach AMCFill. AMCBufferFinish .buffer-finish: 264 * Called from BufferDestroy, this function detaches the buffer 265 * from the group it's attached to, if any. AMCFix .fix: 281 * fix an ambiguous reference to the pool 282 * 283 * Ambiguous references lock down an entire segment by removing it 284 * from old-space and also marking it grey for future scanning. 285 * 286 * fix an exact, final, or weak reference to the pool 287 * 288 * These cases are merged because the action for an already 289 * forwarded object is the same in each case. After that 290 * situation is checked for, the code diverges. 291 * 292 * Weak references are either snapped out or replaced with 293 * ss->weakSplat as appropriate. 294 * 295 * Exact and final references cause the referenced object to be copied t o 296 * new-space and the old copy to be forwarded (broken-heart installed) 297 * so that future references are fixed up to point at the new copy. 298 * 299 * .fix.exact.expose: In order to allocate the new copy the 300 * forwarding buffer must be exposed. This might be done more 301 * efficiently outside the entire scan, since it's likely to happen 302 * a lot. 303 * 304 * .fix.exact.grey: The new copy must be at least as grey as the old one, 305 * as it may have been grey for some other collection. AMCGrey .grey: 453 * Turns everything in the pool which is not condemned for a trace 454 * grey. AMCSegScan .seg-scan: 485 * .seg-scan.blacken: One a group is scanned it is turned black, i.e. 486 * the ti is removed from the grey TraceSet. However, if the 487 * forwarding buffer is still pointing at the group it could 488 * make it grey again when something is fixed, and cause the 489 * group to be scanned again. We can't tolerate this at present, 490 * the the buffer is flushed. The solution might be to scan buffers 491 * explicitly. .seg-scan.loop: 505 /* While the group remains buffered, scan to the limit of */ 506 /* initialized objects in the buffer. Either it will be reached, */ 507 /* or more objects will appear until the segment fills up and the */ 508 /* buffer moves away. */ .seg-scan.finish: 520 /* If the group is unbuffered, or becomes so during scanning */ 521 /* (e.g. if the forwarding buffer gets flushed) then scan to */ 522 /* the limit of the segment. */ .seg-scan.lower: 540 /* The segment is no longer grey for this collection, so */ 541 /* it no longer needs to be shielded. */ AMCScan .scan: 556 * Searches for a group which is grey for the trace and scans it. 557 * If there aren't any, it sets the finished flag to true. AMCReclaim .reclaim: 603 * After a trace, destroy any groups which are still condemned for the 604 * trace, because they must be dead. 605 * 606 * .reclaim.grey: Note that this might delete things which are grey 607 * for other collections. This is OK, because we have conclusively 608 * proved that they are dead -- the other collection must have 609 * assumed they were alive. There might be a problem with the 610 * accounting of grey groups, however. 611 * 612 * .reclaim.buf: If a condemned group still has a buffer attached, we 613 * can't destroy it, even though we know that there are no live objects 614 * there. Even the object the mutator is allocating is dead, because 615 * the buffer is tripped. AMCAccess .access: 648 * This is effectively the read-barrier fault handler. 649 * 650 * .access.buffer: If the page accessed had and still has the 651 * forwarding buffer attached, then trip it. The group will now 652 * be black, and the mutator needs to access it. The forwarding 653 * buffer will be moved onto a fresh grey page. 654 * 655 * .access.error: @@@@ There really ought to be some error recovery. 656 * 657 * .access.multi: @@@@ It shouldn't be necessary to scan more than 658 * once. Instead, should use a multiple-fix thingy. This would 659 * require the ScanState to carry a _set_ of traces rather than 660 * just one. OLD NOTES Group Scanning
A. References
B. Document History
2002-06-07 | RB | Converted from MMInfo database design document. |
2009-08-11 | RHSK | Fix HTML duplicated anchor names (caused by auto-conversion to HTML). |
2009-08-11 | RHSK | Prepend Guide, using design/template-with-guide.html. |
2009-09-14 | RHSK | Guide covers: seg states; pads; retained pages. |
C. Copyright and License
This document is copyright © 1995-2002, 2009 Ravenbrook Limited. All rights reserved. This is an open source license. Contact Ravenbrook for commercial licensing options.
Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met:
- Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer.
- Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution.
- Redistributions in any form must be accompanied by information on how to obtain complete source code for the this software and any accompanying software that uses this software. The source code must either be included in the distribution or be available for no more than the cost of distribution plus a nominal fee, and must be freely redistributable under reasonable conditions. For an executable file, complete source code means the source code for all modules it contains. It does not include source code for modules or files that typically accompany the major components of the operating system on which the executable file runs.
This software is provided by the copyright holders and contributors "as is" and any express or implied warranties, including, but not limited to, the implied warranties of merchantability, fitness for a particular purpose, or non-infringement, are disclaimed. In no event shall the copyright holders and contributors be liable for any direct, indirect, incidental, special, exemplary, or consequential damages (including, but not limited to, procurement of substitute goods or services; loss of use, data, or profits; or business interruption) however caused and on any theory of liability, whether in contract, strict liability, or tort (including negligence or otherwise) arising in any way out of the use of this software, even if advised of the possibility of such damage.