|Publication number||US7310718 B1|
|Application number||US 09/856,779|
|Publication date||Dec 18, 2007|
|Filing date||Nov 24, 1999|
|Priority date||Nov 25, 1998|
|Also published as||EP1135727A1, US7506317, US20040158589, WO2000033192A1|
|Publication number||09856779, 856779, PCT/1999/28089, PCT/US/1999/028089, PCT/US/1999/28089, PCT/US/99/028089, PCT/US/99/28089, PCT/US1999/028089, PCT/US1999/28089, PCT/US1999028089, PCT/US199928089, PCT/US99/028089, PCT/US99/28089, PCT/US99028089, PCT/US9928089, US 7310718 B1, US 7310718B1, US-B1-7310718, US7310718 B1, US7310718B1|
|Inventors||Sheng Liang, Steffen Grarup|
|Original Assignee||Sun Microsystems, Inc.|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (7), Non-Patent Citations (7), Referenced by (12), Classifications (11), Legal Events (5)|
|External Links: USPTO, USPTO Assignment, Espacenet|
This application incorporates-by-reference the computer program listing submitted on two compact discs, entitled “Copy 1” and “Copy 2.” Each compact disc contains one 7,952-byte file, entitled “Compact Disc File.txt,” created on Jun. 14, 2007.
The present invention relates to computer software, and more specifically to enabling comprehensive profiling of garbage-collected memory systems.
Despite the rapid growth in memory sizes of computers, storage in which data is placed in a computer requires careful conservation and recycling. Some programming languages support static allocation of storage space while others support dynamic allocation. Static allocation connotes that storage space is determined at compilation of the computer program. In the case of dynamic allocation, storage space is determined during execution of the programs, known as “run-time”. The area of dynamic memory reserved for storage space that is determined at run-time is a referred to as a “heap”.
Many programming languages place the responsibility with the programmer to explicitly allocate storage and explicitly dispose or free that storage when it is no longer needed. Thus, when the programmer has the responsibility of the allocating and recycling or reclaiming of storage in dynamic memory, the memory management is said to be manually performed or “explicit”.
An individually allocated piece of memory in the heap may be referred to herein as an “object”. The goal is to dispose of unused or dead objects by freeing such objects and returning the space associated with such objects to the “free list” or “free pool”. The “free list” or “free pool” is the set of unallocated usable spaces in the heap. However, with explicit allocation and deallocation, some objects cannot be deallocated or used because they have become “unreachable”. The space occupied by an unreachable object is said to have “leaked” away. In addition, explicit deallocation may result in “dangling references”. To illustrate, consider the following.
Data that is dynamically allocated in memory is usually accessible by the user program through the roots of the computation of the user program, or by following chains of pointers (“references”) from these roots. The roots of the computation of a user program are those locations that hold values that a user program can manipulate directly. Processor registers and the program stack are examples of roots of the computation. Thus, an object in the heap is said to be “live” if its address is held in a root, or there is a reference to it held in another live object in the heap. Otherwise, the object is said to be “dead”. An object becomes unreachable when the last reference to the object is destroyed before the object is deallocated. The object is no longer live if there are no references to it. In addition, such an object cannot be freed (deallocated) because it is unreachable, as there are no references leading to it. Objects that are neither live nor free cannot be used or recycled by explicit allocation and deallocation, and are called “garbage”. Thus, unreachable dead objects are “garbage”.
“Dangling references” occur when an object in the heap is deallocated while there are still references to the object. Dangling references can produce incorrect results. For example, if the memory location of the object that was deallocated while there were still references to it is reallocated to another of the user program's data structures, then the same memory location would represent two different objects.
Automatic dynamic memory management can obviate the problem of unreachable objects and dangling references. In one approach, automatic dynamic memory management involves assigning the responsibility for dynamic memory management to the program's run-time system. The programmer must still request dynamically allocated storage to be reserved but the programmer no longer needs to recycle the storage because the recycling is done automatically through a process called garbage collection. Garbage collection can recover unreachable objects by returning the space associated with such objects to the free list or free pool. In addition, garbage collection avoids dangling references by not deallocating any object while there remains references to it from other live objects.
Furthermore, garbage collection may be useful in correcting the problem of “fragmentation”. Fragmentation occurs in the course of allocating objects of different sizes in the heap. There are various strategies for choosing the optimum space in the heap to allocate to an object. However, such strategies usually result in areas or “fragments” of free pool in the heap. Sometimes the fragments are too small to be useful and are not contiguous and therefore, cannot be combined. For example,
Garbage collection is one of many features associated with profiling a given user program such as an application program. “Profiling” describes the monitoring and tracing of events that occur during run-time. For example, a profiling tool may provide information on how much or how frequently dynamic memory is allocated by each portion of the executing application program. As will be explained in greater detail, some of the dynamic memory allocation information can be extracted from the object allocation process and the garbage collection process.
In one approach, profiling tools are specifically customized and instrumented for a specific platform. Typically, tool vendors specialize in providing tools that are only suitable for certain platforms. For example, different virtual machine vendors provide myriad Java™ virtual machine implementations. Java™ is an object-oriented language, the source code files of which are compiled into a format called bytecode. Bytecode can be executed on any machine on which Java run-time system software is installed. The Java™ run-time system is known as a Java™ virtual machine, a self-contained operating environment that behaves as if it is a separate computer. The Java™ virtual machine translates the bytecode's set of instructions into the underlying computer processor's set of instructions. Thus, a Java™ application will execute in any Java™ virtual machine regardless of the hardware and software underlying the system. It would be highly desirable if users were given the choice of using any profiling tool that is available in the market without worrying if the tool is compatible with the user's particular implementation of Java™ virtual machine.
Based on the foregoing, it is desirable to develop a mechanism that supports a wide variety of profiling tools for garbage collected memory systems.
A method and system are provided for allowing a profiler to communicate with a virtual machine without regard to the specific implementation of the virtual machine. A wide variety of profilers can be accommodated by using a set of virtual machine profiler interface events that are designed to be independent of any method for dynamically managing storage allocation and deallocation in a heap within the virtual machine process. The profiler may request specific information with respect to storage allocation and or deallocation in the heap. In response, the virtual machine can send the requested information using the virtual machine profiler interface events.
In other aspects, the invention encompasses a computer apparatus, a computer-readable medium, and a carrier wave configured to carry out the foregoing techniques.
Many other aspects and features will become apparent from the following description.
The present invention is illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:
Techniques for enabling comprehensive profiling of Java™ virtual machines are described. In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be apparent, however, to one skilled in the art that the present invention may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form in order to avoid unnecessarily obscuring the present invention.
Although certain embodiments have been described using the Java™ programming language, the present invention can be practiced on a variety of programming languages, and as such should not seen as limited to only the Java™ programming language.
One embodiment of the invention is a general-purpose profiling interface between a virtual machine and the front-end of a profiler. The profiling interface is general enough to support a wide variety of profilers and virtual machine implementations. A general-purpose profiling interface offers at least two advantages over direct profiling support in the virtual machine implementation.
First, various profilers are not restricted to one format in presenting profiling information nor are they restricted in the types of profiling information they can present. For example, one profiler may simply record events that occur in the virtual machine trace file. Alternatively, another profiler may display information in response to interactive user input.
Another advantage of a general-purpose profiling interface is that a given profiler can be used with different virtual machine implementations, which support the same general-purpose profiling interface. This allows for profiling tool vendors and virtual machine vendors to leverage each other's products effectively.
Profiler agent 220 can be programmed to delegate resource-intensive tasks to profiler front-end 260. Thus, even though profiler agent 220 runs on virtual machine process 200, profiler agent 220 does not overly distort the profiling information on the user application.
VMPI 215 is a two-way function call interface between virtual machine 210 and profiler agent 220. Typically, profiler agent 220 is implemented as a dynamically loaded library. Virtual machine 210 makes VMPI function calls to inform profiler agent 220 about various events that occur during the execution of the user application program as indicated by an arrow 216 of
In one embodiment of the invention, VMPI events may be data structures consisting of an integer indicating the VMPI event type, an identifier of the thread in which the VMPI event occurred, followed by information that is specific to the VMPI event. Examples of some VMPI events are:
As an illustration, virtual machine 210 sends the VMPI_Event_GC_Start event before the start of garbage collection, and when garbage collection is completed, virtual machine 210 sends the VMPI_Event_GC_Finish event. These and other VMPI events are explained in greater detail in Appendix A.
Allocation and Garbage Collection Events
In one embodiment of the invention, VMPI 215 is a flexible and yet comprehensive interface that uniformly accommodates a wide variety of memory allocation and garbage collection methods. Implicit in garbage collection methods are methods for memory allocation. Appendix A contains a detailed description of representative VMPI events. However, the focus is on the following set of profiling interface events that support all known types of garbage collection methods. Such profiling events are herein referred to as garbage collection algorithm-independent events since they support all known types of garbage collection methods and are further described in greater detail in Appendix A. These garbage collection algorithm-independent events may be used by virtual machine 210 to convey profiling information to profiler agent 220.
The abstract concept of a “heap arena” is used for describing allocation of new objects within the heap. A heap arena describes a logically partitioned portion of the heap. In one embodiment of the invention, all object allocations occur in a given heap arena even when there are a plurality of arenas. After the objects are allocated, they may be moved from one heap arena to another as explained below. To help explain the set of events, each event is described by its components. The first component is the event-type. The second component is the event-specific information contained in a pair of parentheses following the event-type.
new_arena (arena ID)
delete_arena (arena ID)
new_object (arena ID, object ID, class ID)
delete_object (arena ID)
move_object (old arena ID, old object ID, new arena ID, new object ID)
For example, “new_arena” is the event type and “arena ID” is the corresponding event-specific information. The virtual machine interface refers to entities in the virtual machine by various types of IDs. Threads, classes, methods, objects, heap arenas all have unique IDs. The subject of IDs is further detailed in Appendix A. In brief, each ID has a defining event and an undefining event. A defining event provides information related to the ID. For example, the defining event for an “object ID” contains, among other entries, the class ID of the object. The defining event for an object ID is “object allocate”. The object ID remains valid until one of its undefining events arrive. The undefining events for an object ID are as follows:
When an object ID is invalidated either by an “object free” event or an “arena delete” event, the object is said to be “garbage collected”.
Virtual machine 210 sends a “new arena” event to profiler agent 220 via VMPI 215 when a new heap arena for allocating objects is created. In one embodiment of the invention, the “new_arena” event may be described as follows in a Java™ virtual machine implementation.
The “arena_id” is the ID assigned to the new heap arena, and “arena_name” is the name of the new heap arena.
Similarly, virtual machine 210 sends a “delete_arena” event to profiler agent 220 through VMPI 215 when freeing all the objects remaining in the particular heap arena. Thus, profiler agent 220 is apprised of which objects remain in the heap arena by keeping track of the object allocations in the heap arena as well as the movement of objects in and out of the heap arena. In one embodiment of the invention, the “delete_arena” event may be described as follows in a Java™ virtual machine implementation. Note that the arena_id is the ID of the heap arena being deleted.
Virtual machine 210 sends to profiler agent 220 a “new_object” event, also known as an “object_allocate” event when an object is allocated, or when profiler agent 220 requests such an event. In one embodiment of the invention, the “object_allocate” event may be described as follows in a Java™ virtual machine implementation.
“Arena_id” identifies the heap arena where the objects are to be allocated. “Class_id” either identifies the class to which the object that are to be allocated belong, or the array element class is the “is_array” has a value of “JVMPI_Class”. The “is array” can have values that indicate that it is a normal object, an array of objects, an array of Boolean expressions, etc. Further details may be found in Appendix A. “Size” is the number of bytes, and “obj_id” is the unique object ID.
Virtual machine 210 sends to profiler agent 220 a “delete_object” event, also known as an “object_free” event when an object is freed. In one embodiment of the invention, the “object_free” event may be described as follows in a Java™ virtual machine implementation.
“Obj_id” identifies the object being freed.
Virtual machine 210 sends to profiler agent 220 a “move_object” event, also known as an “object_move” event when an object is moved to a new location in the heap. In one embodiment of the invention, the “object_move” event may be described as follows in a Java™ virtual machine implementation.
“Arena_id” identifies the current heap arena where the object resides. “Obj_id” identifies the current object ID of the object to be moved. “New_arena_id” identifies the new heap arena to which the object is to be moved. “New_obj_id” is the new ID assigned to the object after it is moved.
The use of allocation and garbage collection events are described below in connection with various garbage collection methods, such as the Mark-and-Sweep garbage collector, the Mark-Compact garbage collector, the Two-Space Copying garbage collector, the Generational garbage collector, and the Reference-Counting garbage collector.
The “new_arena” and “delete_arena” events, as in the case of Generational garbage collection, provide useful profiling information such as the number of generations of objects, which objects are most likely to die young, and the relative age of the objects.
The “new_object” and “delete_object” events provide profiling information such as when, which and how many objects are allocated and deleted, as well as information on the classes to which the objects belong.
The “move_object” event, as in the case of the Mark-Compact garbage collection, would provide profiling information such as how frequently compaction had to be performed and what are the associated costs. The frequent need for compaction may be a reflection of the unsuitability of the allocation scheme for the particular user program application. In the case of Generational garbage collection, the “move_object” provides information as to survivability of objects. The longer surviving objects are moved to older heap arenas (older generations).
The combination of all five events provide information on the number of live objects that the program is using at any given time. Such information is highly useful in profiling garbage-collected memory systems.
Mark-and-Sweep Garbage Collector
The invention may be used with a Mark-and-Sweep garbage collector. The Mark-and-Sweep collector is an automatic storage and reclamation tracing garbage collection technique. Whether an object is alive may be determined directly or indirectly. The indirect methods are also referred to as “tracing” methods. Direct methods require for each object in the heap that a record be maintained of all references to that object from other heap objects or roots. In contrast, indirect or tracing garbage collectors visit all reachable objects, starting from the roots of the computation and by following pointers. Thus, the objects that were not visited by the garbage collector are made available for recycling.
Under the Mark-and-Sweep method, objects are not reclaimed immediately after they become garbage, but remain unreachable and undetected until all available storage in the heap is exhausted. If a request is then made for storage space, useful processing is temporarily suspended while the garbage collector is called to sweep all currently unused objects from the heap back into the free-pool. The Mark-and-Sweep collector, being a tracing collector, relies on a global traversal of all live objects to determine which objects are available for reclamation. The traversal of the live objects is the “marking” phase and the reclamation of unused objects is the “sweep” phase. During the “marking” phase, a bit associated with each cell, known as the “mark-bit” is reserved for use by the mark-and-sweep collector. The collector uses the “mark-bit” to record if the object is reachable from the roots of the computation. The mark-bit is set as each reachable object is visited. During the “sweep” phase, the unmarked objects are returned to the free-pool.
When used in conjunction with the allocation and garbage collection events described above, the Mark-and-Sweep collector would issue a “new_object” event when allocating objects. Only one heap arena is needed. The collector sends a “delete_object” event when returning dead objects to the free-pool.
Mark-Sweep-Compact Garbage Collector
The invention may be used with a Mark-Compact garbage collector. Under the Mark-Compact scheme, in addition to marking live objects in the heap, objects are relocated within the heap and the values of pointers that referred to the relocated objects are updated. The marked (live) objects are relocated in order to compact the heap. At the end of the compaction phase, the heap will be divided into two contiguous areas. One area will hold the active or live objects while the other area will hold the free-pool. There are several schemes related to the placement of the relocated objects in the heap. For example, under the “arbitrary” scheme, the live objects may be relocated without regard for their original order, or whether they point to one another. However, such a scheme often leads to reduced virtual memory performance. Under the “linearizing” scheme, live objects that originally pointed to one another are moved into adjacent positions. Under the “sliding” scheme, live objects are slid to one end of the heap.
When used in conjunction with the allocation and garbage collection events described above, the Mark-Compact collector would issue a “new_object” event when allocating objects and a “delete_object” event when returning dead objects to the free-pool. In addition, the collector would issue a “move_object” event in order to perform compaction of the heap. Only one heap arena is used in this case because the old and new heap arena IDs in the “move_object” event are the same. In this case, the relocation of the live objects can be viewed as simply reshuffling the positions of the live objects within the same area. Thus, there is no need for the abstraction of separate multiple heap arenas.
Two-Space-Copying Garbage Collector
The invention may be used with a Two-Space-Copying garbage collector. The Two-Space-Copying collector is tracing collector. Under the Two-Space-Copying scheme, the collector first divides the heap equally into two semi-spaces before any allocation of objects is performed. Allocation of objects occurs only in one of the semi spaces, referred to herein as the “Fromspace”. The other semi space, referred to herein as the “Tospace”, is reserved for later use. When garbage collection begins, the collector traverses the live objects in the “Fromspace” and copies each live object to the “Tospace”. After all the live objects in the “Fromspace” have been traced, a replica of the live objects has been created in the “Tospace”, and the user application program is restarted. The garbage objects are abandoned in the “Fromspace”. The collector is said to “scavenge” the live objects from amidst the garbage.
When used in conjunction with the allocation and garbage collection events described above, the Two-Space-Copying collector would issue two “new_arena” events to create two heap arenas representing the two semi-spaces. The collector would issue a “new_object” when allocating objects. During garbage collection, the collector would issue “move_object” events as it scavenges the live objects for placement into the reserved heap arena. Once all the live objects have been scavenged from the old heap arena, the collector sends a “delete_arena” event to delete the old heap arena and thereby freeing all the objects in the old heap arena. Next, in order to restore the two-space scheme, the collector would issue a “new_arena” event using the same arena_id as the old heap arena that was deleted.
Generational Garbage Collector
The invention may be used with a Generational garbage collector. Under the Generational garbage collection scheme, objects are segregated by age into two or more regions of the heap called generations. Different generations can then be garbage collected at different frequencies, with the youngest generation being collected more frequently than the older generations. Objects are first allocated in the youngest generation, but are promoted into older generations if they survive long enough.
When used in conjunction with the allocation and garbage collection events described above, the Generational collector would issue a “new_arena” event to create an heap arena for each generation of objects. In addition, the Generational collector would issue a “new_object” to allocate objects in the youngest generation. The collector would issue a “move_object” event as live objects are scavenged from a younger generation to an older generation. All objects (garbage) in an heap arena are implicitly freed when the collector sends a “delete_arena” event to delete the youngest generation at any given time.
Next, in order to restore the generational scheme, the collector would issue a “new_arena” event using the same arena-id as the youngest generation that was deleted.
Reference-Counting Garbage Collector
The Reference-Counting garbage collector is a direct method of storage reclamation. Under the Reference-Counting scheme, each object has a reference count field used for counting the number of references to the object from other live objects or roots. Free objects have a reference count of zero. When a new object is allocated from the free-pool, its reference count is set to one. Each time a pointer is set to refer to an object, the value of the object's counter is increased by one. Likewise, when a reference to an object is deleted, the object's counter is decreased by one. When an object's counter drops to zero, there are no more pointers to the object and the object becomes garbage.
When used in conjunction with the allocation and garbage collection events described above, the Reference-Counting collector sends a “new_object” event each time an object is allocated, and it sends a “delete_object” event when the reference count of an object drops to zero.
The approach described herein for comprehensive profiling of Java™ virtual machines may be implemented in computer software, in hardware circuitry, or as a combination of computer software and hardware circuitry. Accordingly, the invention is not limited to a particular computer software or hardware circuitry implementation.
The present invention is both flexible and powerful enough to meet the needs of a wide variety of virtual machine implementations and profiling tools. The invention's power lies in its provision for capturing the behavior of memory systems of the virtual machine using a set of events that are independent of any underlying garbage collection algorithm that has been implemented in the virtual machine.
Although certain embodiments have been described using the Java™ programming language, the present invention can be practiced on a variety of programming languages, and as such should not seen as limited to only the Java™ programming language.
B. Implementation Hardware
Computer system 300 may be coupled via bus 302 to a display 312, such as a cathode ray tube (CRT), for displaying information to a computer user. An input device 314, including alphanumeric and other keys, is coupled to bus 302 for communicating information and command selections to processor 304. Another type of user input device is cursor control 316, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 304 and for controlling cursor movement on display 312. This input device typically has two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane.
The invention is related to the use of computer system 300 for implementing comprehensive profiling of Java™ virtual machines. According to one embodiment of the invention, implementing comprehensive profiling of Java™ virtual machines is provided by computer system 300 in response to processor 304 executing one or more sequences of one or more instructions contained in main memory 306. Such instructions may be read into main memory 306 from another computer-readable medium, such as storage device 310. Execution of the sequences of instructions contained in main memory 306 causes processor 304 to perform the process steps described herein. One or more processors in a multi-processing arrangement may also be employed to execute the sequences of instructions contained in main memory 306. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions to implement the invention. Thus, embodiments of the invention are not limited to any specific combination of hardware circuitry and software.
The term “computer-readable medium” as used herein refers to any medium that participates in providing instructions to processor 304 for execution. Such a medium may take many forms, including but not limited to, non-volatile media, volatile media, and transmission media. Non-volatile media includes, for example, optical or magnetic disks, such as storage device 310. Volatile media includes dynamic memory, such as main memory 306. Transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise bus 302. Transmission media can also take the form of acoustic or light waves, such as those generated during radio wave and infrared data communications.
Common forms of computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, or any other magnetic medium, a CD-ROM, any other optical medium, punch cards, paper tape, any other physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, any other memory chip or cartridge, a carrier wave as described hereinafter, or any other medium from which a computer can read.
Various forms of computer readable media may be involved in carrying one or more sequences of one or more instructions to processor 304 for execution. For example, the instructions may initially be carried on a magnetic disk of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to computer system 300 can receive the data on the telephone line and use an infrared transmitter to convert the data to an infrared signal. An infrared detector coupled to bus 302 can receive the data carried in the infrared signal and place the data on bus 302. Bus 302 carries the data to main memory 306, from which processor 304 retrieves and executes the instructions. The instructions received by main memory 306 may optionally be stored on storage device 310 either before or after execution by processor 304.
Computer system 300 also includes a communication interface 318 coupled to bus 302. Communication interface 318 provides a two-way data communication coupling to a network link 320 that is connected to a local network 322. For example, communication interface 318 may be an integrated services digital network (ISDN) card or a modern to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 318 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN. Wireless links may also be implemented. In any such implementation, communication interface 318 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
Network link 320 typically provides data communication through one or more networks to other data devices. For example, network link 320 may provide a connection through local network 322 to a host computer 324 or to data equipment operated by an Internet Service Provider (ISP) 326. ISP 326 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the “Internet” 328. Local network 322 and Internet 328 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks and the signals on network link 320 and through communication interface 318, which carry the digital data to and from computer system 300, are exemplary forms of carrier waves transporting the information.
Computer system 300 can send messages and receive data, including program code, through the network(s), network link 320 and communication interface 318. In the Internet example, a server 330 might transmit a requested code for an application program through Internet 328, ISP 326, local network 322 and communication interface 318. In accordance with the invention, one such downloaded application provides for implementing comprehensive profiling of Java™ virtual machines as described herein.
The received code may be executed by processor 304 as it is received, and/or stored in storage device 310, or other non-volatile storage for later execution. In this manner, computer system 300 may obtain application code in the form of a carrier wave.
In the foregoing specification, the invention has been described with reference to specific embodiments thereof. It will, however, be evident that various modifications and changes may be made thereto without departing from the broader spirit and scope of the invention. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US5193180||Jun 21, 1991||Mar 9, 1993||Pure Software Inc.||System for modifying relocatable object code files to monitor accesses to dynamically allocated memory|
|US5621912||Dec 29, 1994||Apr 15, 1997||International Business Machines Corporation||Method and apparatus for enabling monitoring of guests and native operating systems|
|US6047125||Oct 1, 1997||Apr 4, 2000||Sun Microsystems, Inc.||Garbage collection system for improved use of memory by removal of reference conflicts|
|US6058457||Jun 23, 1997||May 2, 2000||Sun Microsystems, Inc.||Method for storing method frames in multiple stacks|
|US6532531||Jan 23, 1997||Mar 11, 2003||Sun Microsystems, Inc.||Method frame storage using multiple memory circuits|
|EP0777181A1||Nov 27, 1996||Jun 4, 1997||Hewlett-Packard Company||System and method for profiling code on symmetric multiprocessor architectures|
|WO1998048353A1||Apr 21, 1998||Oct 29, 1998||Sun Microsystems, Inc.||Write barrier system and method including pointer-specific instruction variant replacement mechanism|
|1||"Memory Allocation Monitoring Class," IBM Technical Disclosure Bulletin, US IBM Corp. New York, vol. 37, No. 4A, Apr. 1, 1994, p. 445.|
|2||A. Beguelin, et al., "Recent Enhancements to PVM," 3HE International Journal of Supercomputer Applications and High Performance Computing, US, MIT Press, vol. 9, No. 2, Jun. 1, 1995, pp. 108-127.|
|3||*||Java Virtual Machine Profiler Interface (JVMPI). Last modified Nov. 11, 1998, pp. 18-50.|
|4||Liang et al., Comprehensive Profiling Support in the Java Virtual Machine, USENIX Association, 5<SUP>th </SUP>USENIX Conf. on Object-Oriented Technologies and Systems (COOTS '99) (May 1999).|
|5||Office Action dated Jun. 28, 2007, from United States Patent and Trademark Office for U.S. Appl. No. 10/758,266.|
|6||Rosenberg, Jonathan B., How Debuggers Work, Wiley Computer Publishing, pp. 180-182 (1996).|
|7||Spencer, Gordon, "JDK 1.2: Immature, but tons of new abilities," JavaWorld, Inc. (Jan. 1, 1998).|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US7882505 *||Feb 1, 2011||Oracle America, Inc.||Method and apparatus for switching between per-thread and per-processor resource pools in multi-threaded programs|
|US8074025 *||Dec 6, 2011||Intel Corporation||Method and system for copying live entities of source blocks identified by source list for selected destination block to selected destination block of memory heap|
|US8661221 *||Mar 16, 2011||Feb 25, 2014||International Business Machines Corporation||Leasing fragmented storage between processes|
|US8775769 *||Jul 1, 2008||Jul 8, 2014||International Business Machines Corporation||Partition-based method and apparatus for diagnosing memory leak in java systems|
|US8838929 *||Oct 5, 2011||Sep 16, 2014||Arm Limited||Allocation and deallocation of bounded time relative portions of a graphics memory|
|US20060064687 *||Sep 23, 2004||Mar 23, 2006||Jan Dostert||Thread-level resource usage measurement|
|US20060218557 *||Mar 25, 2005||Sep 28, 2006||Sun Microsystems, Inc.||Method and apparatus for switching between per-thread and per-processor resource pools in multi-threaded programs|
|US20090037687 *||Jul 1, 2008||Feb 5, 2009||International Business Machines Corporation||Partition-based method and apparatus for diagnosing memory leak in java systems|
|US20090150613 *||Dec 11, 2007||Jun 11, 2009||Ligang Wang||Memory management|
|US20090319720 *||Dec 24, 2009||Seagate Technology Llc||System and method of garbage collection in a memory device|
|US20120239899 *||Sep 20, 2012||International Business Machines Corporation||Leasing Fragmented Storage Between Processes|
|US20130088501 *||Apr 11, 2013||Arm Limited||Allocating and deallocating portions of memory|
|U.S. Classification||711/170, 714/E11.2, 711/E12.009, 718/1|
|International Classification||G06F12/00, G06F11/34, G06F12/02|
|Cooperative Classification||G06F12/0253, G06F11/3466|
|European Classification||G06F12/02D2G, G06F11/34T|
|Nov 14, 2003||AS||Assignment|
Owner name: SUN MICROSYSTEMS, INC., CALIFORNIA
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LIANG, SHENG;GRARUP, STEFFEN;REEL/FRAME:014828/0880
Effective date: 19990421
|May 13, 2008||CC||Certificate of correction|
|May 18, 2011||FPAY||Fee payment|
Year of fee payment: 4
|Jun 3, 2015||FPAY||Fee payment|
Year of fee payment: 8
|Dec 16, 2015||AS||Assignment|
Owner name: ORACLE AMERICA, INC., CALIFORNIA
Free format text: MERGER AND CHANGE OF NAME;ASSIGNORS:ORACLE USA, INC.;SUN MICROSYSTEMS, INC.;ORACLE AMERICA, INC.;REEL/FRAME:037302/0869
Effective date: 20100212