US7103889B2 - Method, system, and article of manufacture for agent processing - Google Patents

Method, system, and article of manufacture for agent processing Download PDF

Info

Publication number
US7103889B2
US7103889B2 US10/202,662 US20266202A US7103889B2 US 7103889 B2 US7103889 B2 US 7103889B2 US 20266202 A US20266202 A US 20266202A US 7103889 B2 US7103889 B2 US 7103889B2
Authority
US
United States
Prior art keywords
agent
address space
executing
invoking process
agents
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime, expires
Application number
US10/202,662
Other versions
US20040216126A1 (en
Inventor
Daniel K. Hiltgen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Oracle America Inc
Original Assignee
Sun Microsystems Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sun Microsystems Inc filed Critical Sun Microsystems Inc
Priority to US10/202,662 priority Critical patent/US7103889B2/en
Assigned to SUN MICROSYSTEMS, INC. reassignment SUN MICROSYSTEMS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HILTGEN, DANIEL K.
Publication of US20040216126A1 publication Critical patent/US20040216126A1/en
Application granted granted Critical
Publication of US7103889B2 publication Critical patent/US7103889B2/en
Assigned to Oracle America, Inc. reassignment Oracle America, Inc. MERGER AND CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: Oracle America, Inc., ORACLE USA, INC., SUN MICROSYSTEMS, INC.
Adjusted expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/48Program initiating; Program switching, e.g. by interrupt
    • G06F9/4806Task transfer initiation or dispatching
    • G06F9/4843Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
    • G06F9/485Task life-cycle, e.g. stopping, restarting, resuming execution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/02Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/16Implementation or adaptation of Internet protocol [IP], of transmission control protocol [TCP] or of user datagram protocol [UDP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/30Definitions, standards or architectural aspects of layered protocol stacks
    • H04L69/32Architecture of open systems interconnection [OSI] 7-layer type protocol stacks, e.g. the interfaces between the data link level and the physical level

Definitions

  • the present invention is related to a method, system, and article of manufacture for agent processing.
  • a client computer on which client applications run may be connected by a network, such as the Internet, to a server computer.
  • the Internet is a world-wide collection of connected computer networks (i.e., a network of networks).
  • the server computer may be connected to several low level devices.
  • One type of low level device is a redundant array of independent disks (RAID).
  • RAID redundant array of independent disks
  • a RAID device enables storage of the same data on multiple hard disks, thus allowing simultaneous accesses to the copies of data.
  • a client application desires to access a low level device to, for example, retrieve data about the low level device (e.g., whether the low level device is online and available for use or offline and not available for use)
  • the client application submits a request for the data to, for example, the server computer connected to the low level device.
  • the server computer generates a process to handle the client application's request.
  • the process is assigned a portion of memory, which is referred to as the address space of the process.
  • agents implemented in software at the server computer may be used to handle low level device interactions.
  • An agent performs some function, such as retrieving data about a low level device.
  • the server computer may generate an agent to fulfill the client application's request.
  • the agents may access hardware or third party libraries that have unknown stability levels (e.g., crash often).
  • the agents run in the same address space as the process (e.g., a server process) that executed the agents. If agents are permitted to run within the same address space as a process, when one agent fails (e.g., because it is accessing an unstable device), the entire process fails and must be restarted. If the process generates multiple agents, then the failure of one agent causes the termination of all of the agents. This, of course, causes disruptions for a client application.
  • a Common Gateway Interface is a specification for transferring information between a World Wide Web (“Web”) server and a CGI program.
  • a CGI program may send multiple CGI requests (i.e., requests from a CGI program) to a Web server.
  • Each request is handled by a CGI process, and each CGI process uses a separate address space. For example, if three CGI processes are invoked, each CGI process has a separate address space, resulting in the use of three different address spaces.
  • FastCGI is a programming interface that can speed up Web applications that use the Common Gateway Interface (CGI).
  • CGI Common Gateway Interface
  • FastCGI is a plug-in to a Web server computer that processes CGI requests received from a client computer.
  • FastCGI manages multiple CGI requests with a single process that uses a single address space. Thus, if three CGI requests are received a the server computer, one CGI process using one address space handles all three CGI requests.
  • an agent that may process the request is identified. Whether the agent should be executed within an address space of an invoking process that would invoke the agent or within a separate address space is determined. If it is determined that the agent should be executed within a separate address space, the agent is executed within a separate address space.
  • agents there are multiple agents and an agent is identified from the group of agents.
  • determining whether the agent should be executed within an address space of an invoking process or within a separate address space is based on information in a configuration file.
  • the configuration file includes an indication of whether an agent should be executed in a separate address space, a group identifier, and an indication of when to launch the agent.
  • the agent is launched when the invoking process starts. In certain implementations, the agent is launched when the request is received.
  • an agent has a group identifier and it is determined whether the agent is part of a group of agents sharing the group identifier. In certain implementations, when an agent is part of the group of agents sharing the group identifier and when the agent is executed within the separate address space, each of the agents in the group of agents are executed within that separate address space.
  • an indication to switch the agent that is executing within a separate address space to the agent that is executing within the address space of the invoking process that invoked the agent is received.
  • the agent that is to be switched is terminated and executed within the address space of the invoking process.
  • the agent is part of a group of agents and each of the agents in the group is terminated and executed within the address space of the invoking process that invoked the agent.
  • an indication to switch the agent that is executing within an address space of an invoking process that invoked the agent to an agent that is executing in the separate address space is received.
  • the agent that is executing within the address space of a process that invoked the agent is quiesced and executed within the separate address space.
  • an agent may be forked or threaded.
  • the designation of an agent as forked or threaded may be switched while the agent is processing.
  • the process redetermines which agents should be forked based on information stored before the process terminated.
  • FIG. 1A illustrates, in a block diagram, an architecture in which the invention may be implemented in accordance with certain implementations of the invention.
  • FIG. 1B illustrates, in a block diagram, further details of an architecture in which the invention may be implemented in accordance with certain implementations of the invention.
  • FIG. 1C illustrates further details of a secondary process in accordance with certain implementations of the invention.
  • FIG. 1D illustrates a group of forked agents in accordance with certain implementations of the invention.
  • FIGS. 2A and 2B illustrate, in block diagrams, an alternative architecture in which the invention may be implemented in accordance with certain implementations of the invention
  • FIGS. 3A and 3B illustrate logic performed when a process is started in accordance with certain implementations of the invention
  • FIG. 4 illustrates logic performed to return a result to a client from a threaded agent in accordance with certain implementations of the invention.
  • FIG. 5 illustrates logic performed to return a result to a client from a forked agent in accordance with certain implementations of the invention.
  • FIG. 6 illustrates logic performed when an agent fails in accordance with certain implementations of the invention.
  • FIG. 7 illustrates logic performed for identifying agents to be forked after a primary process is restarted in accordance with certain implementations of the invention
  • FIG. 8 illustrates logic for switching an executing agent from one designation to another (e.g., from forked to threaded or threaded to forked) in accordance with certain implementations of the invention.
  • FIG. 9 illustrates, in a block diagram, a CIM environment in accordance with certain implementations of the invention.
  • FIG. 10 illustrates one implementation of the architecture of the computer systems in accordance with certain implementations of the invention
  • an agent may be executed within an address space of a process that is invoking the agent (an “invoking process”) or may be executed within a secondary address space.
  • an agent is executed within an address space of a process that is invoking the agent, the agent is said to be a “threaded agent.”
  • an agent is executed within a secondary address space, rather than the address space of the process that executed the agent, the agent is said to be a “forked agent.”
  • the invention provides a common infrastructure to enable agents to be executed either within the same address space of a server computer process that receives a request from a client computer or executed within a separate address space.
  • the designation of an agent as threaded or forked is based on run time configurable parameters.
  • the designation of an agent as threaded or forked is dynamically determined based on past execution history of the agent. The designation of an agent as threaded or forked may be changed while the agent is executing.
  • a system administrator may designate an agent as threaded or forked. For example, if a system administrator knows (e.g., from past testing experience) that a particular agent often fails (i.e., is unstable), then the system administrator may designate this agent as a forked agent that executes within a separate address space, rather than within the address space of a process that invoked the agent. On the other hand, if an agent is known to often complete successfully (i.e., does not terminate often and is stable), then the system administrator may designate the agent as a threaded agent that executes within the address space of the process that invoked the agent.
  • the process is terminated and restarted.
  • the agent (rather than the process) is terminated and restarted. Since an agent that is unstable may be executed within a separate address space, while an agent that is stable may be executed within the address space of the process that invoked the stable agent, the process will be more stable (i.e., it is less likely that the process will be terminated and restarted).
  • an agent manages a low level device (e.g., a RAID device).
  • the agent may retrieve data on the state of the low level device.
  • the data may indicate, for example, whether the low level device is online or offline, the status of spindles or temperature sensors at the low level device, or the number of volumes in a RAID device.
  • the agent may also change the low level device (e.g., take a volume of a RAID device offline).
  • the agent may retrieve data stored on the low level device.
  • FIG. 1A illustrates, in a block diagram, an architecture in which the invention may be implemented in accordance with certain implementations.
  • a client computer A 100 includes a client application A 110 .
  • the client computer A 100 is connected to a server computer, such as a Hypertext Transfer Protocol (HTTP) server computer 102 .
  • HTTP is an application protocol that provides a set of rules for exchanging files on the World Wide Web, which is a part of the Internet.
  • the HTTP server 102 may include an HTTP daemon that waits for requests from client application 110 .
  • a daemon is a program that runs continuously and exists to handle certain (e.g., periodic) service requests that are expected. The daemon program forwards the requests to other components (e.g., programs or processes) for processing.
  • the HTTP server 102 is connected to RAID devices 154 , 156 , and 192 .
  • the HTTP server 102 includes an operating system 104 , such as a UNIX® operating system.
  • the operating system 104 generates one or more processes to perform various functions.
  • the client application A 110 sends requests to the HTTP server 102 to access RAID devices 154 , 156 , and/or 192 .
  • FIG. 1B illustrates, in a block diagram, further details of an architecture in which the invention may be implemented in accordance with certain implementations
  • FIG. 1C illustrates further details of a secondary process in accordance with certain implementations.
  • Client A 100 , Client B 102 , and Client C 104 may make requests for agent services through a server interface, such as Hypertext Transfer Protocol (HTTP).
  • HTTP Hypertext Transfer Protocol
  • a user may launch a client application 110 , 112 , or 114 .
  • the client application 110 , 112 , or 114 establishes a socket connection to a URL 126 , 128 , or 130 at the HTTP server 124 .
  • a socket is a software object that enables a client application 110 , 112 , and 114 to use a network protocol.
  • a client application 110 , 112 , and 114 can send and receive Transmission Control Protocol/Internet Protocol (TCP/IP) messages by opening a socket and reading and writing data to and from the socket in order to communicate with a forked or threaded agent 150 , 152 , or 190 .
  • TCP/IP Transmission Control Protocol/Internet Protocol
  • the operating system 104 invokes a primary process 118 .
  • the primary process 118 binds to uniform resource locators (URLS) (i.e., “ports”) at the HTTP server 102 .
  • the primary process 118 may be a daemon.
  • the primary process 118 may include an agent daemon 120 for servicing requests for agent services from client applications 110 , 112 , and 114 .
  • the agent daemon 120 includes a main component 122 that starts up the agent daemon 120 .
  • the agent daemon 120 also includes a mapping to URLs 126 , 128 , and 130 at HTTP server 124 (which represents HTTP server 102 for the primary process).
  • the agent daemon 120 also includes agent session managers 132 .
  • Agent session managers 132 includes session managers 134 , 136 , and 138 for managing agent proxies 142 , 144 , and 182 .
  • the agent proxies 142 and 144 are part of agent library 140
  • agent proxy 182 is part of agent library 180 .
  • Each agent proxy 142 , 144 , and 182 manages a forked or threaded agent 150 , 152 , or 190 .
  • Threaded agent-N 152 is the nth agent. In certain implementations, the number of agents is set by a vendor who designed the agents. Each agent 150 , 152 , or 190 may be connected to a low level device, such as RAID devices 154 , 156 , and 192 . In certain implementations, agents 150 , 152 , and 190 are written to an agent application programming interface (agent API). The agent API may be written in any programming language (e.g., a C language API or a Java language API). In FIG. 1A , an agent proxy 142 , 144 , or 182 manages the interaction with the agent 150 , 152 , or 190 through the agent API.
  • agent API agent application programming interface
  • the agent API may be written in any programming language (e.g., a C language API or a Java language API).
  • an agent proxy 142 , 144 , or 182 manages the interaction with the agent 150 , 152 , or 190 through the agent API.
  • An HTTP server 124 process communicates with a session manager 134 , 136 , or 138 .
  • the session manager 134 , 136 , or 138 communicates with an agent 150 , 152 or 190 through an agent proxy 142 , 144 , or 182 . If the agent 190 is forked, the communication occurs over a form of interprocess communication (IPC).
  • IPC is a set of programming interfaces used to manage processes executed by an operating system. IPC enables two or more processes to communicate with each other. If an agent 150 or 152 is threaded, a local procedure may be used to communicate with the agent.
  • Neither the client application 110 , 112 , or 114 nor the agent 150 , 152 , or 190 needs to be aware of whether a particular agent 150 , 152 , or 190 is forked or threaded.
  • the client application 110 , 112 , or 114 or the agent 150 , 152 , or 190 may request information on whether a particular agent 150 , 152 , or 190 is forked or threaded.
  • Client application A 110 may make a request for agent services (e.g., obtaining data describing the state of RAID device 192 or data stored at RAID device 192 ) through a socket to URL 126 .
  • the request is passed on to a session manager, such as session manager 134 .
  • the session manager 134 determines which agent should process the request and whether the agent is a forked agent.
  • the agent retrieves data requested by the client (e.g., data describing a device or data stored on the device).
  • the determination of which agent should process the request may be based on which agent is connected to a device from which the client application has requested data. For example, if client application A 110 has requested data about RAID device 192 , then the session manager would assign agent- 1 190 to process the request.
  • the determination of whether the selected agent is forked or threaded may be made, for example, with reference to a configuration file that specifies which agents are to be threaded and which are to be forked.
  • the session manager 134 determines that the request should be passed to a forked agent. Additionally, the session manager 134 determines when to launch an agent. In particular, an agent may be classified as “lazy” or “immediate” in a configuration file.
  • the session managers 132 review the configuration file and determine whether any of the agents are immediate. Agents classified as immediate are launched when the primary process 118 starts. Agents classified as lazy are launched when a request from a client application 110 , 112 , or 114 is received. Table A illustrates a sample configuration file.
  • Agent-1 Forked Lazy 1
  • Agent-2 Threaded Immediate 2
  • Agent-3 Threaded Lazy 3
  • agent- 1 190 is forked, is lazy (i.e., launched in response to a first client application request), and has group ID 1 .
  • Agent- 2 150 is threaded, immediate (i.e., launched when the primary process 118 is launched), and has group ID 2 .
  • Agent- 3 152 is threaded, lazy, and has group ID 3 .
  • the group ID values are used when forking multiple agents.
  • multiple agents may be forked as a group so that they may logically work together and share data.
  • the session manager 134 , 136 , or 138 determines whether to group two or more agents based on the group ID in the configuration file.
  • the session manager 134 invokes a secondary process 160 .
  • the secondary process 160 includes many of the components of the primary process, but does not include agent proxies 142 and 144 managing threaded agents 150 and 152 , respectively, and does not include the threaded agents 150 and 152 , as illustrated in FIG. 1C .
  • the session manager 134 of the primary process 118 has a link to the agent proxy 182 in the secondary process 160 .
  • the agent proxy 182 manages forked agent- 1 190 , which may access RAID device 192 .
  • client application B 112 may make a request for agent services to access RAID device 154 through a socket to URL 128 .
  • the request is passed on to session manager 136 .
  • the session manager 136 determines which agent should process the request and whether the agent is a forked agent. In this example, the session manager 136 determines that the request should be passed to a threaded agent- 2 150 .
  • the session manager 136 passes the request to agent proxy 142 , which manages threaded agent- 2 150 . Threaded agent- 2 150 processes the request.
  • FIG. 1C illustrates further details of secondary process 160 in accordance with certain implementations.
  • the secondary process 160 illustrates that agent proxy 182 manages forked agent- 1 190 .
  • Agent proxy 182 is managed by session manager 134 in primary process 118 .
  • FIG. 1D illustrates a group of forked agents in accordance with certain implementations.
  • there could be additional session managers in primary process 118 e.g., session manager 133 ), each of which is linked to an agent proxy (e.g., agent proxy 181 ) in the secondary process 160 .
  • Each of the agent proxies in the secondary process 160 may be linked to a forked agent (e.g., forked agent 191 ) that is lined to a device (e.g., device 193 ).
  • the forked agents ( 190 and 191 ) within the secondary process 160 are said to form a group of agents.
  • the group of agents share an address space other than the address space of the primary process.
  • agents may be forked from the secondary process 160 to additional processes (e.g., to a third process).
  • FIGS. 2A and 2B illustrate, in block diagrams, an alternative architecture in which the invention may be implemented in accordance with certain implementations.
  • the functionality of agent session managers 132 and 140 in the primary process may be combined.
  • FIGS. 3A and 3B illustrate logic performed when a process is started in accordance with certain implementations.
  • Control begins at block 300 when a client application 110 , 112 , or 114 sends a request to an HTTP server 124 .
  • the HTTP server 124 passes the request to a session manager 134 , 136 , or 138 (block 302 ).
  • the session manager 134 , 136 , or 138 determines which agent 150 , 152 , or 190 is to be to be used to service the request and whether the agent 150 , 152 , or 190 is threaded or forked (block 304 ).
  • a system administrator designates which agents 150 , 152 , or 190 are to be forked and indicates the designation of threaded or forked in a configuration file, and the session manager 134 , 136 , or 138 uses the configuration file to make its determination. If the agent is forked (block 306 ), processing continues to block 308 , otherwise, processing continues to block 310 .
  • the session manager 136 or 138 within a primary process 118 passes the request to an agent proxy 142 or 144 within the primary process 118 (block 310 ).
  • the agent proxy 142 or 144 within the primary process 118 passes the request to the threaded agent 150 or 152 (block 312 ).
  • the threaded agent 150 or 152 performs the request (block 314 ).
  • the session manager 134 within a primary process 118 creates a secondary process 160 that includes the forked agent 190 (block 308 ).
  • the session manager 134 within the primary process 118 determines whether the forked agent 190 is in a group of agents, by for example, using a configuration file (block 316 ). If so, processing continues to block 318 , otherwise, processing continues to block 320 .
  • the session manager, 134 invokes other agents in the group in the secondary process 160 .
  • the session manager 134 within the primary process 118 passes the request to an agent proxy 182 within the secondary process 160 (block 320 ).
  • the agent proxy 182 within the secondary process 160 passes the request to the forked agent 190 (block 322 ).
  • the forked agent 190 performs the request (block 324 ), by, for example accessing RAID device 192 .
  • FIG. 4 illustrates logic performed to return a result to a client application 112 or 114 from a threaded agent 150 or 152 in accordance with certain implementations of the invention.
  • Control begins at block 400 with a threaded agent 150 or 152 obtaining a result for a client request.
  • the threaded agent 150 or 152 returns the result to an agent proxy 142 or 144 within the primary process 118 (block 402 ).
  • the agent proxy 142 or 144 within the primary process 118 returns the result to a session manager 136 or 138 within the primary process 118 (block 404 ).
  • the session manager 136 or 138 within the primary process 118 returns the result to a URL 128 or 130 of an HTTP server 124 within the primary process 118 (block 406 ).
  • the URL 128 or 130 of HTTP server 124 within the primary process 118 returns the result to the client application 112 or 114 (block 408 ).
  • FIG. 5 illustrates logic performed to return a result to a client application 110 from a forked agent 190 in accordance with certain implementations of the invention.
  • Control begins at block 500 with a forked agent 190 obtaining a result for a client request.
  • the forked agent 190 returns the result to an agent proxy 182 within the secondary process 160 (block 502 ).
  • the agent proxy 182 within the secondary process 160 returns the result to a session manager 134 within the primary process 118 (block 504 ).
  • the session manager 134 within the primary process 118 returns the result to a URL 126 of an HTTP server 124 within the primary process 118 (block 506 ).
  • the URL 126 of HTTP server 124 within the primary process 118 returns the result to the client application 110 (block 508 ).
  • FIG. 6 illustrates logic performed when an agent 110 , 112 , or 114 fails in accordance with certain implementations of the invention.
  • Control begins at block 600 with an agent 110 , 112 , or 114 failing.
  • the forked agent 190 is terminated and the secondary process 160 is restarted. With the described implementation, the termination of the forked agent 190 does not require termination of other agents 150 or 152 or the primary process 118 .
  • block 606 if the agent was threaded, all agents 150 , 152 , and 190 are terminated and the primary process 118 is restarted.
  • FIG. 7 illustrates logic performed for identifying agents 110 , 112 , or 114 to be forked after a primary process 118 is restarted in accordance with certain implementations of the invention.
  • Control begins at block 700 with the primary process 1 18 restarting.
  • the primary process 1 18 is restarted due to failure of the primary process 118 during an earlier execution.
  • the primary process 118 may have failed due to a threaded agent 150 or 152 failing.
  • an agent 150 , 152 , or 190 fails, data is stored on the status of all of the agents 150 , 152 , and 190 .
  • the data indicates which of the threaded and forked agents 150 , 152 , and 190 successfully completed their functions, which failed, and which were continuing processing at time of failure.
  • data may be stored on just the one or more agents 150 , 152 or 190 that failed.
  • the data is stored in a file that is persistent across process terminations (e.g., system crashes).
  • each session manager 134 , 136 , or 138 uses the data to determine whether an agent 110 , 112 or 114 should be forked (block 702 ). For example, if a threaded agent 112 or 114 failed previously, upon restart, whichever session manager 136 or 138 that is determining whether that agent 112 or 114 should be forked, may designate that agent 112 or 114 as a forked agent when the primary process 118 restarts due to that agent's 112 or 114 previously failure.
  • the session manager 134 , 136 or 138 may designate that agent as a threaded agent when the primary process 118 restarts. Then, the agents designated as forked agents are forked (block 704 ). In certain implementations, the determination of which agents should be forked may be made by an agent daemon, a session manager, or an agent library, or by some combination of these components.
  • FIG. 8 illustrates logic for switching an executing agent 150 , 152 or 190 from one designation to another (e.g., from forked to threaded or threaded to forked) in accordance with certain implementations of the invention.
  • Control begins at block 800 with an agent's 150 , 152 or 190 designation being switched.
  • a user interface is provided that allows, for example, a system administrator to switch an agent's 150 , 152 or 190 designation.
  • a system administrator may switch an agent's 150 , 152 or 190 designation based on information provided through the user interface, such as which low level device the agent 150 , 152 or 190 is accessing.
  • a system administrator may switch an agent's 150 , 152 or 190 designation by modifying a configuration file, and, in this case, the session manager 134 , 136 or 138 monitors changes to the configuration file.
  • block 802 it is determined whether the agent 150 , 152 or 190 is being switched from a threaded agent 150 or 152 to a forked agent. If so, processing continues at block 804 , otherwise, processing continues at block 806 .
  • the threaded agent 150 or 152 is quiesced and then restarted as a forked agent.
  • the forked agent 190 is terminated and restarted as a threaded agent.
  • the agent API includes a quiesce method that may be invoked to tell a threaded agent 150 or 152 to temporarily stop processing and prepare data for restarting. When the threaded agent 150 or 152 is restarted as a forked agent, the forked agent uses the prepared data to restart at a point at which processing had been quiesced.
  • a forked agent 190 when a forked agent 190 is to be switched to a threaded agent, if the forked agent 190 is part of a group of agents, only the forked agent 190 is switched. In certain implementations, when a forked agent 190 is to be switched to a threaded agent, if the forked agent 190 is part of a group of agents, all of the forked agents in the group are switched to threaded agents.
  • the agents are Common Information Model (CIM) providers in a CIM environment.
  • CIM is a standard for an object-oriented model for managing information.
  • the CIM standard is provided by the Distributed Management Task Force, Inc. (DMTF).
  • DMTF Distributed Management Task Force, Inc.
  • FIG. 9 illustrates, in a block diagram, a CIM environment in accordance with certain implementations of the invention.
  • a client computer 900 executes a CIM client application 902 .
  • the client computer 900 is connected to a management server 920 .
  • the management server 920 includes a CIM object manager 930 .
  • the CIM object manager 930 passes requests received from a CIM client application 902 for objects stored on RAID device 960 to CIM provider 940 and passes requests for objects stored on RAID device 970 to CIM provider 950 .
  • the CIM object manager 930 may be replaced with the agent daemon 120 and agent library 140 .
  • the threaded and forked agents 150 , 152 , and 190 may replace the CIM providers 940 and 950 .
  • UNIX is a registered trademark of the Open Group.
  • the above described implementations for creating consistent copies may be implemented using standard programming and/or engineering techniques to produce software, firmware, hardware, or any combination thereof.
  • article of manufacture refers to code or logic implemented in hardware logic (e.g., an integrated circuit chip, Programmable Gate Array (PGA), Application Specific Integrated Circuit (ASIC), etc.) or a computer readable medium (e.g., magnetic storage medium (e.g., hard disk drives, floppy disks, tape, etc.), optical storage (CD-ROMs, optical disks, etc.), volatile and non-volatile memory devices (e.g., EEPROMs, ROMs, PROMs, RAMs, DRAMs, SRAMs, firmware, programmable logic, etc.)).
  • hardware logic e.g., an integrated circuit chip, Programmable Gate Array (PGA), Application Specific Integrated Circuit (ASIC), etc.
  • a computer readable medium e.g., magnetic storage medium (e.g., hard disk drives, floppy disks, tape, etc.), optical
  • Code in the computer readable medium is accessed and executed by a processor.
  • the code may further be accessible through a transmission media or from a file server over a network.
  • the article of manufacture in which the code is implemented may comprise a transmission media, such as a network transmission line, wireless transmission media, signals propagating through space, radio waves, infrared signals, etc.
  • a transmission media such as a network transmission line, wireless transmission media, signals propagating through space, radio waves, infrared signals, etc.
  • the HTTP server 124 may be replaced with any socket interface.
  • agent proxies 142 , 144 , and 182 may determine which agent should process a request and/or whether the agent should be forked or threaded.
  • FIGS. 3A , 3 B, and 4 – 8 describe specific operations occurring in a particular order. In alternative implementations, certain of the logic operations may be performed in a different order, modified or removed. Morever, steps may be added to the above described logic and still conform to the described implementations. Further, operations described herein may occur sequentially or certain operations may be processed in parallel, or operations described as performed by a single process may be performed by distributed processes.
  • FIGS. 3A , 3 B, and 4 – 8 The logic of FIGS. 3A , 3 B, and 4 – 8 was described as being implemented in software. This logic may be part of the operating system of the host systems or an application program. In yet further implementations, this logic may be maintained in storage areas managed by the control units or in a read only memory or other hardwired type of device. The preferred logic may be implemented in hard disk drives or in programmable and non-programmable gate array logic.
  • FIG. 10 illustrates one implementation of the architecture of the computer systems 100 and 102 in accordance with certain implementations of the invention.
  • the computer systems 100 and 102 may implement a computer architecture 1000 having a processor 1002 (e.g., a microprocessor), a memory 1004 (e.g., a volatile memory device), and storage 1006 (e.g., a non-volatile storage, such as magnetic disk drives, optical disk drives, a tape drive, etc.).
  • the storage 1006 may comprise an internal storage device or an attached or network accessible storage. Programs in the storage 1006 are loaded into the memory 1004 and executed by the processor 1002 in a manner known in the art.
  • the architecture further includes a network card 1008 to enable communication with a network.
  • An input device 1010 is used to provide user input to the processor 1002 , and may include a keyboard, mouse, pen-stylus, microphone, touch sensitive display screen, or any other activation or input mechanism known in the art.
  • An output device 1012 is capable of rendering information transmitted from the processor 1002 , or other component, such as a display monitor, printer, storage, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer And Data Communications (AREA)

Abstract

Disclosed is a system, method, and article of manufacture for agent processing. In particular, an agent may be executed within an address space of an invoking process that invoked the agent or may be executed in a separate address space. While the agent is processing, the execution of the agent within the address space of the invoking process may be switched to execution of the agent within the separate address space. Similarly, while the agent is processing, the execution of the agent in the separate address space may be switched to execution of the agent in the address space of the invoking process. Moreover, if the invoking process terminates before completing processing, when the invoking process restarts, the invoking process redetermines which agents should be executed in separate address spaces based on information stored before the invoking process terminated.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention is related to a method, system, and article of manufacture for agent processing.
2. Description of the Related Art
A client computer on which client applications run may be connected by a network, such as the Internet, to a server computer. The Internet is a world-wide collection of connected computer networks (i.e., a network of networks). The server computer may be connected to several low level devices. One type of low level device is a redundant array of independent disks (RAID). A RAID device enables storage of the same data on multiple hard disks, thus allowing simultaneous accesses to the copies of data.
When a client application desires to access a low level device to, for example, retrieve data about the low level device (e.g., whether the low level device is online and available for use or offline and not available for use), the client application submits a request for the data to, for example, the server computer connected to the low level device. The server computer generates a process to handle the client application's request. The process is assigned a portion of memory, which is referred to as the address space of the process.
Moreover, agents implemented in software at the server computer may be used to handle low level device interactions. An agent performs some function, such as retrieving data about a low level device. Thus, the server computer may generate an agent to fulfill the client application's request.
In many cases, the agents may access hardware or third party libraries that have unknown stability levels (e.g., crash often). In some systems, the agents run in the same address space as the process (e.g., a server process) that executed the agents. If agents are permitted to run within the same address space as a process, when one agent fails (e.g., because it is accessing an unstable device), the entire process fails and must be restarted. If the process generates multiple agents, then the failure of one agent causes the termination of all of the agents. This, of course, causes disruptions for a client application.
A Common Gateway Interface (CGI) is a specification for transferring information between a World Wide Web (“Web”) server and a CGI program. A CGI program may send multiple CGI requests (i.e., requests from a CGI program) to a Web server. Each request is handled by a CGI process, and each CGI process uses a separate address space. For example, if three CGI processes are invoked, each CGI process has a separate address space, resulting in the use of three different address spaces.
FastCGI is a programming interface that can speed up Web applications that use the Common Gateway Interface (CGI). In particular, FastCGI is a plug-in to a Web server computer that processes CGI requests received from a client computer. FastCGI manages multiple CGI requests with a single process that uses a single address space. Thus, if three CGI requests are received a the server computer, one CGI process using one address space handles all three CGI requests.
Thus, there is a need in the art for improved agent execution.
SUMMARY OF THE INVENTION
Provided are a method, system, and for processing a request. Initially, an agent that may process the request is identified. Whether the agent should be executed within an address space of an invoking process that would invoke the agent or within a separate address space is determined. If it is determined that the agent should be executed within a separate address space, the agent is executed within a separate address space.
In certain implementations, there are multiple agents and an agent is identified from the group of agents.
In additional implementations, determining whether the agent should be executed within an address space of an invoking process or within a separate address space is based on information in a configuration file. The configuration file includes an indication of whether an agent should be executed in a separate address space, a group identifier, and an indication of when to launch the agent.
In yet other implementations, it is determined when the agent should be launched. In certain implementations, the agent is launched when the invoking process starts. In certain implementations, the agent is launched when the request is received.
In further implementations, an agent has a group identifier and it is determined whether the agent is part of a group of agents sharing the group identifier. In certain implementations, when an agent is part of the group of agents sharing the group identifier and when the agent is executed within the separate address space, each of the agents in the group of agents are executed within that separate address space.
In additional implementations, it is determined that the agent has failed. If the agent was executed within the separate address space, the agent is restarted.
In yet other implementations, an indication to switch the agent that is executing within a separate address space to the agent that is executing within the address space of the invoking process that invoked the agent is received. The agent that is to be switched is terminated and executed within the address space of the invoking process. In certain implementations, the agent is part of a group of agents and each of the agents in the group is terminated and executed within the address space of the invoking process that invoked the agent.
In further implementations, an indication to switch the agent that is executing within an address space of an invoking process that invoked the agent to an agent that is executing in the separate address space is received. The agent that is executing within the address space of a process that invoked the agent is quiesced and executed within the separate address space.
In additional implementations, when the invoking process that has invoked the agent is restarted, a determination is made of which agents are to be executed in one or more separate address spaces based on information stored by the invoking process before it was restarted. Then, those agents are executed in the one or more separate address spaces.
The described implementations of the invention provide a method, system, and for agent processing. In particular, an agent may be forked or threaded. The designation of an agent as forked or threaded may be switched while the agent is processing. Moreover, if a process terminates before completing processing, when the process restarts, the process redetermines which agents should be forked based on information stored before the process terminated.
BRIEF DESCRIPTION OF THE DRAWINGS
Referring now to the drawings in which like reference numbers represent corresponding parts throughout:
FIG. 1A illustrates, in a block diagram, an architecture in which the invention may be implemented in accordance with certain implementations of the invention.
FIG. 1B illustrates, in a block diagram, further details of an architecture in which the invention may be implemented in accordance with certain implementations of the invention.
FIG. 1C illustrates further details of a secondary process in accordance with certain implementations of the invention.
FIG. 1D illustrates a group of forked agents in accordance with certain implementations of the invention.
FIGS. 2A and 2B illustrate, in block diagrams, an alternative architecture in which the invention may be implemented in accordance with certain implementations of the invention
FIGS. 3A and 3B illustrate logic performed when a process is started in accordance with certain implementations of the invention
FIG. 4 illustrates logic performed to return a result to a client from a threaded agent in accordance with certain implementations of the invention.
FIG. 5 illustrates logic performed to return a result to a client from a forked agent in accordance with certain implementations of the invention.
FIG. 6 illustrates logic performed when an agent fails in accordance with certain implementations of the invention.
FIG. 7 illustrates logic performed for identifying agents to be forked after a primary process is restarted in accordance with certain implementations of the invention
FIG. 8 illustrates logic for switching an executing agent from one designation to another (e.g., from forked to threaded or threaded to forked) in accordance with certain implementations of the invention.
FIG. 9 illustrates, in a block diagram, a CIM environment in accordance with certain implementations of the invention.
FIG. 10 illustrates one implementation of the architecture of the computer systems in accordance with certain implementations of the invention
DETAILED DESCRIPTION
In the following description, reference is made to the accompanying drawings which form a part hereof and which illustrate several implementations of the present invention. It is understood that other implementations may be utilized and structural and operational changes may be made without departing from the scope of the present invention.
In certain implementations of the invention, an agent may be executed within an address space of a process that is invoking the agent (an “invoking process”) or may be executed within a secondary address space. When an agent is executed within an address space of a process that is invoking the agent, the agent is said to be a “threaded agent.” When an agent is executed within a secondary address space, rather than the address space of the process that executed the agent, the agent is said to be a “forked agent.”
The invention provides a common infrastructure to enable agents to be executed either within the same address space of a server computer process that receives a request from a client computer or executed within a separate address space. In certain implementations, the designation of an agent as threaded or forked is based on run time configurable parameters. In certain implementations, the designation of an agent as threaded or forked is dynamically determined based on past execution history of the agent. The designation of an agent as threaded or forked may be changed while the agent is executing.
With implementations of this invention, a system administrator may designate an agent as threaded or forked. For example, if a system administrator knows (e.g., from past testing experience) that a particular agent often fails (i.e., is unstable), then the system administrator may designate this agent as a forked agent that executes within a separate address space, rather than within the address space of a process that invoked the agent. On the other hand, if an agent is known to often complete successfully (i.e., does not terminate often and is stable), then the system administrator may designate the agent as a threaded agent that executes within the address space of the process that invoked the agent. When an agent that is executed within the address space of the process that invoked the agent fails, the process is terminated and restarted. When an agent that is executed in a separate address space fails, the agent (rather than the process) is terminated and restarted. Since an agent that is unstable may be executed within a separate address space, while an agent that is stable may be executed within the address space of the process that invoked the stable agent, the process will be more stable (i.e., it is less likely that the process will be terminated and restarted).
In certain implementations, an agent manages a low level device (e.g., a RAID device). For example, the agent may retrieve data on the state of the low level device. The data may indicate, for example, whether the low level device is online or offline, the status of spindles or temperature sensors at the low level device, or the number of volumes in a RAID device. The agent may also change the low level device (e.g., take a volume of a RAID device offline). In certain implementations, the agent may retrieve data stored on the low level device.
FIG. 1A illustrates, in a block diagram, an architecture in which the invention may be implemented in accordance with certain implementations. A client computer A 100 includes a client application A 110. The client computer A 100 is connected to a server computer, such as a Hypertext Transfer Protocol (HTTP) server computer 102. HTTP is an application protocol that provides a set of rules for exchanging files on the World Wide Web, which is a part of the Internet. The HTTP server 102 may include an HTTP daemon that waits for requests from client application 110. A daemon is a program that runs continuously and exists to handle certain (e.g., periodic) service requests that are expected. The daemon program forwards the requests to other components (e.g., programs or processes) for processing.
The HTTP server 102 is connected to RAID devices 154, 156, and 192. The HTTP server 102 includes an operating system 104, such as a UNIX® operating system. The operating system 104 generates one or more processes to perform various functions. The client application A 110 sends requests to the HTTP server 102 to access RAID devices 154, 156, and/or 192.
FIG. 1B illustrates, in a block diagram, further details of an architecture in which the invention may be implemented in accordance with certain implementations, while FIG. 1C illustrates further details of a secondary process in accordance with certain implementations. Client A 100, Client B 102, and Client C 104 may make requests for agent services through a server interface, such as Hypertext Transfer Protocol (HTTP). In particular, a user may launch a client application 110, 112, or 114. The client application 110, 112, or 114 establishes a socket connection to a URL 126, 128, or 130 at the HTTP server 124. In some operating systems, such as the UNIX® operating system, a socket is a software object that enables a client application 110, 112, and 114 to use a network protocol. For example, in the UNIX® environment, a client application 110, 112, and 114 can send and receive Transmission Control Protocol/Internet Protocol (TCP/IP) messages by opening a socket and reading and writing data to and from the socket in order to communicate with a forked or threaded agent 150, 152, or 190.
In certain implementations, when the operating system 104 starts up, the operating system 104 invokes a primary process 118. The primary process 118 binds to uniform resource locators (URLS) (i.e., “ports”) at the HTTP server 102. The primary process 118 may be a daemon. The primary process 118 may include an agent daemon 120 for servicing requests for agent services from client applications 110, 112, and 114.
The agent daemon 120 includes a main component 122 that starts up the agent daemon 120. The agent daemon 120 also includes a mapping to URLs 126, 128, and 130 at HTTP server 124 (which represents HTTP server 102 for the primary process). The agent daemon 120 also includes agent session managers 132. Agent session managers 132 includes session managers 134, 136, and 138 for managing agent proxies 142, 144, and 182. The agent proxies 142 and 144 are part of agent library 140, while agent proxy 182 is part of agent library 180. Each agent proxy 142, 144, and 182 manages a forked or threaded agent 150, 152, or 190. Threaded agent-N 152 is the nth agent. In certain implementations, the number of agents is set by a vendor who designed the agents. Each agent 150, 152, or 190 may be connected to a low level device, such as RAID devices 154, 156, and 192. In certain implementations, agents 150, 152, and 190 are written to an agent application programming interface (agent API). The agent API may be written in any programming language (e.g., a C language API or a Java language API). In FIG. 1A, an agent proxy 142, 144, or 182 manages the interaction with the agent 150, 152, or 190 through the agent API.
An HTTP server 124 process communicates with a session manager 134, 136, or 138. The session manager 134, 136, or 138 communicates with an agent 150, 152 or 190 through an agent proxy 142, 144, or 182. If the agent 190 is forked, the communication occurs over a form of interprocess communication (IPC). IPC is a set of programming interfaces used to manage processes executed by an operating system. IPC enables two or more processes to communicate with each other. If an agent 150 or 152 is threaded, a local procedure may be used to communicate with the agent. Neither the client application 110, 112, or 114 nor the agent 150, 152, or 190 needs to be aware of whether a particular agent 150, 152, or 190 is forked or threaded. In certain implementations, the client application 110, 112, or 114 or the agent 150, 152, or 190 may request information on whether a particular agent 150, 152, or 190 is forked or threaded.
Client application A 110 may make a request for agent services (e.g., obtaining data describing the state of RAID device 192 or data stored at RAID device 192) through a socket to URL 126. The request is passed on to a session manager, such as session manager 134. The session manager 134 determines which agent should process the request and whether the agent is a forked agent. To process a request, the agent retrieves data requested by the client (e.g., data describing a device or data stored on the device). The determination of which agent should process the request may be based on which agent is connected to a device from which the client application has requested data. For example, if client application A 110 has requested data about RAID device 192, then the session manager would assign agent-1 190 to process the request.
The determination of whether the selected agent is forked or threaded may be made, for example, with reference to a configuration file that specifies which agents are to be threaded and which are to be forked. In this example, the session manager 134 determines that the request should be passed to a forked agent. Additionally, the session manager 134 determines when to launch an agent. In particular, an agent may be classified as “lazy” or “immediate” in a configuration file. When the primary process 118 starts, the session managers 132 review the configuration file and determine whether any of the agents are immediate. Agents classified as immediate are launched when the primary process 118 starts. Agents classified as lazy are launched when a request from a client application 110, 112, or 114 is received. Table A illustrates a sample configuration file.
TABLE A
Forked or Lazy or
Threaded Immediate
Designation Designation Group ID
Agent-1 Forked Lazy 1
Agent-2 Threaded Immediate 2
Agent-3 Threaded Lazy 3
According to the configuration file of Table A, agent-1 190 is forked, is lazy (i.e., launched in response to a first client application request), and has group ID 1. Agent-2 150 is threaded, immediate (i.e., launched when the primary process 118 is launched), and has group ID 2. Agent-3 152 is threaded, lazy, and has group ID 3.
The group ID values are used when forking multiple agents. In certain implementations, multiple agents may be forked as a group so that they may logically work together and share data. The session manager 134, 136, or 138 determines whether to group two or more agents based on the group ID in the configuration file.
The session manager 134 invokes a secondary process 160. For ease of illustration, one secondary process 160 is shown. However, multiple additional processes may be invoked for additional forked agents. The secondary process 160 includes many of the components of the primary process, but does not include agent proxies 142 and 144 managing threaded agents 150 and 152, respectively, and does not include the threaded agents 150 and 152, as illustrated in FIG. 1C. The session manager 134 of the primary process 118 has a link to the agent proxy 182 in the secondary process 160. The agent proxy 182 manages forked agent-1 190, which may access RAID device 192.
Moreover, client application B 112 may make a request for agent services to access RAID device 154 through a socket to URL 128. The request is passed on to session manager 136. The session manager 136 determines which agent should process the request and whether the agent is a forked agent. In this example, the session manager 136 determines that the request should be passed to a threaded agent-2 150. The session manager 136 passes the request to agent proxy 142, which manages threaded agent-2 150. Threaded agent-2 150 processes the request.
FIG. 1C illustrates further details of secondary process 160 in accordance with certain implementations. In particular, the secondary process 160 illustrates that agent proxy 182 manages forked agent-1 190. Agent proxy 182 is managed by session manager 134 in primary process 118.
For ease of illustration, only one forked agent-1 190 has been illustrated in secondary process 160. FIG. 1D illustrates a group of forked agents in accordance with certain implementations. For example, there could be additional session managers in primary process 118 (e.g., session manager 133), each of which is linked to an agent proxy (e.g., agent proxy 181) in the secondary process 160. Each of the agent proxies in the secondary process 160 may be linked to a forked agent (e.g., forked agent 191) that is lined to a device (e.g., device 193). The forked agents (190 and 191) within the secondary process 160 are said to form a group of agents. In certain implementations, the group of agents share an address space other than the address space of the primary process. In certain implementations, agents may be forked from the secondary process 160 to additional processes (e.g., to a third process).
FIGS. 2A and 2B illustrate, in block diagrams, an alternative architecture in which the invention may be implemented in accordance with certain implementations. In particular, in FIGS. 2A and 2B the functionality of agent session managers 132 and 140 in the primary process may be combined.
FIGS. 3A and 3B illustrate logic performed when a process is started in accordance with certain implementations. Control begins at block 300 when a client application 110, 112, or 114 sends a request to an HTTP server 124. The HTTP server 124 passes the request to a session manager 134, 136, or 138 (block 302). The session manager 134, 136, or 138 determines which agent 150, 152, or 190 is to be to be used to service the request and whether the agent 150, 152, or 190 is threaded or forked (block 304). In certain implementations, a system administrator designates which agents 150, 152, or 190 are to be forked and indicates the designation of threaded or forked in a configuration file, and the session manager 134, 136, or 138 uses the configuration file to make its determination. If the agent is forked (block 306), processing continues to block 308, otherwise, processing continues to block 310.
When an agent is threaded, the session manager 136 or 138 within a primary process 118 passes the request to an agent proxy 142 or 144 within the primary process 118 (block 310). The agent proxy 142 or 144 within the primary process 118 passes the request to the threaded agent 150 or 152 (block 312). The threaded agent 150 or 152 performs the request (block 314).
When an agent is forked, the session manager 134 within a primary process 118 creates a secondary process 160 that includes the forked agent 190 (block 308). The session manager 134 within the primary process 118 determines whether the forked agent 190 is in a group of agents, by for example, using a configuration file (block 316). If so, processing continues to block 318, otherwise, processing continues to block 320. In block 318, the session manager, 134 invokes other agents in the group in the secondary process 160. Then, the session manager 134 within the primary process 118 passes the request to an agent proxy 182 within the secondary process 160 (block 320). The agent proxy 182 within the secondary process 160 passes the request to the forked agent 190 (block 322). The forked agent 190 performs the request (block 324), by, for example accessing RAID device 192.
FIG. 4 illustrates logic performed to return a result to a client application 112 or 114 from a threaded agent 150 or 152 in accordance with certain implementations of the invention. Control begins at block 400 with a threaded agent 150 or 152 obtaining a result for a client request. The threaded agent 150 or 152 returns the result to an agent proxy 142 or 144 within the primary process 118 (block 402). The agent proxy 142 or 144 within the primary process 118 returns the result to a session manager 136 or 138 within the primary process 118 (block 404). The session manager 136 or 138 within the primary process 118 returns the result to a URL 128 or 130 of an HTTP server 124 within the primary process 118 (block 406). The URL 128 or 130 of HTTP server 124 within the primary process 118 returns the result to the client application 112 or 114 (block 408).
FIG. 5 illustrates logic performed to return a result to a client application 110 from a forked agent 190 in accordance with certain implementations of the invention. Control begins at block 500 with a forked agent 190 obtaining a result for a client request. The forked agent 190 returns the result to an agent proxy 182 within the secondary process 160 (block 502). The agent proxy 182 within the secondary process 160 returns the result to a session manager 134 within the primary process 118 (block 504). The session manager 134 within the primary process 118 returns the result to a URL 126 of an HTTP server 124 within the primary process 118 (block 506). The URL 126 of HTTP server 124 within the primary process 118 returns the result to the client application 110 (block 508).
FIG. 6 illustrates logic performed when an agent 110, 112, or 114 fails in accordance with certain implementations of the invention. Control begins at block 600 with an agent 110, 112, or 114 failing. In block 602, it is determined whether the agent that failed is a forked agent 190. If so, processing continues to block 604. Otherwise, processing continues to block 606. In block 604, the forked agent 190 is terminated and the secondary process 160 is restarted. With the described implementation, the termination of the forked agent 190 does not require termination of other agents 150 or 152 or the primary process 118. In block 606, if the agent was threaded, all agents 150, 152, and 190 are terminated and the primary process 118 is restarted.
FIG. 7 illustrates logic performed for identifying agents 110, 112, or 114 to be forked after a primary process 118 is restarted in accordance with certain implementations of the invention. Control begins at block 700 with the primary process 1 18 restarting. The primary process 1 18 is restarted due to failure of the primary process 118 during an earlier execution. For example, the primary process 118 may have failed due to a threaded agent 150 or 152 failing. In certain implementations, if an agent 150, 152, or 190 fails, data is stored on the status of all of the agents 150, 152, and 190. The data indicates which of the threaded and forked agents 150, 152, and 190 successfully completed their functions, which failed, and which were continuing processing at time of failure. In certain implementations, data may be stored on just the one or more agents 150, 152 or 190 that failed. The data is stored in a file that is persistent across process terminations (e.g., system crashes).
When the primary process 1 18 restarts, the primary process 118 is able to access the data. In particular, each session manager 134, 136, or 138 uses the data to determine whether an agent 110, 112 or 114 should be forked (block 702). For example, if a threaded agent 112 or 114 failed previously, upon restart, whichever session manager 136 or 138 that is determining whether that agent 112 or 114 should be forked, may designate that agent 112 or 114 as a forked agent when the primary process 118 restarts due to that agent's 112 or 114 previously failure. Additionally, if a forked agent 190 did not fail, the session manager 134, 136 or 138 may designate that agent as a threaded agent when the primary process 118 restarts. Then, the agents designated as forked agents are forked (block 704). In certain implementations, the determination of which agents should be forked may be made by an agent daemon, a session manager, or an agent library, or by some combination of these components.
FIG. 8 illustrates logic for switching an executing agent 150, 152 or 190 from one designation to another (e.g., from forked to threaded or threaded to forked) in accordance with certain implementations of the invention. Control begins at block 800 with an agent's 150, 152 or 190 designation being switched. In certain implementations, a user interface is provided that allows, for example, a system administrator to switch an agent's 150, 152 or 190 designation. A system administrator may switch an agent's 150, 152 or 190 designation based on information provided through the user interface, such as which low level device the agent 150, 152 or 190 is accessing. In certain implementations, a system administrator may switch an agent's 150, 152 or 190 designation by modifying a configuration file, and, in this case, the session manager 134, 136 or 138 monitors changes to the configuration file.
In block 802, it is determined whether the agent 150, 152 or 190 is being switched from a threaded agent 150 or 152 to a forked agent. If so, processing continues at block 804, otherwise, processing continues at block 806. In block 804, the threaded agent 150 or 152 is quiesced and then restarted as a forked agent. In block 806, since a forked agent 190 is being switched to a threaded agent, the forked agent 190 is terminated and restarted as a threaded agent. In particular, the agent API includes a quiesce method that may be invoked to tell a threaded agent 150 or 152 to temporarily stop processing and prepare data for restarting. When the threaded agent 150 or 152 is restarted as a forked agent, the forked agent uses the prepared data to restart at a point at which processing had been quiesced.
In certain implementations, when a forked agent 190 is to be switched to a threaded agent, if the forked agent 190 is part of a group of agents, only the forked agent 190 is switched. In certain implementations, when a forked agent 190 is to be switched to a threaded agent, if the forked agent 190 is part of a group of agents, all of the forked agents in the group are switched to threaded agents.
In certain implementations, the agents are Common Information Model (CIM) providers in a CIM environment. CIM is a standard for an object-oriented model for managing information. The CIM standard is provided by the Distributed Management Task Force, Inc. (DMTF). For further information on the CIM standard, see “Specifications for CIM Operations over HTTP,” Version 1.1, May 2, 2002, from the DMTF, which is incorporated by reference herein in its entirety.
FIG. 9 illustrates, in a block diagram, a CIM environment in accordance with certain implementations of the invention. A client computer 900 executes a CIM client application 902. The client computer 900 is connected to a management server 920. The management server 920 includes a CIM object manager 930. The CIM object manager 930 passes requests received from a CIM client application 902 for objects stored on RAID device 960 to CIM provider 940 and passes requests for objects stored on RAID device 970 to CIM provider 950. In certain implementations of the invention, the CIM object manager 930 may be replaced with the agent daemon 120 and agent library 140. The threaded and forked agents 150, 152, and 190 may replace the CIM providers 940 and 950.
UNIX is a registered trademark of the Open Group.
Additional Implementation Details
The above described implementations for creating consistent copies may be implemented using standard programming and/or engineering techniques to produce software, firmware, hardware, or any combination thereof. The term “article of manufacture” as used herein refers to code or logic implemented in hardware logic (e.g., an integrated circuit chip, Programmable Gate Array (PGA), Application Specific Integrated Circuit (ASIC), etc.) or a computer readable medium (e.g., magnetic storage medium (e.g., hard disk drives, floppy disks, tape, etc.), optical storage (CD-ROMs, optical disks, etc.), volatile and non-volatile memory devices (e.g., EEPROMs, ROMs, PROMs, RAMs, DRAMs, SRAMs, firmware, programmable logic, etc.)). Code in the computer readable medium is accessed and executed by a processor. The code may further be accessible through a transmission media or from a file server over a network. In such cases, the article of manufacture in which the code is implemented may comprise a transmission media, such as a network transmission line, wireless transmission media, signals propagating through space, radio waves, infrared signals, etc. Of course, those skilled in the art will recognize that many modifications may be made to this configuration without departing from the scope of the present invention, and that the article of manufacture may comprise any information bearing medium known in the art.
In alternative implementations, the HTTP server 124 may be replaced with any socket interface.
In certain implementations, rather than the session managers 134, 136, and 138, other components, such as the agent proxies 142, 144, and 182 may determine which agent should process a request and/or whether the agent should be forked or threaded.
The logic of FIGS. 3A, 3B, and 48 describe specific operations occurring in a particular order. In alternative implementations, certain of the logic operations may be performed in a different order, modified or removed. Morever, steps may be added to the above described logic and still conform to the described implementations. Further, operations described herein may occur sequentially or certain operations may be processed in parallel, or operations described as performed by a single process may be performed by distributed processes.
The logic of FIGS. 3A, 3B, and 48 was described as being implemented in software. This logic may be part of the operating system of the host systems or an application program. In yet further implementations, this logic may be maintained in storage areas managed by the control units or in a read only memory or other hardwired type of device. The preferred logic may be implemented in hard disk drives or in programmable and non-programmable gate array logic.
FIG. 10 illustrates one implementation of the architecture of the computer systems 100 and 102 in accordance with certain implementations of the invention. The computer systems 100 and 102 may implement a computer architecture 1000 having a processor 1002 (e.g., a microprocessor), a memory 1004 (e.g., a volatile memory device), and storage 1006 (e.g., a non-volatile storage, such as magnetic disk drives, optical disk drives, a tape drive, etc.). The storage 1006 may comprise an internal storage device or an attached or network accessible storage. Programs in the storage 1006 are loaded into the memory 1004 and executed by the processor 1002 in a manner known in the art. The architecture further includes a network card 1008 to enable communication with a network. An input device 1010 is used to provide user input to the processor 1002, and may include a keyboard, mouse, pen-stylus, microphone, touch sensitive display screen, or any other activation or input mechanism known in the art. An output device 1012 is capable of rendering information transmitted from the processor 1002, or other component, such as a display monitor, printer, storage, etc.
The foregoing description of the preferred implementations of the invention has been presented for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise form disclosed. Many modifications and variations are possible in light of the above teaching. It is intended that the scope of the invention be limited not by this detailed description, but rather by the claims appended hereto. The above specification, examples and data provide a complete description of the manufacture and use of the composition of the invention. Since many implementations of the invention can be made without departing from the spirit and scope of the invention, the invention resides in the claims hereinafter appended.

Claims (18)

1. A method in a computer system for processing a request for information concerning a hardware device generated at a client with an invoking process executing in a server, the method comprising:
identifying an agent that may process the request by accessing the hardware device;
determining whether the agent should be executed within an address space of the invoking process that would invoke the agent or within a separate address space in the server;
if it is determined that the agent should be executed within a separate address space, executing the agent within a separate address space;
following a failure and restarting of the invoking process, using the invoking process to make a determination whether to execute the identified agent within the invoking process address space or within a separate address space based on the performance of the agent when the agent was executing;
when the invoking process is restarted:
determining which agents are to be executed in one or more separate address spaces based on performance information of those agents when they were executing stored by the invoking process during the execution of those agents; and
executing those agents in the one or more separate address spaces;
wherein the agent is a Common Information Model (CIM) provider; and
wherein the invoking process is a CIM object manager.
2. The method of claim 1, further comprising:
receiving an indication to switch an agent that is executing within the separate address space to an agent that is executing within the address space of the invoking process that invoked the agent;
terminating the agent to be switched; and
executing the agent within the address space of the invoking process.
3. The method of claim 1, wherein the agent is part of a group of agents and further comprising:
terminating each of the agents in the group; and
executing each of the agents in the group within the address space of the invoking process that invoked the agents.
4. The method of claim 1, further comprising:
receiving an indication to switch the agent that is executing within the address space of an invoking process that invoked the agent to the agent that is executing in a separate address space;
quiescing the agent that is executing within the address space of the invoking process; and
executing the agent within the separate address space.
5. The method of claim 1 wherein, following a failure and restarting of the invoking process, the determination whether to execute the identified agent within the invoking process address space or within a separate address space comprises executing the identified agent in the separate address space when the agents fails while executing in the invoking process address space.
6. The method of claim 1 wherein, following a failure and restarting of the invoking process, the determination whether to execute the identified agent within the invoking process address space or within a separate address space comprises executing the identified agent in the invoking process address space when the agent completes successfully while executing in the separate address space.
7. A system operating in a server of a computer system for processing a request generated at a client computer for information concerning a hardware device of the computer system, the system comprising:
a session manager executing in the server;
one or more agents in the server;
one or more invoking processes that can invoke the agents to execute in the server;
means at the session manager for identifying which agent may process the request by accessing the hardware device;
determining at the session manager whether the identified agent should be executed within an address space of an invoking process or within a separate address space in the server;
if it is determined that the agent should be executed within a separate address space, executing the agent within a separate address space; and
means in the session manager and operable following a failure and restarting of the invoking process, for making a determination whether to execute the identified agent within the invoking process address space or within a separate address space based on the performance of the agent when the agent was executing;
when the invoking process is restarted:
means for determining which agents are to be executed in one or more separate address spaces based on performance information of those agents when they were executing stored by the invoking process during the execution of those agents; and
means for executing those agents in the one or more separate address spaces;
wherein each of the agents is a Common Information Model (CIM) provider; and
wherein each of the invoking processes is a CIM object manager.
8. The system of claim 7, further comprising:
means for receiving an indication to switch an agent that is executing within the separate address space to an agent that is executing within the address space of the invoking process that invoked the agent;
means for terminating the agent to be switched; and
means for executing the agent within the address space of the invoking process.
9. The system of claim 7, wherein the agent is part of a group of agents and further comprising:
means for terminating each of the agents in the group; and
means for executing each of the agents in the group within the address space of the invoking process that invoked the agents.
10. The system of claim 7, further comprising:
means for receiving an indication to switch the agent that is executing within the address space of an invoking process that invoked the agent to the agent that is executing in a separate address space;
means for quiescing the agent that is executing within the address space of the invoking process; and
means for executing the agent within the separate address space.
11. The system of claim 7 wherein the means operable following a failure and restarting of the invoking process, for making the determination whether to execute the identified agent within the invoking process address space or within a separate address space comprises means for executing the identified agent in the separate address space when the agents fails while executing in the invoking process address space.
12. The system of claim 7 wherein the means operable following a failure and restarting of the invoking process, for making the determination whether to execute the identified agent within the invoking process address space or within a separate address space comprises means for executing the identified agent in the invoking process address space when the agent completes successfully while executing in the separate address space.
13. An article of manufacture comprising a storage medium having computer readable program code thereon including instructions for processing a request for information concerning a hardware device with an invoking a process executing in a server, wherein the program code is capable of causing operations in the processor of a computer system, the operations comprising:
identifying an agent that may process the request by accessing the hardware device;
determining whether the agent should be executed within an address space of an the invoking process that would invoke the agent or within a separate address space in the server;
if it is determined that the agent should be executed within a separate address space, executing the agent within a separate address space;
following a failure and restarting of the invoking process using the invoking process to make a determination whether to execute the identified agent within the invoking process address space or within a separate address space based on the performance of the agent when the agent was executing; and
when the invoking process is restarted:
determining which agents are to be executed in one or more separate address spaces based on performance information of those agents when they were executing stored by the invoking process during the execution of those agents; and
executing those agents in the one or more separate address spaces; wherein the agent is a Common Information Model (CIM) provider; and wherein the invoking process is a CIM object manager.
14. The article of manufacture of claim 13, the operations further comprising:
receiving an indication to switch an agent that is executing within the separate address space to an agent that is executing within the address space of the invoking process that invoked the agent;
terminating the agent to be switched; and
executing the agent within the address space of the invoking process.
15. The article of manufacture of claim 13, wherein the agent is part of a group of agents and the operations further comprising:
terminating each of the agents in the group; and
executing each of the agents in the group within the address space of the invoking process that invoked the agents.
16. The article of manufacture of claim 13, the operations further comprising:
receiving an indication to switch the agent that is executing within the address space of an invoking process that invoked the agent to the agent that is executing in a separate address space;
quiescing the agent that is executing within the address space of the invoking process; and
executing the agent within the separate address space.
17. The article of manufacture of claim 13 wherein, following a failure and restarting of the invoking process, the determination whether to execute the identified agent within the invoking process address space or within a separate address space comprises executing the identified agent in the separate address space when the agents fails while executing in the invoking process address space.
18. The article of manufacture of claim 13 wherein, following a failure and restarting of the invoking process, the determination whether to execute the identified agent within the invoking process address space or within a separate address space comprises executing the identified agent in the invoking process address space when the agent completes successfully while executing in the separate address space.
US10/202,662 2002-07-23 2002-07-23 Method, system, and article of manufacture for agent processing Expired - Lifetime US7103889B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US10/202,662 US7103889B2 (en) 2002-07-23 2002-07-23 Method, system, and article of manufacture for agent processing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US10/202,662 US7103889B2 (en) 2002-07-23 2002-07-23 Method, system, and article of manufacture for agent processing

Publications (2)

Publication Number Publication Date
US20040216126A1 US20040216126A1 (en) 2004-10-28
US7103889B2 true US7103889B2 (en) 2006-09-05

Family

ID=33298066

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/202,662 Expired - Lifetime US7103889B2 (en) 2002-07-23 2002-07-23 Method, system, and article of manufacture for agent processing

Country Status (1)

Country Link
US (1) US7103889B2 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060047824A1 (en) * 2004-06-30 2006-03-02 Ken Bowler System and method for transferring data in high latency firewalled networks
US7519694B1 (en) * 2005-08-24 2009-04-14 Sun Microsystems, Inc. Method and a system to dynamically update/reload agent configuration data
US20110264824A1 (en) * 2008-08-08 2011-10-27 Jayaraman Venkata Subramanian Enhancement to sip forking for improved user services
US8930475B1 (en) 2012-03-30 2015-01-06 Signiant Inc. Systems and methods for secure cloud-based media file sharing
US9692799B2 (en) 2012-07-30 2017-06-27 Signiant Inc. System and method for sending and/or receiving digital content based on a delivery specification
CN108614799A (en) * 2016-12-13 2018-10-02 通用汽车环球科技运作有限责任公司 The method for carrying out data exchange in real time operating system between main core and secondary core
US10735516B1 (en) 2019-02-15 2020-08-04 Signiant Inc. Cloud-based authority to enhance point-to-point data transfer with machine learning

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3864251B2 (en) * 2002-12-06 2006-12-27 インターナショナル・ビジネス・マシーンズ・コーポレーション Message processing apparatus, message processing method, and message processing program
US20040181529A1 (en) * 2003-03-11 2004-09-16 Sun Microsystems, Inc. Method, system, and program for enabling access to device information
US7725473B2 (en) * 2003-12-17 2010-05-25 International Business Machines Corporation Common information model
US9003059B2 (en) * 2008-03-31 2015-04-07 Microsoft Technology Licensing, Llc Running applications in an online or offline mode based on the availability of the connection to the remote web server
US8776086B1 (en) * 2012-03-08 2014-07-08 Emc Corporation Agent management framework

Citations (109)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2012527A (en) 1931-03-16 1935-08-27 Jr Edward H Batchelder Refrigerator car
US2675228A (en) 1953-02-05 1954-04-13 Edward O Baird Electrical control means for closure devices
US3571677A (en) 1969-12-31 1971-03-23 Itt Single bellows water-cooled vehicle capacitors
US4138692A (en) 1977-09-12 1979-02-06 International Business Machines Corporation Gas encapsulated cooling module
US4228219A (en) 1979-04-26 1980-10-14 Imperial Chemical Industries Limited Aromatic polyether sulfone used as a prime coat for a fluorinated polymer layer
US4558395A (en) 1983-04-27 1985-12-10 Hitachi, Ltd. Cooling module for integrated circuit chips
US4665466A (en) 1983-09-16 1987-05-12 Service Machine Company Low headroom ventilating apparatus for cooling an electrical enclosure
US4721996A (en) 1986-10-14 1988-01-26 Unisys Corporation Spring loaded module for cooling integrated circuit packages directly with a liquid
US4729424A (en) 1985-04-05 1988-03-08 Nec Corporation Cooling system for electronic equipment
US4733331A (en) 1985-09-30 1988-03-22 Jeumont-Schneider Corporation Heat dissipation mechanism for power semiconductor elements
US4791983A (en) 1987-10-13 1988-12-20 Unisys Corporation Self-aligning liquid-cooling assembly
US4809134A (en) 1988-04-18 1989-02-28 Unisys Corporation Low stress liquid cooling assembly
US4870477A (en) 1986-05-23 1989-09-26 Hitachi, Ltd. Integrated circuit chips cooling module having coolant leakage prevention device
US4882654A (en) 1988-12-22 1989-11-21 Microelectronics And Computer Technology Corporation Method and apparatus for adjustably mounting a heat exchanger for an electronic component
US4977444A (en) 1987-10-26 1990-12-11 Hitachi, Ltd. Semiconductor cooling apparatus
US5144531A (en) 1990-01-10 1992-09-01 Hitachi, Ltd. Electronic apparatus cooling system
US5166863A (en) 1991-07-15 1992-11-24 Amdahl Corporation Liquid-cooled assembly of heat-generating devices and method for assembling and disassembling
US5177667A (en) 1991-10-25 1993-01-05 International Business Machines Corporation Thermal conduction module with integral impingement cooling
US5183104A (en) 1989-06-16 1993-02-02 Digital Equipment Corporation Closed-cycle expansion-valve impingement cooling system
US5282847A (en) 1991-02-28 1994-02-01 Medtronic, Inc. Prosthetic vascular grafts with a pleated structure
US5305461A (en) * 1992-04-03 1994-04-19 International Business Machines Corporation Method of transparently interconnecting message passing systems
US5323847A (en) 1990-08-01 1994-06-28 Hitachi, Ltd. Electronic apparatus and method of cooling the same
US5406807A (en) 1992-06-17 1995-04-18 Hitachi, Ltd. Apparatus for cooling semiconductor device and computer having the same
US5465192A (en) 1993-06-21 1995-11-07 Nec Corporation Cooling apparatus for integrated circuit chips for preventing forcible contact between a cooling member and the chips
US5504858A (en) 1993-06-29 1996-04-02 Digital Equipment Corporation Method and apparatus for preserving data integrity in a multiple disk raid organized storage system
US5535094A (en) 1995-04-26 1996-07-09 Intel Corporation Integrated circuit package with an integral heat sink and fan
US5588119A (en) 1993-08-23 1996-12-24 Vincent; Ronald Method for correlating logical device names with a hub port in a local area network
US5659701A (en) * 1991-12-02 1997-08-19 International Business Machines Corporation Apparatus and method for distributed program stack
US5675473A (en) 1996-02-23 1997-10-07 Motorola, Inc. Apparatus and method for shielding an electronic module from electromagnetic radiation
US5701045A (en) 1995-05-31 1997-12-23 Sanyo Denki Co., Ltd. Axial flow air fan having lateral suction and discharge ports for cooling electronic components
US5706668A (en) 1994-12-21 1998-01-13 Hilpert; Bernhard Computer housing with cooling means
US5751933A (en) 1990-09-17 1998-05-12 Dev; Roger H. System for determining the status of an entity in a computer network
US5771388A (en) 1994-05-04 1998-06-23 National Instruments Corporation System and method for mapping driver level event function calls from a process-based driver level program to a session-based instrumentation control driver level system
US5912802A (en) 1994-06-30 1999-06-15 Intel Corporation Ducted opposing bonded fin heat sink blower multi-microprocessor cooling system
US5940269A (en) 1998-02-10 1999-08-17 D-Link Corporation Heat sink assembly for an electronic device
US5950011A (en) 1996-03-01 1999-09-07 Bull S.A. System using designer editor and knowledge base for configuring preconfigured software in an open system in a distributed environment
US5956750A (en) 1996-04-08 1999-09-21 Hitachi, Ltd. Apparatus and method for reallocating logical to physical disk devices using a storage controller, with access frequency and sequential access ratio calculations and display
US6006251A (en) 1995-07-11 1999-12-21 Hitachi, Ltd. Service providing system for providing services suitable to an end user request based on characteristics of a request, attributes of a service and operating conditions of a processor
US6029742A (en) 1994-01-26 2000-02-29 Sun Microsystems, Inc. Heat exchanger for electronic equipment
US6031528A (en) 1996-11-25 2000-02-29 Intel Corporation User based graphical computer network diagnostic tool
US6050327A (en) 1998-03-24 2000-04-18 Lucent Technologies Inc. Electronic apparatus having an environmentally sealed external enclosure
US6058426A (en) * 1997-07-14 2000-05-02 International Business Machines Corporation System and method for automatically managing computing resources in a distributed computing environment
US6067545A (en) 1997-08-01 2000-05-23 Hewlett-Packard Company Resource rebalancing in networked computer systems
US6067559A (en) * 1998-04-23 2000-05-23 Microsoft Corporation Server architecture for segregation of dynamic content generation applications into separate process spaces
US6101616A (en) 1997-03-27 2000-08-08 Bull S.A. Data processing machine network architecture
US6119118A (en) * 1996-05-10 2000-09-12 Apple Computer, Inc. Method and system for extending file system metadata
US6118776A (en) 1997-02-18 2000-09-12 Vixel Corporation Methods and apparatus for fiber channel interconnection of private loop devices
US6125924A (en) 1999-05-03 2000-10-03 Lin; Hao-Cheng Heat-dissipating device
US6130820A (en) 1999-05-04 2000-10-10 Intel Corporation Memory card cooling device
US6135200A (en) 1998-03-11 2000-10-24 Denso Corporation Heat generating element cooling unit with louvers
US6137680A (en) 1998-03-31 2000-10-24 Sanyo Denki Co., Ltd. Electronic component cooling apparatus
US6144379A (en) 1997-11-20 2000-11-07 International Business Machines Corporation Computer controlled user interactive display system for presenting graphs with interactive icons for accessing related graphs
US6151331A (en) 1998-09-23 2000-11-21 Crossroads Systems, Inc. System and method for providing a proxy FARP for legacy storage devices
US6151031A (en) 1996-09-09 2000-11-21 Hewlett-Packard Company Map builder system and method for enabling generic interfacing of an application with a display map generation process in a management system
US6167445A (en) 1998-10-26 2000-12-26 Cisco Technology, Inc. Method and apparatus for defining and implementing high-level quality of service policies in computer networks
US6182142B1 (en) 1998-07-10 2001-01-30 Encommerce, Inc. Distributed access management of information resources
US6205803B1 (en) 1996-04-26 2001-03-27 Mainstream Engineering Corporation Compact avionics-pod-cooling unit thermal control method and apparatus
US6205796B1 (en) 1999-03-29 2001-03-27 International Business Machines Corporation Sub-dew point cooling of electronic systems
US6213194B1 (en) 1997-07-16 2001-04-10 International Business Machines Corporation Hybrid cooling system for electronics module
US6229538B1 (en) 1998-09-11 2001-05-08 Compaq Computer Corporation Port-centric graphic representations of network controllers
US6243747B1 (en) 1995-02-24 2001-06-05 Cabletron Systems, Inc. Method and apparatus for defining and enforcing policies for configuration management in communications networks
US6301605B1 (en) 1997-11-04 2001-10-09 Adaptec, Inc. File array storage architecture having file system distributed across a data processing platform
US6314555B1 (en) * 1997-07-25 2001-11-06 British Telecommunications Public Limited Company Software system generation
US6313990B1 (en) 2000-05-25 2001-11-06 Kioan Cheon Cooling apparatus for electronic devices
US20010044907A1 (en) 2000-05-19 2001-11-22 Fujitsu Limited Information processing apparatus, power saving control method and recording medium for storing power saving control program
US20010043617A1 (en) 2000-05-19 2001-11-22 Mckinnon Martin W. Allocating access across a shared communications medium
US20020019864A1 (en) 1999-12-09 2002-02-14 Mayer J?Uuml;Rgen System and method for managing the configuration of hierarchically networked data processing devices
US6381637B1 (en) 1996-10-23 2002-04-30 Access Co., Ltd. Information apparatus having automatic web reading function
US6392667B1 (en) 1997-06-09 2002-05-21 Aprisma Management Technologies, Inc. Method and apparatus for representing objects as visually discernable entities based on spatial definition and perspective
US6396697B1 (en) 2000-12-07 2002-05-28 Foxconn Precision Components Co., Ltd. Heat dissipation assembly
US20020069377A1 (en) 1998-03-10 2002-06-06 Atsushi Mabuchi Control device and control method for a disk array
US6408336B1 (en) 1997-03-10 2002-06-18 David S. Schneider Distributed administration of access to information
US20020083169A1 (en) 2000-12-21 2002-06-27 Fujitsu Limited Network monitoring system
US6425005B1 (en) 1997-10-06 2002-07-23 Mci Worldcom, Inc. Method and apparatus for managing local resources at service nodes in an intelligent network
US6425007B1 (en) 1995-06-30 2002-07-23 Sun Microsystems, Inc. Network navigation and viewing system for network management system
US20020113816A1 (en) 1998-12-09 2002-08-22 Frederick H. Mitchell Method and apparatus providing a graphical user interface for representing and navigating hierarchical networks
US6438984B1 (en) 2001-08-29 2002-08-27 Sun Microsystems, Inc. Refrigerant-cooled system and method for cooling electronic components
US20020133669A1 (en) 1999-06-11 2002-09-19 Narayan Devireddy Policy based storage configuration
US20020143905A1 (en) 2001-03-30 2002-10-03 Priya Govindarajan Method and apparatus for discovering network topology
US20020143920A1 (en) 2001-03-30 2002-10-03 Opticom, Inc. Service monitoring and reporting system
US6463454B1 (en) 1999-06-17 2002-10-08 International Business Machines Corporation System and method for integrated load distribution and resource management on internet environment
US20020147801A1 (en) 2001-01-29 2002-10-10 Gullotta Tony J. System and method for provisioning resources to users based on policies, roles, organizational information, and attributes
US20020152305A1 (en) 2000-03-03 2002-10-17 Jackson Gregory J. Systems and methods for resource utilization analysis in information management environments
US20020162010A1 (en) 2001-03-15 2002-10-31 International Business Machines Corporation System and method for improved handling of fiber channel remote devices
US6477572B1 (en) 1998-12-17 2002-11-05 International Business Machines Corporation Method for displaying a network topology for a task deployment service
US6487643B1 (en) * 2000-09-29 2002-11-26 Intel Corporation Method and apparatus for preventing starvation in a multi-node architecture
US6496871B1 (en) * 1998-06-30 2002-12-17 Nec Research Institute, Inc. Distributed agent software system and method having enhanced process mobility and communication in a computer network
US6505244B1 (en) 1999-06-29 2003-01-07 Cisco Technology Inc. Policy engine which supports application specific plug-ins for enforcing policies in a feedback-based, adaptive data network
US6526768B2 (en) 2001-07-24 2003-03-04 Kryotech, Inc. Apparatus and method for controlling the temperature of an integrated circuit device
US6542360B2 (en) 2000-06-30 2003-04-01 Kabushiki Kaisha Toshiba Electronic apparatus containing heat generating component, and extension apparatus for extending the function of the electronic apparatus
US6574708B2 (en) * 2001-05-18 2003-06-03 Broadcom Corporation Source controlled cache allocation
US6587343B2 (en) 2001-08-29 2003-07-01 Sun Microsystems, Inc. Water-cooled system and method for cooling electronic components
US6604136B1 (en) 1998-06-27 2003-08-05 Intel Corporation Application programming interfaces and methods enabling a host to interface with a network processor
US6604137B2 (en) 1997-07-31 2003-08-05 Mci Communications Corporation System and method for verification of remote spares in a communications network when a network outage occurs
US6636239B1 (en) 2000-02-24 2003-10-21 Sanavigator, Inc. Method of operating a graphical user interface to selectively enable and disable a datapath in a network
US6658526B2 (en) 1997-03-12 2003-12-02 Storage Technology Corporation Network attached virtual data storage subsystem
US6671776B1 (en) 1999-10-28 2003-12-30 Lsi Logic Corporation Method and system for determining and displaying the topology of a storage array network having multiple hosts and computer readable medium for generating the topology
US6704778B1 (en) 1999-09-01 2004-03-09 International Business Machines Corporation Method and apparatus for maintaining consistency among large numbers of similarly configured information handling servers
US6714936B1 (en) 1999-05-25 2004-03-30 Nevin, Iii Rocky Harry W. Method and apparatus for displaying data stored in linked nodes
US6760761B1 (en) 2000-03-27 2004-07-06 Genuity Inc. Systems and methods for standardizing network devices
US6772204B1 (en) 1996-02-20 2004-08-03 Hewlett-Packard Development Company, L.P. Method and apparatus of providing a configuration script that uses connection rules to produce a configuration file or map for configuring a network device
US6775700B2 (en) 2001-03-27 2004-08-10 Intel Corporation System and method for common information model object manager proxy interface and management
US6799208B1 (en) 2000-05-02 2004-09-28 Microsoft Corporation Resource manager architecture
US6804773B1 (en) * 2000-10-03 2004-10-12 Dell Products L.P. System and method for transferring information over a network
US6823382B2 (en) 2001-08-20 2004-11-23 Altaworks Corporation Monitoring and control engine for multi-tiered service-level management of distributed web-application servers
US6834298B1 (en) 1999-09-21 2004-12-21 Siemens Information And Communication Networks, Inc. System and method for network auto-discovery and configuration
US6845395B1 (en) 1999-06-30 2005-01-18 Emc Corporation Method and apparatus for identifying network devices on a storage network
US6871232B2 (en) 2001-03-06 2005-03-22 International Business Machines Corporation Method and system for third party resource provisioning management
US6978291B2 (en) * 2001-04-30 2005-12-20 Isogon Corporation Method for correlating job-step or execution-process information with corresponding software licensing information

Patent Citations (110)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2012527A (en) 1931-03-16 1935-08-27 Jr Edward H Batchelder Refrigerator car
US2675228A (en) 1953-02-05 1954-04-13 Edward O Baird Electrical control means for closure devices
US3571677A (en) 1969-12-31 1971-03-23 Itt Single bellows water-cooled vehicle capacitors
US4138692A (en) 1977-09-12 1979-02-06 International Business Machines Corporation Gas encapsulated cooling module
US4228219A (en) 1979-04-26 1980-10-14 Imperial Chemical Industries Limited Aromatic polyether sulfone used as a prime coat for a fluorinated polymer layer
US4558395A (en) 1983-04-27 1985-12-10 Hitachi, Ltd. Cooling module for integrated circuit chips
US4665466A (en) 1983-09-16 1987-05-12 Service Machine Company Low headroom ventilating apparatus for cooling an electrical enclosure
US4729424A (en) 1985-04-05 1988-03-08 Nec Corporation Cooling system for electronic equipment
US4733331A (en) 1985-09-30 1988-03-22 Jeumont-Schneider Corporation Heat dissipation mechanism for power semiconductor elements
US4870477A (en) 1986-05-23 1989-09-26 Hitachi, Ltd. Integrated circuit chips cooling module having coolant leakage prevention device
US4721996A (en) 1986-10-14 1988-01-26 Unisys Corporation Spring loaded module for cooling integrated circuit packages directly with a liquid
US4791983A (en) 1987-10-13 1988-12-20 Unisys Corporation Self-aligning liquid-cooling assembly
US4977444A (en) 1987-10-26 1990-12-11 Hitachi, Ltd. Semiconductor cooling apparatus
US4809134A (en) 1988-04-18 1989-02-28 Unisys Corporation Low stress liquid cooling assembly
US4882654A (en) 1988-12-22 1989-11-21 Microelectronics And Computer Technology Corporation Method and apparatus for adjustably mounting a heat exchanger for an electronic component
US5183104A (en) 1989-06-16 1993-02-02 Digital Equipment Corporation Closed-cycle expansion-valve impingement cooling system
US5144531A (en) 1990-01-10 1992-09-01 Hitachi, Ltd. Electronic apparatus cooling system
US5323847A (en) 1990-08-01 1994-06-28 Hitachi, Ltd. Electronic apparatus and method of cooling the same
US5751933A (en) 1990-09-17 1998-05-12 Dev; Roger H. System for determining the status of an entity in a computer network
US5282847A (en) 1991-02-28 1994-02-01 Medtronic, Inc. Prosthetic vascular grafts with a pleated structure
US5166863A (en) 1991-07-15 1992-11-24 Amdahl Corporation Liquid-cooled assembly of heat-generating devices and method for assembling and disassembling
US5177667A (en) 1991-10-25 1993-01-05 International Business Machines Corporation Thermal conduction module with integral impingement cooling
US5659701A (en) * 1991-12-02 1997-08-19 International Business Machines Corporation Apparatus and method for distributed program stack
US5305461A (en) * 1992-04-03 1994-04-19 International Business Machines Corporation Method of transparently interconnecting message passing systems
US5406807A (en) 1992-06-17 1995-04-18 Hitachi, Ltd. Apparatus for cooling semiconductor device and computer having the same
US5465192A (en) 1993-06-21 1995-11-07 Nec Corporation Cooling apparatus for integrated circuit chips for preventing forcible contact between a cooling member and the chips
US5504858A (en) 1993-06-29 1996-04-02 Digital Equipment Corporation Method and apparatus for preserving data integrity in a multiple disk raid organized storage system
US5588119A (en) 1993-08-23 1996-12-24 Vincent; Ronald Method for correlating logical device names with a hub port in a local area network
US6029742A (en) 1994-01-26 2000-02-29 Sun Microsystems, Inc. Heat exchanger for electronic equipment
US5771388A (en) 1994-05-04 1998-06-23 National Instruments Corporation System and method for mapping driver level event function calls from a process-based driver level program to a session-based instrumentation control driver level system
US5912802A (en) 1994-06-30 1999-06-15 Intel Corporation Ducted opposing bonded fin heat sink blower multi-microprocessor cooling system
US5706668A (en) 1994-12-21 1998-01-13 Hilpert; Bernhard Computer housing with cooling means
US6243747B1 (en) 1995-02-24 2001-06-05 Cabletron Systems, Inc. Method and apparatus for defining and enforcing policies for configuration management in communications networks
US5535094A (en) 1995-04-26 1996-07-09 Intel Corporation Integrated circuit package with an integral heat sink and fan
US5701045A (en) 1995-05-31 1997-12-23 Sanyo Denki Co., Ltd. Axial flow air fan having lateral suction and discharge ports for cooling electronic components
US6425007B1 (en) 1995-06-30 2002-07-23 Sun Microsystems, Inc. Network navigation and viewing system for network management system
US6006251A (en) 1995-07-11 1999-12-21 Hitachi, Ltd. Service providing system for providing services suitable to an end user request based on characteristics of a request, attributes of a service and operating conditions of a processor
US6772204B1 (en) 1996-02-20 2004-08-03 Hewlett-Packard Development Company, L.P. Method and apparatus of providing a configuration script that uses connection rules to produce a configuration file or map for configuring a network device
US5675473A (en) 1996-02-23 1997-10-07 Motorola, Inc. Apparatus and method for shielding an electronic module from electromagnetic radiation
US5950011A (en) 1996-03-01 1999-09-07 Bull S.A. System using designer editor and knowledge base for configuring preconfigured software in an open system in a distributed environment
US5956750A (en) 1996-04-08 1999-09-21 Hitachi, Ltd. Apparatus and method for reallocating logical to physical disk devices using a storage controller, with access frequency and sequential access ratio calculations and display
US6205803B1 (en) 1996-04-26 2001-03-27 Mainstream Engineering Corporation Compact avionics-pod-cooling unit thermal control method and apparatus
US6119118A (en) * 1996-05-10 2000-09-12 Apple Computer, Inc. Method and system for extending file system metadata
US6151031A (en) 1996-09-09 2000-11-21 Hewlett-Packard Company Map builder system and method for enabling generic interfacing of an application with a display map generation process in a management system
US6381637B1 (en) 1996-10-23 2002-04-30 Access Co., Ltd. Information apparatus having automatic web reading function
US6031528A (en) 1996-11-25 2000-02-29 Intel Corporation User based graphical computer network diagnostic tool
US6118776A (en) 1997-02-18 2000-09-12 Vixel Corporation Methods and apparatus for fiber channel interconnection of private loop devices
US6408336B1 (en) 1997-03-10 2002-06-18 David S. Schneider Distributed administration of access to information
US6658526B2 (en) 1997-03-12 2003-12-02 Storage Technology Corporation Network attached virtual data storage subsystem
US6101616A (en) 1997-03-27 2000-08-08 Bull S.A. Data processing machine network architecture
US6392667B1 (en) 1997-06-09 2002-05-21 Aprisma Management Technologies, Inc. Method and apparatus for representing objects as visually discernable entities based on spatial definition and perspective
US6058426A (en) * 1997-07-14 2000-05-02 International Business Machines Corporation System and method for automatically managing computing resources in a distributed computing environment
US6213194B1 (en) 1997-07-16 2001-04-10 International Business Machines Corporation Hybrid cooling system for electronics module
US6314555B1 (en) * 1997-07-25 2001-11-06 British Telecommunications Public Limited Company Software system generation
US6604137B2 (en) 1997-07-31 2003-08-05 Mci Communications Corporation System and method for verification of remote spares in a communications network when a network outage occurs
US6067545A (en) 1997-08-01 2000-05-23 Hewlett-Packard Company Resource rebalancing in networked computer systems
US6425005B1 (en) 1997-10-06 2002-07-23 Mci Worldcom, Inc. Method and apparatus for managing local resources at service nodes in an intelligent network
US6301605B1 (en) 1997-11-04 2001-10-09 Adaptec, Inc. File array storage architecture having file system distributed across a data processing platform
US6144379A (en) 1997-11-20 2000-11-07 International Business Machines Corporation Computer controlled user interactive display system for presenting graphs with interactive icons for accessing related graphs
US5940269A (en) 1998-02-10 1999-08-17 D-Link Corporation Heat sink assembly for an electronic device
US20020069377A1 (en) 1998-03-10 2002-06-06 Atsushi Mabuchi Control device and control method for a disk array
US6135200A (en) 1998-03-11 2000-10-24 Denso Corporation Heat generating element cooling unit with louvers
US6050327A (en) 1998-03-24 2000-04-18 Lucent Technologies Inc. Electronic apparatus having an environmentally sealed external enclosure
US6137680A (en) 1998-03-31 2000-10-24 Sanyo Denki Co., Ltd. Electronic component cooling apparatus
US6067559A (en) * 1998-04-23 2000-05-23 Microsoft Corporation Server architecture for segregation of dynamic content generation applications into separate process spaces
US6604136B1 (en) 1998-06-27 2003-08-05 Intel Corporation Application programming interfaces and methods enabling a host to interface with a network processor
US6496871B1 (en) * 1998-06-30 2002-12-17 Nec Research Institute, Inc. Distributed agent software system and method having enhanced process mobility and communication in a computer network
US6182142B1 (en) 1998-07-10 2001-01-30 Encommerce, Inc. Distributed access management of information resources
US6229538B1 (en) 1998-09-11 2001-05-08 Compaq Computer Corporation Port-centric graphic representations of network controllers
US6151331A (en) 1998-09-23 2000-11-21 Crossroads Systems, Inc. System and method for providing a proxy FARP for legacy storage devices
US6167445A (en) 1998-10-26 2000-12-26 Cisco Technology, Inc. Method and apparatus for defining and implementing high-level quality of service policies in computer networks
US6628304B2 (en) 1998-12-09 2003-09-30 Cisco Technology, Inc. Method and apparatus providing a graphical user interface for representing and navigating hierarchical networks
US20020113816A1 (en) 1998-12-09 2002-08-22 Frederick H. Mitchell Method and apparatus providing a graphical user interface for representing and navigating hierarchical networks
US6477572B1 (en) 1998-12-17 2002-11-05 International Business Machines Corporation Method for displaying a network topology for a task deployment service
US6205796B1 (en) 1999-03-29 2001-03-27 International Business Machines Corporation Sub-dew point cooling of electronic systems
US6125924A (en) 1999-05-03 2000-10-03 Lin; Hao-Cheng Heat-dissipating device
US6130820A (en) 1999-05-04 2000-10-10 Intel Corporation Memory card cooling device
US6714936B1 (en) 1999-05-25 2004-03-30 Nevin, Iii Rocky Harry W. Method and apparatus for displaying data stored in linked nodes
US20020133669A1 (en) 1999-06-11 2002-09-19 Narayan Devireddy Policy based storage configuration
US6463454B1 (en) 1999-06-17 2002-10-08 International Business Machines Corporation System and method for integrated load distribution and resource management on internet environment
US6505244B1 (en) 1999-06-29 2003-01-07 Cisco Technology Inc. Policy engine which supports application specific plug-ins for enforcing policies in a feedback-based, adaptive data network
US6845395B1 (en) 1999-06-30 2005-01-18 Emc Corporation Method and apparatus for identifying network devices on a storage network
US6704778B1 (en) 1999-09-01 2004-03-09 International Business Machines Corporation Method and apparatus for maintaining consistency among large numbers of similarly configured information handling servers
US6834298B1 (en) 1999-09-21 2004-12-21 Siemens Information And Communication Networks, Inc. System and method for network auto-discovery and configuration
US6671776B1 (en) 1999-10-28 2003-12-30 Lsi Logic Corporation Method and system for determining and displaying the topology of a storage array network having multiple hosts and computer readable medium for generating the topology
US20020019864A1 (en) 1999-12-09 2002-02-14 Mayer J?Uuml;Rgen System and method for managing the configuration of hierarchically networked data processing devices
US6636239B1 (en) 2000-02-24 2003-10-21 Sanavigator, Inc. Method of operating a graphical user interface to selectively enable and disable a datapath in a network
US20020152305A1 (en) 2000-03-03 2002-10-17 Jackson Gregory J. Systems and methods for resource utilization analysis in information management environments
US6760761B1 (en) 2000-03-27 2004-07-06 Genuity Inc. Systems and methods for standardizing network devices
US6799208B1 (en) 2000-05-02 2004-09-28 Microsoft Corporation Resource manager architecture
US20010044907A1 (en) 2000-05-19 2001-11-22 Fujitsu Limited Information processing apparatus, power saving control method and recording medium for storing power saving control program
US20010043617A1 (en) 2000-05-19 2001-11-22 Mckinnon Martin W. Allocating access across a shared communications medium
US6313990B1 (en) 2000-05-25 2001-11-06 Kioan Cheon Cooling apparatus for electronic devices
US6542360B2 (en) 2000-06-30 2003-04-01 Kabushiki Kaisha Toshiba Electronic apparatus containing heat generating component, and extension apparatus for extending the function of the electronic apparatus
US6487643B1 (en) * 2000-09-29 2002-11-26 Intel Corporation Method and apparatus for preventing starvation in a multi-node architecture
US6804773B1 (en) * 2000-10-03 2004-10-12 Dell Products L.P. System and method for transferring information over a network
US6396697B1 (en) 2000-12-07 2002-05-28 Foxconn Precision Components Co., Ltd. Heat dissipation assembly
US20020083169A1 (en) 2000-12-21 2002-06-27 Fujitsu Limited Network monitoring system
US20020147801A1 (en) 2001-01-29 2002-10-10 Gullotta Tony J. System and method for provisioning resources to users based on policies, roles, organizational information, and attributes
US6871232B2 (en) 2001-03-06 2005-03-22 International Business Machines Corporation Method and system for third party resource provisioning management
US20020162010A1 (en) 2001-03-15 2002-10-31 International Business Machines Corporation System and method for improved handling of fiber channel remote devices
US6775700B2 (en) 2001-03-27 2004-08-10 Intel Corporation System and method for common information model object manager proxy interface and management
US20020143905A1 (en) 2001-03-30 2002-10-03 Priya Govindarajan Method and apparatus for discovering network topology
US20020143920A1 (en) 2001-03-30 2002-10-03 Opticom, Inc. Service monitoring and reporting system
US6978291B2 (en) * 2001-04-30 2005-12-20 Isogon Corporation Method for correlating job-step or execution-process information with corresponding software licensing information
US6574708B2 (en) * 2001-05-18 2003-06-03 Broadcom Corporation Source controlled cache allocation
US6526768B2 (en) 2001-07-24 2003-03-04 Kryotech, Inc. Apparatus and method for controlling the temperature of an integrated circuit device
US6823382B2 (en) 2001-08-20 2004-11-23 Altaworks Corporation Monitoring and control engine for multi-tiered service-level management of distributed web-application servers
US6587343B2 (en) 2001-08-29 2003-07-01 Sun Microsystems, Inc. Water-cooled system and method for cooling electronic components
US6438984B1 (en) 2001-08-29 2002-08-27 Sun Microsystems, Inc. Refrigerant-cooled system and method for cooling electronic components

Non-Patent Citations (19)

* Cited by examiner, † Cited by third party
Title
"Specification for CIM Operations over HTTP", Distributed Management Task Force, Inc., May 2, 2002, pp. 1-102.
Anonymous, "Javaone-Sun's 2000 Worldwide Java Developer Conference", Nov. 24, 2003.
Anonymous, "Jini: New Technology for a Networked World", Jun. 30, 1999, p. pp. 1-5.
Anonymous, "Jiro Technical Overview White Paper, Free White Papers/Tech Bulletins, Sun-Solve Search Result", Jun. 28, 2000.
Anonymous, "Jiro Technology Technical Overview", 2000.
Anonymous, "Managed Storage Networks White Paper, Free White Papers/Tech Bulletins, Sun-Solve Search Result", Jun. 28, 2000.
Anonymous, "Pattern: Blackboard", https://www.vico.org/pages/Patrons%20Blackboard, Jul. 12, 2002, p. pp. 1-2.
Anonymous, "Sun's Managed Storage Networks", 2000.
Blake "An agent-based cross-organizational; workflow architecture in support of web services" 2002 IEEE, pp. 1-6. *
Brent Knight, "Reducing the Complexity of Managing Private Loops", / /, p. 1-11.
Bushman, "Pattern Oriented Software Architecture, A System of Patterns", / /, p. pp. 71-95, John Wiley & Sons, New York.
Business Editors & High-Tech Writers, "Tivoli Simplifies SAN Management", Business Wire, Oct. 9, 2000, New York.
Cazalens et al. "A web site indexing process for an Internet information retrieval agent system" 2000 IEEE, pp. 254-258.□□. *
Dupuy, et al., "Netmate: A Network Management Environment", IEEE Network, Mar. 5, 1991, p. 35-40, 43, New York.
Finke, et al., "Distribution and Inheritance in the HERON Approach to Heterogeneous Computing", Proceedings of the Int'l Conf on Distributed Computing Systems, May 25, 1993, p. 399-408, IEEE Comp Soc Press, Los Alamitos.
Huang "Using Intelligent Agents to manage Fuzzy Business Process" 2001 IEEE, pp. 508-523. *
Martin, Jean-Christophe, "Policy-Based Networks", Date Unknown.
Mukfopadhy et al. "Multi-agent Marko v decision processes with limited agent communication" 2001 IEEE, pp. 7-12.□□. *
Richard, "Fibre Channel as a Network Backbone", WESCON/94, IDEA/Microelectronics, Conference record Anaheim, p. 653-659, New York.

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060047824A1 (en) * 2004-06-30 2006-03-02 Ken Bowler System and method for transferring data in high latency firewalled networks
US7526557B2 (en) * 2004-06-30 2009-04-28 Signiant, Inc. System and method for transferring data in high latency firewalled networks
US20090182846A1 (en) * 2004-06-30 2009-07-16 Signiant, Inc. System and method for transferring data in high latency firewalled networks
US8667145B2 (en) 2004-06-30 2014-03-04 Signiant, Inc. System and method for transferring data in high latency firewalled networks
US7519694B1 (en) * 2005-08-24 2009-04-14 Sun Microsystems, Inc. Method and a system to dynamically update/reload agent configuration data
US20110264824A1 (en) * 2008-08-08 2011-10-27 Jayaraman Venkata Subramanian Enhancement to sip forking for improved user services
US8930475B1 (en) 2012-03-30 2015-01-06 Signiant Inc. Systems and methods for secure cloud-based media file sharing
US9596216B1 (en) 2012-03-30 2017-03-14 Signiant Inc. Systems and methods for secure cloud-based media file sharing
US9830330B2 (en) 2012-03-30 2017-11-28 Signiant Inc. Systems and methods for secure cloud-based media file sharing
US9692799B2 (en) 2012-07-30 2017-06-27 Signiant Inc. System and method for sending and/or receiving digital content based on a delivery specification
CN108614799A (en) * 2016-12-13 2018-10-02 通用汽车环球科技运作有限责任公司 The method for carrying out data exchange in real time operating system between main core and secondary core
CN108614799B (en) * 2016-12-13 2021-10-08 通用汽车环球科技运作有限责任公司 Method for exchanging data between primary core and secondary core in real-time operating system
US10735516B1 (en) 2019-02-15 2020-08-04 Signiant Inc. Cloud-based authority to enhance point-to-point data transfer with machine learning
US11811871B2 (en) 2019-02-15 2023-11-07 Signiant Inc. Cloud-based authority to enhance point-to-point data transfer with machine learning

Also Published As

Publication number Publication date
US20040216126A1 (en) 2004-10-28

Similar Documents

Publication Publication Date Title
CA2279382C (en) Web request broker controlling multiple processes
US6237005B1 (en) Web server mechanism for processing multiple transactions in an interpreted language execution environment
EP0613274B1 (en) Socket structure for concurrent multiple protocol access
US7062516B2 (en) Methods, systems, and articles of manufacture for implementing a runtime logging service storage infrastructure
EP1303096B1 (en) Virtual network with adaptive dispatcher
AU742156B2 (en) Distributed web application server
US8028056B1 (en) Server monitoring framework
US6687831B1 (en) Method and apparatus for multiple security service enablement in a data processing system
US20080256248A1 (en) Single server access in a multiple tcp/ip instance environment
US7003527B1 (en) Methods and apparatus for managing devices within storage area networks
US8060891B2 (en) Management of external hardware appliances in a distributed operating system
US20080294764A1 (en) Storage medium bearing hba information provision program, hba information provision method and hba information provision apparatus
US8412901B2 (en) Making automated use of data volume copy service targets
US7899897B2 (en) System and program for dual agent processes and dual active server processes
US20070112812A1 (en) System and method for writing data to a directory
JP2002505461A (en) Transport processing method and apparatus in event-based distributed system
US7743167B2 (en) Method and system for servicing requests in a dynamic cluster
US7103889B2 (en) Method, system, and article of manufacture for agent processing
US6058425A (en) Single server access in a multiple TCP/IP instance environment
US7107591B1 (en) Task-specific flexible binding in a software system
US20080281969A1 (en) Controlling access to versions of application software by a server, based on site ID
US20100070552A1 (en) Providing a Socket Connection between a Java Server and a Host Environment
US8122114B1 (en) Modular, dynamically extensible, and integrated storage area network management system
US8335820B1 (en) Techniques for persistence of connections
US20080140687A1 (en) Socket structure simultaneously supporting both toe and ethernet network interface card and method of forming the socket structure

Legal Events

Date Code Title Description
AS Assignment

Owner name: SUN MICROSYSTEMS, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HILTGEN, DANIEL K.;REEL/FRAME:013142/0132

Effective date: 20020716

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

AS Assignment

Owner name: ORACLE AMERICA, INC., CALIFORNIA

Free format text: MERGER AND CHANGE OF NAME;ASSIGNORS:ORACLE USA, INC.;SUN MICROSYSTEMS, INC.;ORACLE AMERICA, INC.;REEL/FRAME:037302/0616

Effective date: 20100212

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553)

Year of fee payment: 12