Jump to: navigation, search

EclipseLink/UserGuide/JPA/Basic JPA Development/Caching/Coordination

< EclipseLink‎ | UserGuide‎ | JPA‎ | Basic JPA Development‎ | Caching
Revision as of 15:08, 22 May 2012 by James.sutherland.oracle.com (Talk | contribs) (Clustering and Cache Coordination)

EclipseLink JPA

Clustering and Cache Coordination

An application cluster is a set of middle tier server machines or VMs servicing requests for a single application, or set of applications. Multiple servers are used to increase the scalability of the application and/or to provide fault tolerance and high availability. Typically the same application will be deployed to all of the servers in the cluster and application requests will be load balanced across the set of servers. The application cluster will access a single database, or a database cluster. An application cluster may allow new servers to be added to increase scalability, and for servers to be removed such as for updates and servicing.

Application clusters can consist of Java EE servers, web containers, or Java server applications.

EclipseLink can function in any clustered environment. The main issue in a clustered environment is utilizing a shared persistence unit (L2) cache. If you are using a shared cache (enabled by default in EclipseLink), then each server will maintain its own cache, and each caches data can get out of sync with the other servers and the database.

EclipseLink provides cache coordination in a clustered environment to ensure the servers caches are is sync.

There are also many other solutions to caching in a clustered environment, including:

  • Disable the shared cache (through setting @Cacheable(false), or @Cache(isolation=ISOLATED)).
  • Only cache read-only objects.
  • Set a cache invalidation timeout to reduce stale data.
  • Use refreshing on objects/queries when fresh data is required.
  • Use optimistic locking to ensure write consistency (writes on stale data will fail, and will automatically invalidate the cache).
  • Use a distributed cache (such as Oracle TopLink Grid's integration of EclipseLink with Oracle Coherence).
  • Use database events to invalidate changed data in the cache (such as EclipseLink's support for Oracle DCN/QCN).

Cache coordination enables a set of persistence units deployed to different servers in the cluster (or on the same server) to synchronize their changes. Cache coordination works by each persistence unit on each server in the cluster being able to broadcast notification of transactional object changes to the other persistence units in the cluster. EclipseLink supports cache coordination over RMI and JMS. The cache coordination framework is also extensible so other options could be developed.

Cache coordination works by broadcasting changes for each transaction to the other servers in the cluster. Each other server will receive the change notification, and either invalidate the changed objects in their cache, or update the cached objects state with the changes. Cache coordination occurs after the database commit, so only committed changes are broadcast.

Cache coordination greatly reduces to chance of an application getting stale data, but does not eliminate the possibility. Optimistic locking should still be used to ensure data integrity. Even in a single server application stale data is still possible within a persistence context unless pessimistic locking is used. Optimistic (or pessimistic) locking is always required to ensure data integrity in any multi-user system.

Cache coordination is configured using persistence unit properties. The following cache coordination properties are supported:

Cache Coordination Persistence Unit Properties
Attribute Description Default Required?
eclipselink.cache.coordination.protocol Enable cache coordination using the communication protocol:
  • rmi - Use Java RMI to broadcast changes.
  • rmi-iiop - Use Java RMI over CORBA IIOP to broadcast changes.
  • jms - Use the Java Messaging Service to broadcast changes.
  • jms-publishing - Allows an EJB MessageDrivenBean to be used to broadcast changes. The MDB must be configured separately.
  • <class-name> - The fully qualified class name of a custom implementation of a TransportManager.
no coordination Required
eclipselink.cache.coordination.channel Sets the channel for cache coordination. All persistence units using the same channel will be coordinated. EclipseLinkCommandChannel Optional
eclipselink.cache.coordination.propagate-asynchronously Configures if changes are broadcast using a separate thread. If set to false the transaction will wait for all servers to be notified before returning.

Note that JMS is always asynchronous.

true Optional
eclipselink.cache.coordination.remove-connection-on-error Set if a connection should be removed if a communication error occurs when coordinating with it.

This is normally used for RMI coordination in case a server goes down (it will reconnect when it comes back up).

true Optional
eclipselink.cache.coordination.naming-service Set the naming service to use to look-up and register the cache coordination service, either:
  • jndi - The server's Java Naming and Directory Interface is used. If the server's JNDI is replicated in an application server cluster, then the url option is not required.
  • rmi - The RMI registry is used. This can be used if JNDI is not available, or the JNDI implementation does not support RMI objects.
jndi Optional
eclipselink.cache.coordination.jndi.user Set the application server user name to connect to JNDI with. This is only required if JNDI requires authentication. no authentication Optional
eclipselink.cache.coordination.jndi.password Set the application server user password to connect to JNDI with. This is only required if JNDI requires authentication.

This is normally not required if connecting to a local service.

no authentication Optional
eclipselink.cache.coordination.rmi.url Only required by RMI cache coordination. Sets the URL of the host server. This is the URL that other cluster members should use to connect to this host. This may not be required in a clustered environment where JNDI is replicated. This can also be set as a System property or using a SessionCustomizer to avoid a separate persistence.xml per server. local Optional
eclipselink.cache.coordination.rmi.multicast-group Only used for RMI coordination. Sets the multicast socket group address. The multicast group is used to find other members of the cluster. Optional
eclipselink.cache.coordination.rmi.multicast-group.port Only used for RMI coordination. Sets the multicast socket group port. The multicast group is used to find other members of the cluster. 3121 Optional
eclipselink.cache.coordination.rmi.announcement-delay Only used for RMI coordination. Sets the number of milliseconds to wait for announcements from other cluster members on start-up. 1000 Optional
eclipselink.cache.coordination.rmi.packet-time-to-live Only used for RMI coordination. Sets the multicast socket packet time to live. The multicast group is used to find other members of the cluster. Set the number of hops the data packets of the session announcement will take before expiring. The default is 2, a hub and an interface card to prevent the data packets from leaving the local network.

Note that if sessions are hosted on different LANs that are part of WAN, the announcement sending by one session may not reach other sessions. In this case, consult your network administrator for the right time-to-live value or test your network by increase the value until sessions receive announcement sent by others.

2 Optional

Version: 2.4 DRAFT
Other versions...