herence.spring.coherence-spring-tests.4.3.0.source-code.coherence-server-junit-extension-cache-config.xml Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of coherence-spring-tests Show documentation
Show all versions of coherence-spring-tests Show documentation
Integration Tests for Coherence Spring
The newest version!
<?xml version="1.0"?> <!-- Copyright (c) 2000, 2022, Oracle and/or its affiliates. Licensed under the Universal Permissive License v 1.0 as shown at https://oss.oracle.com/licenses/upl. --> <!-- This XML document is a default Coherence Cache Configuration deployment descriptor that should be customized (or replaced) for your particular caching and deployment requirements. This configuration is usable in servers, proxies, clustered clients, and non-clustered extend clients. When used from within a server such a DefaultCacheServer, the server will automatically host a storage enabled cache service as well as a proxy service to allow extend clients to access the caches. Clients using the configuration are storage disabled by default. This configuration defines a number of inter-related cache schemes: - server - this scheme defines the storage tier for all caches - thin-direct - this scheme is for use by cluster members to access the caches hosted by the "server" scheme - near-direct - this scheme adds near caching to "thin-direct" - thin-remote - conceptually similar to "thin-direct" but for use by extend clients - near-remote - conceptually similar to "near-direct" but for use by extend clients - topic-server - this scheme defines the storage tier for all topics The default scheme for caches is "near-direct". This default can be overridden via two system properties. The "coherence.profile" system property controls the first portion of the scheme name and defines the approach used for in-process caching, i.e. "near" (on-demand) or "thin" (none). The "coherence.client" system property controls how a client connects to the cluster, i.e. "direct" (cluster member) or "remote" (extend client). Note: System properties defined within this cache configuration are specific to this configuration and are not meaningful to other cache configurations unless similarly defined there. --> <cache-config xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns="http://xmlns.oracle.com/coherence/coherence-cache-config" xsi:schemaLocation="http://xmlns.oracle.com/coherence/coherence-cache-config coherence-cache-config.xsd"> <caching-scheme-mapping> <cache-mapping> <cache-name>*</cache-name> <scheme-name>${coherence.profile near}-${coherence.client direct}</scheme-name> </cache-mapping> </caching-scheme-mapping> <topic-scheme-mapping> <topic-mapping> <topic-name>*</topic-name> <scheme-name>topic-server</scheme-name> </topic-mapping> </topic-scheme-mapping> <caching-schemes> <!-- near caching scheme for clustered clients --> <near-scheme> <scheme-name>near-direct</scheme-name> <front-scheme> <local-scheme> <high-units>{front-limit-entries 10000}</high-units> </local-scheme> </front-scheme> <back-scheme> <distributed-scheme> <scheme-ref>thin-direct</scheme-ref> </distributed-scheme> </back-scheme> </near-scheme> <!-- near caching scheme for extend clients --> <near-scheme> <scheme-name>near-remote</scheme-name> <scheme-ref>near-direct</scheme-ref> <back-scheme> <remote-cache-scheme> <scheme-ref>thin-remote</scheme-ref> </remote-cache-scheme> </back-scheme> </near-scheme> <!-- remote caching scheme for accessing the proxy from extend clients --> <remote-cache-scheme> <scheme-name>thin-remote</scheme-name> <service-name>RemoteCache</service-name> <proxy-service-name>Proxy</proxy-service-name> </remote-cache-scheme> <!-- partitioned caching scheme for clustered clients --> <distributed-scheme> <scheme-name>thin-direct</scheme-name> <scheme-ref>server</scheme-ref> <local-storage system-property="coherence.distributed.localstorage">false</local-storage> <autostart>false</autostart> </distributed-scheme> <!-- partitioned caching scheme for servers --> <distributed-scheme> <scheme-name>server</scheme-name> <service-name>PartitionedCache</service-name> <local-storage system-property="coherence.distributed.localstorage">true</local-storage> <backing-map-scheme> <local-scheme> <high-units>{back-limit-bytes 0B}</high-units> </local-scheme> </backing-map-scheme> <autostart>true</autostart> </distributed-scheme> <distributed-scheme> <scheme-name>extend</scheme-name> <lease-granularity>member</lease-granularity> <backing-map-scheme> <local-scheme/> </backing-map-scheme> <autostart>true</autostart> </distributed-scheme> <!-- partitioned topic scheme for servers --> <paged-topic-scheme> <scheme-name>topic-server</scheme-name> <service-name>PartitionedTopic</service-name> <local-storage system-property="coherence.distributed.localstorage">true</local-storage> <autostart>true</autostart> <high-units>{topic-high-units-bytes 0B}</high-units> </paged-topic-scheme> <!-- proxy scheme that allows extend clients to connect to the cluster over TCP/IP --> <proxy-scheme> <service-name>Proxy</service-name> <autostart>true</autostart> </proxy-scheme> <!-- custom named proxy service added --> <proxy-scheme> <service-name>ExtendTcpCacheService</service-name> <autostart>true</autostart> </proxy-scheme> </caching-schemes> </cache-config>
© 2015 - 2025 Weber Informatics LLC | Privacy Policy