参考指南
4. 文档结构
以下章节解释了 Spring Data for Apache Geode 提供的核心功能:
-
使用 Spring 容器引导 Apache Geode 描述了为配置、初始化和访问提供的配置支持 Apache Geode 缓存、区域和相关的分布式系统组件。
-
使用 Apache Geode API 介绍了 Apache Geode API 与各种数据访问功能之间的集成 在 Spring 中可用,例如基于模板的数据访问、异常转换、事务管理和缓存。
-
使用 Apache Geode 序列化介绍了 Apache Geode 对托管对象的序列化和反序列化的增强功能。
-
POJO 映射描述了使用 Spring Data 对存储在 Apache Geode 中的 POJO 的持久性映射。
-
Spring Data for Apache Geode 存储库描述了如何创建和使用 Spring Data Repositories 来访问数据 使用基本 CRUD 和简单的查询作存储在 Apache Geode 中。
-
函数执行的注释支持描述了如何使用注释创建和使用 Apache Geode 函数 在数据所在的位置执行分布式计算。
-
连续查询 (CQ) 介绍如何使用 Apache Geode 的连续查询 (CQ) 功能 根据 Apache Geode 的 OQL(对象查询语言)。
-
在 Apache Geode 中引导 Spring ApplicationContext 描述了如何配置和引导 Spring
ApplicationContext
在 Apache Geode 服务器中运行Gfsh
. -
示例应用程序 描述了随发行版提供的示例,以说明各种功能 在 Spring Data for Apache Geode 中可用。
5. 使用 Spring 容器引导 Apache Geode
Spring Data for Apache Geode 提供 Apache Geode 内存数据网格 (IMDG) 的完整配置和初始化 使用 Spring IoC 容器。该框架包含多个类,以帮助简化 Apache Geode 组件,包括:缓存、区域、索引、DiskStores、函数、WAN 网关、 持久化备份和其他几个分布式系统组件,以支持各种应用程序用例 以最小的努力。
本部分假定您对 Apache Geode 有基本的了解。有关更多信息,请参阅 Apache Geode 产品文档。 |
5.1. 使用 Spring 而不是 Apache Geode 的优势cache.xml
Spring Data for Apache Geode 的 XML 命名空间支持 Apache Geode 内存数据网格 (IMDG) 的完整配置。 XML 命名空间是在 Spring 上下文中配置 Apache Geode 以正确管理 Spring 容器中的 Apache Geode 生命周期。在 Spring 中配置 Apache Geode 的另一种方法 context 使用基于 annotation 的配置。
虽然支持 Apache Geode 的原生cache.xml
由于遗留原因而保留,Apache Geode 应用程序开发人员
鼓励使用 XML 配置的人在 Spring XML 中执行所有作,以利用许多美妙的功能
Spring 必须提供模块化的 XML 配置、属性占位符和覆盖,
SPEL (Spring Expression Language) 和环境配置文件。
在 XML 命名空间的背后,Spring Data for Apache Geode 广泛使用了 Spring 的FactoryBean
pattern 来简化创建,
配置和 Apache Geode 组件的初始化。
Apache Geode 提供了多个回调接口,例如CacheListener
,CacheLoader
和CacheWriter
,
允许开发人员添加自定义事件处理程序。使用 Spring 的 IoC 容器,你可以配置这些回调
作为普通的 Spring bean 并将它们注入到 Apache Geode 组件中。这是
本地cache.xml
,它提供的配置选项相对有限,并且需要回调来实现
Apache Geode 的Declarable
界面(请参阅布线Declarable
组件以查看您仍然可以如何使用Declarables
在 Spring 的容器中)。
此外,IDE (如 Spring Tool Suite(STS))为 Spring XML 名称空间提供了出色的支持。 包括代码补全、弹出注释和实时验证。
5.2. 使用 Core 命名空间
为了简化配置, Spring Data for Apache Geode 提供了一个专用的 XML 命名空间,用于配置核心 Apache Geode
组件。可以使用 Spring 的标准<bean>
定义。然而
所有 Bean 属性都通过 XML 名称空间公开,因此使用原始 Bean 定义几乎没有什么好处。
有关 Spring 中基于 XML Schema 的配置的更多信息,请参阅 Spring Framework 参考文档中的附录。 |
Spring Data Repository 支持使用单独的 XML 命名空间。有关更多信息,请参阅 Spring Data for Apache Geode 存储库 了解如何为 Apache Geode 存储库配置 Spring Data。 |
要使用 Spring Data for Apache Geode XML 命名空间,请在 Spring XML 配置元数据中声明它。 如下例所示:
<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:gfe="https://www.springframework.org/schema/geode" (1)(2)
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
https://www.springframework.org/schema/geode https://www.springframework.org/schema/geode/spring-geode.xsd (3)
">
<bean id ... >
<gfe:cache ...> (4)
</beans>
1 | Spring Data for Apache Geode XML 命名空间前缀。任何名称都有效,但是,在本参考文档中,gfe 被使用。 |
2 | XML 命名空间前缀映射到 URI。 |
3 | XML 命名空间 URI 位置。请注意,即使该位置指向外部地址(确实存在 并且有效),则 Spring 在本地解析模式,因为它包含在 Spring Data for Apache Geode 库中。 |
4 | 使用 XML 命名空间和gfe 前缀。 |
您可以将默认命名空间从
|
5.3. 使用数据访问命名空间
除了核心 XML 命名空间 (gfe
),Spring Data for Apache Geode 提供了一个数据访问 XML 命名空间 (gfe-data
),
这主要是为了简化 Apache Geode 客户端应用程序的开发。此命名空间
目前包含对 Apache Geode 存储库和函数执行的支持,以及<datasource>
标签,它提供了一种便捷的连接到
Apache Geode 集群。
5.3.1. 连接到 Apache Geode 的简单方法
对于许多应用程序,使用默认值与 Apache Geode 数据网格的基本连接就足够了。
适用于 Apache Geode 的 Spring Data<datasource>
标签提供了一种访问数据的简单方法。数据源会创建一个ClientCache
和连接Pool
.此外,它还会查询集群服务器中所有现有的根区域,并在
每个客户端都有一个(空的)客户端区域代理。
<gfe-data:datasource>
<locator host="remotehost" port="1234"/>
</gfe-data:datasource>
这<datasource>
标记在语法上类似于<gfe:pool>
.它可以配置一个或多个嵌套locator
或server
元素连接到现有数据网格。此外,所有属性都可用于配置池
受支持。此配置会自动为集群成员上定义的每个区域创建 client Region bean
连接到 Locator,以便它们可以被 Spring Data 映射注释(GemfireTemplate
)
并自动连接到应用程序类中。
当然,您可以显式配置客户端区域。例如,如果要在本地内存中缓存数据, 如下例所示:
<gfe-data:datasource>
<locator host="remotehost" port="1234"/>
</gfe-data:datasource>
<gfe:client-region id="Example" shortcut="CACHING_PROXY"/>
5.4. 配置缓存
要使用 Apache Geode,您需要创建新缓存或连接到现有缓存。使用当前版本
中,每个 VM 只能有一个打开的缓存(更严格地说,每个ClassLoader
).在大多数情况下,
缓存只能创建一次。
本节介绍对等节点的创建和配置Cache 成员,适用于点对点 (P2P)
拓扑和缓存服务器。一个Cache 成员还可用于独立应用程序和集成测试。
但是,在典型的生产系统中,大多数应用程序进程充当缓存客户端,从而创建一个ClientCache 实例。配置 Apache Geode ClientCache 和 Client Region 部分中对此进行了介绍。 |
对等体Cache
with default 配置可以通过以下简单声明创建:
<gfe:cache/>
在 Spring 容器初始化期间,任何ApplicationContext
包含此缓存定义会注册一个CacheFactoryBean
,这会创建一个名为gemfireCache
,它引用 Apache GeodeCache
实例。
此 Bean 引用现有的Cache
或者,如果尚不存在,则为新创建的 VPN。由于没有
其他属性,新创建的Cache
应用默认缓存配置。
所有依赖于Cache
请遵循此命名约定,因此无需显式声明
这Cache
Dependency。如果您愿意,可以使用cache-ref
属性
通过各种 SDG XML 命名空间元素。此外,您还可以使用id
属性
如下:
<gfe:cache id="myCache"/>
A Apache GeodeCache
可以使用 Spring 进行完全配置。但是,Apache Geode 的本机 XML 配置
文件cache.xml
。对于需要本地配置 Apache Geode 缓存的情况,
您可以使用cache-xml-location
属性
如下:
<gfe:cache id="cacheConfiguredWithNativeCacheXml" cache-xml-location="classpath:cache.xml"/>
在此示例中,如果需要创建缓存,它将使用名为cache.xml
位于 Classpath 根中
进行配置。
该配置利用了 Spring 的Resource abstraction 来查找文件。这Resource abstraction 允许使用各种搜索模式,具体取决于运行时环境
或在资源位置中指定的前缀(如果有)。 |
除了引用外部 XML 配置文件之外,您还可以指定使用 Spring 的任何 Apache Geode System 属性Properties
支持功能。
例如,您可以使用properties
元素定义在util
命名空间来定义Properties
径直
或从属性文件加载属性,如下所示:
<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:gfe="https://www.springframework.org/schema/geode"
xmlns:util="http://www.springframework.org/schema/util"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
https://www.springframework.org/schema/geode https://www.springframework.org/schema/geode/spring-geode.xsd
http://www.springframework.org/schema/util https://www.springframework.org/schema/util/spring-util.xsd
">
<util:properties id="gemfireProperties" location="file:/path/to/gemfire.properties"/>
<gfe:cache properties-ref="gemfireProperties"/>
</beans>
建议使用属性文件来外部化特定于环境的设置 在应用程序配置之外。
仅当需要创建新缓存时,缓存设置才适用。如果 VM 中已存在打开的缓存,则 这些设置将被忽略。 |
5.4.1. 高级缓存配置
对于高级缓存配置,cache
元素提供了许多作为属性公开的配置选项
或子元素,如下清单所示:
(1)
<gfe:cache
cache-xml-location=".."
properties-ref=".."
close="false"
copy-on-read="true"
critical-heap-percentage="90"
eviction-heap-percentage="70"
enable-auto-reconnect="false" (2)
lock-lease="120"
lock-timeout="60"
message-sync-interval="1"
pdx-serializer-ref="myPdxSerializer"
pdx-persistent="true"
pdx-disk-store="diskStore"
pdx-read-serialized="false"
pdx-ignore-unread-fields="true"
search-timeout="300"
use-bean-factory-locator="true" (3)
use-cluster-configuration="false" (4)
>
<gfe:transaction-listener ref="myTransactionListener"/> (5)
<gfe:transaction-writer> (6)
<bean class="org.example.app.gemfire.transaction.TransactionWriter"/>
</gfe:transaction-writer>
<gfe:gateway-conflict-resolver ref="myGatewayConflictResolver"/> (7)
<gfe:jndi-binding jndi-name="myDataSource" type="ManagedDataSource"/> (8)
</gfe:cache>
1 | 属性支持各种缓存选项。有关此示例中所示任何内容的更多信息,
请参阅 Apache Geode 产品文档。
这close 属性确定在关闭 Spring 应用程序上下文时是否应关闭缓存。
默认值为true .但是,对于多个应用程序上下文使用缓存
(常见于 Web 应用程序),请将此值设置为false . |
2 | 设置enable-auto-reconnect 属性设置为true (默认值为false ) 允许断开连接的 Apache Geode 成员
自动重新连接并重新加入 Apache Geode 集群。
有关更多详细信息,请参阅 Apache Geode 产品文档。 |
3 | 设置use-bean-factory-locator 属性设置为true (默认为false ) 仅在
Spring (XML) 配置元数据和 Apache Geodecache.xml 用于配置 Apache Geode 缓存节点
(无论是客户端还是对等体)。此选项允许 Apache Geode 组件(例如CacheLoader ) 表示为cache.xml 与 bean 自动连接(例如DataSource ) 中定义的 )。此选项通常为
与cache-xml-location . |
4 | 设置use-cluster-configuration 属性设置为true (默认值为false ) 使 Apache Geode 成员能够
从 Locator 中检索基于集群的常见共享配置。
有关更多详细信息,请参阅 Apache Geode 产品文档。 |
5 | 示例TransactionListener callback 声明。引用的 Bean 必须实现 TransactionListener。
一个TransactionListener 可以实现来处理与事务相关的事件(例如 afterCommit 和 afterRollback)。 |
6 | 示例TransactionWriter callback 声明。Bean 必须实现 TransactionWriter。
这TransactionWriter 是可以否决事务的回调。 |
7 | 示例GatewayConflictResolver callback 声明。引用的 Bean
必须实现 https://geode.apache.org/releases/latest/javadoc/org/apache/geode/cache/util/GatewayConflictResolver.html [GatewayConflictResolver]。
一个GatewayConflictResolver 是一个Cache -level 插件,该插件被调用以决定如何处理事件
源自其他系统并通过 WAN 网关 到达。
它提供分布式 Region 创建服务。 |
8 | 声明 JNDI 绑定以在 Apache Geode 事务中登记外部 DataSource。 |
启用 PDX 序列化
前面的示例包括许多与 Apache Geode 的增强序列化框架 PDX 相关的属性。
虽然对 PDX 的完整讨论超出了本参考指南的范围,但重要的是要注意 PDX
通过注册PdxSerializer
,该参数通过设置pdx-serializer
属性。
Apache Geode 提供了一个实现类 (org.apache.geode.pdx.ReflectionBasedAutoSerializer
) 使用
Java 反射。但是,开发人员通常会提供自己的实现。属性的值
只是对实现PdxSerializer
接口。
有关序列化支持的更多信息,请参阅使用 Apache Geode 序列化。
启用自动重新连接
在设置<gfe:cache enable-auto-reconnect="[true|false*]>
属性设置为true
.
通常,只有在使用 Spring Data for Apache Geode 的 XML 命名空间来配置
并引导添加到集群的新非应用程序 Apache Geode 服务器。换句话说,“auto-reconnect”
当 Spring Data for Apache Geode 用于开发和构建 Apache Geode 应用程序时,不应启用该
成为同行Cache
Apache Geode 集群的成员。
此限制的主要原因是大多数 Apache Geode 应用程序都使用对 Apache Geode 的引用Cache
或 Regions 以执行数据访问作。这些引用由 Spring 容器 “注入”
到应用程序组件(如 Repositories)中供应用程序使用。当对等成员强制
与集群的其余部分断开连接,可能是因为 Peer 成员变得无响应,或者
网络分区将一个或多个对等成员分成一个组,该组太小而无法作为独立的
分布式系统、对等成员关闭以及所有 Apache Geode 组件引用(缓存、区域、
和其他)变为无效。
本质上,每个 peer member 中的当前 forced disconnect processing logic 从头开始拆除系统。 JGroups 堆栈关闭,分布式系统处于关闭状态,最后,高速缓存关闭。 实际上,所有内存引用都会过时并丢失。
与分布式系统断开连接后,对等成员会进入 “reconnecting” 状态,并定期 尝试重新加入分布式系统。如果对等成员成功重新连接,则该成员将重建其 “视图” 的分布式系统,并接收新的分布式系统 ID。此外,所有缓存 区域和其他 Apache Geode 组件将被重建。因此,所有旧的引用,可能已经 由 Spring 容器注入到应用程序中,现在已过时且不再有效。
Apache Geode 不保证(即使使用 Apache Geode 公共 Java API)应用程序缓存、 区域或其他组件引用由 reconnect作自动刷新。因此,Apache Geode 应用程序必须注意刷新自己的引用。
遗憾的是,无法收到断开连接事件的通知,也无法收到随后的重新连接事件的通知。
如果是这种情况,您将有一种清晰的方法来知道何时调用ConfigurableApplicationContext.refresh()
,
如果它甚至适用于应用程序这样做,这就是为什么 Apache Geode 的这个“功能”不是
推荐给对等方Cache
应用。
有关“自动重新连接”的更多信息,请参阅 Apache Geode 的产品文档。
使用基于集群的配置
Apache Geode 的集群配置服务是加入集群的任何对等成员都可以获取的便捷方式 集群的 “一致视图” ,方法是使用由 Locator 维护的共享持久配置。 使用基于集群的配置可确保对等成员的配置与 Apache Geode 兼容 成员加入时的分布式系统。
Spring Data for Apache Geode 的此功能(将use-cluster-configuration
属性设置为true
) 的工作方式相同
作为cache-xml-location
属性,但 Apache Geode 配置元数据的来源来自
通过定位器从网络,而不是原生cache.xml
文件。
所有 Apache Geode 本机配置元数据,无论是来自cache.xml
或从 Cluster Configuration Service 获取
在任何 Spring (XML) 配置元数据之前应用。因此, Spring 的配置用于“增强”这个
本机 Apache Geode 配置元数据,并且很可能特定于应用程序。
同样,要启用此功能,请在 Spring XML 配置中指定以下内容:
<gfe:cache use-cluster-configuration="true"/>
虽然某些 Apache Geode 工具(例如 Gfsh)在类似架构的更改时会“记录”其作
被制造(例如gfsh>create region --name=Example --type=PARTITION )、Spring Data for Apache Geode 的配置元数据
未录制。直接使用 Apache Geode 的公共 Java API 时也是如此。它也没有被记录下来。 |
有关 Apache Geode 的集群配置服务的更多信息,请参阅产品文档。
5.4.2. 配置 Apache Geode CacheServer
Spring Data for Apache Geode 包括对配置 CacheServer 的专门支持, 允许通过 Spring 容器完成配置,如下例所示:
<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:context="http://www.springframework.org/schema/context"
xmlns:gfe="https://www.springframework.org/schema/geode"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
http://www.springframework.org/schema/context https://www.springframework.org/schema/context/spring-context.xsd
https://www.springframework.org/schema/geode https://www.springframework.org/schema/geode/spring-geode.xsd
">
<gfe:cache/>
<!-- Example depicting serveral Apache Geode CacheServer configuration options -->
<gfe:cache-server id="advanced-config" auto-startup="true"
bind-address="localhost" host-name-for-clients="localhost" port="${gemfire.cache.server.port}"
load-poll-interval="2000" max-connections="22" max-message-count="1000" max-threads="16"
max-time-between-pings="30000" groups="test-server">
<gfe:subscription-config eviction-type="ENTRY" capacity="1000" disk-store="file://${java.io.tmpdir}"/>
</gfe:cache-server>
<context:property-placeholder location="classpath:cache-server.properties"/>
</beans>
前面的配置显示了cache-server
元素和许多可用选项。
此配置不是对端口进行硬编码,而是使用 Spring 的上下文名称空间来声明property-placeholder .属性占位符读取一个或多个属性文件,然后在运行时将属性占位符替换为值。这样做可以让管理员
更改值,而不必触摸主应用程序配置。Spring 还提供了 SPEL 和环境抽象,以支持从主代码库外部化特定于环境的属性,从而简化跨多台机器的部署。 |
为避免初始化问题,CacheServer 由 Spring Data for Apache Geode 启动 在 Spring 容器之后启动
已完全初始化。这样做可以让定义
以声明方式在服务器开始接受连接之前完全初始化和注册。请记住这一点
以编程方式配置这些元素时,因为服务器可能在组件之前启动,因此看不到
通过客户端立即连接。 |
5.4.3. 配置 Apache Geode ClientCache
除了定义 Apache Geode 对等节点Cache
,
Spring Data for Apache Geode 还支持 Apache Geode 的定义ClientCache
在 Spring 容器中。一个ClientCache
定义在配置和使用上类似于 Apache Geode 对等缓存,并且受org.springframework.data.gemfire.client.ClientCacheFactoryBean
.
使用默认配置的 Apache Geode 缓存客户端的最简单定义如下:
<beans>
<gfe:client-cache/>
</beans>
client-cache
支持许多与 Cache 元素相同的选项。然而,与
对成熟的对等体Cache
成员,则缓存客户端会通过 Pool 连接到远程缓存服务器。默认情况下,
创建一个 Pool 以连接到在 上运行的服务器localhost
并监听端口40404
.使用默认 Pool
由所有客户端区域分配,除非该区域配置为使用特定池。
池可以使用pool
元素。此客户端 Pool 可用于配置直接连接到
通过一个或多个 Locator 为单个实体或整个缓存提供的服务器。
例如,要自定义client-cache
,开发人员需要定义一个 Pool 并连接它
添加到缓存定义中,如下例所示:
<beans>
<gfe:client-cache id="myCache" pool-name="myPool"/>
<gfe:pool id="myPool" subscription-enabled="true">
<gfe:locator host="${gemfire.locator.host}" port="${gemfire.locator.port}"/>
</gfe:pool>
</beans>
这<client-cache>
元素还具有ready-for-events
属性。如果该属性设置为true
、客户端缓存
初始化包括对ClientCache.readyForEvents()
.
Client Region 更详细地介绍了客户端配置。
Apache Geode 的 DEFAULT 池和 Spring Data for Apache Geode 池定义
如果 Apache GeodeClientCache
is local-only,则不需要 Pool 定义。例如,您可以定义
以下内容:
<gfe:client-cache/>
<gfe:client-region id="Example" shortcut="LOCAL"/>
在这种情况下,“示例”区域为LOCAL
并且不会在客户端和服务器之间分发任何数据。因此
不需要 Pool。这适用于任何客户端、仅限本地的区域,如 Apache Geode 的ClientRegionShortcut
(全部LOCAL_*
快捷方式)。
但是,如果客户端区域是服务器端区域的(缓存)代理,则需要 Pool。在这种情况下, 有几种方法可以定义和使用池。
当ClientCache
、池和基于代理的区域都已定义但未明确标识,Spring Data for Apache Geode
自动解析引用,如下例所示:
<gfe:client-cache/>
<gfe:pool>
<gfe:locator host="${geode.locator.host}" port="${geode.locator.port}"/>
</gfe:pool>
<gfe:client-region id="Example" shortcut="PROXY"/>
在前面的示例中,ClientCache
被标识为gemfireCache
,将 Pool 设置为gemfirePool
,
将客户区域设置为 “Example”。但是,ClientCache
初始化 Apache Geode 的DEFAULT
池
从gemfirePool
,而客户区域使用gemfirePool
在客户端之间分发数据时
和服务器。
基本上,Spring Data for Apache Geode 将上述配置解析为以下内容:
<gfe:client-cache id="gemfireCache" pool-name="gemfirePool"/>
<gfe:pool id="gemfirePool">
<gfe:locator host="${geode.locator.host}" port="${geode.locator.port}"/>
</gfe:pool>
<gfe:client-region id="Example" cache-ref="gemfireCache" pool-name="gemfirePool" shortcut="PROXY"/>
Apache Geode 仍会创建一个名为DEFAULT
.Spring Data for Apache Geode 会导致DEFAULT
要初始化的池
从gemfirePool
.在定义了多个 Pools 并且 client Region 的情况下,这样做非常有用
使用单独的 Pools,或者根本不声明 Pool。
请考虑以下事项:
<gfe:client-cache pool-name="locatorPool"/>
<gfe:pool id="locatorPool">
<gfe:locator host="${geode.locator.host}" port="${geode.locator.port}"/>
</gfe:pool>
<gfe:pool id="serverPool">
<gfe:server host="${geode.server.host}" port="${geode.server.port}"/>
</gfe:pool>
<gfe:client-region id="Example" pool-name="serverPool" shortcut="PROXY"/>
<gfe:client-region id="AnotherExample" shortcut="CACHING_PROXY"/>
<gfe:client-region id="YetAnotherExample" shortcut="LOCAL"/>
在此设置中,Apache Geodeclient-cache
DEFAULT
pool 从locatorPool
,
由pool-name
属性。没有定义的 Spring Data for Apache GeodegemfirePool
,因为两个池
被明确识别(命名)——locatorPool
和serverPool
分别。
“示例”区域明确引用并专门使用serverPool
.这AnotherExample
区域用途
Apache Geode 的DEFAULT
Pool,同样是从locatorPool
基于客户端缓存
bean 定义的pool-name
属性。
最后,YetAnotherExample
Region 不使用 Pool,因为它是LOCAL
.
这AnotherExample Region 将首先查找名为gemfirePool ,但这需要
匿名 Pool Bean 的定义(即<gfe:pool/> ) 或显式命名为gemfirePool (例如,<gfe:pool id="gemfirePool"/> ). |
如果我们更改了locatorPool 自gemfirePool 或使 Pool bean 定义匿名,
它将具有与前面的配置相同的效果。 |
5.5. 配置区域
需要一个区域来存储和检索缓存中的数据。org.apache.geode.cache.Region
是一个接口
扩展java.util.Map
并使用熟悉的键值语义实现基本数据访问。这Region
接口
连接到需要它的应用程序类中,因此实际的 Region 类型与编程模型分离。
通常,每个区域都与一个域对象相关联,类似于关系数据库中的表。
Apache Geode 实现以下类型的区域:
-
REPLICATE – 在定义区域的集群中的所有缓存成员之间复制数据。这提供了 读取性能非常高,但执行复制需要更长的时间。
-
PARTITION - 数据在集群中的许多缓存成员中分区(分片),这些缓存成员定义 区域。这提供了较高的读写性能,适用于过大的大型数据集 对于单个节点。
-
LOCAL - 数据仅存在于本地节点上。
-
客户端 – 从技术上讲,客户区域是充当 REPLICATE 或 PARTITION 区域的代理的 LOCAL 区域 托管在集群中的缓存服务器上。它可以保存在本地创建或获取的数据。或者,它可以为空。 本地更新将同步到缓存服务器。此外,客户区域可以订阅事件,以便 与来自访问同一服务器区域的远程进程的更改保持同步(同步)。
有关各种区域类型及其功能以及配置选项的更多信息, 请参阅 Apache Geode 的 Region Types 文档。
5.5.1. 使用外部配置的 Region
引用已在 Apache Geode 本机中配置的区域cache.xml
文件中,请使用lookup-region
元素。
只需使用name
属性。例如,要声明标识的 bean 定义
如ordersRegion
对于名为Orders
,你可以使用以下 bean 定义:
<gfe:lookup-region id="ordersRegion" name="Orders"/>
如果name
未指定,则 Bean 的id
将用作区域的名称。上面的示例变为:
<!-- lookup for a Region called 'Orders' -->
<gfe:lookup-region id="Orders"/>
如果 Region 不存在,则会引发初始化异常。要配置新区域, 请继续阅读下面的相应部分。 |
在前面的示例中,由于没有明确定义缓存名称,因此默认命名约定 (gemfireCache
)
被使用。或者,可以使用cache-ref
属性:
<gfe:cache id="myCache"/>
<gfe:lookup-region id="ordersRegion" name="Orders" cache-ref="myCache"/>
lookup-region
允许您检索现有的预配置区域,而无需公开区域语义
或设置基础设施。
5.5.2. 自动区域查找
auto-region-lookup
允许您导入 Apache Geode 原生中定义的所有区域cache.xml
文件转换为
弹簧ApplicationContext
当您使用cache-xml-location
属性<gfe:cache>
元素。
例如,请考虑以下cache.xml
文件:
<?xml version="1.0" encoding="UTF-8"?>
<cache xmlns="https://geode.apache.org/schema/cache"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://geode.apache.org/schema/cache https://geode.apache.org/schema/cache/cache-1.0.xsd"
version="1.0">
<region name="Parent" refid="REPLICATE">
<region name="Child" refid="REPLICATE"/>
</region>
</cache>
您可以导入前面的cache.xml
文件,如下所示:
<gfe:cache cache-xml-location="cache.xml"/>
然后,您可以使用<gfe:lookup-region>
元素(例如<gfe:lookup-region id="Parent"/>
) 引用
特定 Region 作为 Spring 容器中的 bean 进行 bean 的 bean 调用,或者您可以选择导入cache.xml
通过使用以下内容:
<gfe:auto-region-lookup/>
Spring Data for Apache Geode 会自动为 中定义的所有 Apache Geode 区域创建 beancache.xml
那些还没有
显式添加到 Spring 容器中<gfe:lookup-region>
bean 声明。
重要的是要认识到 Spring Data for Apache Geode 在创建和初始化缓存后使用 Spring BeanPostProcessor 对缓存进行后处理,以确定 Apache Geode 中定义的区域
在 Spring 中添加为 beanApplicationContext
.
您可以注入这些 “auto-looked-up” Region,就像在 Spring 中定义的任何其他 bean 一样ApplicationContext
,
但有一个例外:您可能需要定义一个depends-on
与 'gemfireCache' bean 的关联,如下所示:
package example;
@Repository("appDao")
@DependsOn("gemfireCache")
public class ApplicationDao extends DaoSupport {
@Resource(name = "Parent")
private Region<?, ?> parent;
@Resource(name = "/Parent/Child")
private Region<?, ?> child;
...
}
The preceding example only applies when you use Spring’s component-scan
functionality.
If you declare your components by using Spring XML config, then you would do the following:
<bean class="example.ApplicationDao" depends-on="gemfireCache"/>
Doing so ensures that the Apache Geode cache and all the Regions defined in cache.xml
get created before
any components with auto-wire references when using the <gfe:auto-region-lookup>
element.
5.5.3. Configuring Regions
Spring Data for Apache Geode provides comprehensive support for configuring any type of Region through the following elements:
-
LOCAL Region: <local-region>
-
PARTITION Region: <partitioned-region>
-
REPLICATE Region: <replicated-region>
-
Client Region: <client-region>
See the Apache Geode documentation for a comprehensive description of
Region Types.
Common Region Attributes
The following table lists the attributes available for all Region types:
Table 1. Common Region Attributes
Name
Values
Description
cache-ref
Apache Geode Cache bean reference
The name of the bean defining the Apache Geode Cache (by default, 'gemfireCache').
cloning-enabled
boolean (default: false
)
When true
, the updates are applied to a clone of the value and then the clone is saved to the cache. When false
, the value is modified in place in the cache.
close
boolean (default: false
)
Determines whether the region should be closed at shutdown.
concurrency-checks-enabled
boolean (default: true
)
Determines whether members perform checks to provide consistent handling for concurrent or out-of-order updates to distributed regions.
data-policy
See Apache Geode’s data policy.
The region’s data policy. Note that not all data policies are supported for every Region type.
destroy
boolean (default: false
)
Determines whether the region should be destroyed at shutdown.
disk-store-ref
The name of a configured disk store.
A reference to a bean created through the disk-store
element.
disk-synchronous
boolean (default: true
)
Determines whether disk store writes are synchronous.
id
Any valid bean name.
The default region name if no name
attribute is specified.
ignore-if-exists
boolean (default: false
)
Ignores this bean definition if the region already exists in the cache, resulting in a lookup instead.
ignore-jta
boolean (default: false
)
Determines whether this Region participates in JTA (Java Transaction API) transactions.
index-update-type
synchronous
or asynchronous
(default: synchronous
)
Determines whether Indices are updated synchronously or asynchronously on entry creation.
initial-capacity
integer (default: 16)
The initial memory allocation for the number of Region entries.
key-constraint
Any valid, fully-qualified Java class name.
Expected key type.
load-factor
float (default: .75)
Sets the initial parameters on the underlying java.util.ConcurrentHashMap
used for storing region entries.
name
Any valid region name.
The name of the region. If not specified, it assumes the value of the id
attribute (that is, the bean name).
persistent
*boolean (default: false
)
Determines whether the region persists entries to local disk (disk store).
shortcut
See https://geode.apache.org/releases/latest/javadoc/org/apache/geode/cache/RegionShortcut.html
The RegionShortcut
for this region. Allows easy initialization of the region based on pre-defined defaults.
statistics
boolean (default: false
)
Determines whether the region reports statistics.
template
The name of a region template.
A reference to a bean created through one of the *region-template
elements.
value-constraint
Any valid, fully-qualified Java class name.
Expected value type.
CacheListener
instances
CacheListener
instances are registered with a Region to handle Region events, such as when entries are created,
updated, destroyed, and so on. A CacheListener
can be any bean that implements the
CacheListener
interface.
A Region may have multiple listeners, declared with the cache-listener
element nested in the containing
*-region
element.
The following example has two declared CacheListener’s
. The first references a named, top-level Spring bean.
The second is an anonymous inner bean definition.
<bean id="myListener" class="org.example.app.geode.cache.SimpleCacheListener"/>
<gfe:replicated-region id="regionWithListeners">
<gfe:cache-listener>
<!-- nested CacheListener bean reference -->
<ref bean="myListener"/>
<!-- nested CacheListener bean definition -->
<bean class="org.example.app.geode.cache.AnotherSimpleCacheListener"/>
</gfe:cache-listener>
</gfe:replicated-region>
The following example uses an alternate form of the cache-listener
element with the ref
attribute.
Doing so allows for more concise configuration when defining a single CacheListener
.
Note: The XML namespace allows only a single cache-listener
element, so either the style shown in
the preceding example or the style in the following example must be used.
<beans>
<gfe:replicated-region id="exampleReplicateRegionWithCacheListener">
<gfe:cache-listener ref="myListener"/>
</gfe:replicated-region>
<bean id="myListener" class="example.CacheListener"/>
</beans>
Using ref
and a nested declaration in the cache-listener
element is illegal.
The two options are mutually exclusive and using both in the same element results in an exception.
Bean Reference Conventions
The cache-listener
element is an example of a common pattern used in the XML namespace anywhere Apache Geode
provides a callback interface to be implemented in order to invoke custom code in response to cache or Region events.
When you use Spring’s IoC container, the implementation is a standard Spring bean. In order to simplify the configuration,
the schema allows a single occurrence of the cache-listener
element, but, if multiple instances are permitted,
it may contain nested bean references and inner bean definitions in any combination. The convention is to use
the singular form (that is, cache-listener
vs cache-listeners
), reflecting that the most common scenario is,
in fact, a single instance. We have already seen examples of this pattern in the advanced cache
configuration example.
CacheLoaders and CacheWriters
Similar to cache-listener
, the XML namespace provides cache-loader
and cache-writer
elements to register
these Apache Geode components for a Region.
A CacheLoader
is invoked on a cache miss to let an entry be loaded from an external data source, such as a database.
A CacheWriter
is invoked before an entry is created or updated, to allow the entry to be synchronized to an external
data source. The main difference is that Apache Geode supports, at most, a single instance of CacheLoader
and CacheWriter
per Region. However, either declaration style may be used.
The following example declares a Region with both a CacheLoader
and a CacheWriter
:
<beans>
<gfe:replicated-region id="exampleReplicateRegionWithCacheLoaderAndCacheWriter">
<gfe:cache-loader ref="myLoader"/>
<gfe:cache-writer>
<bean class="example.CacheWriter"/>
</gfe:cache-writer>
</gfe:replicated-region>
<bean id="myLoader" class="example.CacheLoader">
<property name="dataSource" ref="mySqlDataSource"/>
</bean>
<!-- DataSource bean definition -->
</beans>
See CacheLoader
and CacheWriter
in the Apache Geode documentation for more details.
5.5.4. Compression
Apache Geode Regions may also be compressed in order to reduce JVM memory consumption and pressure to possibly
avoid global GCs. When you enable compression for a Region, all values stored in memory for the Region are compressed,
while keys and indexes remain uncompressed. New values are compressed when put into the Region and all values
are decompressed automatically when read back from the Region. Values are not compressed when persisted to disk
or when sent over the wire to other peer members or clients.
The following example shows a Region with compression enabled:
<beans>
<gfe:replicated-region id="exampleReplicateRegionWithCompression">
<gfe:compressor>
<bean class="org.apache.geode.compression.SnappyCompressor"/>
</gfe:compressor>
</gfe:replicated-region>
</beans>
See Apache Geode’s documentation for more information on
Region Compression.
5.5.5. Off-Heap
Apache Geode Regions may also be configured to store Region values in off-heap memory, which is a portion of
JVM memory that is not subject to Garbage Collection (GC). By avoid expensive GC cycles, your application
can spend more of its time on things that matter, like processing requests.
Using off-heap memory is as simple as declaring the amount of memory to use and then enabling your Regions
to use off-heap memory, as shown in the following configuration:
<util:properties id="gemfireProperties">
<prop key="off-heap-memory-size">200G</prop>
</util:properties>
<gfe:cache properties-ref="gemfireProperties"/>
<gfe:partitioned-region id="ExampleOffHeapRegion" off-heap="true"/>
You can control other aspects of off-heap memory management by setting the following Apache Geode configuration
properties using the <gfe:cache>
element:s
<gfe:cache critical-off-heap-percentage="90" eviction-off-heap-percentage"80"/>
Apache Geode’s ResourceManager
will use these two threshold values (critical-off-heap-percentage
& eviction-off-heap-percentage
) to more effectively manage the off-heap memory in much the same way
as the JVM does when managing heap memory. Apache Geode ResourceManager
will prevent the cache
from consuming too much off-heap memory by evicting old data. If the off-heap manager is unable to keep up,
then the ResourceManager
refuses additions to the cache until the off-heap memory manager has freed up
an adequate amount of memory.
See Apache Geode’s documentation for more information on
Managing Heap and Off-Heap Memory.
Specifically, read the section,
Managing Off-Heap Memory.
5.5.6. Subregions
Spring Data for Apache Geode also supports Sub-Regions, allowing Regions to be arranged in a hierarchical relationship.
For example, Apache Geode allows for a /Customer/Address
Region and a different /Employee/Address
Region.
Additionally, a Sub-Region may have its own Sub-Regions and configuration. A Sub-Region does not inherit attributes
from its parent Region. Regions types may be mixed and matched subject to Apache Geode constraints. A Sub-Region
is naturally declared as a child element of a Region. The Sub-Region’s name
attribute is the simple name.
The preceding example might be configured as follows:
<beans>
<gfe:replicated-region name="Customer">
<gfe:replicated-region name="Address"/>
</gfe:replicated-region>
<gfe:replicated-region name="Employee">
<gfe:replicated-region name="Address"/>
</gfe:replicated-region>
</beans>
Note that the Monospaced ([id])
attribute is not permitted for a Sub-Region. Sub-Regions are created with bean names
(/Customer/Address and /Employee/Address, respectively, in this case). So they may be injected into other application
beans, such as a GemfireTemplate
, that need them by using the full path name of the Region. The full pathname
of the Region should also be used in OQL query strings.
5.5.7. Region Templates
Spring Data for Apache Geode also supports Region templates.
This feature allows developers to define common Region configuration and attributes once and reuse the configuration
among many Region bean definitions declared in the Spring ApplicationContext
.
Spring Data for Apache Geode includes five Region template tags in its namespace:
Table 2. Region Template Tags
Tag Name
Description
<gfe:region-template>
Defines common generic Region attributes. Extends regionType
in the XML namespace.
<gfe:local-region-template>
Defines common 'Local' Region attributes. Extends localRegionType
in the XML namespace.
<gfe:partitioned-region-template>
Defines common 'PARTITION' Region attributes. Extends partitionedRegionType
in the XML namespace.
<gfe:replicated-region-template>
Defines common 'REPLICATE' Region attributes. Extends replicatedRegionType
in the XML namespace.
<gfe:client-region-template>
Defines common 'Client' Region attributes. Extends clientRegionType
in the XML namespace.
In addition to the tags, concrete <gfe:*-region>
elements (along with the abstract <gfe:*-region-template>
elements)
have a template
attribute used to define the Region template from which the Region inherits its configuration.
Region templates may even inherit from other Region templates.
The following example shows one possible configuration:
<beans>
<gfe:async-event-queue id="AEQ" persistent="false" parallel="false" dispatcher-threads="4">
<gfe:async-event-listener>
<bean class="example.AeqListener"/>
</gfe:async-event-listener>
</gfe:async-event-queue>
<gfe:region-template id="BaseRegionTemplate" initial-capacity="51" load-factor="0.85" persistent="false" statistics="true"
key-constraint="java.lang.Long" value-constraint="java.lang.String">
<gfe:cache-listener>
<bean class="example.CacheListenerOne"/>
<bean class="example.CacheListenerTwo"/>
</gfe:cache-listener>
<gfe:entry-ttl timeout="600" action="DESTROY"/>
<gfe:entry-tti timeout="300 action="INVLIDATE"/>
</gfe:region-template>
<gfe:region-template id="ExtendedRegionTemplate" template="BaseRegionTemplate" load-factor="0.55">
<gfe:cache-loader>
<bean class="example.CacheLoader"/>
</gfe:cache-loader>
<gfe:cache-writer>
<bean class="example.CacheWriter"/>
</gfe:cache-writer>
<gfe:async-event-queue-ref bean="AEQ"/>
</gfe:region-template>
<gfe:partitioned-region-template id="PartitionRegionTemplate" template="ExtendedRegionTemplate"
copies="1" load-factor="0.70" local-max-memory="1024" total-max-memory="16384" value-constraint="java.lang.Object">
<gfe:partition-resolver>
<bean class="example.PartitionResolver"/>
</gfe:partition-resolver>
<gfe:eviction type="ENTRY_COUNT" threshold="8192000" action="OVERFLOW_TO_DISK"/>
</gfe:partitioned-region-template>
<gfe:partitioned-region id="TemplateBasedPartitionRegion" template="PartitionRegionTemplate"
copies="2" local-max-memory="8192" persistent="true" total-buckets="91"/>
</beans>
Region templates work for Sub-Regions as well. Notice that 'TemplateBasedPartitionRegion' extends 'PartitionRegionTemplate',
which extends 'ExtendedRegionTemplate', which extends 'BaseRegionTemplate'. Attributes and sub-elements defined in
subsequent, inherited Region bean definitions override what is in the parent.
How Templating Works
Spring Data for Apache Geode applies Region templates when the Spring ApplicationContext
configuration metadata is parsed, and therefore,
Region templates must be declared in the order of inheritance. In other words, parent templates must be defined before
child templates. Doing so ensures that the proper configuration is applied, especially when element attributes
or sub-elements are overridden.
It is equally important to remember that the Region types must only inherit from other similarly typed Regions.
For instance, it is not possible for a <gfe:replicated-region>
to inherit from a <gfe:partitioned-region-template>
.
Region templates are single-inheritance.
Caution concerning Regions, Sub-Regions and Lookups
Previously, one of the underlying properties of the replicated-region
, partitioned-region
, local-region
,
and client-region
elements in the Spring Data for Apache Geode XML namespace was to perform a lookup first before attempting to
create a Region. This was done in case the Region already existed, which would be the case if the Region was defined
in an imported Apache Geode native cache.xml
configuration file. Therefore, the lookup was performed first
to avoid any errors. This was by design and subject to change.
This behavior has been altered and the default behavior is now to create the Region first. If the Region
already exists, then the creation logic fails-fast and an appropriate exception is thrown. However, much like the
CREATE TABLE IF NOT EXISTS …
DDL syntax, the Spring Data for Apache Geode <gfe:*-region>
XML namespace elements now include
a ignore-if-exists
attribute, which reinstates the old behavior by first performing a lookup of an existing Region
identified by name before attempting to create the Region. If an existing Region is found by name and ignore-if-exists
is set to true
, then the Region bean definition defined in Spring configuration is ignored.
The Spring team highly recommends that the replicated-region
, partitioned-region
, local-region
,
and client-region
XML namespace elements be strictly used for defining new Regions only. One problem that could arise
when the Regions defined by these elements already exist and the Region elements perform a lookup first is, if
you defined different Region semantics and behaviors for eviction, expiration, subscription, and so on in your
application config, then the Region definition might not match and could exhibit contrary behaviors to those required
by the application. Even worse, you might want to define the Region as a distributed Region (for example, PARTITION
)
when, in fact, the existing Region definition is local only.
Recommended Practice - Use only replicated-region
, partitioned-region
, local-region
, and client-region
XML namespace elements to define new Regions.
Consider the following native Apache Geode cache.xml
configuration file:
<?xml version="1.0" encoding="UTF-8"?>
<cache xmlns="https://geode.apache.org/schema/cache"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://geode.apache.org/schema/cache https://geode.apache.org/schema/cache/cache-1.0.xsd"
version="1.0">
<region name="Customers" refid="REPLICATE">
<region name="Accounts" refid="REPLICATE">
<region name="Orders" refid="REPLICATE">
<region name="Items" refid="REPLICATE"/>
</region>
</region>
</region>
</cache>
Further, consider that you may have defined an application DAO as follows:
public class CustomerAccountDao extends GemDaoSupport {
@Resource(name = "Customers/Accounts")
private Region customersAccounts;
...
}
Here, we inject a reference to the Customers/Accounts
Region in our application DAO. Consequently, it is
not uncommon for a developer to define beans for some or all of these Regions in Spring XML configuration
metadata as follows:
<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:gfe="https://www.springframework.org/schema/geode"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
https://www.springframework.org/schema/geode https://www.springframework.org/schema/geode/spring-geode.xsd
">
<gfe:cache cache-xml-location="classpath:cache.xml"/>
<gfe:lookup-region name="Customers/Accounts"/>
<gfe:lookup-region name="Customers/Accounts/Orders"/>
</beans>
The Customers/Accounts
and Customers/Accounts/Orders
Regions are referenced as beans in the Spring container
as Customers/Accounts
and Customers/Accounts/Orders
, respectively. The nice thing about using the lookup-region
element and the corresponding syntax (described earlier) is that it lets you reference a Sub-Region directly without
unnecessarily defining a bean for the parent Region (Customers
, in this case).
Consider the following bad example, which changes the configuration metadata syntax to use the nested format:
<gfe:lookup-region name="Customers">
<gfe:lookup-region name="Accounts">
<gfe:lookup-region name="Orders"/>
</gfe:lookup-region>
</gfe:lookup-region>
Now consider another bad example which uses the top-level replicated-region
element along with
the ignore-if-exists
attribute set to perform a lookup first:
<gfe:replicated-region name="Customers" persistent="true" ignore-if-exists="true">
<gfe:replicated-region name="Accounts" persistent="true" ignore-if-exists="true">
<gfe:replicated-region name="Orders" persistent="true" ignore-if-exists="true"/>
</gfe:replicated-region>
</gfe:replicated-region>
The Region beans defined in the Spring ApplicationContext
consist of the following:
{ "Customers", "/Customers/Accounts", "/Customers/Accounts/Orders" }.
This means the dependency injected reference
shown in the earlier example (that is, @Resource(name = "Customers/Accounts")
) is now broken, since no bean with name
Customers/Accounts
is actually defined. For this reason, you should not configure Regions as shown in
the two preceding examples.
Apache Geode is flexible in referencing both parent Regions and Sub-Regions with or without the leading forward
slash. For example, the parent can be referenced as /Customers
or Customers
and the child as /Customers/Accounts
or Customers/Accounts
. However, Spring Data for Apache Geode is very specific when it comes to naming beans after Regions. It always
uses the forward slash (/) to represent Sub-Regions (for example, /Customers/Accounts
).
Therefore, you should use the non-nested lookup-region
syntax shown earlier or define direct references with
a leading forward slash (/), as follows:
<gfe:lookup-region name="/Customers/Accounts"/>
<gfe:lookup-region name="/Customers/Accounts/Orders"/>
The earlier example, where the nested replicated-region
elements were used to reference the Sub-Regions, shows
the problem stated earlier. Are the Customers, Accounts and Orders Regions and Sub-Regions persistent or not?
They are not persistent, because the Regions were defined in the native Apache Geode cache.xml
configuration file
as REPLICATE
and exist before the cache bean is initialized (once the <gfe:cache>
element is processed).
5.5.8. Data Eviction (with Overflow)
Based on various constraints, each Region can have an eviction policy in place for evicting data from memory.
Currently, in Apache Geode, eviction applies to the Least Recently Used entry (also known as
LRU). Evicted entries are either destroyed
or paged to disk (referred to as “overflow to disk”).
Spring Data for Apache Geode supports all eviction policies (entry count, memory, and heap usage) for PARTITION Regions, REPLICATE Regions,
and client, local Regions by using the nested eviction
element.
For example, to configure a PARTITION Region to overflow to disk if the memory size exceeds more than 512 MB,
you can specify the following configuration:
<gfe:partitioned-region id="examplePartitionRegionWithEviction">
<gfe:eviction type="MEMORY_SIZE" threshold="512" action="OVERFLOW_TO_DISK"/>
</gfe:partitioned-region>
Replicas cannot use local destroy
eviction since that would invalidate them.
See the Apache Geode docs for more information.
When configuring Regions for overflow, you should configure the storage through the disk-store
element
for maximum efficiency.
For a detailed description of eviction policies, see the Apache Geode documentation on
Eviction.
5.5.9. Data Expiration
Apache Geode lets you control how long entries exist in the cache. Expiration is driven by elapsed time,
as opposed to eviction, which is driven by the entry count or heap or memory usage. Once an entry expires,
it may no longer be accessed from the cache.
Apache Geode supports the following expiration types:
-
Time-to-Live (TTL): The amount of time in seconds that an object may remain in the cache after the last creation
or update. For entries, the counter is set to zero for create and put operations. Region counters are reset when
the Region is created and when an entry has its counter reset.
-
Idle Timeout (TTI): The amount of time in seconds that an object may remain in the cache after the last access.
The Idle Timeout counter for an object is reset any time its TTL counter is reset. In addition, an entry’s
Idle Timeout counter is reset any time the entry is accessed through a get operation or a netSearch
.
The Idle Timeout counter for a Region is reset whenever the Idle Timeout is reset for one of its entries.
Each of these may be applied to the Region itself or to entries in the Region. Spring Data for Apache Geode provides <region-ttl>
,
<region-tti>
, <entry-ttl>
, and <entry-tti>
Region child elements to specify timeout values and expiration actions.
The following example shows a PARTITION
Region with expiration values set:
<gfe:partitioned-region id="examplePartitionRegionWithExpiration">
<gfe:region-ttl timeout="30000" action="INVALIDATE"/>
<gfe:entry-tti timeout="600" action="LOCAL_DESTROY"/>
</gfe:replicated-region>
For a detailed description of expiration policies, see the Apache Geode documentation on
expiration.
Annotation-based Data Expiration
With Spring Data for Apache Geode, you can define expiration policies and settings on individual Region entry values (or, to put
it differently, directly on application domain objects). For instance, you can define expiration policies on
a Session-based application domain object as follows:
@Expiration(timeout = "1800", action = "INVALIDATE")
public class SessionBasedApplicationDomainObject {
...
}
You can also specify expiration type specific settings on Region entries by using the @IdleTimeoutExpiration
and @TimeToLiveExpiration
annotations for Idle Timeout (TTI) and Time-to-Live (TTL) expiration, respectively,
as the following example shows:
@TimeToLiveExpiration(timeout = "3600", action = "LOCAL_DESTROY")
@IdleTimeoutExpiration(timeout = "1800", action = "LOCAL_INVALIDATE")
@Expiration(timeout = "1800", action = "INVALIDATE")
public class AnotherSessionBasedApplicationDomainObject {
...
}
Both @IdleTimeoutExpiration
and @TimeToLiveExpiration
take precedence over the generic @Expiration
annotation
when more than one expiration annotation type is specified, as shown in the preceding example. Neither
@IdleTimeoutExpiration
nor @TimeToLiveExpiration
overrides the other. Rather, they compliment each other
when different Region entry expiration policies, such as TTL and TTI, are configured.
All @Expiration
-based annotations apply only to Region entry values. Expiration for a Region is not covered by
Spring Data for Apache Geode’s expiration annotation support. However, Apache Geode and Spring Data for Apache Geode do let you set Region expiration
by using the SDG XML namespace, as follows:
<gfe:*-region id="Example" persistent="false">
<gfe:region-ttl timeout="600" action="DESTROY"/>
<gfe:region-tti timeout="300" action="INVALIDATE"/>
</gfe:*-region>
Spring Data for Apache Geode’s @Expiration
annotation support is implemented with Apache Geode’s
CustomExpiry
interface.
See Apache Geode’s documentation on configuring data expiration
for more details
The Spring Data for Apache Geode AnnotationBasedExpiration
class (and CustomExpiry
implementation) is responsible for processing
the SDG @Expiration
annotations and applying the expiration policy configuration appropriately for Region
entry expiration on request.
To use Spring Data for Apache Geode to configure specific Apache Geode Regions to appropriately apply the expiration policy to
your application domain objects annotated with @Expiration
-based annotations, you must:
-
Define a bean in the Spring ApplicationContext
of type AnnotationBasedExpiration
by using the appropriate
constructor or one of the convenient factory methods. When configuring expiration for a specific expiration type,
such as Idle Timeout (TTI) or Time-to-Live (TTL), you should use one of the factory methods in
the AnnotationBasedExpiration
class, as follows:
<bean id="ttlExpiration" class="org.springframework.data.gemfire.expiration.AnnotationBasedExpiration"
factory-method="forTimeToLive"/>
<gfe:partitioned-region id="Example" persistent="false">
<gfe:custom-entry-ttl ref="ttlExpiration"/>
</gfe:partitioned-region>
To configure Idle Timeout (TTI) Expiration instead, use the forIdleTimeout
factory method
along with the <gfe:custom-entry-tti ref="ttiExpiration"/>
element to set TTI.
-
(optional) Annotate your application domain objects that are stored in the Region with expiration policies
and custom settings by using one of Spring Data for Apache Geode’s @Expiration
annotations: @Expiration
, @IdleTimeoutExpiration
,
or @TimeToLiveExpiration
-
(optional) In cases where particular application domain objects have not been annotated with Spring Data for Apache Geode’s
@Expiration
annotations at all, but the Apache Geode Region is configured to use SDG’s custom
AnnotationBasedExpiration
class to determine the expiration policy and settings for objects stored in the Region,
you can set “default” expiration attributes on the AnnotationBasedExpiration
bean by doing the following:
<bean id="defaultExpirationAttributes" class="org.apache.geode.cache.ExpirationAttributes">
<constructor-arg value="600"/>
<constructor-arg value="#{T(org.apache.geode.cache.ExpirationAction).DESTROY}"/>
</bean>
<bean id="ttiExpiration" class="org.springframework.data.gemfire.expiration.AnnotationBasedExpiration"
factory-method="forIdleTimeout">
<constructor-arg ref="defaultExpirationAttributes"/>
</bean>
<gfe:partitioned-region id="Example" persistent="false">
<gfe:custom-entry-tti ref="ttiExpiration"/>
</gfe:partitioned-region>
You may have noticed that Spring Data for Apache Geode’s @Expiration
annotations use a String
as the attribute type rather
than, and perhaps more appropriately, being strongly typed — for example, int
for 'timeout' and SDG’s
ExpirationActionType
for 'action'. Why is that?
Well, enter one of Spring Data for Apache Geode’s other features, leveraging Spring’s core infrastructure for configuration convenience:
property placeholders and Spring Expression Language (SpEL) expressions.
For instance, a developer can specify both the expiration 'timeout' and 'action' by using property placeholders
in the @Expiration
annotation attributes, as the following example shows:
@TimeToLiveExpiration(timeout = "${geode.region.entry.expiration.ttl.timeout}"
action = "${geode.region.entry.expiration.ttl.action}")
public class ExampleApplicationDomainObject {
...
}
Then, in your Spring XML config or in JavaConfig, you can declare the following beans:
<util:properties id="expirationSettings">
<prop key="geode.region.entry.expiration.ttl.timeout">600</prop>
<prop key="geode.region.entry.expiration.ttl.action">INVALIDATE</prop>
...
</util:properties>
<context:property-placeholder properties-ref="expirationProperties"/>
This is convenient both when multiple application domain objects might share similar expiration policies and when
you wish to externalize the configuration.
However, you may want more dynamic expiration configuration determined by the state of the running system. This is where
the power of SpEL comes into play and is the recommended approach, actually. Not only can you refer to beans
in the Spring container and access bean properties, invoke methods, and so on, but the values for expiration 'timeout'
and 'action' can be strongly typed. Consider the following example (which builds on the preceding example):
<util:properties id="expirationSettings">
<prop key="geode.region.entry.expiration.ttl.timeout">600</prop>
<prop key="geode.region.entry.expiration.ttl.action">#{T(org.springframework.data.gemfire.expiration.ExpirationActionType).DESTROY}</prop>
<prop key="geode.region.entry.expiration.tti.action">#{T(org.apache.geode.cache.ExpirationAction).INVALIDATE}</prop>
...
</util:properties>
<context:property-placeholder properties-ref="expirationProperties"/>
Then, on your application domain object, you can define a timeout and an action as follows:
@TimeToLiveExpiration(timeout = "@expirationSettings['geode.region.entry.expiration.ttl.timeout']"
action = "@expirationSetting['geode.region.entry.expiration.ttl.action']")
public class ExampleApplicationDomainObject {
...
}
You can imagine that the 'expirationSettings' bean could be a more interesting and useful object than a simple
instance of java.util.Properties
. In the preceding example, the properties
element (expirationSettings
) uses SpEL
to base the action value on the actual ExpirationAction
enumerated type, quickly leading to identified failures
if the enumerated type ever changes.
As an example, all of this has been demonstrated and tested in the Spring Data for Apache Geode test suite. See the
source for further details.
5.5.10. Data Persistence
Regions can be persistent. Apache Geode ensures that all the data you put into a Region that is configured
for persistence is written to disk in a way that is recoverable the next time you recreate the Region. Doing so lets
data be recovered after machine or process failure or even after an orderly shutdown and subsequent restart of
the Apache Geode data node.
To enable persistence with Spring Data for Apache Geode, set the persistent
attribute to true
on any of the <*-region>
elements,
as the following example shows:
<gfe:partitioned-region id="examplePersitentPartitionRegion" persistent="true"/>
Persistence may also be configured by setting the data-policy
attribute. To do so, set the attribute’s value to one of
Apache Geode’s DataPolicy settings,
as the following example shows:
<gfe:partitioned-region id="anotherExamplePersistentPartitionRegion" data-policy="PERSISTENT_PARTITION"/>
The DataPolicy
must match the Region type and must also agree with the persistent
attribute if it is also
explicitly set. If the persistent
attribute is set to false
but a persistent DataPolicy
was specified (such as PERSISTENT_REPLICATE
or PERSISTENT_PARTITION
), an initialization exception is thrown.
For maximum efficiency when persisting Regions, you should configure the storage through the disk-store
element.
The DiskStore
is referenced by using the disk-store-ref
attribute. Additionally, the Region may perform disk writes
synchronously or asynchronously. The following example shows a synchronous DiskStore
:
<gfe:partitioned-region id="yetAnotherExamplePersistentPartitionRegion" persistent="true"
disk-store-ref="myDiskStore" disk-synchronous="true"/>
This is discussed further in Configuring a DiskStore.
5.5.11. Subscription Policy
Apache Geode allows configuration of peer-to-peer (P2P) event messaging
to control the entry events that the Region receives. Spring Data for Apache Geode provides the <gfe:subscription/>
sub-element to set
the subscription policy on REPLICATE
and PARTITION
Regions to either ALL
or CACHE_CONTENT
. The following example
shows a region with its subscription policy set to CACHE_CONTENT
:
<gfe:partitioned-region id="examplePartitionRegionWithCustomSubscription">
<gfe:subscription type="CACHE_CONTENT"/>
</gfe:partitioned-region>
5.5.12. Local Region
Spring Data for Apache Geode offers a dedicated local-region
element for creating local Regions. Local Regions, as the name implies,
are standalone, meaning that they do not share data with any other distributed system member. Other than that, all
common Region configuration options apply.
The following example shows a minimal declaration (again, the example relies on the Spring Data for Apache Geode XML namespace naming
conventions to wire the cache):
<gfe:local-region id="exampleLocalRegion"/>
In the preceding example, a local Region is created (if a Region by the same name does not already exist). The name of
the Region is the same as the bean ID (exampleLocalRegion
), and the bean assumes the existence of a Apache Geode
cache named gemfireCache
.
5.5.13. Replicated Region
One of the common Region types is a REPLICATE
Region or “replica”. In short, when a Region is configured to be
a REPLICATE
, every member that hosts the Region stores a copy of the Region’s entries locally. Any update to
a REPLICATE
Region is distributed to all copies of the Region. When a replica is created, it goes through
an initialization stage, in which it discovers other replicas and automatically copies all the entries.
While one replica is initializing, you can still continue to use the other replicas.
All common configuration options are available for REPLICATE Regions. Spring Data for Apache Geode offers a replicated-region
element.
The following example shows a minimal declaration:
<gfe:replicated-region id="exampleReplica"/>
See Apache Geode’s documentation on
Distributed and Replicated Regions
for more details.
5.5.14. Partitioned Region
The Spring Data for Apache Geode XML namespace also supports PARTITION
Regions.
To quote the Apache Geode docs:
“A partitioned region is a region where data is divided between peer servers hosting the region so that each peer
stores a subset of the data. When using a partitioned region, applications are presented with a logical view
of the region that looks like a single map containing all of the data in the region. Reads or writes to this map
are transparently routed to the peer that hosts the entry that is the target of the operation. Apache Geode
divides the domain of hashcodes into buckets. Each bucket is assigned to a specific peer, but may be relocated
at any time to another peer in order to improve the utilization of resources across the cluster.”
A PARTITION
Region is created by using the partitioned-region
element. Its configuration options are similar to
that of the replicated-region
with the addition of partition-specific features, such as the number of redundant copies,
total maximum memory, number of buckets, partition resolver, and so on.
The following example shows how to set up a PARTITION
Region with two redundant copies:
<gfe:partitioned-region id="examplePartitionRegion" copies="2" total-buckets="17">
<gfe:partition-resolver>
<bean class="example.PartitionResolver"/>
</gfe:partition-resolver>
</gfe:partitioned-region>
See Apache Geode’s documentation on
Partitioned Regions
for more details.
Partitioned Region Attributes
The following table offers a quick overview of configuration options specific to PARTITION
Regions.
These options are in addition to the common Region configuration options
described earlier.
Table 3. partitioned-region attributes
Name
Values
Description
copies
0..4
The number of copies for each partition for high-availability. By default, no copies are created,
meaning there is no redundancy. Each copy provides extra backup at the expense of extra storage.
colocated-with
valid region name
The name of the PARTITION
region with which this newly created PARTITION
Region is collocated.
local-max-memory
positive integer
The maximum amount of memory (in megabytes) used by the region in this process.
total-max-memory
any integer value
The maximum amount of memory (in megabytes) used by the region in all processes.
partition-listener
bean name
The name of the PartitionListener
used by this region for handling partition events.
partition-resolver
bean name
The name of the PartitionResolver
used by this region for custom partitioning.
recovery-delay
any long value
The delay in milliseconds that existing members wait before satisfying redundancy after another member crashes.
-1 (the default) indicates that redundancy is not recovered after a failure.
startup-recovery-delay
any long value
The delay in milliseconds that new members wait before satisfying redundancy.
-1 indicates that adding new members does not trigger redundancy recovery. The default is to recover redundancy
immediately when a new member is added.
5.5.15. Client Region
Apache Geode supports various deployment topologies for managing and distributing data. The topic of
Apache Geode topologies is beyond the scope of this documentation. However, to quickly recap, Apache Geode’s
supported topologies can be classified as: peer-to-peer (p2p), client-server, and wide area network (WAN). In the last
two configurations, it is common to declare client Regions that connect to a cache server.
Spring Data for Apache Geode offers dedicated support for each configuration through its client-cache elements:
client-region
and pool
. As the names imply, client-region
defines a client Region, while pool
defines
a Pool of connections used and shared by the various client Regions.
The following example shows a typical client Region configuration:
<bean id="myListener" class="example.CacheListener"/>
<!-- client Region using the default SDG gemfirePool Pool -->
<gfe:client-region id="Example">
<gfe:cache-listener ref="myListener"/>
</gfe:client-region>
<!-- client Region using its own dedicated Pool -->
<gfe:client-region id="AnotherExample" pool-name="myPool">
<gfe:cache-listener ref="myListener"/>
</gfe:client-region>
<!-- Pool definition -->
<gfe:pool id="myPool" subscription-enabled="true">
<gfe:locator host="remoteHost" port="12345"/>
</gfe:pool>
As with the other Region types, client-region
supports CacheListener
instances as well as a CacheLoader
and a CacheWriter
. It also requires a connection Pool
for connecting to a set of either Locators or servers.
Each client Region can have its own Pool
, or they can share the same one. If a Pool is not specified, then
the "DEFAULT" Pool will be used.
In the preceding example, the Pool
is configured with a Locator. A Locator is a separate process used to discover
cache servers and peer data members in the distributed system and is recommended for production systems. It is also
possible to configure the Pool
to connect directly to one or more cache servers by using the server
element.
For a full list of options to set on the client and especially on the Pool
, see
the Spring Data for Apache Geode schema (“Spring Data for Apache Geode Schema”) and Apache Geode’s documentation on
Client-Server Configuration.
Client Interests
To minimize network traffic, each client can separately define its own 'interests' policies, indicating to
Apache Geode the data it actually requires. In Spring Data for Apache Geode, 'interests' can be defined for each client Region
separately. Both key-based and regular expression-based interest types are supported.
The following example shows both key-based and regular expression-based interest
types:
<gfe:client-region id="Example" pool-name="myPool">
<gfe:key-interest durable="true" result-policy="KEYS">
<bean id="key" class="java.lang.String">
<constructor-arg value="someKey"/>
</bean>
</gfe:key-interest>
<gfe:regex-interest pattern=".*" receive-values="false"/>
</gfe:client-region>
A special key, ALL_KEYS
, means 'interest' is registered for all keys. The same can be accomplished
by using the regular expression, ".\*"
.
The <gfe:*-interest>
key and regular expression elements support three attributes: durable
, receive-values
,
and result-policy
.
durable
indicates whether the 'interest' policy and subscription queue created for the client when the client connects
to one or more servers in the cluster is maintained across client sessions. If the client goes away and comes back,
a durable
subscription queue on the servers for the client is maintained while the client is disconnected. When the
client reconnects, the client receives any events that occurred while the client was disconnected from the servers
in the cluster.
A subscription queue on the servers in the cluster is maintained for each Pool
of connections defined in the client
where a subscription has also been “enabled” for that Pool
. The subscription queue is used to store (and possibly
conflate) events sent to the client. If the subscription queue is durable, it persists between client sessions
(that is, connections), potentially up to a specified timeout. If the client does not return within a given time frame
the client Pool subscription queue is destroyed in order to reduce resource consumption on servers in the cluster.
If the subscription queue is not durable
, it is destroyed immediately when the client disconnects. You need to decide
whether your client should receive events that came while it was disconnected or if it needs to receive only the latest
events after it reconnects.
The receive-values
attribute indicates whether or not the entry values are received for create and update events.
If true
, values are received. If false
, only invalidation events are received.
And finally, the 'result-policy` is an enumeration of: KEYS
, KEYS_VALUE
, and NONE
. The default is KEYS_VALUES
.
The result-policy
controls the initial dump when the client first connects to initialize the local cache,
essentially seeding the client with events for all the entries that match the interest policy.
Client-side interest registration does not do much good without enabling subscription on the Pool
, as mentioned earlier.
In fact, it is an error to attempt interest registration without subscription enabled. The following example shows
how to do so:
<gfe:pool ... subscription-enabled="true">
...
</gfe:pool>
In addition to subscription-enabled
, can you also set subscription-ack-interval
,
subscription-message-tracking-timeout
, and subscription-redundancy
. subscription-redundancy
is used to control how
many copies of the subscription queue should be maintained by the servers in the cluster. If redundancy is greater than
one, and the “primary” subscription queue (that is, the server) goes down, then a “secondary” subscription queue
takes over, keeping the client from missing events in a HA scenario.
In addition to the Pool
settings, the server-side Regions use an additional attribute, enable-subscription-conflation
,
to control the conflation of events that are sent to the clients. This can also help further minimize network traffic
and is useful in situations where the application only cares about the latest value of an entry. However, when the
application keeps a time series of events that occurred, conflation is going to hinder that use case. The default value
is false
. The following example shows a Region configuration on the server, for which the client contains a
corresponding client [CACHING_]PROXY
Region with interests in keys in this server Region:
<gfe:partitioned-region name="ServerSideRegion" enable-subscription-conflation="true">
...
</gfe:partitioned-region>
To control the amount of time (in seconds) that a “durable” subscription queue is maintained after a client is
disconnected from the servers in the cluster, set the durable-client-timeout
attribute on the <gfe:client-cache>
element as follows:
<gfe:client-cache durable-client-timeout="600">
...
</gfe:client-cache>
A full, in-depth discussion of how client interests work and capabilities is beyond the scope of this document.
See Apache Geode’s documentation on
Client-to-Server Event Distribution
for more details.
5.5.16. JSON Support
Apache Geode has support for caching JSON documents in Regions, along with the ability to query stored JSON
documents using the Apache Geode OQL (Object Query Language). JSON documents are stored internally as
PdxInstance types
by using the JSONFormatter class
to perform conversion to and from JSON documents (as a String
).
Spring Data for Apache Geode provides the <gfe-data:json-region-autoproxy/>
element to enable an
AOP component to advise appropriate, proxied Region operations,
which effectively encapsulates the JSONFormatter
, thereby letting your applications work directly with JSON Strings.
In addition, Java objects written to JSON configured Regions are automatically converted to JSON using Jackson’s
ObjectMapper
. When these values are read back, they are returned as a JSON String.
By default, <gfe-data:json-region-autoproxy/>
performs the conversion for all Regions. To apply this feature
to selected Regions, provide a comma-delimited list of Region bean IDs in the region-refs
attribute.
Other attributes include a pretty-print
flag (defaults to false
) and convert-returned-collections
.
Also, by default, the results of the getAll()
and values()
Region operations are converted for configured Regions.
This is done by creating a parallel data structure in local memory. This can incur significant overhead for large
collections, so set the convert-returned-collections
to false
if you would like to disable automatic conversion
for these Region operations.
Certain Region operations (specifically those that use Apache Geode’s proprietary Region.Entry
, such as:
entries(boolean)
, entrySet(boolean)
and getEntry()
type) are not targeted for AOP advice. In addition,
the entrySet()
method (which returns a Set<java.util.Map.Entry<?, ?>>
) is also not affected.
The following example configuration shows how to set the pretty-print
and convert-returned-collections
attributes:
<gfe-data:json-region-autoproxy region-refs="myJsonRegion" pretty-print="true" convert-returned-collections="false"/>
This feature also works seamlessly with GemfireTemplate
operations, provided that the template is declared
as a Spring bean. Currently, the native QueryService
operations are not supported.
5.6. Configuring an Index
Apache Geode allows indexes (also sometimes pluralized as indices) to be created on Region data
to improve the performance of OQL (Object Query Language) queries.
In Spring Data for Apache Geode, indexes are declared with the index
element, as the following example shows:
<gfe:index id="myIndex" expression="someField" from="/SomeRegion" type="HASH"/>
In Spring Data for Apache Geode’s XML schema (also called the SDG XML namespace), index
bean declarations are not bound
to a Region, unlike Apache Geode’s native cache.xml
. Rather, they are top-level elements similar to
<gfe:cache>
element. This lets you declare any number of indexes on any Region, whether they were just created
or already exist — a significant improvement over Apache Geode’s native cache.xml
format.
An Index
must have a name. You can give the Index
an explicit name by using the name
attribute.
Otherwise, the bean name (that is, the value of the id
attribute) of the index
bean definition is used as
the Index
name.
The expression
and from
clause form the main components of an Index
, identifying the data to index
(that is, the Region identified in the from
clause) along with what criteria (that is, expression
) is used
to index the data. The expression
should be based on what application domain object fields are used in the predicate
of application-defined OQL queries used to query and look up the objects stored in the Region.
Consider the following example, which has a lastName
property:
@Region("Customers")
class Customer {
@Id
Long id;
String lastName;
String firstName;
...
}
Now consider the following example, which has an application-defined SDG Repository
to query for Customer
objects:
interface CustomerRepository extends GemfireRepository<Customer, Long> {
Customer findByLastName(String lastName);
...
}
The SDG Repository finder/query method results in the following OQL statement being generated and ran:
SELECT * FROM /Customers c WHERE c.lastName = '$1'
Therefore, you might want to create an Index
with a statement similar to the following:
<gfe:index id="myIndex" name="CustomersLastNameIndex" expression="lastName" from="/Customers" type="HASH"/>
The from
clause must refer to a valid, existing Region and is how an Index
gets applied to a Region.
This is not specific to Spring Data for Apache Geode. It is a feature of Apache Geode.
The Index
type
may be one of three enumerated values defined by Spring Data for Apache Geode’s
IndexType
enumeration:
FUNCTIONAL
, HASH
, and PRIMARY_KEY
.
Each of the enumerated values corresponds to one of the QueryService
create[|Key|Hash]Index
methods invoked when the actual Index
is to be created (or “defined” — you can find
more on “defining” indexes in the next section). For instance, if the IndexType
is PRIMARY_KEY
, then the
QueryService.createKeyIndex(..)
is invoked to create a KEY
Index
.
The default is FUNCTIONAL
and results in one of the QueryService.createIndex(..)
methods being invoked. See the
Spring Data for Apache Geode XML schema for a full set of options.
For more information on indexing in Apache Geode, see “Working with Indexes”
in Apache Geode’s User Guide.
5.6.1. Defining Indexes
In addition to creating indexes up front as Index
bean definitions are processed by Spring Data for Apache Geode on Spring container
initialization, you may also define all of your application indexes prior to creating them by using the define
attribute, as follows:
<gfe:index id="myDefinedIndex" expression="someField" from="/SomeRegion" define="true"/>
When define
is set to true
(it defaults to false
), it does not actually create the Index
at that moment.
All “defined” Indexes are created all at once, when the Spring ApplicationContext
is “refreshed” or, to put it
differently, when a ContextRefreshedEvent
is published by the Spring container. Spring Data for Apache Geode registers itself as
an ApplicationListener
listening for the ContextRefreshedEvent
. When fired, Spring Data for Apache Geode calls
QueryService.createDefinedIndexes()
.
Defining indexes and creating them all at once boosts speed and efficiency when creating indexes.
See “Creating Multiple Indexes at Once”
for more details.
5.6.2. IgnoreIfExists
and Override
Two Spring Data for Apache Geode Index
configuration options warrant special mention: ignoreIfExists
and override
.
These options correspond to the ignore-if-exists
and override
attributes on the <gfe:index>
element
in Spring Data for Apache Geode’s XML namespace, respectively.
Make sure you absolutely understand what you are doing before using either of these options. These options can
affect the performance and resources (such as memory) consumed by your application at runtime. As a result, both of
these options are disabled (set to false
) in SDG by default.
These options are only available in Spring Data for Apache Geode and exist to workaround known limitations with Apache Geode.
Apache Geode has no equivalent options or functionality.
Each option significantly differs in behavior and entirely depends on the type of Apache Geode Index
exception
thrown. This also means that neither option has any effect if a Apache Geode Index-type exception is not thrown.
These options are meant to specifically handle Apache Geode IndexExistsException
and IndexNameConflictException
instances, which can occur for various, sometimes obscure reasons. The exceptions have the following causes:
-
An IndexExistsException
is thrown when there exists another Index
with the same definition but a different name when attempting to
create an Index
.
-
An IndexNameConflictException
is thrown when there exists another Index
with the same name but possibly different definition when attempting to
create an Index
.
Spring Data for Apache Geode’s default behavior is to fail-fast, always. So, neither Index
Exception are “handled” by default.
These Index
exceptions are wrapped in a SDG GemfireIndexException
and rethrown. If you wish for Spring Data for Apache Geode
to handle them for you, you can set either of these Index
bean definition options to true
.
IgnoreIfExists
always takes precedence over Override
, primarily because it uses fewer resources, simply because
it returns the “existing” Index
in both exceptional cases.
IgnoreIfExists
Behavior
When an IndexExistsException
is thrown and ignoreIfExists
is set to true
(or <gfe:index ignore-if-exists="true">
),
then the Index
that would have been created by this index
bean definition or declaration is simply ignored,
and the existing Index
is returned.
There is little consequence in returning the existing Index
, since the index
bean definition is the same,
as determined by Apache Geode itself, not SDG.
However, this also means that no Index
with the “name” specified in your index
bean definition or declaration
actually exists from Apache Geode’s perspective (that is, with
QueryService.getIndexes()
).
Therefore, you should be careful when writing OQL query statements that use query hints, especially query hints
that refer to the application Index
being ignored. Those query hints need to be changed.
When an IndexNameConflictException
is thrown and ignoreIfExists
is set to true
(or <gfe:index ignore-if-exists="true">
),
the Index
that would have been created by this index
bean definition or declaration is also ignored,
and the "existing" Index
is again returned, as when an IndexExistsException
is thrown.
However, there is more risk in returning the existing Index
and ignoring the application’s definition of the Index
when an IndexNameConflictException
is thrown. For a IndexNameConflictException
, while the names of the conflicting
indexes are the same, the definitions could be different. This situation could have implications for OQL queries
specific to the application, where you would presume the indexes were defined specifically with the application
data access patterns and queries in mind. However, if like-named indexes differ in definition, this might not be
the case. Consequently, you should verify your Index
names.
SDG makes a best effort to inform the user when the Index
being ignored is significantly different
in its definition from the existing Index
. However, in order for SDG to accomplish this, it must be able to
find the existing Index
, which is looked up by using the Apache Geode API (the only means available).
Override
Behavior
When an IndexExistsException
is thrown and override
is set to true
(or <gfe:index override="true">
),
the Index
is effectively renamed. Remember, IndexExistsExceptions
are thrown when multiple indexes exist that
have the same definition but different names.
Spring Data for Apache Geode can only accomplish this by using Apache Geode’s API, by first removing the existing Index
and then recreating the Index
with the new name. It is possible that either the remove or subsequent create invocation
could fail. There is no way to execute both actions atomically and rollback this joint operation if either fails.
However, if it succeeds, then you have the same problem as before with the ignoreIfExists
option. Any existing OQL
query statement using query hints that refer to the old Index
by name must be changed.
When an IndexNameConflictException
is thrown and override
is set to true
(or <gfe:index override="true">
),
the existing Index
can potentially be re-defined. We say “potentially” because it is possible for the like-named,
existing Index
to have exactly the same definition and name when an IndexNameConflictException
is thrown.
If so, SDG is smart and returns the existing Index
as is, even on override
. There is no harm
in this behavior, since both the name and the definition are exactly the same. Of course, SDG can only
accomplish this when SDG is able to find the existing Index
, which is dependent on Apache Geode’s APIs.
If it cannot be found, nothing happens and a SDG GemfireIndexException
is thrown that wraps the
IndexNameConflictException
.
However, when the definition of the existing Index
is different, SDG attempts to re-create the Index
by using the Index
definition specified in the index
bean definition. Make sure this is what you want and make sure
the index
bean definition matches your expectations and application requirements.
How Does IndexNameConflictExceptions
Actually Happen?
It is probably not all that uncommon for IndexExistsExceptions
to be thrown, especially when multiple configuration
sources are used to configure Apache Geode (Spring Data for Apache Geode, Apache Geode Cluster Config, Apache Geode native
cache.xml
, the API, and so on). You should definitely prefer one configuration method and stick with it.
However, when does an IndexNameConflictException
get thrown?
One particular case is an Index
defined on a PARTITION
Region (PR). When an Index
is defined on a PARTITION
Region
(for example, X
), Apache Geode distributes the Index
definition (and name) to other peer members
in the cluster that also host the same PARTITION
Region (that is, "X"). The distribution of this Index
definition
to, and subsequent creation of, this Index
by peer members is on a need-to-know basis (that is, by peer member hosting
the same PR) is performed asynchronously.
During this window of time, it is possible that these pending PR Indexes
cannot be identified by Apache Geode — such as with a call to QueryService.getIndexes()
with QueryService.getIndexes(:Region)
,
or even with QueryService.getIndex(:Region, indexName:String)
.
As a result, the only way for SDG or other Apache Geode cache client applications (not involving Spring)
to know for sure is to attempt to create the Index
. If it fails with either an IndexNameConflictException
or even
an IndexExistsException
, the application knows there is a problem. This is because the QueryService
Index
creation
waits on pending Index
definitions, whereas the other Apache Geode API calls do not.
In any case, SDG makes a best effort and attempts to inform you what has happened or is happening and tell you
the corrective action. Given that all Apache Geode QueryService.createIndex(..)
methods are synchronous,
blocking operations, the state of Apache Geode should be consistent and accessible after either of these index-type
exceptions are thrown. Consequently, SDG can inspect the state of the system and act accordingly,
based on your configuration.
In all other cases, SDG embraces a fail-fast strategy.
5.7. Configuring a DiskStore
Spring Data for Apache Geode supports DiskStore
configuration and creation through the disk-store
element,
as the following example shows:
<gfe:disk-store id="Example" auto-compact="true" max-oplog-size="10"
queue-size="50" time-interval="9999">
<gfe:disk-dir location="/disk/location/one" max-size="20"/>
<gfe:disk-dir location="/disk/location/two" max-size="20"/>
</gfe:disk-store>
DiskStore
instances are used by Regions for file system persistent backup and overflow of evicted entries
as well as persistent backup for WAN Gateways. Multiple Apache Geode components may share the same DiskStore
.
Additionally, multiple file system directories may be defined for a single DiskStore
, as shown in
the preceding example.
See Apache Geode’s documentation for a complete explanation of
Persistence and Overflow
and configuration options on DiskStore
instances.
5.8. Configuring the Snapshot Service
Spring Data for Apache Geode supports cache and Region snapshots by using
Apache Geode’s Snapshot Service.
The out-of-the-box Snapshot Service support offers several convenient features to simplify the use of Apache Geode’s
Cache
and Region
Snapshot Service APIs.
As the Apache Geode documentation explains,
snapshots let you save and subsequently reload the cached data later, which can be useful for moving data between
environments, such as from production to a staging or test environment in order to reproduce data-related issues
in a controlled context. You can combine Spring Data for Apache Geode’s Snapshot Service support
with Spring’s bean definition profiles
to load snapshot data specific to the environment as necessary.
Spring Data for Apache Geode’s support for Apache Geode’s Snapshot Service begins with the <gfe-data:snapshot-service>
element
from the <gfe-data>
XML namespace.
For example, you can define cache-wide snapshots to be loaded as well as saved by using a couple of snapshot imports
and a data export definition, as follows:
<gfe-data:snapshot-service id="gemfireCacheSnapshotService">
<gfe-data:snapshot-import location="/absolute/filesystem/path/to/import/fileOne.snapshot"/>
<gfe-data:snapshot-import location="relative/filesystem/path/to/import/fileTwo.snapshot"/>
<gfe-data:snapshot-export
location="/absolute/or/relative/filesystem/path/to/export/directory"/>
</gfe-data:snapshot-service>
You can define as many imports and exports as you like. You can define only imports or only exports. The file locations
and directory paths can be absolute or relative to the Spring Data for Apache Geode application, which is the JVM process’s
working directory.
The preceding example is pretty simple, and the Snapshot Service defined in this case refers to the Apache Geode
cache instance with the default name of gemfireCache
(as described in Configuring a Cache). If you name your cache
bean definition something other than the default, you can use the cache-ref
attribute to refer to the cache bean
by name, as follows:
<gfe:cache id="myCache"/>
...
<gfe-data:snapshot-service id="mySnapshotService" cache-ref="myCache">
...
</gfe-data:snapshot-service>
You can also define a Snapshot Service for a particular Region by specifying the region-ref
attribute, as follows:
<gfe:partitioned-region id="Example" persistent="false" .../>
...
<gfe-data:snapshot-service id="gemfireCacheRegionSnapshotService" region-ref="Example">
<gfe-data:snapshot-import location="relative/path/to/import/example.snapshot/>
<gfe-data:snapshot-export location="/absolute/path/to/export/example.snapshot/>
</gfe-data:snapshot-service>
When the region-ref
attribute is specified, Spring Data for Apache Geode’s SnapshotServiceFactoryBean
resolves the region-ref
attribute value to a Region bean defined in the Spring container and creates a
RegionSnapshotService
.
The snapshot import and export definitions function the same way. However, the location
must refer to a file
on an export.
Apache Geode is strict about imported snapshot files actually existing before they are referenced.
For exports, Apache Geode creates the snapshot file. If the snapshot file for export already exists,
the data is overwritten.
Spring Data for Apache Geode includes a suppress-import-on-init
attribute on the <gfe-data:snapshot-service>
element
to suppress the configured Snapshot Service from trying to import data into the cache or Region on initialization.
Doing so is useful, for example, when data exported from one Region is used to feed the import of another Region.
5.8.1. Snapshot Location
With the cache-based Snapshot Service
(that is, a CacheSnapshotService
)
you would typically pass it a directory containing all the snapshot files to load rather than individual snapshot files,
as the overloaded load
method in the CacheSnapshotService
API indicates.
Of course, you can use the overloaded load(:File[], :SnapshotFormat, :SnapshotOptions)
method to get specific
about which snapshot files to load into the Apache Geode cache.
However, Spring Data for Apache Geode recognizes that a typical developer workflow might be to extract and export data
from one environment into several snapshot files, zip all of them up, and then conveniently move the zip file
to another environment for import.
Therefore, Spring Data for Apache Geode lets you specify a jar or zip file on import for a cache
-based Snapshot Service, as follows:
<gfe-data:snapshot-service id="cacheBasedSnapshotService" cache-ref="gemfireCache">
<gfe-data:snapshot-import location="/path/to/snapshots.zip"/>
</gfe-data:snapshot-service>
Spring Data for Apache Geode conveniently extracts the provided zip file and treats it as a directory import (load).
5.8.2. Snapshot Filters
The real power of defining multiple snapshot imports and exports is realized through the use of snapshot filters.
Snapshot filters implement Apache Geode’s SnapshotFilter
interface
and are used to filter Region entries for inclusion into the Region on import and for inclusion into the snapshot
on export.
Spring Data for Apache Geode lets you use snapshot filters on import and export by using the filter-ref
attribute or an anonymous,
nested bean definition, as the following example shows:
<gfe:cache/>
<gfe:partitioned-region id="Admins" persistent="false"/>
<gfe:partitioned-region id="Guests" persistent="false"/>
<bean id="activeUsersFilter" class="example.gemfire.snapshot.filter.ActiveUsersFilter/>
<gfe-data:snapshot-service id="adminsSnapshotService" region-ref="Admins">
<gfe-data:snapshot-import location="/path/to/import/users.snapshot">
<bean class="example.gemfire.snapshot.filter.AdminsFilter/>
</gfe-data:snapshot-import>
<gfe-data:snapshot-export location="/path/to/export/active/admins.snapshot" filter-ref="activeUsersFilter"/>
</gfe-data:snapshot-service>
<gfe-data:snapshot-service id="guestsSnapshotService" region-ref="Guests">
<gfe-data:snapshot-import location="/path/to/import/users.snapshot">
<bean class="example.gemfire.snapshot.filter.GuestsFilter/>
</gfe-data:snapshot-import>
<gfe-data:snapshot-export location="/path/to/export/active/guests.snapshot" filter-ref="activeUsersFilter"/>
</gfe-data:snapshot-service>
In addition, you can express more complex snapshot filters by using the ComposableSnapshotFilter
class.
This class implements Apache Geode’s SnapshotFilter interface
as well as the Composite software design pattern.
In a nutshell, the Composite software design pattern lets you
compose multiple objects of the same type and treat the aggregate as single instance of the object type — a
powerful and useful abstraction.
ComposableSnapshotFilter
has two factory methods, and
and or
. They let you logically combine individual snapshot
filters using the AND and OR logical operators, respectively. The factory methods take a list of SnapshotFilters
.
The following example shows a definition for a ComposableSnapshotFilter
:
<bean id="activeUsersSinceFilter" class="org.springframework.data.gemfire.snapshot.filter.ComposableSnapshotFilter"
factory-method="and">
<constructor-arg index="0">
<list>
<bean class="org.example.app.gemfire.snapshot.filter.ActiveUsersFilter"/>
<bean class="org.example.app.gemfire.snapshot.filter.UsersSinceFilter"
p:since="2015-01-01"/>
</list>
</constructor-arg>
</bean>
You could then go on to combine the activesUsersSinceFilter
with another filter by using or
, as follows:
<bean id="covertOrActiveUsersSinceFilter" class="org.springframework.data.gemfire.snapshot.filter.ComposableSnapshotFilter"
factory-method="or">
<constructor-arg index="0">
<list>
<ref bean="activeUsersSinceFilter"/>
<bean class="example.gemfire.snapshot.filter.CovertUsersFilter"/>
</list>
</constructor-arg>
</bean>
5.8.3. Snapshot Events
By default, Spring Data for Apache Geode uses Apache Geode’s Snapshot Services on startup to import data and on shutdown
to export data. However, you may want to trigger periodic, event-based snapshots, for either import or export,
from within your Spring application.
For this purpose, Spring Data for Apache Geode defines two additional Spring application events, extending Spring’s
ApplicationEvent
class for imports and exports, respectively: ImportSnapshotApplicationEvent
and ExportSnapshotApplicationEvent
.
The two application events can be targeted for the entire Apache Geode cache or for individual Apache Geode
Regions. The constructors in these classes accept an optional Region pathname (such as /Example
) as well as zero
or more SnapshotMetadata
instances.
The array of SnapshotMetadata
overrides the snapshot metadata defined by <gfe-data:snapshot-import>
and <gfe-data:snapshot-export>
sub-elements, which are used in cases where snapshot application events do not
explicitly provide SnapshotMetadata
. Each individual SnapshotMetadata
instance can define its own location
and filters
properties.
All snapshot service beans defined in the Spring ApplicationContext
receive import and export snapshot
application events. However, only matching Snapshot Service beans process import and export events.
A Region-based [Import|Export]SnapshotApplicationEvent
matches if the Snapshot Service bean defined
is a RegionSnapshotService
and its Region reference (as determined by the region-ref
attribute) matches
the Region’s pathname, as specified by the snapshot application event.
A Cache-based [Import|Export]SnapshotApplicationEvent
(that is, a snapshot application event without a Region pathname)
triggers all Snapshot Service beans, including any RegionSnapshotService
beans, to perform either an import or export,
respectively.
You can use Spring’s
ApplicationEventPublisher
interface to fire import and export snapshot application events from your application as follows:
@Component
public class ExampleApplicationComponent {
@Autowired
private ApplicationEventPublisher eventPublisher;
@Resource(name = "Example")
private Region<?, ?> example;
public void someMethod() {
...
File dataSnapshot = new File(System.getProperty("user.dir"), "/path/to/export/data.snapshot");
SnapshotFilter myFilter = ...;
SnapshotMetadata exportSnapshotMetadata =
new SnapshotMetadata(dataSnapshot, myFilter, null);
ExportSnapshotApplicationEvent exportSnapshotEvent =
new ExportSnapshotApplicationEvent(this, example.getFullPath(), exportSnapshotMetadata)
eventPublisher.publishEvent(exportSnapshotEvent);
...
}
}
In the preceding example, only the /Example
Region’s Snapshot Service bean picks up and handles the export event,
saving the filtered, “/Example” Region’s data to the data.snapshot
file in a sub-directory of the application’s
working directory.
Using the Spring application events and messaging subsystem is a good way to keep your application loosely coupled.
You can also use Spring’s Scheduling services to fire
snapshot application events on a periodic basis.
5.9. Configuring the Function Service
Spring Data for Apache Geode provides annotation support for implementing, registering and executing
Apache Geode Functions.
Spring Data for Apache Geode also provides XML namespace support for registering Apache Geode
Functions
for remote function execution.
See Apache Geode’s documentation
for more information on the Function execution framework.
Apache Geode Functions are declared as Spring beans and must implement the org.apache.geode.cache.execute.Function
interface or extend org.apache.geode.cache.execute.FunctionAdapter
.
The namespace uses a familiar pattern to declare Functions, as the following example shows:
<gfe:function-service>
<gfe:function>
<bean class="example.FunctionOne"/>
<ref bean="function2"/>
</gfe:function>
</gfe:function-service>
<bean id="function2" class="example.FunctionTwo"/>
5.10. Configuring WAN Gateways
WAN Gateways provides a way to synchronize Apache Geode Distributed Systems across geographic locations.
Spring Data for Apache Geode provides XML namespace support for configuring WAN Gateways as illustrated in the following examples.
5.10.1. WAN Configuration in Apache Geode 7.0
In the following example, GatewaySenders
are configured for a PARTITION
Region by adding child elements
(gateway-sender
and gateway-sender-ref
) to the Region. A GatewaySender
may register EventFilters
and TransportFilters
.
The following example also shows a sample configuration of an AsyncEventQueue
, which must also be auto-wired
into a Region (not shown):
<gfe:partitioned-region id="region-with-inner-gateway-sender" >
<gfe:gateway-sender remote-distributed-system-id="1">
<gfe:event-filter>
<bean class="org.springframework.data.gemfire.example.SomeEventFilter"/>
</gfe:event-filter>
<gfe:transport-filter>
<bean class="org.springframework.data.gemfire.example.SomeTransportFilter"/>
</gfe:transport-filter>
</gfe:gateway-sender>
<gfe:gateway-sender-ref bean="gateway-sender"/>
</gfe:partitioned-region>
<gfe:async-event-queue id="async-event-queue" batch-size="10" persistent="true" disk-store-ref="diskstore"
maximum-queue-memory="50">
<gfe:async-event-listener>
<bean class="example.AsyncEventListener"/>
</gfe:async-event-listener>
</gfe:async-event-queue>
<gfe:gateway-sender id="gateway-sender" remote-distributed-system-id="2">
<gfe:event-filter>
<ref bean="event-filter"/>
<bean class="org.springframework.data.gemfire.example.SomeEventFilter"/>
</gfe:event-filter>
<gfe:transport-filter>
<ref bean="transport-filter"/>
<bean class="org.springframework.data.gemfire.example.SomeTransportFilter"/>
</gfe:transport-filter>
</gfe:gateway-sender>
<bean id="event-filter" class="org.springframework.data.gemfire.example.AnotherEventFilter"/>
<bean id="transport-filter" class="org.springframework.data.gemfire.example.AnotherTransportFilter"/>
On the other end of a GatewaySender
is a corresponding GatewayReceiver
to receive Gateway events.
The GatewayReceiver
may also be configured with EventFilters
and TransportFilters
, as follows:
<gfe:gateway-receiver id="gateway-receiver" start-port="12345" end-port="23456" bind-address="192.168.0.1">
<gfe:transport-filter>
<bean class="org.springframework.data.gemfire.example.SomeTransportFilter"/>
</gfe:transport-filter>
</gfe:gateway-receiver>
See the Apache Geode
documentation
for a detailed explanation of all the configuration options.
6. Bootstrapping Apache Geode with the Spring Container using Annotations
Spring Data for Apache Geode (SDG) 2.0 introduces a new annotation-based configuration model
to configure and bootstrap Apache Geode using the Spring container.
The primary motivation for introducing an annotation-based approach to the configuration of Apache Geode
in a Spring context is to enable Spring application developers to get up and running as quickly
and as easily as possible.
Let’s get started!
If you would like to get started even faster, refer to
the Quick Start section.
6.1. Introduction
Apache Geode, along with Spring Data for Apache Geode, offer many configuration options:
In addition, Apache Geode and Spring Data for Apache Geode both support different topologies:
-
-
-
-
distributed system design patterns
(such as shared-nothing architecture).
All of these configuration options and topology arrangements can pose challenges when setting up and using
Apache Geode properly. The Spring Data for Apache Geode annotation-based configuration model aims to
simplify configuration in the context of topology, plus more.
The annotation-based configuration model is an alternative to XML-based configuration using Spring Data for Apache Geode’s XML namespace.
With XML, you could use the gfe
XML schema for configuration and the gfe-data
XML schema for data access.
See "Bootstrapping Apache Geode with the Spring Container" for more details.
As of SDG 2.0, the annotation-based configuration model does not yet support the configuration of
Apache Geode’s WAN components and topology.
Like Spring Boot, Spring Data for Apache Geode’s annotation-based configuration model was designed as an opinionated,
convention-over-configuration approach for using Apache Geode. Indeed, this annotation-based configuration model
was inspired by Spring Boot as well as several other Spring and Spring Data projects, collectively.
By following convention, all annotations provide reasonable and sensible defaults for all configuration attributes.
The default value for a given annotation attribute directly corresponds to the default value provided in
Apache Geode for the same configuration property.
The intention is to let you enable Apache Geode features or an embedded services by declaring the appropriate
annotation on your Spring @Configuration
or @SpringBootApplication
class without needing to unnecessarily configure
a large number of properties just to use the feature or service.
Again, getting started, quickly and as easily, is the primary objective.
However, the option to customize the configuration metadata and behavior of Apache Geode is there if you need it,
and Spring Data for Apache Geode’s annotation-based configuration quietly backs away. You need only specify the configuration attributes
you wish to adjust. Also, as we will see later in this document, there are several ways to configure a Apache Geode
feature or embedded service by using the annotations.
You can find all the new SDG Java Annotations
in the org.springframework.data.gemfire.config.annotation
package.
6.2. Configuring Apache Geode Applications with Spring
Like all Spring Boot applications that begin by annotating the application class with @SpringBootApplication
,
a Spring Boot application can easily become a Apache Geode cache application by declaring any one of three
main annotations:
-
@ClientCacheApplication
-
@PeerCacheApplication
-
@CacheServerApplication
These three annotations are the Spring application developer’s starting point when working with Apache Geode.
To realize the intent behind these annotations, you must understand that there are two types of cache instances
that can be created with Apache Geode: a client cache or a peer cache.
You can configure a Spring Boot application as a Apache Geode cache client with an instance of ClientCache
,
which can communicate with an existing cluster of Apache Geode servers used to manage the application’s data.
The client-server topology is the most common system architecture employed when using Apache Geode and you can
make your Spring Boot application a cache client, with a ClientCache
instance, simply by annotating it with
@ClientCacheApplication
.
Alternatively, a Spring Boot application may be a peer member of a Apache Geode cluster. That is, the application
itself is just another server in a cluster of servers that manages data. The Spring Boot application creates
an "embedded", peer Cache
instance when you annotate your application class with @PeerCacheApplication
.
By extension, a peer cache application may also serve as a CacheServer
too, allowing cache clients to connect
and perform data access operations on the server. This is accomplished by annotating the application class with
@CacheServerApplication
in place of @PeerCacheApplication
, which creates a peer Cache
instance along with
the CacheServer
that allows cache clients to connect.
A Apache Geode server is not necessarily a cache server by default. That is, a server is not necessarily
set up to serve cache clients just because it is a server. A Apache Geode server can be a peer member (data node)
of the cluster managing data without serving any clients while other peer members in the cluster are indeed set up
to serve clients in addition to managing data. It is also possible to set up certain peer members in the cluster as
non-data nodes, called data accessors,
which do not store data, but act as a proxy to service clients as CacheServers
. Many different topologies
and cluster arrangements are supported by Apache Geode, but are beyond the scope of this document.
By way of example, if you want to create a Spring Boot cache client application, start with the following:
Spring-based Apache Geode ClientCache
application
@SpringBootApplication
@ClientCacheApplication
class ClientApplication { .. }
Or, if you want to create a Spring Boot application with an embedded peer Cache
instance, where your application
will be a server and peer member of a cluster (distributed system) formed by Apache Geode,
start with the following:
Spring-based Apache Geode embedded peer Cache
application
@SpringBootApplication
@PeerCacheApplication
class ServerApplication { .. }
Alternatively, you can use the @CacheServerApplication
annotation in place of @PeerCacheApplication
to create
both an embedded peer Cache
instance along with a CacheServer
running on localhost
, listening on the default
cache server port, 40404
, as follows:
Spring-based Apache Geode embedded peer Cache
application with CacheServer
@SpringBootApplication
@CacheServerApplication
class ServerApplication { .. }
6.3. Client/Server Applications In-Detail
There are multiple ways that a client can connect to and communicate with servers in a Apache Geode cluster.
The most common and recommended approach is to use Apache Geode Locators.
A cache client can connect to one or more Locators in the Apache Geode cluster instead of directly to a
CacheServer
. The advantage of using Locators over direct CacheServer
connections is that Locators provide metadata
about the cluster to which the client is connected. This metadata includes information such as which servers contain
the data of interest or which servers have the least amount of load. A client Pool
in conjunction with a Locator
also provides fail-over capabilities in case a CacheServer
crashes. By enabling the PARTITION
Region (PR)
single-hop feature in the client Pool
, the client is routed directly to the server containing the data requested
and needed by the client.
Locators are also peer members in a cluster. Locators actually constitute what makes up a cluster of
Apache Geode nodes. That is, all nodes connected by a Locator are peers in the cluster, and new members
use Locators to join a cluster and find other members.
By default, Apache Geode sets up a "DEFAULT" Pool
connected to a CacheServer
running on localhost
,
listening on port 40404
when a ClientCache
instance is created. A CacheServer
listens on port 40404
,
accepting connections on all system NICs. You do not need to do anything special to use the client-server topology.
Simply annotate your server-side Spring Boot application with @CacheServerApplication
and your client-side
Spring Boot application with @ClientCacheApplication
, and you are ready to go.
If you prefer, you can even start your servers with Gfsh’s start server
command. Your Spring Boot @ClientCacheApplication
can still connect to the server regardless of how it was started. However, you may prefer to configure and start your
servers by using the Spring Data for Apache Geode approach since a properly annotated Spring Boot application class is far more intuitive
and easier to debug.
As an application developer, you will no doubt want to customize the "DEFAULT" Pool
set up by Apache Geode
to possibly connect to one or more Locators, as the following example demonstrates:
Spring-based Apache Geode ClientCache
application using Locators
@SpringBootApplication
@ClientCacheApplication(locators = {
@Locator(host = "boombox" port = 11235),
@Locator(host = "skullbox", port = 12480)
})
class ClientApplication { .. }
Along with the locators
attribute, the @ClientCacheApplication
annotation has a servers
attribute as well.
The servers
attribute can be used to specify one or more nested @Server
annotations that let the cache client
connect directly to one or more servers, if necessary.
You can use either the locators
or servers
attribute, but not both (this is enforced by Apache Geode).
You can also configure additional Pool
instances (other than the "DEFAULT" Pool
provided by Apache Geode
when a ClientCache
instance is created with the @ClientCacheApplication
annotation) by using the @EnablePool
and @EnablePools
annotations.
@EnablePools
is a composite annotation that aggregates several nested @EnablePool
annotations on
a single class. Java 8 and earlier does not allow more than one annotation of the same type to be declared
on a single class.
The following example uses the @EnablePool
and @EnablePools
annotations:
Spring-based Apache Geode ClientCache
application using multiple named Pools
@SpringBootApplication
@ClientCacheApplication(logLevel = "info")
@EnablePool(name = "VenusPool", servers = @Server(host = "venus", port = 48484),
min-connections = 50, max-connections = 200, ping-internal = 15000,
prSingleHopEnabled = true, readTimeout = 20000, retryAttempts = 1,
subscription-enable = true)
@EnablePools(pools = {
@EnablePool(name = "SaturnPool", locators = @Locator(host="skullbox", port=20668),
subsription-enabled = true),
@EnablePool(name = "NeptunePool", severs = {
@Server(host = "saturn", port = 41414),
@Server(host = "neptune", port = 42424)
}, min-connections = 25))
})
class ClientApplication { .. }
The name
attribute is the only required attribute of the @EnablePool
annotation. As we will see later, the value
of the name
attribute corresponds to both the name of the Pool
bean created in the Spring container as well as
the name used to reference the corresponding configuration properties. It is also the name of the Pool
registered
and used by Apache Geode.
Similarly, on the server, you can configure multiple CacheServers
that a client can connect to, as follows:
Spring-based Apache Geode CacheServer
application using multiple named CacheServers
@SpringBootApplication
@CacheSeverApplication(logLevel = "info", autoStartup = true, maxConnections = 100)
@EnableCacheServer(name = "Venus", autoStartup = true,
hostnameForClients = "venus", port = 48484)
@EnableCacheServers(servers = {
@EnableCacheServer(name = "Saturn", hostnameForClients = "saturn", port = 41414),
@EnableCacheServer(name = "Neptune", hostnameForClients = "neptune", port = 42424)
})
class ServerApplication { .. }
Like @EnablePools
, @EnableCacheServers
is a composite annotation for aggregating multiple @EnableCacheServer
annotations on a single class. Again, Java 8 and earlier does not allow more than one annotation of the same type
to be declared on a single class.
One thing an observant reader may have noticed is that, in all cases, you have specified hard-coded values for all
hostnames, ports, and configuration-oriented annotation attributes. This is not ideal when the application gets
promoted and deployed to different environments, such as from DEV to QA to STAGING to PROD.
The next section covers how to handle dynamic configuration determined at runtime.
6.4. Configuring and Bootstrapping Locators
Besides Apache Geode Cache applications, you may also create Apache Geode Locator applications.
A Apache Geode Locator is a JVM process that allows nodes to join a Apache Geode cluster as peer members.
Locators also enable clients to discover servers in a cluster. A Locator provides meta-data to the clients to uniformly
balance the load across the members in the cluster, enables single-hop data access operations, along with other things.
A complete discussion of Locators is beyond the scope of this document. Readers are encouraged to read
the Apache Geode User Guide
for more details on Locators and their role in the cluster.
To configure and bootstrap a standalone Locator process, do the following:
Spring Boot, Apache Geode Locator Application
@SpringBootApplication
@LocatorApplication(port = 12345)
class LocatorApplication { ... }
You can start multiple Locators in your cluster. The only requirement is that the member name must be unique
in the cluster. Use the name
attribute of the @LocatorApplication
annotation to name the member Locator
in the cluster accordingly. Alternatively, you can set the spring.data.gemfire.locator.name
property in Spring Boot’s
application.properties
.
Additionally, you must ensure that each Locator starts on a unique port if you fork multiple Locators on the same
machine. Set either the port
annotation attribute or the spring.data.gemfire.locator.port
property.
You may then start 1 or more Apache Geode peer cache members in the cluster joined by the Locator, or Locators,
also configured and bootstrapped with Spring, like so:
Spring Boot, Apache Geode CacheServer
Application joined by the Locator on localhost
, port 12345
@SpringBootApplication
@CacheServerApplication(locators = "localhost[12345]")
class ServerApplication { ... }
Again, you can start as many of the ServerApplication
classes, joined by our Locator above, as you want.
You just need to make sure the member is uniquely named.
@LocatorApplication
is for configuring and bootstrapping standalone, Apache Geode Locator application processes.
This process can only be a Locator and nothing else. If you try to start a Locator with a cache instance, SDG will
throw an error.
If you want to simultaneously start a cache instance along with an embedded Locator, then you should use
the @EnableLocator
annotation instead.
Starting an embedded Locator is convenient during development. However, it is highly recommended that you run standalone
Locator processes in production for high availability. If all your Locators in the cluster go down, then the cluster
will remain intact, however, no new members will be able to join the cluster, which is important to scale-out linearly
in order to satisfy demand.
See the section on Configuring an Embedded Locator
for more details.
6.5. Runtime configuration using Configurers
Another goal when designing the annotation-based configuration model was to preserve type safety in the annotation
attributes. For example, if the configuration attribute could be expressed as an int
(such as a port number),
then the attribute’s type should be an int
.
Unfortunately, this is not conducive to dynamic and resolvable configuration at runtime.
One of the finer features of Spring is the ability to use property placeholders and SpEL expressions
in properties or attributes of the configuration metadata when configuring beans in the Spring container.
However, this would require all annotation attributes to be of type String
, thereby giving up type safety,
which is not desirable.
So, Spring Data for Apache Geode borrows from another commonly used pattern in Spring, Configurers
. Many different Configurer
interfaces are provided in Spring Web MVC, including the
org.springframework.web.servlet.config.annotation.ContentNegotiationConfigurer
.
The Configurers
design pattern enables application developers to receive a callback to customize the configuration
of a component or bean on startup. The framework calls back to user-provided code to adjust the configuration
at runtime. One of the more common uses of this pattern is to supply conditional configuration based on
the application’s runtime environment.
Spring Data for Apache Geode provides several Configurer
callback interfaces to customize different aspects of the annotation-based
configuration metadata at runtime, before the Spring managed beans that the annotations create are initialized:
-
CacheServerConfigurer
-
ClientCacheConfigurer
-
ContinuousQueryListenerContainerConfigurer
-
DiskStoreConfigurer
-
IndexConfigurer
-
PeerCacheConfigurer
-
PoolConfigurer
-
RegionConfigurer
-
GatewayReceiverConfigurer
-
GatewaySenderConfigurer
For example, you can use the CacheServerConfigurer
and ClientCacheConfigurer
to customize the port numbers
used by your Spring Boot CacheServer
and ClientCache
applications, respectively.
Consider the following example from a server application:
Customizing a Spring Boot CacheServer
application with a CacheServerConfigurer
@SpringBootApplication
@CacheServerApplication(name = "SpringServerApplication")
class ServerApplication {
@Bean
CacheServerConfigurer cacheServerPortConfigurer(
@Value("${gemfire.cache.server.host:localhost}") String cacheServerHost
@Value("${gemfire.cache.server.port:40404}") int cacheServerPort) {
return (beanName, cacheServerFactoryBean) -> {
cacheServerFactoryBean.setBindAddress(cacheServerHost);
cacheServerFactoryBean.setHostnameForClients(cacheServerHost);
cacheServerFactoryBean.setPort(cacheServerPort);
};
}
}
Next, consider the following example from a client application:
Customizing a Spring Boot ClientCache
application with a ClientCacheConfigurer
@SpringBootApplication
@ClientCacheApplication
class ClientApplication {
@Bean
ClientCacheConfigurer clientCachePoolPortConfigurer(
@Value("${gemfire.cache.server.host:localhost}") String cacheServerHost
@Value("${gemfire.cache.server.port:40404}") int cacheServerPort) {
return (beanName, clientCacheFactoryBean) ->
clientCacheFactoryBean.setServers(Collections.singletonList(
new ConnectionEndpoint(cacheServerHost, cacheServerPort)));
}
}
By using the provided Configurers
, you can receive a callback to further customize the configuration
that is enabled by the associated annotation at runtime, during startup.
In addition, when the Configurer
is declared as a bean in the Spring container, the bean definition can take advantage
of other Spring container features, such as property placeholders, SpEL expressions by using the @Value
annotation
on factory method parameters, and so on.
All Configurers
provided by Spring Data for Apache Geode take two bits of information in the callback: the name of the bean created
in the Spring container by the annotation and a reference to the FactoryBean
used by the annotation to
create and configure the Apache Geode component (for example, a ClientCache
instance is created
and configured with ClientCacheFactoryBean
).
SDG FactoryBeans
are part of the SDG public API and are what you would use in Spring’s
Java-based container configuration if this new annotation-based
configuration model were not provided. Indeed, the annotations themselves are using these same FactoryBeans
for their configuration. So, in essence, the annotations are a facade providing an extra layer of abstraction
for convenience.
Given that a Configurer
can be declared as a regular bean definition like any other POJO, you can combine different
Spring configuration options, such as the use of Spring Profiles with Conditions
that use both property placeholders
and SpEL expressions. These and other nifty features let you create even more sophisticated and flexible configurations.
However, Configurers
are not the only option.
6.6. Runtime configuration using Properties
In addition to Configurers
, each annotation attribute in the annotation-based configuration model is associated with
a corresponding configuration property (prefixed with spring.data.gemfire.
), which can be declared in a Spring Boot
application.properties
file.
Building on the earlier examples, the client’s application.properties
file would define the following
set of properties:
Client application.properties
spring.data.gemfire.cache.log-level=info
spring.data.gemfire.pool.Venus.servers=venus[48484]
spring.data.gemfire.pool.Venus.max-connections=200
spring.data.gemfire.pool.Venus.min-connections=50
spring.data.gemfire.pool.Venus.ping-interval=15000
spring.data.gemfire.pool.Venus.pr-single-hop-enabled=true
spring.data.gemfire.pool.Venus.read-timeout=20000
spring.data.gemfire.pool.Venus.subscription-enabled=true
spring.data.gemfire.pool.Saturn.locators=skullbox[20668]
spring.data.gemfire.pool.Saturn.subscription-enabled=true
spring.data.gemfire.pool.Neptune.servers=saturn[41414],neptune[42424]
spring.data.gemfire.pool.Neptune.min-connections=25
The corresponding server’s application.properties
file would define the following properties:
Server application.properties
spring.data.gemfire.cache.log-level=info
spring.data.gemfire.cache.server.port=40404
spring.data.gemfire.cache.server.Venus.port=43434
spring.data.gemfire.cache.server.Saturn.port=41414
spring.data.gemfire.cache.server.Neptune.port=41414
Then you can simplify the @ClientCacheApplication
class to the following:
Spring @ClientCacheApplication
class
@SpringBootApplication
@ClientCacheApplication
@EnablePools(pools = {
@EnablePool(name = "Venus"),
@EnablePool(name = "Saturn"),
@EnablePool(name = "Neptune")
})
class ClientApplication { .. }
Also, the @CacheServerApplication
class becomes the following:
Spring @CacheServerApplication
class
@SpringBootApplication
@CacheServerApplication(name = "SpringServerApplication")
@EnableCacheServers(servers = {
@EnableCacheServer(name = "Venus"),
@EnableCacheServer(name = "Saturn"),
@EnableCacheServer(name = "Neptune")
})
class ServerApplication { .. }
The preceding example shows why it is important to "name" your annotation-based beans (other than because it is required
in certain cases). Doing so makes it possible to reference the bean in the Spring container from XML, properties,
and Java. It is even possible to inject annotation-defined beans into an application class, for whatever purpose,
as the following example demonstrates:
@Component
class MyApplicationComponent {
@Resource(name = "Saturn")
CacheServer saturnCacheServer;
...
}
Likewise, naming an annotation-defined bean lets you code a Configurer
to customize a specific, "named" bean
since the beanName
is 1 of 2 arguments passed to the callback.
Oftentimes, an associated annotation attribute property takes two forms: a "named" property along with
an "unnamed" property.
The following example shows such an arrangement:
spring.data.gemfire.cache.server.bind-address=10.105.20.1
spring.data.gemfire.cache.server.Venus.bind-address=10.105.20.2
spring.data.gemfire.cache.server.Saturn...
spring.data.gemfire.cache.server.Neptune...
While there are three named CacheServers
above, there is also one unnamed CacheServer
property providing
the default value for any unspecified value of that property, even for "named" CacheServers
. So, while "Venus"
sets and overrides its own bind-address
, "Saturn" and "Neptune" inherit from the "unnamed"
spring.data.gemfire.cache.server.bind-address
property.
See an annotation’s Javadoc for which annotation attributes support property-based configuration
and whether they support "named" properties over default, "unnamed" properties.
6.6.1. Properties
of Properties
In the usual Spring fashion, you can even express Properties
in terms of other Properties
, whether that is by
The following example shows a nested property being set in an application.properties
file:
Properties of Properties
spring.data.gemfire.cache.server.port=${gemfire.cache.server.port:40404}
The following example shows a nested property being set in Java:
Property placehodler nesting
@Bean
CacheServerConfigurer cacheServerPortConfigurer(
@Value("${gemfire.cache.server.port:${some.other.property:40404}}")
int cacheServerPort) {
...
}
Property placeholder nesting can be arbitrarily deep.
6.7. Configuring Embedded Services
Apache Geode provides the ability to start many different embedded services that are required by an application,
depending on the use case.
6.7.1. Configuring an Embedded Locator
As mentioned previously, Apache Geode Locators are used by clients to connect to and find servers in a cluster.
In addition, new members joining an existing cluster use Locators to find their peers.
It is often convenient for application developers as they are developing their Spring Boot and Spring Data for Apache Geode
applications to startup up a small cluster of two or three Apache Geode servers. Rather than starting
a separate Locator process, you can annotate your Spring Boot @CacheServerApplication
class with @EnableLocator
,
as follows:
Spring, Apache Geode CacheServer
application running an embedded Locator
@SpringBootApplication
@CacheServerApplication
@EnableLocator
class ServerApplication { .. }
The @EnableLocator
annotation starts an embedded Locator in the Spring Apache Geode CacheServer
application
running on localhost
, listening on the default Locator port, 10334
. You can customize the host
(bind address)
and port
that the embedded Locator binds to by using the corresponding annotation attributes.
Alternatively, you can set the @EnableLocator
attributes by setting the corresponding
spring.data.gemfire.locator.host
and spring.data.gemfire.locator.port
properties in application.properties
.
Then you can start other Spring Boot @CacheServerApplication
-enabled applications by connecting to this
Locator with the following:
Spring, Apache Geode CacheServer
application connecting to a Locator
@SpringBootApplication
@CacheServerApplication(locators = "localhost[10334]")
class ServerApplication { .. }
You can even combine both application classes shown earlier into a single class and use your IDE to create different
run profile configurations to launch different instances of the same class with slightly modified configuration
by using Java system properties, as follows:
Spring CacheServer
application running an embedded Locator and connecting to the Locator
@SpringBootApplication
@CacheServerApplication(locators = "localhost[10334]")
public class ServerApplication {
public static void main(String[] args) {
SpringApplication.run(ServerApplication.class);
}
@EnableLocator
@Profile("embedded-locator")
static class Configuration { }
}
Then, for each run profile, you can set and change the following system properties:
IDE run profile configuration
spring.data.gemfire.name=SpringCacheServerOne
spring.data.gemfire.cache.server.port=41414
spring.profiles.active=embedded-locator
Only 1 of the run profiles for the ServerApplication
class should set the -Dspring.profiles.active=embedded-locator
Java system property. Then you can change the ..name
and ..cache.server.port
for each of the other run profiles
and have a small cluster (distributed system) of Apache Geode servers running on your local system.
The @EnableLocator
annotation was meant to be a development-time annotation only and not something
an application developer would use in production. We strongly recommend running Locators as standalone,
independent processes in the cluster.
More details on how Apache Geode Locators work can be found
here.
6.7.2. Configuring an Embedded Manager
A Apache Geode Manager is another peer member or node in the cluster that is responsible for cluster "management".
Management involves creating Regions
, Indexes
, DiskStores
, among other things, along with monitoring the runtime
operations and behavior of the cluster components.
The Manager lets a JMX-enabled client (such as the Gfsh shell tool) connect to the Manager to manage the cluster.
It is also possible to connect to a Manager with JDK-provided tools such as JConsole or JVisualVM, given that these are
both JMX-enabled clients as well.
Perhaps you would also like to enable the Spring @CacheServerApplication
shown earlier as a Manager as well. To do so,
annotate your Spring @Configuration
or @SpringBootApplication
class with @EnableManager
.
By default, the Manager binds to localhost
, listening on the default Manager port of 1099
. Several aspects of
the Manager can be configured with annotation attributes or the corresponding properties.
The following example shows how to create an embedded Manager in Java:
Spring CacheServer
application running an embedded Manager
@SpringBootApplication
@CacheServerApplication(locators = "localhost[10334]")
public class ServerApplication {
public static void main(String[] args) {
SpringApplication.run(ServerApplication.class);
}
@EnableLocator
@EnableManager
@Profile("embedded-locator-manager")
static class Configuration { }
}
With the preceding class, you can even use Gfsh to connect to the small cluster and manage it, as follows:
$ gfsh
_________________________ __
/ _____/ ______/ ______/ /____/ /
/ / __/ /___ /_____ / _____ /
/ /__/ / ____/ _____/ / / / /
/______/_/ /______/_/ /_/ 1.2.1
Monitor and Manage {data-store-name}
gfsh>connect
Connecting to Locator at [host=localhost, port=10334] ..
Connecting to Manager at [host=10.99.199.5, port=1099] ..
Successfully connected to: [host=10.99.199.5, port=1099]
gfsh>list members
Name | Id
---------------------- | ----------------------------------------------------
SpringCacheServerOne | 10.99.199.5(SpringCacheServerOne:14842)<ec><v0>:1024
SpringCacheServerTwo | 10.99.199.5(SpringCacheServerTwo:14844)<v1>:1025
SpringCacheServerThree | 10.99.199.5(SpringCacheServerThree:14846)<v2>:1026
Because we also have the embedded Locator enabled, we can connect indirectly to the Manager through the Locator.
A Locator lets JMX clients connect and find a Manager in the cluster. If none exists, the Locator assumes the role
of a Manager. However, if no Locator exists, we would need to connect directly to the Manager by using the following:
Gfsh connect
command connecting directly to the Manager
gfsh>connect --jmx-manager=localhost[1099]
Like the @EnableLocator
annotation, the @EnableManager
annotation is also meant to be a development-time
only annotation and not something an application developer would use in production. We strongly recommend
that Managers, like Locators, be standalone, independent and dedicated processes in the cluster.
More details on Apache Geode management and monitoring can be found
here.
6.7.3. Configuring the Embedded HTTP Server
Apache Geode is also capable of running an embedded HTTP server. The current implementation is backed by
Eclipse Jetty.
The embedded HTTP server is used to host Apache Geode’s Management (Admin) REST API (not a publicly advertised API),
the Developer REST API,
and the Pulse Monitoring Web Application.
However, to use any of these Apache Geode-provided web applications, you must have a full installation of
Apache Geode installed on your system, and you must set the GEODE_HOME
environment variable to
your installation directory.
To enable the embedded HTTP server, add the @EnableHttpService
annotation to any @PeerCacheApplication
or @CacheServerApplication
annotated class, as follows:
Spring CacheServer
application running the embedded HTTP server
@SpringBootApplication
@CacheServerApplication
@EnableHttpService
public class ServerApplication { .. }
By default, the embedded HTTP server listens on port 7070
for HTTP client requests. Of course, you can use
the annotation attributes or corresponding configuration properties to adjust the port as needed.
Follow the earlier links for more details on HTTP support and the services provided.
6.7.4. Configuring the Embedded Memcached Server (Gemcached)
Apache Geode also implements the Memcached protocol with the ability to service Memcached clients. That is,
Memcached clients can connect to a Apache Geode cluster and perform Memcached operations as if
the Apache Geode servers in the cluster were actual Memcached servers.
To enable the embedded Memcached service, add the @EnableMemcachedServer
annotation to any @PeerCacheApplication
or @CacheServerApplication
annotated class, as follows:
Spring CacheServer
application running an embedded Memcached server
@SpringBootApplication
@CacheServerApplication
@EnabledMemcachedServer
public class ServerApplication { .. }
More details on Apache Geode’s Memcached service (called "Gemcached") can be found
here.
6.7.5. Configuring the Embedded Redis Server
Apache Geode also implements the Redis server protocol, which enables Redis clients to connect to and communicate
with a cluster of Apache Geode servers to issue Redis commands. As of this writing, the Redis server protocol
support in Apache Geode is still experimental.
To enable the embedded Redis service, add the @EnableRedisServer
annotation to any @PeerCacheApplication
or @CacheServerApplication
annotated class, as follows:
Spring CacheServer
application running an embedded Redis server
@SpringBootApplication
@CacheServerApplication
@EnableRedisServer
public class ServerApplication { .. }
You must explicitly declare the org.apache.geode:geode-redis
module on your Spring [Boot] application
classpath.
More details on Apache Geode’s Redis adapter can be found
here.
6.8. Configuring Logging
Oftentimes, it is necessary to turn up logging in order to understand exactly what Apache Geode is doing and when.
To enable Logging, annotate your application class with @EnableLogging
and set the appropriate attributes
or associated properties, as follows:
Spring ClientCache
application with Logging enabled
@SpringBootApplication
@ClientCacheApplication
@EnableLogging(logLevel="info", logFile="/absolute/file/system/path/to/application.log)
public class ClientApplication { .. }
While the logLevel
attribute can be specified with all the cache-based application annotations
(for example, @ClientCacheApplication(logLevel="info")
), it is easier to customize logging behavior with
the @EnableLogging
annotation.
Additionally, you can configure the log-level
by setting the spring.data.gemfire.logging.level
property
in application.properties
.
See the @EnableLogging
annotation Javadoc
for more details.
6.9. Configuring Statistics
To gain even deeper insight into Apache Geode at runtime, you can enable statistics. Gathering statistical data
facilitates system analysis and troubleshooting when complex problems, which are often distributed in nature
and where timing is a crucial factor, occur.
When statistics are enabled, you can use Apache Geode’s
VSD (Visual Statistics Display) tool
to analyze the statistical data that is collected.
To enable statistics, annotate your application class with @EnableStatistics
, as follows:
Spring ClientCache
application with Statistics enabled
@SpringBootApplication
@ClientCacheApplication
@EnableStatistics
public class ClientApplication { .. }
Enabling statistics on a server is particularly valuable when evaluating performance. To do so,
annotate your @PeerCacheApplication
or @CacheServerApplication
class with @EnableStatistics
.
You can use the @EnableStatistics
annotation attributes or associated properties to customize
the statistics gathering and collection process.
See the @EnableStatistics
annotation Javadoc
for more details.
More details on Apache Geode’s statistics can be found
here.
6.10. Configuring PDX
One of the more powerful features of Apache Geode is
PDX serialization.
While a complete discussion of PDX is beyond the scope of this document, serialization using PDX is a much better
alternative to Java serialization, with the following benefits:
-
PDX uses a centralized type registry to keep the serialized bytes of an object more compact.
-
PDX is a neutral serialization format, allowing both Java and Native clients to operate on the same data set.
-
PDX supports versioning and lets object fields be added or removed without affecting existing applications
using either older or newer versions of the PDX serialized objects that have changed, without data loss.
-
PDX lets object fields be accessed individually in OQL query projections and predicates without the object
needing to be de-serialized first.
In general, serialization in Apache Geode is required any time data is transferred to or from clients and servers
or between peers in a cluster during normal distribution and replication processes as well as when data is overflowed
or persisted to disk.
Enabling PDX serialization is much simpler than modifying all of your application domain object types to implement
java.io.Serializable
, especially when it may be undesirable to impose such restrictions on your
application domain model or you do not have any control over the objects your are serializing, which is especially
true when using a 3rd party library (e.g. think of a geo-spatial API with Coordinate
types).
To enable PDX, annotate your application class with @EnablePdx
, as follows:
Spring ClientCache
application with PDX enabled
@SpringBootApplication
@ClientCacheApplication
@EnablePdx
public class ClientApplication { .. }
Typically, an application’s domain object types either implements the
org.apache.geode.pdx.PdxSerializable
interface or you can implement and register a non-invasive implementation of the
org.apache.geode.pdx.PdxSerializer
interface to handle all the application domain object types that need to be serialized.
Unfortunately, Apache Geode only lets one PdxSerializer
be registered, which suggests that all application
domain object types need to be handled by a single PdxSerializer
instance. However, that is a serious anti-pattern
and an unmaintainable practice.
Even though only a single PdxSerializer
instance can be registered with Apache Geode, it makes sense to create a
single PdxSerializer
implementation per application domain object type.
By using the Composite Software Design Pattern, you can provide
an implementation of the PdxSerializer
interface that aggregates all of the application domain object type-specific
PdxSerializer
instances, but acts as a single PdxSerializer
instance and register it.
You can declare this composite PdxSerializer
as a managed bean in the Spring container and refer to this composite
PdxSerializer
by its bean name in the @EnablePdx
annotation using the serializerBeanName
attribute. Spring Data for Apache Geode
takes care of registering it with Apache Geode on your behalf.
The following example shows how to create a custom composite PdxSerializer
:
Spring ClientCache
application with PDX enabled, using a custom composite PdxSerializer
@SpringBootApplication
@ClientCacheApplication
@EnablePdx(serializerBeanName = "compositePdxSerializer")
public class ClientApplication {
@Bean
PdxSerializer compositePdxSerializer() {
return new CompositePdxSerializerBuilder()...
}
}
It is also possible to declare Apache Geode’s
org.apache.geode.pdx.ReflectionBasedAutoSerializer
as a bean definition in a Spring context.
Alternatively, you should use Spring Data for Apache Geode’s more robust
org.springframework.data.gemfire.mapping.MappingPdxSerializer
,
which uses Spring Data mapping metadata and infrastructure applied to the serialization process for more efficient
handling than reflection alone.
Many other aspects and features of PDX can be adjusted with the @EnablePdx
annotation attributes
or associated configuration properties.
See the @EnablePdx
annotation Javadoc
for more details.
6.11. Configuring Apache Geode Properties
While many of the gemfire.properties
are conveniently encapsulated in and abstracted with an annotation in the SDG annotation-based
configuration model, the less commonly used Apache Geode properties are still accessible from
the @EnableGemFireProperties
annotation.
Annotating your application class with @EnableGemFireProperties
is convenient and a nice alternative to creating
a gemfire.properties
file or setting Apache Geode properties as Java system properties on the command line
when launching your application.
We recommend that these Apache Geode properties be set in a gemfire.properties
file when deploying
your application to production. However, at development time, it can be convenient to set these properties individually,
as needed, for prototyping, debugging and testing purposes.
A few examples of some of the less common Apache Geode properties that you usually need not worry about include,
but are not limited to: ack-wait-threshold
, disable-tcp
, socket-buffer-size
, and others.
To individually set any Apache Geode property, annotate your application class with @EnableGemFireProperties
and set the Apache Geode properties you want to change from the default value set by Apache Geode
with the corresponding attribute, as follows:
Spring ClientCache
application with specific Apache Geode properties set
@SpringBootApplication
@ClientCacheApplication
@EnableGemFireProperties(conflateEvents = true, socketBufferSize = 16384)
public class ClientApplication { .. }
Keep in mind that some of the Apache Geode properties are client-specific (for example, conflateEvents
),
while others are server-specific (for example distributedSystemId
, enableNetworkPartitionDetection
,
enforceUniqueHost
, memberTimeout
, redundancyZone
, and others).
More details on Apache Geode properties can be found
here.
6.12. Configuring Regions
So far, outside of PDX, our discussion has centered around configuring Apache Geode’s more administrative functions:
creating a cache instance, starting embedded services, enabling logging and statistics, configuring PDX, and using
gemfire.properties
to affect low-level configuration and behavior. While all these configuration options are important,
none of them relate directly to your application. In other words, we still need some place to store our application data
and make it generally available and accessible.
Apache Geode organizes data in a cache into Regions.
You can think of a Region as a table in a relational database. Generally, a Region should only store a single type of object,
which makes it more conducive for building effective indexes and writing queries. We cover indexing
later.
Previously, Spring Data for Apache Geode users needed to explicitly define and declare the Regions used by their applications to store data
by writing very verbose Spring configuration metadata, whether using SDG’s FactoryBeans
from the API
with Spring’s Java-based container configuration
or using XML.
The following example demonstrates how to configure a Region bean in Java:
Example Region bean definition using Spring’s Java-based container configuration
@Configuration
class GemFireConfiguration {
@Bean("Example")
PartitionedRegionFactoryBean exampleRegion(GemFireCache gemfireCache) {
PartitionedRegionFactoryBean<Long, Example> exampleRegion =
new PartitionedRegionFactoryBean<>();
exampleRegion.setCache(gemfireCache);
exampleRegion.setClose(false);
exampleRegion.setPersistent(true);
return exampleRegion;
}
...
}
The following example demonstrates how to configure the same Region bean in XML:
Example Region bean definition using SDG’s XML Namespace
<gfe:partitioned-region id="exampleRegion" name="Example" persistent="true">
...
</gfe:partitioned-region>
While neither Java nor XML configuration is all that difficult to specify, either one can be cumbersome, especially if
an application requires a large number of Regions. Many relational database-based applications can have hundreds
or even thousands of tables.
Defining and declaring all these Regions by hand would be cumbersome and error prone. Well, now there is a better way.
Now you can define and configure Regions based on their application domain objects (entities) themselves. No longer do
you need to explicitly define Region
bean definitions in Spring configuration metadata, unless you require
finer-grained control.
To simplify Region creation, Spring Data for Apache Geode combines the use of Spring Data Repositories with the expressive
power of annotation-based configuration using the new @EnableEntityDefinedRegions
annotation.
Most Spring Data application developers should already be familiar with the
Spring Data Repository abstraction
and Spring Data for Apache Geode’s implementation/extension,
which has been specifically customized to optimize data access operations for Apache Geode.
First, an application developer starts by defining the application’s domain objects (entities), as follows:
Application domain object type modeling a Book
@Region("Books")
class Book {
@Id
private ISBN isbn;
private Author author;
private Category category;
private LocalDate releaseDate;
private Publisher publisher;
private String title;
}
Next, you define a basic repository for Books
by extending Spring Data Commons
org.springframework.data.repository.CrudRepository
interface, as follows:
Repository for Books
interface BookRepository extends CrudRepository<Book, ISBN> { .. }
The org.springframe.data.repository.CrudRepository
is a Data Access Object (DAO) providing basic data access
operations (CRUD) along with support for simple queries (such as findById(..)
). You can define additional,
more sophisticated queries by declaring query methods on the repository interface
(for example, List<BooK> findByAuthor(Author author);
).
Under the hood, Spring Data for Apache Geode provides an implementation of your application’s repository interfaces when
the Spring container is bootstrapped. SDG even implements the query methods you define so long as you follow
the conventions.
Now, when you defined the Book
class, you also specified the Region in which instances of Book
are mapped (stored)
by declaring the Spring Data for Apache Geode mapping annotation, @Region
on the entity’s type. Of course, if the entity type (Book
,
in this case) referenced in the type parameter of the repository interface (BookRepository
, in this case)
is not annotated with @Region
, the name is derived from the simple class name of the entity type (also Book
,
in this case).
Spring Data for Apache Geode uses the mapping context, which contains mapping metadata for all the entities defined in your application,
to determine all the Regions that are needed at runtime.
To enable and use this feature, annotate the application class with @EnableEntityDefinedRegions
, as follows:
Entity-defined Region Configuration
@SpringBootApplication
@ClientCacheApplication
@EnableEntityDefinedRegions(basePackages = "example.app.domain")
@EnableGemfireRepositories(basePackages = "example.app.repo")
class ClientApplication { .. }
Creating Regions from entity classes is most useful when using Spring Data Repositories in your application.
Spring Data for Apache Geode’s Repository support is enabled with the @EnableGemfireRepositories
annotation, as shown in
the preceding example.
Currently, only entity classes explicitly annotated with @Region
are picked up by the scan
and will have Regions created. If an entity class is not explicitly mapped with @Region
no Region will be created.
By default, the @EnableEntityDefinedRegions
annotation scans for entity classes recursively, starting from
the package of the configuration class on which the @EnableEntityDefinedRegions
annotation is declared.
However, it is common to limit the search during the scan by setting the basePackages
attribute with
the package names containing your application entity classes.
Alternatively, you can use the more type-safe basePackageClasses
attribute for specifying the package to scan
by setting the attribute to an entity type in the package that contains the entity’s class, or by using a non-entity
placeholder class specifically created for identifying the package to scan.
The following example shows how to specify the entity types to scan:
Entity-defined Region Configuration using the Entity class type
@SpringBootApplication
@ClientCacheApplication
@EnableGemfireRepositories
@EnableEntityDefinedRegions(basePackageClasses = {
example.app.books.domain.Book.class,
example.app.customers.domain.Customer.class
})
class ClientApplication { .. }
In addition to specifying where to begin the scan, like Spring’s @ComponentScan
annotation, you can specify include
and exclude
filters with all the same semantics of the org.springframework.context.annotation.ComponentScan.Filter
annotation.
See the @EnableEntityDefinedRegions
annotation Javadoc
for more details.
6.12.1. Configuring Type-specific Regions
Apache Geode supports many different types of Regions.
Each type corresponds to the Region’s DataPolicy
,
which determines exactly how the data in the Region will be managed (i.e. distributed, replicated, and so on).
Other configuration settings (such as the Region’s scope
) can also affect how data is managed.
See “Storage and Distribution Options”
in the Apache Geode User Guide for more details.
When you annotate your application domain object types with the generic @Region
mapping annotation, Spring Data for Apache Geode decides
which type of Region to create. SDG’s default strategy takes the cache type into consideration when
determining the type of Region to create.
For example, if you declare the application as a ClientCache
by using the @ClientCacheApplication
annotation,
SDG creates a client PROXY
Region
by default. Alternatively, if you declare the application as a
peer Cache
by using either the @PeerCacheApplication
or @CacheServerApplication
annotations,
SDG creates a server PARTITION
Region
by default.
Of course, you can always override the default when necessary. To override the default applied by Spring Data for Apache Geode,
four new Region mapping annotations have been introduced:
-
@ClientRegion
-
@LocalRegion
-
@PartitionRegion
-
@ReplicateRegion
The @ClientRegion
mapping annotation is specific to client applications. All of the other Region mapping annotations
listed above can only be used in server applications that have an embedded peer Cache
.
It is sometimes necessary for client applications to create and use local-only Regions, perhaps to aggregate data
from other Regions in order to analyze the data locally and carry out some function performed by the application
on the user’s behalf. In this case, the data does not need to be distributed back to the server unless other
applications need access to the results. This Region might even be temporary and discarded after use, which could be
accomplished with Idle-Timeout (TTI) and Time-To-Live (TTL) expiration policies on the Region itself.
(See “Configuring Expiration” for more on expiration policies.)
Region-level Idle-Timeout (TTI) and Time-To-Live (TTL) expiration policies are independent of and different from
entry-level TTI and TTL expiration policies.
In any case, if you want to create a local-only client Region where the data is not going to be distributed back to
a corresponding Region on the server with the same name, you can declare the @ClientRegion
mapping annotation
and set the shortcut
attribute to ClientRegionShortcut.LOCAL
, as follows:
Spring ClientCache
application with a local-only, client Region
@ClientRegion(shortcut = ClientRegionShortcut.LOCAL)
class ClientLocalEntityType { .. }
All Region type-specific annotations provide additional attributes that are both common across Region types
as well as specific to only that type of Region. For example, the collocatedWith
and redundantCopies
attributes
in the PartitionRegion
annotation apply to server-side, PARTITION
Regions only.
More details on Apache Geode Region types can be found
here.
6.12.2. Configured Cluster-defined Regions
In addition to the @EnableEntityDefinedRegions
annotation, Spring Data for Apache Geode also provides the inverse annotation,
@EnableClusterDefinedRegions
. Rather than basing your Regions on the entity classes defined and driven from
your application use cases (UC) and requirements (the most common and logical approach), alternatively, you can
declare your Regions from the Regions already defined in the cluster to which your ClientCache
application
will connect.
This allows you to centralize your configuration using the cluster of servers as the primary source of data definitions
and ensure that all client applications of the cluster have a consistent configuration. This is particularly useful
when quickly scaling up a large number instances of the same client application to handle the increased load
in a cloud-managed environment.
The idea is, rather than the client application(s) driving the data dictionary, the user defines Regions
using Apache Geode’s Gfsh CLI shell tool. This has the added advantage that when additional peers are added
to the cluster, they too will also have and share the same configuration since it is remembered by Apache Geode’s
Cluster Configuration Service.
By way of example, a user might defined a Region in Gfsh, as follows:
Defining a Region with Gfsh
gfsh>create region --name=Books --type=PARTITION
Member | Status
--------- | --------------------------------------
ServerOne | Region "/Books" created on "ServerOne"
ServerTwo | Region "/Books" created on "ServerTwo"
gfsh>list regions
List of regions
---------------
Books
gfsh>describe region --name=/Books
..........................................................
Name : Books
Data Policy : partition
Hosting Members : ServerTwo
ServerOne
Non-Default Attributes Shared By Hosting Members
Type | Name | Value
------ | ----------- | ---------
Region | size | 0
| data-policy | PARTITION
With Apache Geode’s Cluster Configuration Service, any additional peer members added to the cluster of servers
to handle the increased load (on the backend) will also have the same configuration, for example:
Adding an additional peer member to the cluster
gfsh>list members
Name | Id
--------- | ----------------------------------------------
Locator | 10.0.0.121(Locator:68173:locator)<ec><v0>:1024
ServerOne | 10.0.0.121(ServerOne:68242)<v3>:1025
ServerTwo | 10.0.0.121(ServerTwo:68372)<v4>:1026
gfsh>start server --name=ServerThree --log-level=config --server-port=41414
Starting a Geode Server in /Users/you/geode/cluster/ServerThree...
...
Server in /Users/you/geode/cluster/ServerThree... on 10.0.0.121[41414] as ServerThree is currently online.
Process ID: 68467
Uptime: 3 seconds
Geode Version: 1.2.1
Java Version: 1.8.0_152
Log File: /Users/you/geode/cluster/ServerThree/ServerThree.log
JVM Arguments: -Dgemfire.default.locators=10.0.0.121[10334]
-Dgemfire.use-cluster-configuration=true
-Dgemfire.start-dev-rest-api=false
-Dgemfire.log-level=config
-XX:OnOutOfMemoryError=kill -KILL %p
-Dgemfire.launcher.registerSignalHandlers=true
-Djava.awt.headless=true
-Dsun.rmi.dgc.server.gcInterval=9223372036854775806
Class-Path: /Users/you/geode/cluster/apache-geode-1.2.1/lib/geode-core-1.2.1.jar
:/Users/you/geode/cluster/apache-geode-1.2.1/lib/geode-dependencies.jar
gfsh>list members
Name | Id
----------- | ----------------------------------------------
Locator | 10.0.0.121(Locator:68173:locator)<ec><v0>:1024
ServerOne | 10.0.0.121(ServerOne:68242)<v3>:1025
ServerTwo | 10.0.0.121(ServerTwo:68372)<v4>:1026
ServerThree | 10.0.0.121(ServerThree:68467)<v5>:1027
gfsh>describe member --name=ServerThree
Name : ServerThree
Id : 10.0.0.121(ServerThree:68467)<v5>:1027
Host : 10.0.0.121
Regions : Books
PID : 68467
Groups :
Used Heap : 37M
Max Heap : 3641M
Working Dir : /Users/you/geode/cluster/ServerThree
Log file : /Users/you/geode/cluster/ServerThree/ServerThree.log
Locators : 10.0.0.121[10334]
Cache Server Information
Server Bind :
Server Port : 41414
Running : true
Client Connections : 0
As you can see, "ServerThree" now has the "Books" Region. If the any or all of the server go down, they will have
the same configuration along with the "Books" Region when they come back up.
On the client-side, many Book Store client application instances might be started to process books against
the Book Store online service. The "Books" Region might be 1 of many different Regions needed to implement
the Book Store application service. Rather than have to create and configure each Region individually, SDG
conveniently allows the client application Regions to be defined from the cluster, as follows:
Defining Client Regions from the Cluster with @EnableClusterDefinedRegions
@ClientCacheApplication
@EnableClusterDefinedRegions
class BookStoreClientApplication {
public static void main(String[] args) {
....
}
...
}
@EnableClusterDefinedRegions
can only used on the client.
You can use the clientRegionShortcut
annotation attribute to control the type of Region created on the client.
By default, a client PROXY
Region is created. Set clientRegionShortcut
to ClientRegionShortcut.CACHING_PROXY
to implement "near caching". This setting applies to all client Regions created from Cluster-defined Regions.
If you want to control individual settings (like data policy) of the client Regions created from Regions defined
on the Cluster, then you can implement a
RegionConfigurer
with custom logic based on the Region name.
Then, it becomes a simple matter to use the "Books" Region in your application. You can inject the "Books" Region
directly, as follows:
Using the "Books" Region
@org.springframework.stereotype.Repository
class BooksDataAccessObject {
@Resource(name = "Books")
private Region<ISBN, Book> books;
// implement CRUD and queries with the "Books" Region
}
Or, even define a Spring Data Repository definition based on the application domain type (entity), Book
,
mapped to the "Books" Region, as follows:
Using the "Books" Region with a SD Repository
interface BookRepository extends CrudRepository<Book, ISBN> {
...
}
You can then either inject your custom BooksDataAccessObject
or the BookRepository
into your application service
components to carry out whatever business function required.
6.12.3. Configuring Eviction
Managing data with Apache Geode is an active task. Tuning is generally required, and you must employ a combination
of features (for example, both eviction and expiration)
to effectively manage your data in memory with Apache Geode.
Given that Apache Geode is an In-Memory Data Grid (IMDG), data is managed in-memory and distributed to other nodes
that participate in a cluster in order to minimize latency, maximize throughput and ensure that data is highly available.
Since not all of an application’s data is going to typically fit in memory (even across an entire cluster of nodes,
much less on a single node), you can increase capacity by adding new nodes to the cluster. This is commonly referred to
as linear scale-out (rather than scaling up, which means adding more memory, more CPU, more disk,
or more network bandwidth — basically more of every system resource in order to handle the load).
Still, even with a cluster of nodes, it is usually imperative that only the most important data be kept in memory.
Running out of memory, or even venturing near full capacity, is rarely, if ever, a good thing. Stop-the-world GCs
or worse, OutOfMemoryErrors
, will bring your application to a screaming halt.
So, to help manage memory and keep the most important data around, Apache Geode supports Least Recently Used (LRU)
eviction. That is, Apache Geode evicts Region entries based on when those entries were last accessed by using
the Least Recently Used algorithm.
To enable eviction, annotate the application class with @EnableEviction
, as follows:
Spring application with eviction enabled
@SpringBootApplication
@PeerCacheApplication
@EnableEviction(policies = {
@EvictionPolicy(regionNames = "Books", action = EvictionActionType.INVALIDATE),
@EvictionPolicy(regionNames = { "Customers", "Orders" }, maximum = 90,
action = EvictionActionType.OVERFLOW_TO_DISK,
type = EvictonPolicyType.HEAP_PERCENTAGE)
})
class ServerApplication { .. }
Eviction policies are usually set on the Regions in the servers.
As shown earlier, the policies
attribute can specify one or more nested @EvictionPolicy
annotations, with each one
being individually catered to one or more Regions where the eviction policy needs to be applied.
Additionally, you can reference a custom implementation of Apache Geode’s
org.apache.geode.cache.util.ObjectSizer
interface,
which can be defined as a bean in the Spring container and referenced by name by using the objectSizerName
attribute.
An ObjectSizer
lets you define the criteria used to evaluate and determine the the size of objects stored in a Region.
See the @EnableEviction
annotation Javadoc
for a complete list of eviction configuration options.
More details on Apache Geode eviction can be found
here.
6.12.4. Configuring Expiration
Along with eviction, expiration can also be used to manage memory
by allowing entries stored in a Region to expire. Apache Geode supports both Time-to-Live (TTL)
and Idle-Timeout (TTI) entry expiration policies.
Spring Data for Apache Geode’s annotation-based expiration configuration is based on the
earlier and existing entry expiration annotation support
added in Spring Data for Apache Geode version 1.5.
Essentially, Spring Data for Apache Geode’s expiration annotation support is based on a custom implementation of Apache Geode’s
org.apache.geode.cache.CustomExpiry
interface.
This o.a.g.cache.CustomExpiry
implementation inspects the user’s application domain objects stored in a Region
for the presence of type-level expiration annotations.
Spring Data for Apache Geode provides the following expiration annotations:
-
Expiration
-
IdleTimeoutExpiration
-
TimeToLiveExpiration
An application domain object type can be annotated with one or more of the expiration annotations, as follows:
Applicaton domain object specific expiration policy
@Region("Books")
@TimeToLiveExpiration(timeout = 30000, action = "INVALIDATE")
class Book { .. }
To enable expiration, annotate the application class with @EnableExpiration
, as follows:
Spring application with expiration enabled
@SpringBootApplication
@PeerCacheApplication
@EnableExpiration
class ServerApplication { .. }
In addition to application domain object type-level expiration policies, you can directly and individually configure
expiration policies on a Region by Region basis using the @EnableExpiration
annotation, as follows:
Spring application with region-specific expiration policies
@SpringBootApplication
@PeerCacheApplication
@EnableExpiration(policies = {
@ExpirationPolicy(regionNames = "Books", types = ExpirationType.TIME_TO_LIVE),
@ExpirationPolicy(regionNames = { "Customers", "Orders" }, timeout = 30000,
action = ExpirationActionType.LOCAL_DESTROY)
})
class ServerApplication { .. }
The preceding example sets expiration policies for the Books
, Customers
, and Orders
Regions.
Expiration policies are usually set on the Regions in the servers.
See the @EnableExpiration
annotation Javadoc
for a complete list of expiration configuration options.
More details on Apache Geode expiration can be found
here.
6.12.5. Configuring Compression
In addition to eviction
and expiration, you can also configure your data Regions
with compression to reduce memory consumption.
Apache Geode lets you compress in memory Region values by using pluggable
Compressors
, or different compression codecs.
Apache Geode uses Google’s Snappy compression library by default.
To enable compression, annotate the application class with @EnableCompression
, as follows:
Spring application with Region compression enabled
@SpringBootApplication
@ClientCacheApplication
@EnableCompression(compressorBeanName = "MyCompressor", regionNames = { "Customers", "Orders" })
class ClientApplication { .. }
Neither the compressorBeanName
nor the regionNames
attributes are required.
The compressorBeanName
defaults to SnappyCompressor
, enabling Apache Geode’s
SnappyCompressor
.
The regionNames
attribute is an array of Region names that specify the Regions that have compression enabled.
By default, all Regions compress values if the regionNames
attribute is not explicitly set.
Alternatively, you can use the spring.data.gemfire.cache.compression.compressor-bean-name
and spring.data.gemfire.cache.compression.region-names
properties in the application.properties
file
to set and configure the values of these @EnableCompression
annotation attributes.
To use Apache Geode’s Region compression feature, you must include the org.iq80.snappy:snappy
dependency
in your application’s pom.xml
file (for Maven) or build.gradle
file (for Gradle). This is necessary only if you use
Apache Geode’s default support for Region compression, which uses the
SnappyCompressor
by default.
Of course, if you use another compression library, you need to include dependencies for that compression library
on your application’s classpath. Additionally, you need to implement Apache Geode’s
Compressor
interface to adapt your compression
library of choice, define it as a bean in the Spring compressor, and set the compressorBeanName
to this custom bean definition.
See the @EnableCompression
annotation Javadoc
for more details.
More details on Apache Geode compression can be found
here.
6.12.6. Configuring Off-Heap Memory
Another effective means for reducing pressure on the JVM’s Heap memory and minimizing GC activity is to use
Apache Geode’s off-heap memory support.
Rather than storing Region entries on the JVM Heap, entries are stored in the system’s main memory. Off-heap memory
generally works best when the objects being stored are uniform in size, are mostly less than 128K, and do not need
to be deserialized frequently, as explained in the Apache Geode
User Guide.
To enable off-heap, annotate the application class with @EnableOffHeap
, as follows:
Spring application with Off-Heap enabled
@SpringBootApplication
@PeerCacheApplication
@EnableOffHeap(memorySize = 8192m regionNames = { "Customers", "Orders" })
class ServerApplication { .. }
The memorySize
attribute is required. The value for the memorySize
attribute specifies the amount of main memory
a Region can use in either megabytes (m
) or gigabytes (g
).
The regionNames
attribute is an array of Region names that specifies the Regions that store entries in main memory.
By default, all Regions use main memory if the regionNames
attribute is not explicitly set.
Alternatively, you can use the spring.data.gemfire.cache.off-heap.memory-size
and spring.data.gemfire.cache.off-heap.region-names
properties in the application.properties
file to set
and configure the values of these @EnableOffHeap
annotation attributes.
See the @EnableOffHeap
annotation Javadoc
for more details.
6.12.7. Configuring Disk Stores
Alternatively, you can configure Regions to persist data to disk. You can also configure Regions to overflow
data to disk when Region entries are evicted. In both cases, a DiskStore
is required to persist and/or overflow
the data. When an explicit DiskStore
has not been configured for a Region with persistence or overflow,
Apache Geode uses the DEFAULT
DiskStore
.
We recommend defining Region-specific DiskStores
when persisting and/or overflowing data to disk.
Spring Data for Apache Geode provides annotation support for defining and creating application Region DiskStores
by annotating the application class with the @EnableDiskStore
and @EnableDiskStores
annotations.
@EnableDiskStores
is a composite annotation for aggregating one or more @EnableDiskStore
annotations.
For example, while Book
information might mostly consist of reference data from some external data source
(such as Amazon), Order
data is most likely going to be transactional in nature and something the application
is going to need to retain (and maybe even overflow to disk if the transaction volume is high enough) — or so any book publisher and author hopes, anyway.
Using the @EnableDiskStore
annotation, you can define and create a DiskStore
as follows:
Spring application defining a DiskStore
@SpringBootApplication
@PeerCacheApplication
@EnableDiskStore(name = "OrdersDiskStore", autoCompact = true, compactionThreshold = 70,
maxOplogSize = 512, diskDirectories = @DiskDiretory(location = "/absolute/path/to/order/disk/files"))
class ServerApplication { .. }
Again, more than one DiskStore
can be defined by using the composite, @EnableDiskStores
annotation.
As with other annotations in Spring Data for Apache Geode’s annotation-based configuration model, both @EnableDiskStore
and @EnableDiskStores
have many attributes along with associated configuration properties to customize
the DiskStores
created at runtime.
Additionally, the @EnableDiskStores
annotation defines certain common DiskStore
attributes that apply to all
DiskStores
created from @EnableDiskStore
annotations composed with the @EnableDiskStores
annotation itself.
Individual DiskStore
configuration override a particular global setting, but the @EnableDiskStores
annotation
conveniently defines common configuration attributes that apply across all DiskStores
aggregated by the annotation.
Spring Data for Apache Geode also provides the DiskStoreConfigurer
callback interface, which can be declared in Java configuration
and used instead of configuration properties to customize a DiskStore
at runtime, as the following example shows:
Spring application with custom DiskStore configuration
@SpringBootApplication
@PeerCacheApplication
@EnableDiskStore(name = "OrdersDiskStore", autoCompact = true, compactionThreshold = 70,
maxOplogSize = 512, diskDirectories = @DiskDiretory(location = "/absolute/path/to/order/disk/files"))
class ServerApplication {
@Bean
DiskStoreConfigurer ordersDiskStoreDiretoryConfigurer(
@Value("${orders.disk.store.location}") String location) {
return (beanName, diskStoreFactoryBean) -> {
if ("OrdersDiskStore".equals(beanName) {
diskStoreFactoryBean.setDiskDirs(Collections.singletonList(new DiskDir(location));
}
}
}
}
See the @EnableDiskStore
and @EnableDiskStores
annotation
Javadoc for more details on the available attributes as well as associated configuration properties.
More details on Apache Geode Region persistence and overflow (using DiskStores) can be found
here.
6.12.8. Configuring Indexes
There is not much use in storing data in Regions unless the data can be accessed.
In addition to Region.get(key)
operations, particularly when the key is known in advance, data is commonly retrieved
by executing queries on the Regions that contain the data. With Apache Geode, queries are written by using
the Object Query Language (OQL), and the specific data set that a client wishes to access is expressed
in the query’s predicate (for example, SELECT * FROM /Books b WHERE b.author.name = 'Jon Doe'
).
Generally, querying without indexes is inefficient. When executing queries without an index, Apache Geode
performs the equivalent of a full table scan.
Indexes are created and maintained for fields on objects used in query predicates to match the data of interest, as
expressed by the query’s projection. Different types of indexes, such as
key
and hash indexes, can be created.
Spring Data for Apache Geode makes it easy to create indexes on Regions where the data is stored and accessed. Rather than explicitly
declaring Index
bean definitions by using Spring config as before, we can create an Index
bean definition in Java,
as follows:
Index bean definition using Java config
@Bean("BooksIsbnIndex")
IndexFactoryBean bookIsbnIndex(GemFireCache gemfireCache) {
IndexFactoryBean bookIsbnIndex = new IndexFactoryBean();
bookIsbnIndex.setCache(gemfireCache);
bookIsbnIndex.setName("BookIsbnIndex");
bookIsbnIndex.setExpression("isbn");
bookIsbnIndex.setFrom("/Books"));
bookIsbnIndex.setType(IndexType.KEY);
return bookIsbnIndex;
}
Alternatively, we can use XML to create an Index
bean definition, as follows:
Index bean definition using XML
<gfe:index id="BooksIsbnIndex" expression="isbn" from="/Books" type="KEY"/>
However, now you can directly define indexes on the fields of your application domain object types for which you know
will be used in query predicates to speed up those queries. You can even apply indexes for OQL queries generated
from user-defined query methods on an application’s repository interfaces.
Re-using the example Book
entity class from earlier, we can annotate the fields on Book
that we know are used
in queries that we define with query methods in the BookRepository
interface, as follows:
Application domain object type modeling a book using indexes
@Region("Books")
class Book {
@Id
private ISBN isbn;
@Indexed
private Author author;
private Category category;
private LocalDate releaseDate;
private Publisher publisher;
@LuceneIndexed
private String title;
}
In our new Book
class definition, we annotated the author
field with @Indexed
and the title
field
with @LuceneIndexed
. Also, the isbn
field had previously been annotated with Spring Data’s @Id
annotation,
which identifies the field containing the unique identifier for Book
instances, and, in Spring Data for Apache Geode, the @Id
annotated field or property is used as the key in the Region when storing the entry.
-
@Id
annotated fields or properties result in the creation of an Apache Geode KEY
Index.
-
@Indexed
annotated fields or properties result in the creation of an Apache Geode HASH
Index (the default).
-
@LuceneIndexed
annotated fields or properties result in the creation of an Apache Geode Lucene Index, used in
text-based searches with Apache Geode’s Lucene integration and support.
When the @Indexed
annotation is used without setting any attributes, the index name
, expression
, and fromClause
are derived from the field or property of the class on which the @Indexed
annotation has been added. The expression
is exactly the name of the field or property. The fromClause
is derived from the @Region
annotation on
the domain object’s class, or the simple name of the domain object class if the @Region
annotation was not specified.
Of course, you can explicitly set any of the @Indexed
annotation attributes to override the default values
provided by Spring Data for Apache Geode.
Application domain object type modeling a Book with customized indexes
@Region("Books")
class Book {
@Id
private ISBN isbn;
@Indexed(name = "BookAuthorNameIndex", expression = "author.name", type = "FUNCTIONAL")
private Author author;
private Category category;
private LocalDate releaseDate;
private Publisher publisher;
@LuceneIndexed(name = "BookTitleIndex", destory = true)
private String title;
}
The name
of the index, which is auto-generated when not explicitly set, is also used as the name of the bean
registered in the Spring container for the index. If necessary, this index bean can even be injected by name
into another application component.
The generated name of the index follows this pattern: <Region Name><Field/Property Name><Index Type>Idx
.
For example, the name of the author
index would be, BooksAuthorHashIdx
.
To enable indexing, annotate the application class with @EnableIndexing
, as follows:
Spring application with Indexing enabled
@SpringBootApplication
@PeerCacheApplication
@EnableEntityDefinedRegions
@EnableIndexing
class ServerApplication { .. }
The @EnablingIndexing
annotation has no effect unless the @EnableEntityDefinedRegions
is also declared.
Essentially, indexes are defined from fields or properties on the entity class types, and entity classes must be scanned
to inspect the entity’s fields and properties for the presence of index annotations. Without this scan, index annotations
cannot be found. We also strongly recommend that you limit the scope of the scan.
While Lucene queries are not (yet) supported on Spring Data for Apache Geode repositories, SDG does provide comprehensive
support for Apache Geode
Lucene queries by using the familiar Spring template design pattern.
Finally, we close this section with a few extra tips to keep in mind when using indexes:
-
While OQL indexes are not required to execute OQL Queries, Lucene Indexes are required to execute Lucene
text-based searches.
-
OQL indexes are not persisted to disk. They are only maintained in memory. So, when an Apache Geode
node is restarted, the index must be rebuilt.
-
You also need to be aware of the overhead associated in maintaining indexes, particularly since an index is stored
exclusively in memory and especially when Region entries are updated. Index "maintenance" can be
configured
as an asynchronous task.
Another optimization that you can use when restarting your Spring application where indexes have to be rebuilt
is to first define all the indexes up front and then create them all at once, which, in Spring Data for Apache Geode, happens
when the Spring container is refreshed.
You can define indexes up front and then create them all at once by setting the define
attribute on
the @EnableIndexing
annotation to true
.
See “Creating Multiple Indexes at Once”
in Apache Geode’s User Guide for more details.
Creating sensible indexes is an important task, since it is possible for a poorly designed index
to do more harm than good.
See both the @Indexed
annotation
and @LuceneIndexed
annotation
Javadoc for complete list of configuration options.
More details on Apache Geode OQL queries can be found
here.
More details on Apache Geode indexes can be found
here.
More details on Apache Geode Lucene queries can be found
here.
6.13. Configuring Continuous Queries
Another very important and useful feature of Apache Geode is
Continuous Queries.
In a world of Internet-enabled things, events and streams of data come from everywhere. Being able to handle
and process a large stream of data and react to events in real time is an increasingly important requirement
for many applications. One example is self-driving vehicles. Being able to receive, filter, transform, analyze,
and act on data in real time is a key differentiator and characteristic of real time applications.
Fortunately, Apache Geode was ahead of its time in this regard. By using Continuous Queries (CQ),
a client application can express the data or events it is interested in and register listeners to handle and process
the events as they occur. The data that a client application may be interested in is expressed as an OQL query,
where the query predicate is used to filter or identify the data of interest. When data is changed or added
and it matches the criteria defined in the query predicate of the registered CQ, the client application is notified.
Spring Data for Apache Geode makes it easy to define and register CQs, along with an associated listener to handle and process CQ
events without all the cruft of Apache Geode’s plumbing. SDG’s new annotation-based configuration
for CQs builds on the existing Continuous Query support in the continuous query listener container.
For instance, say a banking application registers interest in every customers' checking acccount to detect overdraft
withdrawls and handle this event by either applying overdraft protection or notifyinfg the customer. Then, the
application might register the following CQ:
Spring ClientCache
application with registered CQ and listener.
@SpringBootApplication
@ClientCacheApplication(subcriptionEnabled = true)
@EnableContinuousQueries
class PublisherPrintApplication {
@ContinuousQuery(name = "OverdraftProtection", query = "SELECT * FROM /CheckingAccount ca WHERE ca.balance < 0.0")
void handleOverdraft(CqEvent event) {
// Quick!!! Put more money into the checking account or notify the customer of the checking account!
}
}
To enable Continuous Queries, annotate your application class with @EnableContinuousQueries
.
Defining Continuous Queries consists of annotating any Spring @Component
-annotated POJO class methods
with the @ContinuousQuery
annotation (in similar fashion to SDG’s Function-annotated POJO methods).
A POJO method defined with a CQ by using the @ContinuousQuery
annotation is called any time data matching
the query predicate is added or changed.
Additionally, the POJO method signature should adhere to the requirements outlined in the section on
the ContinuousQueryListener
and the ContinuousQueryListenerAdapter
.
See the @EnableContinuousQueries
and @ContinuousQuery
annotation
Javadoc for more details on available attributes and configuration settings.
More details on Spring Data for Apache Geode’s continuous query support can be found
here.
More details on Apache Geode’s Continuous Queries can be found
here.
6.14. Configuring Spring’s Cache Abstraction
With Spring Data for Apache Geode, Apache Geode can be used as a caching provider in Spring’s
cache abstraction.
In Spring’s Cache Abstraction, the caching annotations (such as @Cacheable
) identify the cache on which a cache lookup
is performed before invoking a potentially expensive operation. The results of an application service method are cached
after the operation is invoked.
In Spring Data for Apache Geode, a Spring Cache
corresponds directly to a Apache Geode Region. The Region must exist before
any caching annotated application service methods are called. This is true for any of Spring’s caching annotations
(that is, @Cacheable
, @CachePut
and @CacheEvict
) that identify the cache to use in the service operation.
For instance, our publisher’s Point-of-Sale (PoS) application might have a feature to determine or lookup the Price
of a Book
during a sales transaction, as the following example shows:
@Service
class PointOfSaleService
@Cacheable("BookPrices")
Price runPriceCheckFor(Book book) {
...
}
@Transactional
Receipt checkout(Order order) {
...
}
...
}
To make your work easier when you use Spring Data for Apache Geode with Spring’s Cache Abstraction, two new features have been added
to the annotation-based configuration model.
Consider the following Spring caching configuration:
Enabling Caching using Apache Geode as the caching provider
@EnableCaching
class CachingConfiguration {
@Bean
GemfireCacheManager cacheManager(GemFireCache gemfireCache) {
GemfireCacheManager cacheManager = new GemfireCacheManager();
cacheManager.setCache(gemfireCache);
return cacheManager;
}
@Bean("BookPricesCache")
ReplicatedRegionFactoryBean<Book, Price> bookPricesRegion(GemFireCache gemfireCache) {
ReplicatedRegionFactoryBean<Book, Price> bookPricesRegion =
new ReplicatedRegionFactoryBean<>();
bookPricesRegion.setCache(gemfireCache);
bookPricesRegion.setClose(false);
bookPricesRegion.setPersistent(false);
return bookPricesRegion;
}
@Bean("PointOfSaleService")
PointOfSaleService pointOfSaleService(..) {
return new PointOfSaleService(..);
}
}
Using Spring Data for Apache Geode’s new features, you can simplify the same caching configuration to the following:
Enabling Apache Geode Caching
@EnableGemfireCaching
@EnableCachingDefinedRegions
class CachingConfiguration {
@Bean("PointOfSaleService")
PointOfSaleService pointOfSaleService(..) {
return new PointOfSaleService(..);
}
}
First, the @EnableGemfireCaching
annotation replaces both the Spring @EnableCaching
annotation and the need
to declare an explicit CacheManager
bean definition (named "cacheManager") in the Spring config.
Second, the @EnableCachingDefinedRegions
annotation, like the @EnableEntityDefinedRegions
annotation described in
“Configuring Regions”, inspects the entire Spring application, caching
annotated service components to identify all the caches that are needed by the application at runtime and creates
Regions in Apache Geode for these caches on application startup.
The Regions created are local to the application process that created the Regions. If the application is a peer Cache
,
the Regions exist only on the application node. If the application is a ClientCache
, then SDG creates
client PROXY
Regions and expects those Regions with the same name to already exist on the servers in the cluster.
SDG cannot determine the cache required by a service method using a Spring CacheResolver
to resolve the cache used in the operation at runtime.
SDG also supports JCache (JSR-107) cache annotations on application service components.
See the core Spring Framework Reference Guide
for the equivalent Spring caching annotation to use in place of JCache caching annotations.
Refer to the “Support for the Spring Cache Abstraction” section for more details
on using Apache Geode as a caching provider in Spring’s Cache Abstraction.
More details on Spring’s Cache Abstraction can be found
here.
6.15. Configuring Cluster Configuration Push
This may be the most exciting new feature in Spring Data for Apache Geode.
When a client application class is annotated with @EnableClusterConfiguration
, any Regions or Indexes defined
and declared as beans in the Spring Container by the client application are “pushed” to the cluster of servers
to which the client is connected. Not only that, but this “push” is performed in such a way that Apache Geode
remembers the configuration pushed by the client when using HTTP. If all the nodes in the cluster go down, they
come back up with the same configuration as before. If a new server is added to the cluster, it will acquire
identical configuration.
In a sense, this feature is not much different than if you were to use Gfsh to manually create the Regions and Indexes
on all the servers in the cluster. Except that now, with Spring Data for Apache Geode, you no longer need to use Gfsh to create Regions
and Indexes. Your Spring Boot application, enabled with the power of Spring Data for Apache Geode, already contains all the configuration
metadata needed to create Regions and Indexes for you.
When you use the Spring Data Repository abstraction, we know all the Regions (such as those defined by the @Region
annotated entity classes) and Indexes (such as those defined by the @Indexed
-annotated entity fields and properties)
that your application will need.
When you use Spring’s Cache Abstraction, we also know all the Regions for all the caches identified in the caching
annotations needed by your application’s service components.
Essentially, you are already telling us everything we need to know simply by developing your application with the
Spring Framework simply by using all of its API and features, whether expressed in annotation metadata, Java, XML
or otherwise, and whether for configuration, mapping, or whatever the purpose.
The point is, you can focus on your application’s business logic while using the framework’s features and supporting
infrastructure (such as Spring’s Cache Abstraction, Spring Data Repositories, Spring’s Transaction Management,
and so on) and Spring Data for Apache Geode takes care of all the Apache Geode plumbing required by those framework features
on your behalf.
Pushing configuration from the client to the servers in the cluster and having the cluster remember it is made possible
in part by the use of Apache Geode’s Cluster Configuration
service. Apache Geode’s Cluster Configuration service is also the same service used by Gfsh to record
schema-related changes (for example, gfsh> create region --name=Example --type=PARTITION
) issued by the user
to the cluster from the shell.
Of course, since the cluster may “remember” the prior configuration pushed by a client from a previous run,
Spring Data for Apache Geode is careful not to stomp on any existing Regions and Indexes already defined in the servers.
This is especially important, for instance, when Regions already contain data!
Currently, there is no option to overwrite any existing Region or Index definitions. To re-create a Region
or Index, you must use Gfsh to first destroy the Region or Index and then restart the client application
so that configuration is pushed up to the server again. Alternatively, you can use Gfsh to (re-)define the Regions
and Indexes manually.
Unlike Gfsh, Spring Data for Apache Geode supports the creation of Regions and Indexes only on the servers from a client.
For advanced configuration and use cases, you should use Gfsh to manage the (server-side) cluster.
To use this feature you must explicitly declare the org.springframework:spring-web
dependency on the
classpath of your Spring, Apache Geode ClientCache
application.
Consider the power expressed in the following configuration:
Spring ClientCache
application
@SpringBootApplication
@ClientCacheApplication
@EnableCachingDefinedRegions
@EnableEntityDefinedRegions
@EnableIndexing
@EnableGemfireCaching
@EnableGemfireRepositories
@EnableClusterConfiguration
class ClientApplication { .. }
You instantly get a Spring Boot application with a Apache Geode ClientCache
instance, Spring Data Repositories,
Spring’s Cache Abstraction with Apache Geode as the caching provider (where Regions and Indexes
are not only created on the client but pushed to the servers in the cluster).
From there, you only need to do the following:
-
Define the application’s domain model objects annotated with mapping and index annotations.
-
Define Repository interfaces to support basic data access operations and simple queries for each of your entity types.
-
Define the service components containing the business logic transacting the entities.
-
Declare the appropriate annotations on service methods that require caching, transactional behavior, and so on.
Nothing in this case pertains to the infrastructure and plumbing required in the application’s back-end services
(such as Apache Geode). Database users have similar features. Now Spring and Apache Geode developers do too.
When combined with the following Spring Data for Apache Geode annotations, this application really starts to take flight,
with very little effort:
-
@EnableContinuousQueries
-
@EnableGemfireFunctionExecutions
-
@EnableGemfireCacheTransactions
See the @EnableClusterConfiguration
annotation
Javadoc for more details.
6.16. Configuring SSL
Equally important to serializing data to be transferred over the wire is securing the data while in transit.
Of course, the common way to accomplish this in Java is by using the Secure Sockets Extension (SSE)
and Transport Layer Security (TLS).
To enable SSL, annotate your application class with @EnableSsl
, as follows:
Spring ClientCache
application with SSL enabled
@SpringBootApplication
@ClientCacheApplication
@EnableSsl
public class ClientApplication { .. }
Then you need to set the necessary SSL configuration attributes or properties: keystores, usernames/passwords, and so on.
You can individually configure different Apache Geode components (GATEWAY
, HTTP
, JMX
, LOCATOR
, and SERVER
)
with SSL, or you can collectively configure them to use SSL by using the CLUSTER
enumerated value.
You can specify which Apache Geode components the SSL configuration settings should applied by using
the nested @EnableSsl
annotation, components
attribute with enumerated values from the Component
enum,
as follows:
Spring ClientCache
application with SSL enabled by component
@SpringBootApplication
@ClientCacheApplication
@EnableSsl(components = { GATEWAY, LOCATOR, SERVER })
public class ClientApplication { .. }
In addition, you can also specify component-level SSL configuration (ciphers
, protocols
and keystore
/truststore
information) by using the corresponding annotation attribute or associated configuration properties.
See the @EnableSsl
annotation Javadoc
for more details.
More details on Apache Geode SSL support can be found
here.
6.17. Configuring Security
Without a doubt, application security is extremely important, and Spring Data for Apache Geode provides comprehensive support
for securing both Apache Geode clients and servers.
Recently, Apache Geode introduced a new Integrated Security framework
(replacing its old authentication and authorization security model) for handling authentication and authorization.
One of the main features and benefits of this new security framework is that it integrates with
Apache Shiro and can therefore delegate both authentication and authorization requests
to Apache Shiro to enforce security.
The remainder of this section demonstrates how Spring Data for Apache Geode can simplify Apache Geode’s security story even further.
6.17.1. Configuring Server Security
There are several different ways in which you can configure security for servers in a Apache Geode cluster.
-
Implement the Apache Geode org.apache.geode.security.SecurityManager
interface and set Apache Geode’s
security-manager
property to refer to your application SecurityManager
implementation using the fully qualified
class name. Alternatively, users can construct and initialize an instance of their SecurityManager
implementation
and set it with the CacheFactory.setSecurityManager(:SecurityManager)
method when creating a Apache Geode peer Cache
.
-
Create an Apache Shiro shiro.ini
file with the users, roles,
and permissions defined for your application and then set the Apache Geode security-shiro-init
property
to refer to this shiro.ini
file, which must be available in the CLASSPATH
.
-
Using only Apache Shiro, annotate your Spring Boot application class with Spring Data for Apache Geode’s new
@EnableSecurity
annotation and define one or more Apache Shiro Realms
as beans in the Spring container for accessing your application’s security metadata (that is, authorized users, roles,
and permissions).
The problem with the first approach is that you must implement your own SecurityManager
, which can be quite tedious
and error-prone. Implementing a custom SecurityManager
offers some flexibility in accessing security metadata from
whatever data source stores the metadata, such as LDAP or even a proprietary, internal
data source. However, that problem has already been solved by configuring and using Apache Shiro Realms
,
which is more universally known and non-Apache Geode-specific.
See Apache Geode’s security examples for Authentication
and Authorization as one possible way
to implement your own custom, application-specific SecurityManager
. However, we strongly recommend against doing so.
The second approach, using an Apache Shiro INI file, is marginally better, but you still need to be familiar with
the INI file format in the first place. Additionally, an INI file is static and not easily updatable at runtime.
The third approach is the most ideal, since it adheres to widely known and industry-accepted concepts
(that is, Apache Shiro’s Security framework) and is easy to setup, as the following example shows:
Spring server application using Apache Shiro
@SpringBootApplication
@CacheServerApplication
@EnableSecurity
class ServerApplication {
@Bean
PropertiesRealm shiroRealm() {
PropertiesRealm propertiesRealm = new PropertiesRealm();
propertiesRealm.setResourcePath("classpath:shiro.properties");
propertiesRealm.setPermissionResolver(new GemFirePermissionResolver());
return propertiesRealm;
}
}
The configured Realm
shown in the preceding example could easily have been any of Apache Shiro’s supported Realms
:
-
-
-
-
-
A Realm
supporting the INI format.
You could even create a custom implementation of an Apache Shiro Realm
.
See Apache Shiro’s documentation on Realms for more details.
When Apache Shiro is on the CLASSPATH
of the servers in the cluster and one or more Apache Shiro Realms
have been defined as beans in the Spring container, Spring Data for Apache Geode detects this configuration and uses Apache Shiro
as the security provider to secure your Apache Geode servers when the @EnableSecurity
annotation is used.
You can find more information about Spring Data for Apache Geode’s support for Apache Geode’s new integrated security
framework using Apache Shiro in this
spring.io blog post.
See the @EnableSecurity
annotation
Javadoc for more details on available attributes and associated configuration properties.
More details on Apache Geode security can be found
here.
6.17.2. Configuring Client Security
The security story would not be complete without discussing how to secure Spring-based, Apache Geode cache client
applications as well.
Apache Geode’s process for securing a client application is, honestly, rather involved. In a nutshell, you need to:
-
Provide an implementation of the
org.apache.geode.security.AuthInitialize
interface.
-
Set the Apache Geode security-client-auth-init
(System) property to refer to the custom, application-provided
AuthInitialize
interface.
-
Specify the user credentials in a proprietary, Apache Geode gfsecurity.properties
file.
Spring Data for Apache Geode simplifies all of those steps by using the same @EnableSecurity
annotation that was used in the server
applications. In other words, the same @EnableSecurity
annotation handles security for both client and server
applications. This feature makes it easier for users when they decide to switch their applications from an embedded,
peer Cache
application to a ClientCache
application, for instance. Simply change the SDG annotation
from @PeerCacheApplication
or @CacheServerApplication
to @ClientCacheApplication
, and you are done.
Effectively, all you need to do on the client is the following:
Spring client application using @EnableSecurity
@SpringBootApplication
@ClientCacheApplication
@EnableSecurity
class ClientApplication { .. }
Then you can define the familiar Spring Boot application.properties
file containing the required username and password,
as the following example shows, and you are all set:
Spring Boot application.properties
file with the required Security credentials
spring.data.gemfire.security.username=jackBlack
spring.data.gemfire.security.password=b@cK!nB1@cK
By default, Spring Boot can find your application.properties
file when it is placed in the root of
the application’s CLASSPATH
. Of course, Spring supports many ways to locate resources by using its
Resource abstraction.
See the @EnableSecurity
annotation
Javadoc for more details on available attributes and associated configuration properties.
More details on Apache Geode Security can be found
here.
6.18. Configuration Tips
The following tips can help you get the most out of using the new annotation-based configuration model:
6.18.1. Configuration Organization
As we saw in the section on “Configuring Cluster Configuration Push”,
when many Apache Geode or Spring Data for Apache Geode features are enabled by using annotations, we begin to stack a lot of
annotations on the Spring @Configuration
or @SpringBootApplication
class. In this situation, it makes sense
to start compartmentalizing the configuration a bit.
For instance, consider the following declaration:
Spring ClientCache
application with the kitchen sink
@SpringBootApplication
@ClientCacheApplication
@EnableContinuousQueries
@EnableCachingDefinedRegions
@EnableEntityDefinedRegions
@EnableIndexing
@EnableGemfireCacheTransactions
@EnableGemfireCaching
@EnableGemfireFunctionExecutions
@EnableGemfireRepositories
@EnableClusterConfiguration
class ClientApplication { .. }
We could break this configuration down by concern, as follows:
Spring ClientCache
application with the kitcken sink to boot
@SpringBootApplication
@Import({ GemFireConfiguration.class, CachingConfiguration.class,
FunctionsConfiguration.class, QueriesConfiguration.class,
RepositoriesConfiguration.class })
class ClientApplication { .. }
@ClientCacheApplication
@EnableClusterConfiguration
@EnableGemfireCacheTransactions
class GemFireConfiguration { .. }
@EnableGemfireCaching
@EnableCachingDefinedRegions
class CachingConfiguration { .. }
@EnableGemfireFunctionExecutions
class FunctionsConfiguration { .. }
@EnableContinuousQueries
class QueriesConfiguration {
@ContinuousQuery(..)
void processCqEvent(CqEvent event) {
...
}
}
@EnableEntityDefinedRegions
@EnableGemfireRepositories
@EnableIndexing
class RepositoriesConfiguration { .. }
While it does not matter to the Spring Framework, we generally recommend aiming for readability, for the sake of
the next person who has to maintain the code (which might be you at some point in the future).
6.18.2. Additional Configuration-based Annotations
The following SDG Annotations were not discussed in this reference documentation, either because
the annotation supports a deprecated feature of Apache Geode or because there are better, alternative ways
to accomplishing the function that the annotation provides:
-
@EnableAuth
: Enables Apache Geode’s old authentication and authorization security model. (Deprecated.
Apache Geode’s new integrated security framework can be enabled on both clients and servers by using SDG’s
@EnableSecurity
annotation, as described in “Configuring Security”.)
-
@EnableAutoRegionLookup
: Not recommended. Essentially, this annotation supports finding Regions defined in external
configuration metadata (such as cache.xml
or Cluster Configuration when applied to a server) and automatically
registers those Regions as beans in the Spring container. This annotation corresponds with the <gfe:auto-region-lookup>
element in SDG’s XML namespace. More details can found here. Users should generally
prefer Spring configuration when using Spring and Spring Data for Apache Geode. See “Configuring Regions”
and “Configuring Cluster Configuration Push” instead.
-
@EnableBeanFactoryLocator
: Enables the SDG GemfireBeanFactoryLocator
feature, which is only useful
when using external configuration metadata (for example, cache.xml
). For example, if you define a CacheLoader
on
a Region defined in cache.xml
, you can still autowire this CacheLoader
with, say, a relational database
DataSource
bean defined in Spring configuration. This annotation takes advantage of this SDG
feature and might be useful if you have a large amount of legacy configuration metadata,
such as cache.xml
files.
-
@EnableGemFireAsLastResource
: Discussed in Global - JTA Transaction Management
with Apache Geode.
-
@EnableMcast
: Enables Apache Geode’s old peer discovery mechanism that uses UDP-based multi-cast networking.
(Deprecated. Use Apache Geode Locators instead. See “Configuring an Embedded Locator”.
-
@EnableRegionDataAccessTracing
: Useful for debugging purposes. This annotation enables tracing for all data access
operations performed on a Region by registering an AOP Aspect that proxies all Regions declared as beans
in the Spring container, intercepting the Region operation and logging the event.
6.19. Conclusion
As we learned in the previous sections, Spring Data for Apache Geode’s new annotation-based configuration model provides a tremendous
amount of power. Hopefully, it lives up to its goal of making it easier for you to get started quickly and easily
when using Apache Geode with Spring.
Keep in mind that, when you use the new annotations, you can still use Java configuration or XML configuration.
You can even combine all three approaches by using Spring’s @Import
and @ImportResource
annotations on a Spring @Configuration
or @SpringBootApplication
class. The moment you explicitly provide
a bean definition that would otherwise be provided by Spring Data for Apache Geode using 1 of the annotations, the annotation-based
configuration backs away.
In certain cases, you may even need to fall back to Java configuration, as in the Configurers
case, to handle
more complex or conditional configuration logic that is not easily expressed in or cannot be accomplished by
using annotations alone. Do not be alarmed. This behavior is to be expected.
For example, another case where you need Java or XML configuration is when configuring Apache Geode WAN components,
which currently do not have any annotation configuration support. However, defining and registering WAN components
requires only using the org.springframework.data.gemfire.wan.GatewayReceiverFactoryBean
and org.springframework.data.gemfire.wan.GatewaySenderFactoryBean
API classes in the Java configuration of your Spring
@Configuration
or @SpringBootApplication
classes (recommended).
The annotations were not meant to handle every situation. The annotations were meant to help you get up and running
as quickly and as easily as possible, especially during development.
We hope you will enjoy these new capabilities!
6.20. Annotation-based Configuration Quick Start
The following sections provide an overview to the SDG annotations in order to get started quickly.
All annotations provide additional configuration attributes along with associated properties
to conveniently customize the configuration and behavior of Apache Geode at runtime. However, in general,
none of the attributes or associated properties are required to use a particular Apache Geode feature.
Simply declare the annotation to enable the feature and you are done. Refer to the individual Javadoc of
each annotation for more details.
6.20.1. Configure a ClientCache
Application
To configure and bootstrap a Apache Geode ClientCache
application, use the following:
@SpringBootApplication
@ClientCacheApplication
public class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
}
See Configuring Apache Geode Applications with Spring for more details.
6.20.2. Configure a Peer Cache
Application
To configure and bootstrap a Apache Geode Peer Cache
application, use the following:
@SpringBootApplication
@PeerCacheApplication
public class ServerApplication {
public static void main(String[] args) {
SpringApplication.run(ServerApplication.class, args);
}
}
If you would like to enable a CacheServer
that allows ClientCache
applications to connect to this server,
then simply replace the @PeerCacheApplication
annotation with the @CacheServerApplication
annotation. This will
start a CacheServer
running on “localhost”, listening on the default CacheServer
port of 40404
.
See Configuring Apache Geode Applications with Spring for more details.
6.20.3. Configure an Embedded Locator
Annotate your Spring @PeerCacheApplication
or @CacheServerApplication
class with @EnableLocator
to start
an embedded Locator bound to all NICs listening on the default Locator port, 10334
, as follows:
@SpringBootApplication
@CacheServerApplication
@EnableLocator
public class ServerApplication {
public static void main(String[] args) {
SpringApplication.run(ServerApplication.class, args);
}
}
@EnableLocator
can only be used with Apache Geode server applications.
See Configuring an Embedded Locator for more details.
6.20.4. Configure an Embedded Manager
Annotate your Spring @PeerCacheApplication
or @CacheServerApplication
class with @EnableManager
to start
an embedded Manager bound to all NICs listening on the default Manager port, 1099
, as follows:
@SpringBootApplication
@CacheServerApplication
@EnableManager
public class ServerApplication {
public static void main(String[] args) {
SpringApplication.run(ServerApplication.class, args);
}
}
@EnableManager
can only be used with Apache Geode server applications.
See Configuring an Embedded Manager for more details.
6.20.5. Configure the Embedded HTTP Server
Annotate your Spring @PeerCacheApplication
or @CacheServerApplication
class with @EnableHttpService
to start
the embedded HTTP server (Jetty) listening on port 7070
, as follows:
@SpringBootApplication
@CacheServerApplication
@EnableHttpService
public class ServerApplication {
public static void main(String[] args) {
SpringApplication.run(ServerApplication.class, args);
}
}
@EnableHttpService
can only be used with Apache Geode server applications.
See Configuring the Embedded HTTP Server for more details.
6.20.6. Configure the Embedded Memcached Server
Annotate your Spring @PeerCacheApplication
or @CacheServerApplication
class with @EnableMemcachedServer
to start
the embedded Memcached server (Gemcached) listening on port 11211
, as follows:
@SpringBootApplication
@CacheServerApplication
@EnableMemcachedServer
public class ServerApplication {
public static void main(String[] args) {
SpringApplication.run(ServerApplication.class, args);
}
}
@EnableMemcachedServer
can only be used with Apache Geode server applications.
See Configuring the Embedded Memcached Server (Gemcached) for more details.
6.20.7. Configure the Embedded Redis Server
Annotate your Spring @PeerCacheApplication
or @CacheServerApplication
class with @EnableRedisServer
to start
the embedded Redis server listening on port 6379
, as follows:
@SpringBootApplication
@CacheServerApplication
@EnableRedisServer
public class ServerApplication {
public static void main(String[] args) {
SpringApplication.run(ServerApplication.class, args);
}
}
@EnableRedisServer
can only be used with Apache Geode server applications.
You must explicitly declare the org.apache.geode:geode-redis
module on your Spring [Boot] application
classpath.
See Configuring the Embedded Redis Server for more details.
6.20.8. Configure Logging
To configure or adjust Apache Geode logging, annotate your Spring, Apache Geode client or server
application class with @EnableLogging
, as follows:
@SpringBootApplication
@ClientCacheApplication
@EnableLogging(logLevel="trace")
public class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
}
Default log-level
is “config”. Also, this annotation will not adjust log levels in your application,
only for Apache Geode.
See Configuring Logging for more details.
6.20.9. Configure Statistics
To gather Apache Geode statistics at runtime, annotate your Spring, Apache Geode client or server
application class with @EnableStatistics
, as follows:
@SpringBootApplication
@ClientCacheApplication
@EnableStatistics
public class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
}
See Configuring Statistics for more details.
6.20.10. Configure PDX
To enable Apache Geode PDX serialization, annotate your Spring, Apache Geode client or server
application class with @EnablePdx
, as follows:
@SpringBootApplication
@ClientCacheApplication
@EnablePdx
public class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
}
Apache Geode PDX Serialization is an alternative to Java Serialization with many added benefits. For one,
it makes short work of making all of your application domain model types serializable without having to implement
java.io.Serializable
.
By default, SDG configures the MappingPdxSerializer
to serialize your application domain model types,
which does not require any special configuration out-of-the-box in order to properly identify application domain objects
that need to be serialized and then perform the serialization since, the logic in MappingPdxSerializer
is based on
Spring Data’s mapping infrastructure. See MappingPdxSerializer for more details.
See @EnablePdx
Javadoc.
See Configuring PDX for more details.
6.20.11. Configure SSL
To enable Apache Geode SSL, annotate your Spring, Apache Geode client or server application class
with @EnableSsl
, as follows:
@SpringBootApplication
@ClientCacheApplication
@EnableSsl(components = SERVER)
public class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
}
Minimally, Apache Geode requires you to specify a keystore & truststore using the appropriate configuration
attributes or properties. Both keystore & truststore configuration attributes or properties may refer to the same
KeyStore
file. Additionally, you will need to specify a username and password to access the KeyStore
file
if the file has been secured.
Apache Geode SSL allows you to configure the specific components of the system that require TLS, such as
client/server, Locators, Gateways, etc. Optionally, you can specify that all components of Apache Geode
use SSL with “ALL”.
See @EnableSsl
Javadoc.
See Configuring SSL for more details.
6.20.12. Configure Security
To enable Apache Geode security, annotate your Spring, Apache Geode client or server application class
with @EnableSecurity
, as follows:
@SpringBootApplication
@ClientCacheApplication
@EnableSecurity
public class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
}
On the server, you must configure access to the auth credentials. You may either implement the Apache Geode
SecurityManager
interface or declare
1 or more Apache Shiro Realms
. See Configuring Server Security for more details.
On the client, you must configure a username and password. See Configuring Client Security
for more details.
See Configuring Security for more details.
6.20.13. Configure Apache Geode Properties
To configure other, low-level Apache Geode properties not covered by the feature-oriented, SDG
configuration annotations, annotate your Spring, Apache Geode client or server application class
with @GemFireProperties
, as follows:
@SpringBootApplication
@PeerCacheApplication
@EnableGemFireProperties(
cacheXmlFile = "/path/to/cache.xml",
conserveSockets = true,
groups = "GroupOne",
remoteLocators = "lunchbox[11235],mailbox[10101],skullbox[12480]"
)
public class ServerApplication {
public static void main(String[] args) {
SpringApplication.run(ServerApplication.class, args);
}
}
Some Apache Geode properties are client-side only while others are server-side only. Please review the
Apache Geode docs for the appropriate use
of each property.
See Configuring Apache Geode Properties for more details.
6.20.14. Configure Caching
To use Apache Geode as a caching provider in Spring’s Cache Abstraction,
and have SDG automatically create Apache Geode Regions for the caches required by your application
service components, then annotate your Spring, Apache Geode client or server application class
with @EnableGemfireCaching
and @EnableCachingDefinedRegions
, as follows:
@SpringBootApplication
@ClientCacheApplication
@EnableCachingDefinedRegions
@EnableGemfireCaching
public class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
}
Then, simply go on to define the application services that require caching, as follows:
@Service
public class BookService {
@Cacheable("Books")
public Book findBy(ISBN isbn) {
...
}
}
@EnableCachingDefinedRegions
is optional. That is, you may manually define your Regions if you desire.
See Configuring Spring’s Cache Abstraction for more details.
6.20.15. Configure Regions, Indexes, Repositories and Entities for Persistent Applications
To make short work of creating Spring, Apache Geode persistent client or server applications, annotate your
application class with @EnableEntityDefinedRegions
, @EnableGemfireRepositories
and @EnableIndexing
, as follows:
@SpringBootApplication
@ClientCacheApplication
@EnableEntityDefinedRegions(basePackageClasses = Book.class)
@EnableGemfireRepositories(basePackageClasses = BookRepository.class)
@EnableIndexing
public class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
}
The @EnableEntityDefinedRegions
annotation is required when using the @EnableIndexing
annotation.
See Configuring Indexes for more details.
Next, define your entity class and use the @Region
mapping annotation to specify the Region in which your entity
will be stored. Use the @Indexed
annotation to define Indexes on entity fields used in your application queries,
as follows:
package example.app.model;
@Region("Books")
public class Book {
@Id
private ISBN isbn;
@Indexed;
private Author author;
@Indexed
private LocalDate published;
@LuceneIndexed
private String title;
}
The @Region("Books")
entity class annotation is used by the @EnableEntityDefinedRegions
to determine
the Regions required by the application. See Configuring Type-specific Regions and POJO Mapping
for more details.
Finally, define your CRUD Repository with simple queries to persist and access Books
, as follows:
package example.app.repo;
public interface BookRepository extends CrudRepository {
List<Book> findByAuthorOrderByPublishedDesc(Author author);
}
See Spring Data for Apache Geode Repositories for more details.
See @Region
Javadoc.
See @Indexed
Javadoc.
See Configuring Regions for more details.
See Spring Data for Apache Geode Repositories for more details.
6.20.16. Configure Client Regions from Cluster-defined Regions
Alternatively, you can define client [*PROXY] Regions from Regions already defined in the cluster
using @EnableClusterDefinedRegions
, as follows:
@SpringBootApplication
@ClientCacheApplication
@EnableClusterDefinedRegions
@EnableGemfireRepositories
public class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
...
}
See Configured Cluster-defined Regions for more details.
6.20.17. Configure Functions
Apache Geode Functions are useful in distributed compute scenarios where a potentially expensive computation
requiring data can be performed in parallel across the nodes in the cluster. In this case, it is more efficient
to bring the logic to where the data is located (stored) rather than requesting and fetching the data to be processed
by the computation.
Use the @EnableGemfireFunctions
along with the @GemfireFunction
annotation to enable Apache Geode Functions
definitions implemented as methods on POJOs, as follows:
@PeerCacheApplication
@EnableGemfireFunctions
class ServerApplication {
public static void main(String[] args) {
SpringApplication.run(ServerApplication.class, args);
}
@GemfireFunction
Integer computeLoyaltyPoints(Customer customer) {
...
}
}
Use the @EnableGemfireFunctionExecutions
along with 1 of the Function calling annotations: @OnMember
, @OnMembers
,
@OnRegion
, @OnServer
and @OnServers
.
@ClientCacheApplication
@EnableGemfireFunctionExecutions(basePackageClasses = CustomerRewardsFunction.class)
class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
}
@OnRegion("Customers")
interface CustomerRewardsFunctions {
Integer computeLoyaltyPoints(Customer customer);
}
See @OnMember
Javadoc,
@OnMembers
Javadoc,
@OnRegion
Javadoc,
@OnServer
Javadoc,
and @OnServers
Javadoc.
See Annotation Support for Function Execution for more details.
6.20.18. Configure Continuous Query
Real-time, event stream processing is becoming an increasingly important task for data-intensive applications,
primarily in order to respond to user requests in a timely manner. Apache Geode Continuous Query (CQ)
will help you achieve this rather complex task quite easily.
Enable CQ by annotating your application class with @EnableContinuousQueries
and define your CQs along with
the associated event handlers, as follows:
@ClientCacheApplication
@EnableContinuousQueries
class ClientApplication {
public static void main(String[] args) {
SpringApplication.run(ClientApplication.class, args);
}
}
Then, define your CQs by annotating the associated handler method with @ContinousQuery
, as follows:
@Service
class CustomerService {
@ContinuousQuery(name = "CustomerQuery", query = "SELECT * FROM /Customers c WHERE ...")
public void process(CqEvent event) {
...
}
}
Anytime an event occurs changing the Customer
data to match the predicate in your continuous OQL query (CQ),
the process
method will be called.
Apache Geode CQ is a client-side feature only.
See Continuous Query (CQ) and Configuring Continuous Queries for more details.
6.20.19. Configure Cluster Configuration
When developing Spring Data applications using Apache Geode as Apache Geode ClientCache
applications, it is
useful during development to configure the server to match the client in a client/server topology. In fact,
Apache Geode expects that when you have a "/Example" PROXY Region
on the client, that a matching Region
by name
(i.e. "Example") exists in the server.
You could use Gfsh to create every Region and Index that your application requires, or, you could simply push
the configuration meta-data already expressed when developing your Spring Data application using Apache Geode
when you run it.
This is as simple as annotation your main application class with @EnableClusterConfiguration(..)
:
Using @EnableClusterConfiguration
@ClientCacheApplication
@EnableClusterConfiguration(useHttp = true)
class ClientApplication {
...
}
Most of the time, when using a client/server topology, particularly in production environments, the servers
of the cluster will be started using Gfsh. In which case, it customary to use HTTP(S) to send the configuration
metadata (e.g. Region & Index definitions) to the cluster. When HTTP is used, the configuration metadata is sent
to the Manager in the cluster and distributed across the server nodes in the cluster consistently.
In order to use @EnableClusterConfiguration
you must declare the org.springframework:spring-web
dependency
in your Spring application classpath.
See Configuring Cluster Configuration Push for more details.
6.20.20. Configure GatewayReceivers
The replication of data between different Apache Geode clusters is an increasingly important fault-tolerance
and high-availability (HA) mechanism. Apache Geode WAN replication is a mechanism that allows one
Apache Geode cluster to replicate its data to another Apache Geode cluster in a reliable and fault-tolerant
manner.
Apache Geode WAN replication requires two components to be configured:
-
GatewayReceiver
- The WAN replication component that receives data from a remote Apache Geode cluster’s GatewaySender
.
-
GatewaySender
- The WAN replication component that sends data to a remote Apache Geode cluster’s GatewayReceiver
.
To enable a GatewayReceiver
, the application class needs to be annotated with @EnableGatewayReceiver
as follows:
@CacheServerApplication
@EnableGatewayReceiver(manualStart = false, startPort = 10000, endPort = 11000, maximumTimeBetweenPings = 1000,
socketBufferSize = 16384, bindAddress = "localhost",transportFilters = {"transportBean1", "transportBean2"},
hostnameForSenders = "hostnameLocalhost"){
...
...
}
}
class MySpringApplication { .. }
Apache Geode GatewayReceiver
is a server-side feature only and can only be configured on a CacheServer
or peer Cache
node.
6.20.21. Configure GatewaySenders
To enable GatewaySender
, the application class needs to be annotated with @EnableGatewaySenders
and @EnableGatewaySender
as follows:
@CacheServerApplication
@EnableGatewaySenders(gatewaySenders = {
@EnableGatewaySender(name = "GatewaySender", manualStart = true,
remoteDistributedSystemId = 2, diskSynchronous = true, batchConflationEnabled = true,
parallel = true, persistent = false,diskStoreReference = "someDiskStore",
orderPolicy = OrderPolicyType.PARTITION, alertThreshold = 1234, batchSize = 100,
eventFilters = "SomeEventFilter", batchTimeInterval = 2000, dispatcherThreads = 22,
maximumQueueMemory = 400,socketBufferSize = 16384,
socketReadTimeout = 4000, regions = { "Region1"}),
@EnableGatewaySender(name = "GatewaySender2", manualStart = true,
remoteDistributedSystemId = 2, diskSynchronous = true, batchConflationEnabled = true,
parallel = true, persistent = false, diskStoreReference = "someDiskStore",
orderPolicy = OrderPolicyType.PARTITION, alertThreshold = 1234, batchSize = 100,
eventFilters = "SomeEventFilter", batchTimeInterval = 2000, dispatcherThreads = 22,
maximumQueueMemory = 400, socketBufferSize = 16384,socketReadTimeout = 4000,
regions = { "Region2" })
}){
class MySpringApplication { .. }
}
Apache Geode GatewaySender
is a server-side feature only and can only be configured on a CacheServer
or a peer Cache
node.
In the above example, the application is configured with 2 Regions, Region1
and Region2
. In addition,
two GatewaySenders
will be configured to service both Regions. GatewaySender1
will be configured to replicate
Region1’s data and `GatewaySender2
will be configured to replicate `Region2’s data.
As demonstrated each GatewaySender
property can be configured on each EnableGatewaySender
annotation.
It is also possible to have a more generic, "defaulted" properties approach, where all properties are configured on
the EnableGatewaySenders
annotation. This way, a set of generic, defaulted values can be set on the parent annotation
and then overridden on the child if required, as demonstrated below:
@CacheServerApplication
@EnableGatewaySenders(gatewaySenders = {
@EnableGatewaySender(name = "GatewaySender", transportFilters = "transportBean1", regions = "Region2"),
@EnableGatewaySender(name = "GatewaySender2")},
manualStart = true, remoteDistributedSystemId = 2,
diskSynchronous = false, batchConflationEnabled = true, parallel = true, persistent = true,
diskStoreReference = "someDiskStore", orderPolicy = OrderPolicyType.PARTITION, alertThreshold = 1234, batchSize = 1002,
eventFilters = "SomeEventFilter", batchTimeInterval = 2000, dispatcherThreads = 22, maximumQueueMemory = 400,
socketBufferSize = 16384, socketReadTimeout = 4000, regions = { "Region1", "Region2" },
transportFilters = { "transportBean2", "transportBean1" })
class MySpringApplication { .. }
When the regions
attribute is left empty or not populated, the GatewaySender
(s) will automatically attach
itself to every configured Region
within the application.
7. Working with Apache Geode APIs
Once the Apache Geode Cache and Regions have been configured, they can be injected and used inside application objects.
This chapter describes the integration with Spring’s Transaction Management functionality and DAO exception hierarchy.
This chapter also covers support for dependency injection of Apache Geode managed objects.
7.1. GemfireTemplate
As with many other high-level abstractions provided by Spring, Spring Data for Apache Geode provides a template
to simplify Apache Geode data access operations. The class provides several methods containing common Region operations,
but also provides the capability to execute code against native Apache Geode APIs without having to deal with
Apache Geode checked exceptions by using a GemfireCallback
.
The template class requires a Apache Geode Region
, and once configured, is thread-safe and is reusable
across multiple application classes:
<bean id="gemfireTemplate" class="org.springframework.data.gemfire.GemfireTemplate" p:region-ref="SomeRegion"/>
Once the template is configured, a developer can use it alongside GemfireCallback
to work directly with
the Apache Geode Region
without having to deal with checked exceptions, threading or resource management concerns:
template.execute(new GemfireCallback<Iterable<String>>() {
public Iterable<String> doInGemfire(Region region)
throws GemFireCheckedException, GemFireException {
Region<String, String> localRegion = (Region<String, String>) region;
localRegion.put("1", "one");
localRegion.put("3", "three");
return localRegion.query("length < 5");
}
});
For accessing the full power of the Apache Geode query language, a developer can use the find
and findUnique
methods, which, compared to the query
method, can execute queries across multiple Regions, execute projections,
and the like.
The find
method should be used when the query selects multiple items (through SelectResults
) and the latter,
findUnique
, as the name suggests, when only one object is returned.
7.2. Exception Translation
Using a new data access technology requires not only accommodating a new API but also handling exceptions
specific to that technology.
To accommodate the exception handling case, the Spring Framework provides a technology agnostic and consistent
exception hierarchy
that abstracts the application from proprietary, and usually "checked", exceptions to a set of focused runtime
exceptions.
As mentioned in Spring Framework’s documentation,
Exception translation
can be applied transparently to your Data Access Objects (DAO) through the use of the @Repository
annotation and AOP
by defining a PersistenceExceptionTranslationPostProcessor
bean. The same exception translation functionality
is enabled when using Apache Geode as long as the CacheFactoryBean
is declared, e.g. using either a <gfe:cache/>
or <gfe:client-cache>
declaration, which acts as an exception translator and is automatically detected by
the Spring infrastructure and used accordingly.
7.3. Local, Cache Transaction Management
One of the most popular features of the Spring Framework is
Transaction Management.
If you are not familiar with Spring’s transaction abstraction then we strongly recommend
reading
about Spring’s Transaction Management infrastructure as it offers a consistent programming model that works
transparently across multiple APIs and can be configured either programmatically or declaratively
(the most popular choice).
For Apache Geode, Spring Data for Apache Geode provides a dedicated, per-cache, PlatformTransactionManager
that, once declared,
allows Region operations to be executed atomically through Spring:
Enable Transaction Management using XML
<gfe:transaction-manager id="txManager" cache-ref="myCache"/>
The example above can be simplified even further by eliminating the cache-ref
attribute if the Apache Geode
cache is defined under the default name, gemfireCache
. As with the other Spring Data for Apache Geode namespace elements, if the cache
bean name is not configured, the aforementioned naming convention will be used. Additionally, the transaction manager
name is “gemfireTransactionManager” if not explicitly specified.
Currently, Apache Geode supports optimistic transactions with read committed isolation. Furthermore, to guarantee
this isolation, developers should avoid making in-place changes that manually modify values present in the cache.
To prevent this from happening, the transaction manager configures the cache to use copy on read semantics by default,
meaning a clone of the actual value is created each time a read is performed. This behavior can be disabled if needed
through the copyOnRead
property.
Since a copy of the value for a given key is made when copy on read is enabled, you must subsequently call
Region.put(key, value)
inorder for the value to be updated, transactionally.
For more information on the semantics and behavior of the underlying Geode transaction manager, please refer to the Geode
CacheTransactionManager Javadoc
as well as the documentation.
7.4. Global, JTA Transaction Management
It is also possible for Apache Geode to participate in Global, JTA-based transactions, such as a transaction
managed by an Java EE Application Server (e.g. WebSphere Application Server (WAS)) using Container Managed Transactions
(CMT) along with other JTA resources.
However, unlike many other JTA "compliant" resources (e.g. JMS Message Brokers like ActiveMQ), Apache Geode is not
an XA compliant resource. Therefore, Apache Geode must be positioned as the "Last Resource" in a JTA transaction
(prepare phase) since it does not implement the 2-phase commit protocol, or rather does not handle distributed
transactions.
Many managed environments capable of CMT maintain support for "Last Resource", non-XA compliant resources in JTA-based
transactions, though it is not actually required in the JTA spec. More information on what a non-XA compliant,
"Last Resource" means can be found in Red Hat’s documentation.
In fact, Red Hat’s JBoss project, Narayana is one such LGPL Open Source implementation. Narayana
refers to this as "Last Resource Commit Optimization" (LRCO). More details can be found here.
However, whether you are using Apache Geode in a standalone environment with an Open Source JTA Transaction
Management implementation that supports "Last Resource", or a managed environment (e.g. Java EE AS such as WAS),
Spring Data for Apache Geode has you covered.
There are a series of steps you must complete to properly use Apache Geode as a "Last Resource" in a JTA
transaction involving more than 1 transactional resource. Additionally, there can only be 1 non-XA compliant resource
(e.g. Apache Geode) in such an arrangement.
1) First, you must complete Steps 1-4 in Apache Geode’s documentation
here.
#1 above is independent of your Spring [Boot] and/or [Data for Apache Geode] application and must be
completed successfully.
2) Referring to Step 5 in Apache Geode’s documentation,
Spring Data for Apache Geode’s Annotation support will attempt to set the GemFireCache
, copyOnRead
property for you when using the @EnableGemFireAsLastResource
annotation.
However, if SDG’s auto-configuration is unsuccessful in this regard, then you must explicitly set the copy-on-read
attribute in the <gfe:cache>
or <gfe:client-cache>
XML element or set the copyOnRead
property of
the CacheFactoryBean
class in JavaConfig to true. For example:
ClientCache
XML:
Set copy-on-read using XML (client)
<gfe:client-cache ... copy-on-read="true"/>
ClientCache
JavaConfig:
Set copyOnRead using JavaConfig (client)
@Bean
ClientCacheFactoryBean gemfireCache() {
ClientCacheFactoryBean gemfireCache = new ClientCacheFactoryBean();
gemfireCache.setCopyOnRead(true);
return gemfireCache;
}
Peer Cache
XML:
Set copy-on-read using XML (server)
<gfe:cache ... copy-on-read="true"/>
Peer Cache
JavaConfig:
Set copyOnRead using JavaConfig (server)
@Bean
CacheFactoryBean gemfireCache() {
CacheFactoryBean gemfireCache = new CacheFactoryBean();
gemfireCache.setCopyOnRead(true);
return gemfireCache;
}
Explicitly setting the copy-on-read
attribute or the copyOnRead
property is really not necessary. Enabling
transaction management takes case of copying on reads.
3) At this point, you skip Steps 6-8 in Apache Geode’s documentation
and let Spring Data Geode work its magic. All you need to do is annotate your Spring @Configuration
class
with Spring Data for Apache Geode’s new @EnableGemFireAsLastResource
annotation and a combination of Spring’s
Transaction Management infrastructure and Spring Data for Apache Geode’s
@EnableGemFireAsLastResource
annotation configuration does the trick.
The configuration looks like this…
@Configuration
@EnableGemFireAsLastResource
@EnableTransactionManagement(order = 1)
class GeodeConfiguration {
...
}
The only requirements are…
3.1) The @EnableGemFireAsLastResource
annotation must be declared on the same Spring @Configuration
class
where Spring’s @EnableTransactionManagement
annotation is also specified.
3.2) The order
attribute of the @EnableTransactionManagement
annotation must be explicitly set to an integer value
that is not Integer.MAX_VALUE
or Integer.MIN_VALUE
(defaults to Integer.MAX_VALUE
).
Of course, hopefully you are aware that you also need to configure Spring’s JtaTransactionManager
when using JTA transactions like so..
@Bean
public JtaTransactionManager transactionManager(UserTransaction userTransaction) {
JtaTransactionManager transactionManager = new JtaTransactionManager();
transactionManager.setUserTransaction(userTransaction);
return transactionManager;
}
The configuration in section Local, Cache Transaction Management does not apply here.
The use of Spring Data for Apache Geode’s GemfireTransactionManager
is applicable in "Local-only", Cache Transactions,
not "Global", JTA Transactions. Therefore, you do not configure the SDG GemfireTransactionManager
in this case.
You configure Spring’s JtaTransactionManager
as shown above.
For more details on using Spring’s Transaction Management with JTA,
see here.
Effectively, Spring Data for Apache Geode’s @EnableGemFireAsLastResource
annotation imports configuration containing 2 Aspect
bean definitions that handles the Apache Geode o.a.g.ra.GFConnectionFactory.getConnection()
and o.a.g.ra.GFConnection.close()
operations at the appropriate points during the transactional operation.
Specifically, the correct sequence of events follow:
-
jtaTransation.begin()
-
GFConnectionFactory.getConnection()
-
Call the application’s @Transactional
service method
-
Either jtaTransaction.commit()
or jtaTransaction.rollback()
-
Finally, GFConnection.close()
This is consistent with how you, as the application developer, would code this manually if you had to use the JTA API
+ Apache Geode API yourself, as shown in the
Apache Geode example.
Thankfully, Spring does the heavy lifting for you and all you need to do after applying the appropriate configuration
(shown above) is:
Declaring a service method as @Transactional
@Service
class MyTransactionalService {
@Transactional
public <Return-Type> someTransactionalServiceMethod() {
// perform business logic interacting with and accessing multiple JTA resources atomically
}
...
}
#1 & #4 above are appropriately handled for you by Spring’s JTA based PlatformTransactionManager
once the
@Transactional
boundary is entered by your application (i.e. when the MyTransactionService.someTransactionalServiceMethod()
is called).
#2 & #3 are handled by Spring Data for Apache Geode’s new Aspects enabled with the @EnableGemFireAsLastResource
annotation.
#3 of course is the responsibility of your application.
Indeed, with the appropriate logging configured, you will see the correct sequence of events…
Transaction Log Output
2017-Jun-22 11:11:37 TRACE TransactionInterceptor - Getting transaction for [example.app.service.MessageService.send]
2017-Jun-22 11:11:37 TRACE GemFireAsLastResourceConnectionAcquiringAspect - Acquiring {data-store-name} Connection
from {data-store-name} JCA ResourceAdapter registered at [gfe/jca]
2017-Jun-22 11:11:37 TRACE MessageService - PRODUCER [ Message :
[{ @type = example.app.domain.Message, id= MSG0000000000, message = SENT }],
JSON : [{"id":"MSG0000000000","message":"SENT"}] ]
2017-Jun-22 11:11:37 TRACE TransactionInterceptor - Completing transaction for [example.app.service.MessageService.send]
2017-Jun-22 11:11:37 TRACE GemFireAsLastResourceConnectionClosingAspect - Closed {data-store-name} Connection @ [Reference [...]]
For more details on using Apache Geode cache-level transactions, see here.
For more details on using Apache Geode in JTA transactions,
see here.
For more details on configuring Apache Geode as a "Last Resource",
see here.
7.5. Using @TransactionalEventListener
When using transactions, it may be desirable to register a listener to perform certain actions before or after the
transaction commits, or after a rollback occurs.
Spring Data for Apache Geode makes it easy to create listeners that will be invoked during specific phases of a transaction with the
@TransactionalEventListener
annotation. Methods annotated with @TransactionalEventListener
(as shown below) will be
notified of events published from transactional methods, during the specified phase
.
After Transaction Commit Event Listener
@TransactionalEventListener(phase = TransactionPhase.AFTER_COMMIT)
public void handleAfterCommit(MyEvent event) {
// do something after transaction is committed
}
Inorder for the above method to be invoked, you must publish an event from within your transaction, like below:
Publishing a Transactional Event
@Service
class MyTransactionalService {
@Autowired
private final ApplicationEventPublisher applicationEventPublisher;
@Transactional
public <Return-Type> someTransactionalServiceMethod() {
// Perform business logic interacting with and accessing multiple transactional resources atomically, then...
applicationEventPublisher.publishEvent(new MyApplicationEvent(...));
}
...
}
The @TransactionalEventListener
annotation allows you to specify the transaction phase
in which the event handler
method will be invoked. Options include: AFTER_COMMIT
, AFTER_COMPLETION
, AFTER_ROLLBACK
, and BEFORE_COMMIT
.
If not specified, the phase
defaults to AFTER_COMMIT
. If you wish the listener to be called even when no transaction
is present, you may set fallbackExecution
to true
.
7.6. Auto Transaction Event Publishing
As of Spring Data for Apache Geode Neumann/2.3
, it is now possible to enable auto transaction event publishing.
Using the @EnableGemfireCacheTransactions
annotation, set the enableAutoTransactionEventPublishing
attribute
to true. The default is false.
Enable auto transaction event publishing
@EnableGemfireCacheTransactions(enableAutoTransactionEventPublishing = true)
class GeodeConfiguration { ... }
Then you can create @TransactionalEventListener
annotated POJO methods to handle transaction events during either
the AFTER_COMMIT
or AFTER_ROLLBACK
transaction phases.
@Component
class TransactionEventListeners {
@TransactionalEventListener(phase = TransactionPhase.AFTER_COMMIT)
public void handleAfterCommit(TransactionApplicationEvent event) {
...
}
@TransactionalEventListener(phase = TransactionPhase.AFTER_ROLLBACK)
public void handleAfterRollback(TransactionApplicationEvent event) {
...
}
}
Only TransactionPhase.AFTER_COMMIT
and TransactionPhase.AFTER_ROLLBACK
are supported.
TransactionPhase.BEFORE_COMMIT
is not supported because 1) SDG adapts Apache Geode’s TransactionListener
and TransactionWriter
interfaces to implement auto transaction event publishing, and 2) when Apache Geode’s
TransactionWriter.beforeCommit(:TransactionEvent)
is called, it is already after the
AbstractPlatformTransactionManager.triggerBeforeCommit(:TransactionStatus)
call where @TranactionalEventListener
annotated POJO methods are called during the transaction lifecycle.
With auto transaction event publishing, you do not need to explicitly call the
applicationEventPublisher.publishEvent(..)
method inside your application @Transactional
@Service
methods.
However, if you still want to receive transaction events "before commit", then you must still call the
applicationEventPublisher.publishEvent(..)
method within your application @Transactional
@Service
methods.
See the note above for more details.
7.7. Continuous Query (CQ)
A powerful functionality offered by Apache Geode is
Continuous Query (or CQ).
In short, CQ allows a developer to create and register an OQL query, and then automatically be notified when new data
that gets added to Apache Geode matches the query predicate. Spring Data for Apache Geode provides dedicated
support for CQs through the org.springframework.data.gemfire.listener
package and its listener container;
very similar in functionality and naming to the JMS integration in the Spring Framework; in fact, users familiar with
the JMS support in Spring, should feel right at home.
Basically Spring Data for Apache Geode allows methods on POJOs to become end-points for CQ. Simply define the query
and indicate the method that should be called to be notified when there is a match. Spring Data for Apache Geode takes care
of the rest. This is very similar to Java EE’s message-driven bean style, but without any requirement for base class
or interface implementations, based on Apache Geode.
Currently, Continuous Query is only supported in Apache Geode’s client/server topology. Additionally, the client Pool
used is required to have the subscription enabled. Please refer to the Apache Geode
documentation
for more information.
7.7.1. Continuous Query Listener Container
Spring Data for Apache Geode simplifies creation, registration, life-cycle and dispatch of CQ events by taking care of
the infrastructure around CQ with the use of SDG’s ContinuousQueryListenerContainer
, which does all the heavy lifting
on behalf of the user. Users familiar with EJB and JMS should find the concepts familiar as it is designed
as close as possible to the support provided in the Spring Framework with its Message-driven POJOs (MDPs).
The SDG ContinuousQueryListenerContainer
acts as an event (or message) listener container; it is used to
receive the events from the registered CQs and invoke the POJOs that are injected into it. The listener container
is responsible for all threading of message reception and dispatches into the listener for processing. It acts as
the intermediary between an EDP (Event-driven POJO) and the event provider and takes care of creation and registration
of CQs (to receive events), resource acquisition and release, exception conversion and the like. This allows you,
as an application developer, to write the (possibly complex) business logic associated with receiving an event
(and reacting to it), and delegate the boilerplate Apache Geode infrastructure concerns to the framework.
The listener container is fully customizable. A developer can chose either to use the CQ thread to perform the dispatch
(synchronous delivery) or a new thread (from an existing pool) for an asynchronous approach by defining the suitable
java.util.concurrent.Executor
(or Spring’s TaskExecutor
). Depending on the load, the number of listeners
or the runtime environment, the developer should change or tweak the executor to better serve her needs. In particular,
in managed environments (such as app servers), it is highly recommended to pick a proper TaskExecutor
to take advantage of its runtime.
7.7.2. The ContinuousQueryListener
and ContinuousQueryListenerAdapter
The ContinuousQueryListenerAdapter
class is the final component in Spring Data for Apache Geode CQ support. In a nutshell,
class allows you to expose almost any implementing class as an EDP with minimal constraints.
ContinuousQueryListenerAdapter
implements the ContinuousQueryListener
interface, a simple listener interface
similar to Apache Geode’s CqListener.
Consider the following interface definition. Notice the various event handling methods and their parameters:
public interface EventDelegate {
void handleEvent(CqEvent event);
void handleEvent(Operation baseOp);
void handleEvent(Object key);
void handleEvent(Object key, Object newValue);
void handleEvent(Throwable throwable);
void handleQuery(CqQuery cq);
void handleEvent(CqEvent event, Operation baseOp, byte[] deltaValue);
void handleEvent(CqEvent event, Operation baseOp, Operation queryOp, Object key, Object newValue);
}
package example;
class DefaultEventDelegate implements EventDelegate {
// implementation elided for clarity...
}
In particular, note how the above implementation of the EventDelegate
interface has no Apache Geode dependencies at all.
It truly is a POJO that we can and will make into an EDP via the following configuration.
the class does not have to implement an interface; an interface is only used to better showcase the decoupling
between the contract and the implementation.
<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:gfe="https://www.springframework.org/schema/geode"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
https://www.springframework.org/schema/geode https://www.springframework.org/schema/geode/spring-geode.xsd
">
<gfe:client-cache/>
<gfe:pool subscription-enabled="true">
<gfe:server host="localhost" port="40404"/>
</gfe:pool>
<gfe:cq-listener-container>
<!-- default handle method -->
<gfe:listener ref="listener" query="SELECT * FROM /SomeRegion"/>
<gfe:listener ref="another-listener" query="SELECT * FROM /AnotherRegion" name="myQuery" method="handleQuery"/>
</gfe:cq-listener-container>
<bean id="listener" class="example.DefaultMessageDelegate"/>
<bean id="another-listener" class="example.DefaultMessageDelegate"/>
...
<beans>
The example above shows a few of the various forms that a listener can have; at its minimum, the listener
reference and the actual query definition are required. It’s possible, however, to specify a name for
the resulting Continuous Query (useful for monitoring) but also the name of the method (the default is handleEvent
).
The specified method can have various argument types, the EventDelegate
interface lists the allowed types.
The example above uses the Spring Data for Apache Geode namespace to declare the event listener container
and automatically register the listeners. The full blown, beans definition is displayed below:
<!-- this is the Event Driven POJO (MDP) -->
<bean id="eventListener" class="org.springframework.data.gemfire.listener.adapter.ContinuousQueryListenerAdapter">
<constructor-arg>
<bean class="gemfireexample.DefaultEventDelegate"/>
</constructor-arg>
</bean>
<!-- and this is the event listener container... -->
<bean id="gemfireListenerContainer" class="org.springframework.data.gemfire.listener.ContinuousQueryListenerContainer">
<property name="cache" ref="gemfireCache"/>
<property name="queryListeners">
<!-- set of CQ listeners -->
<set>
<bean class="org.springframework.data.gemfire.listener.ContinuousQueryDefinition" >
<constructor-arg value="SELECT * FROM /SomeRegion" />
<constructor-arg ref="eventListener"/>
</bean>
</set>
</property>
</bean>
Each time an event is received, the adapter automatically performs type translation between the Apache Geode event
and the required method argument(s) transparently. Any exception caused by the method invocation is caught
and handled by the container (by default, being logged).
7.8. Wiring Declarable
Components
Apache Geode XML configuration (usually referred to as cache.xml
) allows user objects to be declared
as part of the configuration. Usually these objects are CacheLoaders
or other pluggable callback components
supported by Apache Geode. Using native Apache Geode configuration, each user type declared through XML must implement
the Declarable
interface, which allows arbitrary parameters to be passed to the declared class
through a Properties
instance.
In this section, we describe how you can configure these pluggable components when defined in cache.xml
using Spring while keeping your Cache/Region configuration defined in cache.xml
. This allows your
pluggable components to focus on the application logic and not the location or creation of DataSources
or other collaborators.
However, if you are starting a green field project, it is recommended that you configure Cache, Region,
and other pluggable Apache Geode components directly in Spring. This avoids inheriting from the Declarable
interface
or the base class presented in this section.
See the following sidebar for more information on this approach.
As an example of configuring a Declarable
component using Spring, consider the following declaration
(taken from the Declarable
Javadoc):
<cache-loader>
<class-name>com.company.app.DBLoader</class-name>
<parameter name="URL">
<string>jdbc://12.34.56.78/mydb</string>
</parameter>
</cache-loader>
To simplify the task of parsing, converting the parameters and initializing the object, Spring Data for Apache Geode offers
a base class (WiringDeclarableSupport
) that allows Apache Geode user objects to be wired through a template bean definition
or, in case that is missing, perform auto-wiring through the Spring IoC container. To take advantage of this feature,
the user objects need to extend WiringDeclarableSupport
, which automatically locates the declaring BeanFactory
and performs wiring as part of the initialization process.
7.8.1. Configuration using template bean definitions
When used, WiringDeclarableSupport
tries to first locate an existing bean definition and use that
as the wiring template. Unless specified, the component class name will be used as an implicit bean definition name.
Let’s see how our DBLoader
declaration would look in that case:
class DBLoader extends WiringDeclarableSupport implements CacheLoader {
private DataSource dataSource;
public void setDataSource(DataSource dataSource){
this.dataSource = dataSource;
}
public Object load(LoaderHelper helper) { ... }
}
<cache-loader>
<class-name>com.company.app.DBLoader</class-name>
<!-- no parameter is passed (use the bean's implicit name, which is the class name) -->
</cache-loader>
<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:p="http://www.springframework.org/schema/p"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
">
<bean id="dataSource" ... />
<!-- template bean definition -->
<bean id="com.company.app.DBLoader" abstract="true" p:dataSource-ref="dataSource"/>
</beans>
In the scenario above, as no parameter was specified, a bean with the id/name com.company.app.DBLoader
was used
as a template for wiring the instance created by Apache Geode. For cases where the bean name uses a different convention,
one can pass in the bean-name
parameter in the Apache Geode configuration:
<cache-loader>
<class-name>com.company.app.DBLoader</class-name>
<!-- pass the bean definition template name as parameter -->
<parameter name="bean-name">
<string>template-bean</string>
</parameter>
</cache-loader>
<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:p="http://www.springframework.org/schema/p"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
">
<bean id="dataSource" ... />
<!-- template bean definition -->
<bean id="template-bean" abstract="true" p:dataSource-ref="dataSource"/>
</beans>
The template bean definitions do not have to be declared in XML.
Any format is allowed (Groovy, annotations, etc).
7.8.2. Configuration using auto-wiring and annotations
By default, if no bean definition is found, WiringDeclarableSupport
will
autowire
the declaring instance. This means that unless any dependency injection metadata is offered by the instance,
the container will find the object setters and try to automatically satisfy these dependencies.
However, a developer can also use JDK 5 annotations to provide additional information to the auto-wiring process.
We strongly recommend reading the dedicated
chapter
in the Spring documentation for more information on the supported annotations and enabling factors.
For example, the hypothetical DBLoader
declaration above can be injected with a Spring-configured DataSource
in the following way:
class DBLoader extends WiringDeclarableSupport implements CacheLoader {
// use annotations to 'mark' the needed dependencies
@javax.inject.Inject
private DataSource dataSource;
public Object load(LoaderHelper helper) { ... }
}
<cache-loader>
<class-name>com.company.app.DBLoader</class-name>
<!-- no need to declare any parameters since the class is auto-wired -->
</cache-loader>
<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:context="http://www.springframework.org/schema/context"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
http://www.springframework.org/schema/context https://www.springframework.org/schema/context/spring-context.xsd
">
<!-- enable annotation processing -->
<context:annotation-config/>
</beans>
By using the JSR-330 annotations, the CacheLoader
code has been simplified since the location and creation
of the DataSource
has been externalized and the user code is concerned only with the loading process.
The DataSource
might be transactional, created lazily, shared between multiple objects or retrieved from JNDI.
These aspects can easily be configured and changed through the Spring container without touching
the DBLoader
code.
7.9. Support for the Spring Cache Abstraction
Spring Data for Apache Geode provides an implementation of the Spring
Cache Abstraction
to position Apache Geode as a caching provider in Spring’s caching infrastructure.
To use Apache Geode as a backing implementation, a "caching provider" in Spring’s Cache Abstraction,
simply add GemfireCacheManager
to your configuration:
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:cache="http://www.springframework.org/schema/cache"
xmlns:gfe="https://www.springframework.org/schema/geode"
xmlns:p="http://www.springframework.org/schema/p"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
http://www.springframework.org/schema/cache https://www.springframework.org/schema/cache/spring-cache.xsd
https://www.springframework.org/schema/geode https://www.springframework.org/schema/geode/spring-geode.xsd
">
<!-- enable declarative caching -->
<cache:annotation-driven/>
<gfe:cache id="gemfire-cache"/>
<!-- declare GemfireCacheManager; must have a bean ID of 'cacheManager' -->
<bean id="cacheManager" class="org.springframework.data.gemfire.cache.GemfireCacheManager"
p:cache-ref="gemfire-cache">
</beans>
The cache-ref
attribute on the CacheManager
bean definition is not necessary if the default cache bean name
is used (i.e. "gemfireCache"), i.e. <gfe:cache>
without an explicit ID.
When the GemfireCacheManager
(Singleton) bean instance is declared and declarative caching is enabled
(either in XML with <cache:annotation-driven/>
or in JavaConfig with Spring’s @EnableCaching
annotation),
the Spring caching annotations (e.g. @Cacheable
) identify the "caches" that will cache data in-memory
using Apache Geode Regions.
These caches (i.e. Regions) must exist before the caching annotations that use them otherwise an error will occur.
By way of example, suppose you have a Customer Service application with a CustomerService
application component
that performs caching…
@Service
class CustomerService {
@Cacheable(cacheNames="Accounts", key="#customer.id")
Account createAccount(Customer customer) {
...
}
Then you will need the following config.
XML:
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:cache="http://www.springframework.org/schema/cache"
xmlns:gfe="https://www.springframework.org/schema/geode"
xmlns:p="http://www.springframework.org/schema/p"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
http://www.springframework.org/schema/cache https://www.springframework.org/schema/cache/spring-cache.xsd
https://www.springframework.org/schema/geode https://www.springframework.org/schema/geode/spring-geode.xsd
">
<!-- enable declarative caching -->
<cache:annotation-driven/>
<bean id="cacheManager" class="org.springframework.data.gemfire.cache.GemfireCacheManager">
<gfe:cache/>
<gfe:partitioned-region id="accountsRegion" name="Accounts" persistent="true" ...>
...
</gfe:partitioned-region>
</beans>
JavaConfig:
@Configuration
@EnableCaching
class ApplicationConfiguration {
@Bean
CacheFactoryBean gemfireCache() {
return new CacheFactoryBean();
}
@Bean
GemfireCacheManager cacheManager() {
GemfireCacheManager cacheManager = GemfireCacheManager();
cacheManager.setCache(gemfireCache());
return cacheManager;
}
@Bean("Accounts")
PartitionedRegionFactoryBean accountsRegion() {
PartitionedRegionFactoryBean accounts = new PartitionedRegionFactoryBean();
accounts.setCache(gemfireCache());
accounts.setClose(false);
accounts.setPersistent(true);
return accounts;
}
}
Of course, you are free to choose whatever Region type you like (e.g. REPLICATE, PARTITION, LOCAL, etc).
For more details on Spring’s Cache Abstraction, again, please refer to the
documentation.
8. Working with Apache Geode Serialization
To improve overall performance of the Apache Geode In-memory Data Grid, Apache Geode supports a dedicated
serialization protocol, called PDX, that is both faster and offers more compact results over standard Java serialization
in addition to working transparently across various language platforms (Java, C++, and .NET).
See PDX Serialization Features
and PDX Serialization Internals for more details.
This chapter discusses the various ways in which Spring Data for Apache Geode simplifies and improves Apache Geode’s
custom serialization in Java.
8.1. Wiring deserialized instances
It is fairly common for serialized objects to have transient data. Transient data is often dependent on the system
or environment where it lives at a certain point in time. For instance, a DataSource
is environment specific.
Serializing such information is useless and potentially even dangerous, since it is local to a certain VM or machine.
For such cases, Spring Data for Apache Geode offers a special Instantiator
that performs wiring for each new instance created by Apache Geode during deserialization.
Through such a mechanism, you can rely on the Spring container to inject and manage certain dependencies, making it easy
to split transient from persistent data and have rich domain objects in a transparent manner.
Spring users might find this approach similar to that of @Configurable
).
The WiringInstantiator
works similarly to WiringDeclarableSupport
, trying to first locate a bean definition
as a wiring template and otherwise falling back to auto-wiring.
See the previous section (Wiring Declarable
Components) for more details on wiring functionality.
To use the SDG Instantiator
, declare it as a bean, as the following example shows:
<bean id="instantiator" class="org.springframework.data.gemfire.serialization.WiringInstantiator">
<!-- DataSerializable type -->
<constructor-arg>org.pkg.SomeDataSerializableClass</constructor-arg>
<!-- type id -->
<constructor-arg>95</constructor-arg>
</bean>
During the Spring container startup, once it has been initialized, the Instantiator
, by default, registers itself with
the Apache Geode serialization system and performs wiring on all instances of SomeDataSerializableClass
created
by Apache Geode during deserialization.
8.2. Auto-generating Custom Instantiators
For data intensive applications, a large number of instances might be created on each machine as data flows in.
Apache Geode uses reflection to create new types, but, for some scenarios, this might prove to be expensive.
As always, it is good to perform profiling to quantify whether this is the case or not. For such cases, Spring Data for Apache Geode
allows the automatic generation of Instatiator
classes, which instantiate a new type (using the default constructor)
without the use of reflection. The following example shows how to create an instantiator:
<bean id="instantiatorFactory" class="org.springframework.data.gemfire.serialization.InstantiatorFactoryBean">
<property name="customTypes">
<map>
<entry key="org.pkg.CustomTypeA" value="1025"/>
<entry key="org.pkg.CustomTypeB" value="1026"/>
</map>
</property>
</bean>
The preceding definition automatically generates two Instantiators
for two classes (CustomTypeA
and CustomTypeB
)
and registers them with Apache Geode under user ID 1025
and 1026
. The two Instantiators
avoid the use of
reflection and create the instances directly through Java code.
9. POJO Mapping
This section covers:
9.1. Object Mapping Fundamentals
This section covers the fundamentals of Spring Data object mapping, object creation, field and property access, mutability and immutability.
Note, that this section only applies to Spring Data modules that do not use the object mapping of the underlying data store (like JPA).
Also be sure to consult the store-specific sections for store-specific object mapping, like indexes, customizing column or field names or the like.
Core responsibility of the Spring Data object mapping is to create instances of domain objects and map the store-native data structures onto those.
This means we need two fundamental steps:
-
Instance creation by using one of the constructors exposed.
-
Instance population to materialize all exposed properties.
9.1.1. Object creation
Spring Data automatically tries to detect a persistent entity’s constructor to be used to materialize objects of that type.
The resolution algorithm works as follows:
-
If there is a single static factory method annotated with @PersistenceCreator
then it is used.
-
If there is a single constructor, it is used.
-
If there are multiple constructors and exactly one is annotated with @PersistenceCreator
, it is used.
-
If there’s a no-argument constructor, it is used.
Other constructors will be ignored.
The value resolution assumes constructor/factory method argument names to match the property names of the entity, i.e. the resolution will be performed as if the property was to be populated, including all customizations in mapping (different datastore column or field name etc.).
This also requires either parameter names information available in the class file or an @ConstructorProperties
annotation being present on the constructor.
The value resolution can be customized by using Spring Framework’s @Value
value annotation using a store-specific SpEL expression.
Please consult the section on store specific mappings for further details.
9.1.2. Property population
Once an instance of the entity has been created, Spring Data populates all remaining persistent properties of that class.
Unless already populated by the entity’s constructor (i.e. consumed through its constructor argument list), the identifier property will be populated first to allow the resolution of cyclic object references.
After that, all non-transient properties that have not already been populated by the constructor are set on the entity instance.
For that we use the following algorithm:
-
If the property is immutable but exposes a with…
method (see below), we use the with…
method to create a new entity instance with the new property value.
-
If property access (i.e. access through getters and setters) is defined, we’re invoking the setter method.
-
If the property is mutable we set the field directly.
-
If the property is immutable we’re using the constructor to be used by persistence operations (see Object creation) to create a copy of the instance.
-
By default, we set the field value directly.
Let’s have a look at the following entity:
Example 2. A sample entity
class Person {
private final @Id Long id; (1)
private final String firstname, lastname; (2)
private final LocalDate birthday;
private final int age; (3)
private String comment; (4)
private @AccessType(Type.PROPERTY) String remarks; (5)
static Person of(String firstname, String lastname, LocalDate birthday) { (6)
return new Person(null, firstname, lastname, birthday,
Period.between(birthday, LocalDate.now()).getYears());
}
Person(Long id, String firstname, String lastname, LocalDate birthday, int age) { (6)
this.id = id;
this.firstname = firstname;
this.lastname = lastname;
this.birthday = birthday;
this.age = age;
}
Person withId(Long id) { (1)
return new Person(id, this.firstname, this.lastname, this.birthday, this.age);
}
void setRemarks(String remarks) { (5)
this.remarks = remarks;
}
}
1
The identifier property is final but set to null
in the constructor.
The class exposes a withId(…)
method that’s used to set the identifier, e.g. when an instance is inserted into the datastore and an identifier has been generated.
The original Person
instance stays unchanged as a new one is created.
The same pattern is usually applied for other properties that are store managed but might have to be changed for persistence operations.
The wither method is optional as the persistence constructor (see 6) is effectively a copy constructor and setting the property will be translated into creating a fresh instance with the new identifier value applied.
2
The firstname
and lastname
properties are ordinary immutable properties potentially exposed through getters.
3
The age
property is an immutable but derived one from the birthday
property.
With the design shown, the database value will trump the defaulting as Spring Data uses the only declared constructor.
Even if the intent is that the calculation should be preferred, it’s important that this constructor also takes age
as parameter (to potentially ignore it) as otherwise the property population step will attempt to set the age field and fail due to it being immutable and no with…
method being present.
4
The comment
property is mutable and is populated by setting its field directly.
5
The remarks
property is mutable and is populated by invoking the setter method.
6
The class exposes a factory method and a constructor for object creation.
The core idea here is to use factory methods instead of additional constructors to avoid the need for constructor disambiguation through @PersistenceCreator
.
Instead, defaulting of properties is handled within the factory method.
If you want Spring Data to use the factory method for object instantiation, annotate it with @PersistenceCreator
.
9.1.3. General recommendations
-
Try to stick to immutable objects — Immutable objects are straightforward to create as materializing an object is then a matter of calling its constructor only.
Also, this avoids your domain objects to be littered with setter methods that allow client code to manipulate the objects state.
If you need those, prefer to make them package protected so that they can only be invoked by a limited amount of co-located types.
Constructor-only materialization is up to 30% faster than properties population.
-
Provide an all-args constructor — Even if you cannot or don’t want to model your entities as immutable values, there’s still value in providing a constructor that takes all properties of the entity as arguments, including the mutable ones, as this allows the object mapping to skip the property population for optimal performance.
-
Use factory methods instead of overloaded constructors to avoid @PersistenceCreator
— With an all-argument constructor needed for optimal performance, we usually want to expose more application use case specific constructors that omit things like auto-generated identifiers etc.
It’s an established pattern to rather use static factory methods to expose these variants of the all-args constructor.
-
Make sure you adhere to the constraints that allow the generated instantiator and property accessor classes to be used —
-
For identifiers to be generated, still use a final field in combination with an all-arguments persistence constructor (preferred) or a with…
method —
-
Use Lombok to avoid boilerplate code — As persistence operations usually require a constructor taking all arguments, their declaration becomes a tedious repetition of boilerplate parameter to field assignments that can best be avoided by using Lombok’s @AllArgsConstructor
.
Overriding Properties
Java’s allows a flexible design of domain classes where a subclass could define a property that is already declared with the same name in its superclass.
Consider the following example:
public class SuperType {
private CharSequence field;
public SuperType(CharSequence field) {
this.field = field;
}
public CharSequence getField() {
return this.field;
}
public void setField(CharSequence field) {
this.field = field;
}
}
public class SubType extends SuperType {
private String field;
public SubType(String field) {
super(field);
this.field = field;
}
@Override
public String getField() {
return this.field;
}
public void setField(String field) {
this.field = field;
// optional
super.setField(field);
}
}
Both classes define a field
using assignable types. SubType
however shadows SuperType.field
.
Depending on the class design, using the constructor could be the only default approach to set SuperType.field
.
Alternatively, calling super.setField(…)
in the setter could set the field
in SuperType
.
All these mechanisms create conflicts to some degree because the properties share the same name yet might represent two distinct values.
Spring Data skips super-type properties if types are not assignable.
That is, the type of the overridden property must be assignable to its super-type property type to be registered as override, otherwise the super-type property is considered transient.
We generally recommend using distinct property names.
Spring Data modules generally support overridden properties holding different values.
From a programming model perspective there are a few things to consider:
-
Which property should be persisted (default to all declared properties)?
You can exclude properties by annotating these with @Transient
.
-
How to represent properties in your data store?
Using the same field/column name for different values typically leads to corrupt data so you should annotate least one of the properties using an explicit field/column name.
-
Using @AccessType(PROPERTY)
cannot be used as the super-property cannot be generally set without making any further assumptions of the setter implementation.
9.1.4. Kotlin support
Spring Data adapts specifics of Kotlin to allow object creation and mutation.
Kotlin object creation
Kotlin classes are supported to be instantiated , all classes are immutable by default and require explicit property declarations to define mutable properties.
Consider the following data
class Person
:
data class Person(val id: String, val name: String)
The class above compiles to a typical class with an explicit constructor.We can customize this class by adding another constructor and annotate it with @PersistenceCreator
to indicate a constructor preference:
data class Person(var id: String, val name: String) {
@PersistenceCreator
constructor(id: String) : this(id, "unknown")
}
Kotlin supports parameter optionality by allowing default values to be used if a parameter is not provided.
When Spring Data detects a constructor with parameter defaulting, then it leaves these parameters absent if the data store does not provide a value (or simply returns null
) so Kotlin can apply parameter defaulting.Consider the following class that applies parameter defaulting for name
data class Person(var id: String, val name: String = "unknown")
Every time the name
parameter is either not part of the result or its value is null
, then the name
defaults to unknown
.
Property population of Kotlin data classes
In Kotlin, all classes are immutable by default and require explicit property declarations to define mutable properties.
Consider the following data
class Person
:
data class Person(val id: String, val name: String)
This class is effectively immutable.
It allows creating new instances as Kotlin generates a copy(…)
method that creates new object instances copying all property values from the existing object and applying property values provided as arguments to the method.
Kotlin Overriding Properties
Kotlin allows declaring property overrides to alter properties in subclasses.
open class SuperType(open var field: Int)
class SubType(override var field: Int = 1) :
SuperType(field) {
}
Such an arrangement renders two properties with the name field
.
Kotlin generates property accessors (getters and setters) for each property in each class.
Effectively, the code looks like as follows:
public class SuperType {
private int field;
public SuperType(int field) {
this.field = field;
}
public int getField() {
return this.field;
}
public void setField(int field) {
this.field = field;
}
}
public final class SubType extends SuperType {
private int field;
public SubType(int field) {
super(field);
this.field = field;
}
public int getField() {
return this.field;
}
public void setField(int field) {
this.field = field;
}
}
Getters and setters on SubType
set only SubType.field
and not SuperType.field
.
In such an arrangement, using the constructor is the only default approach to set SuperType.field
.
Adding a method to SubType
to set SuperType.field
via this.SuperType.field = …
is possible but falls outside of supported conventions.
Property overrides create conflicts to some degree because the properties share the same name yet might represent two distinct values.
We generally recommend using distinct property names.
Spring Data modules generally support overridden properties holding different values.
From a programming model perspective there are a few things to consider:
-
Which property should be persisted (default to all declared properties)?
You can exclude properties by annotating these with @Transient
.
-
How to represent properties in your data store?
Using the same field/column name for different values typically leads to corrupt data so you should annotate least one of the properties using an explicit field/column name.
-
Using @AccessType(PROPERTY)
cannot be used as the super-property cannot be set.
9.2. Entity Mapping
Spring Data for Apache Geode provides support to map entities that are stored in a Region. The mapping metadata is defined by
using annotations on application domain classes, as the following example shows:
Example 3. Mapping a domain class to a Apache Geode Region
@Region("People")
public class Person {
@Id Long id;
String firstname;
String lastname;
@PersistenceConstructor
public Person(String firstname, String lastname) {
// …
}
…
}
The @Region
annotation can be used to customize the Region in which an instance of the Person
class is stored.
The @Id
annotation can be used to annotate the property that should be used as the cache Region key, identifying
the Region entry. The @PersistenceConstructor
annotation helps to disambiguate multiple potentially available
constructors, taking parameters and explicitly marking the constructor annotated as the constructor to be used to
construct entities. In an application domain class with no or only a single constructor, you can omit the annotation.
In addition to storing entities in top-level Regions, entities can be stored in Sub-Regions as well,
as the following example shows:
@Region("/Users/Admin")
public class Admin extends User {
…
}
@Region("/Users/Guest")
public class Guest extends User {
…
}
Be sure to use the full path of the Apache Geode Region, as defined with the Spring Data for Apache Geode XML namespace
by using the id
or name
attributes of the <*-region>
element.
9.2.1. Entity Mapping by Region Type
In addition to the @Region
annotation, Spring Data for Apache Geode also recognizes type-specific Region mapping annotations:
@ClientRegion
, @LocalRegion
, @PartitionRegion
, and @ReplicateRegion
.
Functionally, these annotations are treated exactly the same as the generic @Region
annotation in the SDG
mapping infrastructure. However, these additional mapping annotations are useful in Spring Data for Apache Geode’s
annotation configuration model. When combined with the @EnableEntityDefinedRegions
configuration annotation
on a Spring @Configuration
annotated class, it is possible to generate Regions in the local cache, whether
the application is a client or peer.
These annotations let you be more specific about what type of Region your application entity class should be mapped to
and also has an impact on the data management policies of the Region (for example, partition — also known as sharding — versus replicating data).
Using these type-specific Region mapping annotations with the SDG annotation configuration model saves you
from having to explicitly define these Regions in configuration.
9.3. Repository Mapping
As an alternative to specifying the Region in which the entity is stored by using the @Region
annotation
on the entity class, you can also specify the @Region
annotation on the entity’s Repository
interface.
See Spring Data for Apache Geode Repositories for more details.
However, suppose you want to store a Person
record in multiple Apache Geode Regions (for example, People
and Customers
). Then you can define your corresponding Repository
interface extensions as follows:
@Region("People")
public interface PersonRepository extends GemfireRepository<Person, String> {
…
}
@Region("Customers")
public interface CustomerRepository extends GemfireRepository<Person, String> {
...
}
Then, using each Repository individually, you can store the entity in multiple Apache Geode Regions,
as the following example shows:
@Service
class CustomerService {
CustomerRepository customerRepo;
PersonRepository personRepo;
Customer update(Customer customer) {
customerRepo.save(customer);
personRepo.save(customer);
return customer;
}
You can even wrap the update
service method in a Spring managed transaction, either as a local cache transaction
or a global transaction.
9.4. MappingPdxSerializer
Spring Data for Apache Geode provides a custom PdxSerializer
implementation, called MappingPdxSerializer
, that uses Spring Data mapping metadata to customize entity serialization.
The serializer also lets you customize entity instantiation by using the Spring Data EntityInstantiator
abstraction.
By default, the serializer use the ReflectionEntityInstantiator
, which uses the persistence constructor of
the mapped entity. The persistence constructor is either the default constructor, a singly declared constructor,
or a constructor explicitly annotated with @PersistenceConstructor
.
To provide arguments for constructor parameters, the serializer reads fields with the named constructor parameter,
explicitly identified by using Spring’s @Value
annotation, from the supplied
PdxReader
,
as shown in the following example:
Example 4. Using @Value
on entity constructor parameters
public class Person {
public Person(@Value("#root.thing") String firstName, @Value("bean") String lastName) {
…
}
}
An entity class annotated in this way has the “thing” field read from the PdxReader
and passed as the argument value
for the constructor parameter, firstname
. The value for lastName
is a Spring bean with the name “bean”.
In addition to the custom instantiation logic and strategy provided by EntityInstantiators
,
the MappingPdxSerializer
also provides capabilities well beyond Apache Geode’s own
ReflectionBasedAutoSerializer
.
While Apache Geode’s ReflectionBasedAutoSerializer
conveniently uses Java Reflection to populate entities
and uses regular expressions to identify types that should be handled (serialized and deserialized) by the serializer,
it cannot, unlike MappingPdxSerializer
, perform the following:
-
Register custom PdxSerializer
objects per entity field or property names and types.
-
Conveniently identifies ID properties.
-
Automatically handles read-only properties.
-
Automatically handles transient properties.
-
Allows more robust type filtering in a null
and type-safe manner (for example, not limited to
only expressing types with regex).
We now explore each feature of the MappingPdxSerializer
in a bit more detail.
9.4.1. Custom PdxSerializer Registration
The MappingPdxSerializer
gives you the ability to register custom PdxSerializers
based on an entity’s field
or property names and types.
For example, suppose you have defined an entity type modeling a User
as follows:
package example.app.security.auth.model;
public class User {
private String name;
private Password password;
...
}
While the user’s name probably does not require any special logic to serialize the value, serializing the password
on the other hand might require additional logic to handle the sensitive nature of the field or property.
Perhaps you want to protect the password when sending the value over the network, between a client and a server,
beyond TLS alone, and you only want to store the salted hash. When using the MappingPdxSerializer
, you can register
a custom PdxSerializer
to handle the user’s password, as follows:
Example 5. Registering custom PdxSerializers
by POJO field/property type
Map<?, PdxSerializer> customPdxSerializers = new HashMap<>();
customPdxSerializers.put(Password.class, new SaltedHashPasswordPdxSerializer());
mappingPdxSerializer.setCustomPdxSerializers(customPdxSerializers);
After registering the application-defined SaltedHashPasswordPdxSerializer
instance with the Password
application domain model type, the MappingPdxSerializer
will then consult the custom PdxSerializer
to serialize and deserialize all Password
objects regardless of the containing object (for example, User
).
However, suppose you want to customize the serialization of Passwords
only on User
objects.
To do so, you can register the custom PdxSerializer
for the User
type by specifying the fully qualified name
of the Class’s
field or property, as the following example shows:
Example 6. Registering custom PdxSerializers
by POJO field/property name
Map<?, PdxSerializer> customPdxSerializers = new HashMap<>();
customPdxSerializers.put("example.app.security.auth.model.User.password", new SaltedHashPasswordPdxSerializer());
mappingPdxSerializer.setCustomPdxSerializers(customPdxSerializers);
Notice the use of the fully-qualified field or property name (that is example.app.security.auth.model.User.password
)
as the custom PdxSerializer
registration key.
You could construct the registration key by using a more logical code snippet, such as the following:
User.class.getName().concat(".password");
. We recommended this over the example shown earlier.
The preceding example tried to be as explicit as possible about the semantics of registration.
9.4.2. Mapping ID Properties
Like Apache Geode’s ReflectionBasedAutoSerializer
, SDG’s MappingPdxSerializer
is also able to
determine the identifier of the entity. However, MappingPdxSerializer
does so by using Spring Data’s mapping metadata,
specifically by finding the entity property designated as the identifier using Spring Data’s
@Id
annotation.
Alternatively, any field or property named “id”, not explicitly annotated with @Id
, is also designated as
the entity’s identifier.
For example:
class Customer {
@Id
Long id;
...
}
In this case, the Customer
id
field is marked as the identifier field in the PDX type metadata by using
PdxWriter.markIdentifierField(:String)
when the PdxSerializer.toData(..)
method is called during serialization.
9.4.3. Mapping Read-only Properties
What happens when your entity defines a read-only property?
First, it is important to understand what a “read-only” property is. If you define a POJO by following the
JavaBeans specification (as Spring does),
you might define a POJO with a read-only property, as follows:
package example;
class ApplicationDomainType {
private AnotherType readOnly;
public AnotherType getReadOnly() [
this.readOnly;
}
...
}
The readOnly
property is read-only because it does not provide a setter method. It only has a getter method.
In this case, the readOnly
property (not to be confused with the readOnly
DomainType
field)
is considered read-only.
As a result, the MappingPdxSerializer
will not try to set a value for this property when populating an instance of
ApplicationDomainType
in the PdxSerializer.fromData(:Class<ApplicationDomainType>, :PdxReader)
method
during deserialization, particularly if a value is present in the PDX serialized bytes.
This is useful in situations where you might be returning a view or projection of some entity type and you only want
to set state that is writable. Perhaps the view or projection of the entity is based on authorization or some other
criteria. The point is, you can leverage this feature as is appropriate for your application’s use cases
and requirements. If you want the field or property to always be written, simply define a setter method.
9.4.4. Mapping Transient Properties
Likewise, what happens when your entity defines transient
properties?
You would expect the transient
fields or properties of your entity not to be serialized to PDX when serializing
the entity. That is exactly what happens, unlike Apache Geode’s own ReflectionBasedAutoSerializer
,
which serializes everything accessible from the object through Java Reflection.
The MappingPdxSerializer
will not serialize any fields or properties that are qualified as being transient, either
by using Java’s own transient
keyword (in the case of class instance fields) or by using the
@Transient
Spring Data annotation on either fields or properties.
For example, you might define an entity with transient fields and properties as follows:
package example;
class Process {
private transient int id;
private File workingDirectory;
private String name;
private Type type;
@Transient
public String getHostname() {
...
}
...
}
Neither the Process
id
field nor the readable hostname
property are written to PDX.
9.4.5. Filtering by Class Type
Similar to Apache Geode’s ReflectionBasedAutoSerializer
, SDG’s MappingPdxSerializer
lets you filter
the types of objects that are serialized and deserialized.
However, unlike Apache Geode’s ReflectionBasedAutoSerializer
, which uses complex regular expressions to express
which types the serializer handles, SDG’s MappingPdxSerializer
uses the much more robust
java.util.function.Predicate
interface
and API to express type-matching criteria.
If you like to use regular expressions, you can implement a Predicate
using Java’s
regular expression support.
The nice part about Java’s Predicate
interface is that you can compose Predicates
by using convenient
and appropriate API methods, including:
and(:Predicate)
,
or(:Predicate)
,
and negate()
.
The following example shows the Predicate
API in action:
Predicate<Class<?>> customerTypes =
type -> Customer.class.getPackage().getName().startsWith(type.getName()); // Include all types in the same package as `Customer`
Predicate includedTypes = customerTypes
.or(type -> User.class.isAssignble(type)); // Additionally, include User sub-types (e.g. Admin, Guest, etc)
mappingPdxSerializer.setIncludeTypeFilters(includedTypes);
mappingPdxSerializer.setExcludeTypeFilters(
type -> !Reference.class.getPackage(type.getPackage()); // Exclude Reference types
Any Class
object passed to your Predicate
is guaranteed not to be null
.
SDG’s MappingPdxSerializer
includes support for both include and exclude class type filters.
Exclude Type Filtering
By default, SDG’s MappingPdxSerializer
registers pre-defined Predicates
that filter, or exclude types
from the folliowing packages:
-
java.*
-
com.gemstone.gemfire.*
-
org.apache.geode.*
-
org.springframework.*
In addition, the MappingPdxSerializer
filters null
objects when calling PdxSerializer.toData(:Object, :PdxWriter)
and null
class types when calling PdxSerializer.fromData(:Class<?>, :PdxReader)
methods.
It is very easy to add exclusions for other class types, or an entire package of types, by simply defining a Predicate
and adding it to the MappingPdxSerializer
as shown earlier.
The MappingPdxSerializer.setExcludeTypeFilters(:Predicate<Class<?>>)
method is additive, meaning it composes
your application-defined type filters with the existing, pre-defined type filter Predicates
indicated above
using the Predicate.and(:Predicate<Class<?>>)
method.
However, what if you want to include a class type (for example, java.security Principal
) implicitly excluded by
the exclude type filters? See Include Type Filtering.
Include Type Filtering
If you want to include a class type explicitly, or override a class type filter that implicitly excludes a class type
required by your application (for example, java.security.Principal
, which is excluded by default with the java.*
package exclude type filter on MappingPdxSerializer
), then just define the appropriate Predicate
and add it to
the serializer using MappingPdxSerializer.setIncludeTypeFilters(:Predicate<Class<?>>)
method, as follows:
Predicate<Class<?>> principalTypeFilter =
type -> java.security.Principal.class.isAssignableFrom(type);
mappingPdxSerializer.setIncludeTypeFilters(principalTypeFilters);
Again, the MappingPdxSerializer.setIncludeTypeFilters(:Predicate<Class<?>>)
method,
like setExcludeTypeFilters(:Predicate<Class<?>>)
, is additive and therefore composes any passed type filter
using Predicate.or(:Predicate<Class<?>>)
. This means you may call setIncludeTypeFilters(:Predicate<Class<?>>)
as many time as necessary.
When include type filters are present, then the MappingPdxSerializer
makes a decision of whether to de/serialize
an instance of a class type when the class type is either not implicitly excluded OR when the class type
is explicitly included, whichever returns true. Then, an instance of the class type will be serialized
or deserialized appropriately.
For example, when a type filter of Predicate<Class<Principal>>
is explicitly registered as shown previously,
it cancels out the implicit exclude type filter on java.*
package types.
10. Spring Data for Apache Geode Repositories
Spring Data for Apache Geode provides support for using the Spring Data Repository abstraction to easily persist entities into
Apache Geode along with executing queries. A general introduction to the Repository programming model
is provided here.
10.1. Spring XML Configuration
To bootstrap Spring Data Repositories, use the <repositories/>
element from the Spring Data for Apache Geode Data namespace,
as the following example shows:
Example 7. Bootstrap Spring Data for Apache Geode Repositories in XML
<beans xmlns="http://www.springframework.org/schema/beans"
xmlns:gfe-data="https://www.springframework.org/schema/data/geode"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="
http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans.xsd
https://www.springframework.org/schema/data/geode https://www.springframework.org/schema/data/geode/spring-data-geode.xsd
">
<gfe-data:repositories base-package="com.example.acme.repository"/>
</beans>
The preceding configuration snippet looks for interfaces below the configured base package and creates Repository instances
for those interfaces backed by a SimpleGemFireRepository
.
The bootstrap process fails unless you have your application domain classes correctly mapped
to configured Regions.
10.2. Spring Java-based Configuration
Alternatively, many developers prefer to use Spring’s Java-based container configuration.
Using this approach, you can bootstrap Spring Data Repositories by using the SDG @EnableGemfireRepositories
annotation, as the following example shows:
Example 8. Bootstrap Spring Data for Apache Geode Repositories with @EnableGemfireRepositories
@SpringBootApplication
@EnableGemfireRepositories(basePackages = "com.example.acme.repository")
class SpringDataApplication {
...
}
Rather than use the basePackages
attribute, you may prefer to use the type-safe basePackageClasses
attribute instead.
The basePackageClasses
lets you specify the package that contains all your application Repository classes by
specifying only one of your application Repository interface types. Consider creating a special no-op marker class
or interface in each package that serves no purpose other than to identify the location of application Repositories
referenced by this attribute.
In addition to the basePackages and basePackageClasses
attributes, like Spring’s
@ComponentScan
annotation,
the @EnableGemfireRepositories
annotation provides include and exclude filters, based on Spring’s
ComponentScan.Filter
type.
You can use the filterType
attribute to filter by different aspects, such as whether an application Repository type
is annotated with a particular annotation or extends a particular class type and so on. See the
FilterType
Javadoc
for more details.
The @EnableGemfireRepositories
annotation also lets you specify the location of named OQL queries, which reside in
a Java Properties
file, by using the namedQueriesLocation
attribute. The property name must match the name
of a Repository query method and the property value is the OQL query you want executed when the Repository query method
is called.
The repositoryImplementationPostfix
attribute can be set to an alternate value (defaults to Impl
) if your
application requires one or more custom repository implementations.
This feature is commonly used to extend the Spring Data Repository infrastructure to implement a feature not provided by
the data store (for example, SDG).
One example of where custom repository implementations are needed with Apache Geode is when performing joins.
Joins are not supported by SDG Repositories. With a Apache Geode PARTITION
Region, the join must be
performed on collocated PARTITION
Regions, since Apache Geode does not support “distributed” joins.
In addition, the Equi-Join OQL Query must be performed inside a Apache Geode Function.
See here
for more details on Apache Geode Equi-Join Queries.
Many other aspects of the SDG’s Repository infrastructure extension may be customized as well. See the
@EnableGemfireRepositories
Javadoc for more details on all configuration settings.
10.3. Executing OQL Queries
Spring Data for Apache Geode Repositories enable the definition of query methods to easily execute Apache Geode OQL queries
against the Region the managed entity maps to, as the following example shows:
Example 9. Sample Repository
@Region("People")
public class Person { … }
public interface PersonRepository extends CrudRepository<Person, Long> {
Person findByEmailAddress(String emailAddress);
Collection<Person> findByFirstname(String firstname);
@Query("SELECT * FROM /People p WHERE p.firstname = $1")
Collection<Person> findByFirstnameAnnotated(String firstname);
@Query("SELECT * FROM /People p WHERE p.firstname IN SET $1")
Collection<Person> findByFirstnamesAnnotated(Collection<String> firstnames);
}
The first query method listed in the preceding example causes the following OQL query to be derived:
SELECT x FROM /People x WHERE x.emailAddress = $1
. The second query method works the same way except
it returns all entities found, whereas the first query method expects a single result to be found.
If the supported keywords are not sufficient to declare and express your OQL query, or the method name becomes too
verbose, then you can annotate the query methods with @Query
as shown on the third and fourth methods.
The following table gives brief samples of the supported keywords that you can use in query methods:
Table 4. Supported keywords for query methods
Keyword
Sample
Logical result
GreaterThan
findByAgeGreaterThan(int age)
x.age > $1
GreaterThanEqual
findByAgeGreaterThanEqual(int age)
x.age >= $1
LessThan
findByAgeLessThan(int age)
x.age < $1
LessThanEqual
findByAgeLessThanEqual(int age)
x.age ⇐ $1
IsNotNull
, NotNull
findByFirstnameNotNull()
x.firstname =! NULL
IsNull
, Null
findByFirstnameNull()
x.firstname = NULL
In
findByFirstnameIn(Collection<String> x)
x.firstname IN SET $1
NotIn
findByFirstnameNotIn(Collection<String> x)
x.firstname NOT IN SET $1
IgnoreCase
findByFirstnameIgnoreCase(String firstName)
x.firstname.equalsIgnoreCase($1)
(No keyword)
findByFirstname(String name)
x.firstname = $1
Like
findByFirstnameLike(String name)
x.firstname LIKE $1
Not
findByFirstnameNot(String name)
x.firstname != $1
IsTrue
, True
findByActiveIsTrue()
x.active = true
IsFalse
, False
findByActiveIsFalse()
x.active = false
10.4. OQL Query Extensions Using Annotations
Many query languages, such as Apache Geode’s OQL (Object Query Language), have extensions that are not directly
supported by Spring Data Commons' Repository infrastructure.
One of Spring Data Commons' Repository infrastructure goals is to function as the lowest common denominator to maintain
support for and portability across the widest array of data stores available and in use for application development
today. Technically, this means developers can access multiple different data stores supported by Spring Data Commons
within their applications by reusing their existing application-specific Repository interfaces — a convenient
and powerful abstraction.
To support Apache Geode’s OQL Query language extensions and preserve portability across different data stores,
Spring Data for Apache Geode adds support for OQL Query extensions by using Java annotations. These annotations are ignored by other
Spring Data Repository implementations (such as Spring Data JPA or Spring Data Redis) that do not have similar
query language features.
For instance, many data stores most likely do not implement Apache Geode’s OQL IMPORT
keyword. Implementing IMPORT
as an annotation (that is, @Import
) rather than as part of the query method signature (specifically, the method 'name')
does not interfere with the parsing infrastructure when evaluating the query method name to construct another data store
language appropriate query.
Currently, the set of Apache Geode OQL Query language extensions that are supported by Spring Data for Apache Geode include:
Table 5. Supported Apache Geode OQL extensions for Repository query methods
Keyword
Annotation
Description
Arguments
@Hint
OQL query index hints
String[]
(e.g. @Hint({ "IdIdx", "TxDateIdx" }))
@Import
Qualify application-specific types.
String
(e.g. @Import("org.example.app.domain.Type"))
@Limit
Limit the returned query result set.
Integer
(e.g. @Limit(10); default is Integer.MAX_VALUE)
@Trace
Enable OQL query-specific debugging.
NA
As an example, suppose you have a Customers
application domain class and corresponding Apache Geode Region
along with a CustomerRepository
and a query method to lookup Customers
by last name, as follows:
Example 10. Sample Customers Repository
package ...;
...
@Region("Customers")
public class Customer ... {
@Id
private Long id;
...
}
package ...;
public interface CustomerRepository extends GemfireRepository<Customer, Long> {
@Trace
@Limit(10)
@Hint("LastNameIdx")
@Import("org.example.app.domain.Customer")
List<Customer> findByLastName(String lastName);
...
}
The preceding example results in the following OQL Query:
<TRACE> <HINT 'LastNameIdx'> IMPORT org.example.app.domain.Customer; SELECT * FROM /Customers x WHERE x.lastName = $1 LIMIT 10
Spring Data for Apache Geode’s Repository extension is careful not to create conflicting declarations when the OQL annotation extensions
are used in combination with the @Query
annotation.
As another example, suppose you have a raw @Query
annotated query method defined in your CustomerRepository
,
as follows:
Example 11. CustomerRepository
public interface CustomerRepository extends GemfireRepository<Customer, Long> {
@Trace
@Limit(10)
@Hint("CustomerIdx")
@Import("org.example.app.domain.Customer")
@Query("<TRACE> <HINT 'ReputationIdx'> SELECT DISTINCT * FROM /Customers c WHERE c.reputation > $1 ORDER BY c.reputation DESC LIMIT 5")
List<Customer> findDistinctCustomersByReputationGreaterThanOrderByReputationDesc(Integer reputation);
}
The preceding query method results in the following OQL query:
IMPORT org.example.app.domain.Customer; <TRACE> <HINT 'ReputationIdx'> SELECT DISTINCT * FROM /Customers x
WHERE x.reputation > $1 ORDER BY c.reputation DESC LIMIT 5
The @Limit(10)
annotation does not override the LIMIT
explicitly defined in the raw query.
Also, the @Hint("CustomerIdx")
annotation does not override the HINT
explicitly defined in the raw query.
Finally, the @Trace
annotation is redundant and has no additional effect.
The ReputationIdx
index is probably not the most sensible index, given the number of customers who may possibly have
the same value for their reputation, which reduces the effectiveness of the index. Please choose indexes and other
optimizations wisely, as an improper or poorly chosen index can have the opposite effect on your performance because
of the overhead in maintaining the index. The ReputationIdx
was used only to serve the purpose of the example.
10.5. Query Post Processing
Thanks to using the Spring Data Repository abstraction, the query method convention for defining data store specific
queries (e.g. OQL) is easy and convenient. However, it is sometimes desirable to still want to inspect or even possibly
modify the query generated from the Repository query method.
Since 2.0.x, Spring Data for Apache Geode includes the o.s.d.gemfire.repository.query.QueryPostProcessor
functional interface.
The interface is loosely defined as follows:
Example 12. QueryPostProcessor
package org.springframework.data.gemfire.repository.query;
@FunctionalInterface
interface QueryPostProcessor<T extends Repository, QUERY> extends Ordered {
QUERY postProcess(QueryMethod queryMethod, QUERY query, Object... arguments);
}
There are additional default methods provided that let you compose instances of QueryPostProcessor
similar to how
java.util.function.Function.andThen(:Function)
and java.util.function.Function.compose(:Function)
work.
Additionally, the QueryPostProcessor
interface implements the
org.springframework.core.Ordered
interface,
which is useful when multiple QueryPostProcessors
are declared and registered in the Spring container and used to
create a pipeline of processing for a group of generated query method queries.
Finally, the QueryPostProcessor
accepts type arguments corresponding to the type parameters, T
and QUERY
,
respectively. Type T
extends the Spring Data Commons marker interface,
org.springframework.data.repository.Repository
.
We discuss this further later in this section. All QUERY
type parameter arguments in Spring Data for Apache Geode’s case are of type
java.lang.String
.
It is useful to define the query as type QUERY
, since this QueryPostProcessor
interface may be ported to
Spring Data Commons and therefore must handle all forms of queries by different data stores (such as JPA, MongoDB,
or Redis).
You can implement this interface to receive a callback with the query that was generated from the application
Repository
interface method when the method is called.
For example, you might want to log all queries from all application Repository interface definitions. You could do so
by using the following QueryPostProcessor
implementation:
Example 13. LoggingQueryPostProcessor
package example;
class LoggingQueryPostProcessor implements QueryPostProcessor<Repository, String> {
private Logger logger = Logger.getLogger("someLoggerName");
@Override
public String postProcess(QueryMethod queryMethod, String query, Object... arguments) {
String message = String.format("Executing query [%s] with arguments [%s]", query, Arrays.toString(arguments));
this.logger.info(message);
}
}
The LoggingQueryPostProcessor
was typed to the Spring Data org.springframework.data.repository.Repository
marker interface, and, therefore, logs all application Repository interface query method generated queries.
You could limit the scope of this logging to queries only from certain types of application Repository interfaces,
such as, say, a CustomerRepository
, as the following example shows:
Example 14. CustomerRepository
interface CustomerRepository extends CrudRepository<Customer, Long> {
Customer findByAccountNumber(String accountNumber);
List<Customer> findByLastNameLike(String lastName);
}
Then you could have typed the LoggingQueryPostProcessor
specifically to the CustomerRepository
, as follows:
Example 15. CustomerLoggingQueryPostProcessor
class LoggingQueryPostProcessor implements QueryPostProcessor<CustomerRepository, String> { .. }
As a result, only queries defined in the CustomerRepository
interface, such as findByAccountNumber
, are logged.
You might want to create a QueryPostProcessor
for a specific query defined by a Repository query method. For example,
suppose you want to limit the OQL query generated from the CustomerRepository.findByLastNameLike(:String)
query method
to only return five results along with ordering the Customers
by firstName
, in ascending order . To do so,
you can define a custom QueryPostProcessor
, as the following example shows:
Example 16. OrderedLimitedCustomerByLastNameQueryPostProcessor
class OrderedLimitedCustomerByLastNameQueryPostProcessor implements QueryPostProcessor<CustomerRepository, String> {
private final int limit;
public OrderedLimitedCustomerByLastNameQueryPostProcessor(int limit) {
this.limit = limit;
}
@Override
public String postProcess(QueryMethod queryMethod, String query, Object... arguments) {
return "findByLastNameLike".equals(queryMethod.getName())
? query.trim()
.replace("SELECT", "SELECT DISTINCT")
.concat(" ORDER BY firstName ASC")
.concat(String.format(" LIMIT %d", this.limit))
: query;
}
}
While the preceding example works, you can achieve the same effect by using the Spring Data Repository convention
provided by Spring Data for Apache Geode. For instance, the same query could be defined as follows:
Example 17. CustomerRepository using the convention
interface CustomerRepository extends CrudRepository<Customer, Long> {
@Limit(5)
List<Customer> findDistinctByLastNameLikeOrderByFirstNameDesc(String lastName);
}
However, if you do not have control over the application CustomerRepository
interface definition,
then the QueryPostProcessor
(that is, OrderedLimitedCustomerByLastNameQueryPostProcessor
) is convenient.
If you want to ensure that the LoggingQueryPostProcessor
always comes after the other application-defined
QueryPostProcessors
that may have bean declared and registered in the Spring ApplicationContext
, you can set
the order
property by overriding the o.s.core.Ordered.getOrder()
method, as the following example shows:
Example 18. Defining the order
property
class LoggingQueryPostProcessor implements QueryPostProcessor<Repository, String> {
@Override
int getOrder() {
return 1;
}
}
class CustomerQueryPostProcessor implements QueryPostProcessor<CustomerRepository, String> {
@Override
int getOrder() {
return 0;
}
}
This ensures that you always see the effects of the post processing applied by other QueryPostProcessors
before the LoggingQueryPostProcessor
logs the query.
You can define as many QueryPostProcessors
in the Spring ApplicationContext
as you like and apply them in any order,
to all or specific application Repository interfaces, and be as granular as you like by using the provided arguments
to the postProcess(..)
method callback.
11. Annotation Support for Function Execution
Spring Data for Apache Geode includes annotation support to simplify working with Apache Geode
Function execution.
Under the hood, the Apache Geode API provides classes to implement and register Apache Geode
Functions that are deployed on Apache Geode
servers, which may then be invoked by other peer member applications or remotely from cache clients.
Functions can execute in parallel, distributed among multiple Apache Geode servers in the cluster, aggregating the
results using the map-reduce pattern and sent back to the caller. Functions can also be targeted to run on a single
server or Region. The Apache Geode API supports remote execution of Functions targeted by using various predefined
scopes: on Region, on members (in groups), on servers, and others. The implementation and execution of remote Functions,
as with any RPC protocol, requires some boilerplate code.
Spring Data for Apache Geode, true to Spring’s core value proposition, aims to hide the mechanics of remote Function execution and let you
focus on core POJO programming and business logic. To this end, Spring Data for Apache Geode introduces annotations to declaratively
register the public methods of a POJO class as Apache Geode Functions along with the ability to invoke registered
Functions (including remotely) by using annotated interfaces.
11.1. Implementation Versus Execution
There are two separate concerns to address: implementation and execution.
The first is Function implementation (server-side), which must interact with the
FunctionContext
to access the invocation arguments,
ResultsSender
to send results,
and other execution context information. The Function implementation typically accesses the cache and Regions
and is registered with the
FunctionService
under a unique ID.
A cache client application invoking a Function does not depend on the implementation. To invoke a Function,
the application instantiates an
Execution
providing the Function ID, invocation arguments, and the Function target, which defines its scope:
Region, server, servers, member, or members. If the Function produces a result, the invoker uses a
ResultCollector
to aggregate and acquire the execution results. In certain cases, a custom ResultCollector
implementation
is required and may be registered with the Execution
.
'Client' and 'Server' are used here in the context of Function execution, which may have a different meaning
than client and server in Apache Geode’s client-server topology. While it is common for an application using
a ClientCache
instance to invoke a Function on one or more Apache Geode servers in a cluster, it is also
possible to execute Functions in a peer-to-peer (P2P) configuration, where the application is a member of the cluster
hosting a peer Cache
instance. Keep in mind that a peer member cache application is subject to all the constraints
of being a peer member of the cluster.
11.2. Implementing a Function
Using Apache Geode APIs, the FunctionContext
provides a runtime invocation context that includes the client’s
calling arguments and a ResultSender
implementation to send results back to the client. Additionally, if the Function
is executed on a Region, the FunctionContext
is actually an instance of RegionFunctionContext
, which provides
additional information, such as the target Region on which the Function was invoked, any filter (a set of specific keys)
associated with the Execution
, and so on. If the Region is a PARTITION
Region, the Function should use
the PartitionRegionHelper
to extract the local data set.
By using Spring, you can write a simple POJO and use the Spring container to bind one or more of your POJO’s
public methods to a Function. The signature for a POJO method intended to be used as a Function must generally conform
to the client’s execution arguments. However, in the case of a Region execution, the Region data may also be provided
(presumably the data is held in the local partition if the Region is a PARTITION
Region).
Additionally, the Function may require the filter that was applied, if any. This suggests that the client and server
share a contract for the calling arguments but that the method signature may include additional parameters to pass values
provided by the FunctionContext
. One possibility is for the client and server to share a common interface, but this
is not strictly required. The only constraint is that the method signature includes the same sequence of calling arguments
with which the Function was invoked after the additional parameters are resolved.
For example, suppose the client provides a String
and an int
as the calling arguments. These are provided
in the FunctionContext
as an array, as the following example shows:
Object[] args = new Object[] { "test", 123 };
The Spring container should be able to bind to any method signature similar to the following (ignoring the return type
for the moment):
public Object method1(String s1, int i2) { ... }
public Object method2(Map<?, ?> data, String s1, int i2) { ... }
public Object method3(String s1, Map<?, ?> data, int i2) { ... }
public Object method4(String s1, Map<?, ?> data, Set<?> filter, int i2) { ... }
public void method4(String s1, Set<?> filter, int i2, Region<?,?> data) { ... }
public void method5(String s1, ResultSender rs, int i2) { ... }
public void method6(FunctionContest context) { ... }
The general rule is that once any additional arguments (that is, Region data and filter) are resolved,
the remaining arguments must correspond exactly, in order and type, to the expected Function method parameters.
The method’s return type must be void or a type that may be serialized (as a java.io.Serializable
, DataSerializable
,
or PdxSerializable
). The latter is also a requirement for the calling arguments.
The Region data should normally be defined as a Map
, to facilitate unit testing, but may also be of type Region,
if necessary. As shown in the preceding example, it is also valid to pass the FunctionContext
itself
or the ResultSender
if you need to control over how the results are returned to the client.
11.2.1. Annotations for Function Implementation
The following example shows how SDG’s Function annotations are used to expose POJO methods
as Apache Geode Functions:
@Component
public class ApplicationFunctions {
@GemfireFunction
public String function1(String value, @RegionData Map<?, ?> data, int i2) { ... }
@GemfireFunction(id = "myFunction", batchSize=100, HA=true, optimizedForWrite=true)
public List<String> function2(String value, @RegionData Map<?, ?> data, int i2, @Filter Set<?> keys) { ... }
@GemfireFunction(hasResult=true)
public void functionWithContext(FunctionContext functionContext) { ... }
}
Note that the class itself must be registered as a Spring bean and each Apache Geode Function is annotated with
@GemfireFunction
. In the preceding example, Spring’s @Component
annotation was used, but you can register the bean
by using any method supported by Spring (such as XML configuration or with a Java configuration class when using
Spring Boot). This lets the Spring container create an instance of this class and wrap it in a
PojoFunctionWrapper
.
Spring creates a wrapper instance for each method annotated with @GemfireFunction
. Each wrapper instance shares
the same target object instance to invoke the corresponding method.
The fact that the POJO Function class is a Spring bean may offer other benefits. Since it shares
the ApplicationContext
with Apache Geode components, such as the cache and Regions, these may be injected into
the class if necessary.
Spring creates the wrapper class and registers the Functions with Apache Geode’s FunctionService
. The Function ID
used to register each Function must be unique. By using convention, it defaults to the simple (unqualified) method name.
The name can be explicitly defined by using the id
attribute of the @GemfireFunction
annotation.
The @GemfireFunction
annotation also provides other configuration attributes: HA
and optimizedForWrite
,
which correspond to properties defined by Apache Geode’s
Function
interface.
If the POJO Function method’s return type is void
, then the hasResult
attribute is automatically set to false
.
Otherwise, if the method returns a value, the hasResult
attributes is set to true
. Even for void
method return
types, the GemfireFunction
annotation’s hasResult
attribute can be set to true
to override this convention,
as shown in the functionWithContext
method shown previously. Presumably, the intention is that you will use
the ResultSender
directly to send results to the caller.
Finally, the GemfireFunction
annotation supports the requiredPermissions
attribute, which specifies the permissions
required to execute the Function. By default, all Functions require the DATA:WRITE
permission. The attribute
accepts an array of Strings allowing you to modify the permissions as required by your application and/or Function UC.
Each resource permission is expected to be in the following format: <RESOURCE>:<OPERATION>:[Target]:[Key]
.
RESOURCE
can be 1 of the {data-store-javadoc]/org/apache/geode/security/ResourcePermission.Resource.html[ResourcePermission.Resource
]
enumerated values. OPERATION
can be 1 of the {data-store-javadoc}/org/apache/geode/security/ResourcePermission.Operation.html[ResourcePermission.Operation
]
enumerated values. Optionally, Target
can be the name of a Region or 1 of the
{data-store-javadoc}/org/apache/geode/security/ResourcePermission.Target.html[ResourcePermission.Target
]
enumerated values. And finally, optionally, Key
is a valid Key in the Target
Region if specified.
The PojoFunctionWrapper
implements Apache Geode’s Function
interface, binds method parameters, and invokes
the target method in its execute()
method. It also sends the method’s return value back to the caller
by using the ResultSender
.
11.2.2. Batching Results
If the return type is an array or Collection
, then some consideration must be given to how the results are returned.
By default, the PojoFunctionWrapper
returns the entire array or Collection
at once. If the number of elements
in the array or Collection
is quite large, it may incur a performance penalty. To divide the payload into smaller,
more manageable chunks, you can set the batchSize
attribute, as illustrated in function2
, shown earlier.
If you need more control of the ResultSender
, especially if the method itself would use too much memory
to create the Collection
, you can pass in the ResultSender
or access it through the FunctionContext
and use it directly within the method to sends results back to the caller.
11.2.3. Enabling Annotation Processing
In accordance with Spring standards, you must explicitly activate annotation processing for @GemfireFunction
annotations. The following example activates annotation processing with XML:
<gfe:annotation-driven/>
The following example activates annotation processing by annotating a Java configuration class:
@Configuration
@EnableGemfireFunctions
class ApplicationConfiguration { ... }
11.3. Executing a Function
A process that invokes a remote Function needs to provide the Function’s ID, calling arguments, the execution target
(onRegion
, onServers
, onServer
, onMember
, or onMembers
) and (optionally) a filter set. By using Spring Data for Apache Geode,
all you need do is define an interface supported by annotations. Spring creates a dynamic proxy for the interface,
which uses the FunctionService
to create an Execution
, invoke the Execution
, and (if necessary) coerce
the results to the defined return type. This technique is similar to the way Spring Data for Apache Geode’s Repository extension works.
Thus, some of the configuration and concepts should be familiar.
Generally, a single interface definition maps to multiple Function executions, one corresponding to each method
defined in the interface.
11.3.1. Annotations for Function Execution
To support client-side Function execution, the following SDG Function annotations are provided: @OnRegion
,
@OnServer
, @OnServers
, @OnMember
, and @OnMembers
. These annotations correspond to the Execution
implementations provided by Apache Geode’s
FunctionService
class.
Each annotation exposes the appropriate attributes. These annotations also provide an optional resultCollector
attribute
whose value is the name of a Spring bean implementing the
ResultCollector
interface
to use for the execution.
The proxy interface binds all declared methods to the same execution configuration. Although it is expected
that single method interfaces are common, all methods in the interface are backed by the same proxy instance
and therefore all share the same configuration.
The following listing shows a few examples:
@OnRegion(region="SomeRegion", resultCollector="myCollector")
public interface FunctionExecution {
@FunctionId("function1")
String doIt(String s1, int i2);
String getString(Object arg1, @Filter Set<Object> keys);
}
By default, the Function ID is the simple (unqualified) method name. The @FunctionId
annotation can be used
to bind this invocation to a different Function ID.
11.3.2. Enabling Annotation Processing
The client-side uses Spring’s classpath component scanning capability to discover annotated interfaces. To enable
Function execution annotation processing in XML, insert the following element in your XML configuration:
<gfe-data:function-executions base-package="org.example.myapp.gemfire.functions"/>
The function-executions
element is provided in the gfe-data
XML namespace. The base-package
attribute is required
to avoid scanning the entire classpath. Additional filters can be provided as described in the Spring
reference documentation.
Optionally, you can annotate your Java configuration class as follows:
@EnableGemfireFunctionExecutions(basePackages = "org.example.myapp.gemfire.functions")
11.4. Programmatic Function Execution
Using the Function execution annotated interface defined in the previous section, simply auto-wire your interface
into an application bean that will invoke the Function:
@Component
public class MyApplication {
@Autowired
FunctionExecution functionExecution;
public void doSomething() {
functionExecution.doIt("hello", 123);
}
}
Alternately, you can use a Function execution template directly. In the following example,
the GemfireOnRegionFunctionTemplate
creates an onRegion
Function Execution
:
Example 19. Using the GemfireOnRegionFunctionTemplate
Set<?, ?> myFilter = getFilter();
Region<?, ?> myRegion = getRegion();
GemfireOnRegionOperations template = new GemfireOnRegionFunctionTemplate(myRegion);
String result = template.executeAndExtract("someFunction", myFilter, "hello", "world", 1234);
Internally, Function Executions
always return a List
. executeAndExtract
assumes a singleton List
containing the result and attempts to coerce that value into the requested type. There is also an execute
method
that returns the List
as is. The first parameter is the Function ID. The filter argument is optional. The remaining
arguments are a variable argument List
.
11.5. Function Execution with PDX
When using Spring Data for Apache Geode’s Function annotation support combined with Apache Geode’s
PDX Serialization,
there are a few logistical things to keep in mind.
As explained earlier in this section, and by way of example, you should typically define Apache Geode Functions
by using POJO classes annotated with Spring Data for Apache Geode
Function annotations,
as follows:
public class OrderFunctions {
@GemfireFunction(...)
Order process(@RegionData data, Order order, OrderSource orderSourceEnum, Integer count) { ... }
}
The Integer
typed count
parameter is arbitrary, as is the separation of the Order
class
and the OrderSource
enum, which might be logical to combine. However, the arguments were setup this way
to demonstrate the problem with Function executions in the context of PDX.
Your Order
class and OrderSource
enum might be defined as follows:
public class Order ... {
private Long orderNumber;
private LocalDateTime orderDateTime;
private Customer customer;
private List<Item> items
...
}
public enum OrderSource {
ONLINE,
PHONE,
POINT_OF_SALE
...
}
Of course, you can define a Function Execution
interface to call the 'process' Apache Geode server Function,
as follows:
@OnServer
public interface OrderProcessingFunctions {
Order process(Order order, OrderSource orderSourceEnum, Integer count);
}
Clearly, this process(..)
Order
Function is being called from the client-side with a ClientCache
instance
(that is <gfe:client-cache/>
). This implies that the Function arguments must also be serializable. The same is true
when invoking peer-to-peer member Functions (such as @OnMember(s)
) between peers in the cluster. Any form of
distribution
requires the data transmitted between client and server (or peers) to be serialized.
Now, if you have configured Apache Geode to use PDX for serialization (instead of Java serialization, for instance)
you can also set the pdx-read-serialized
attribute to true
in your configuration of the Apache Geode server(s),
as follows:
<gfe:cache pdx-read-serialized="true"/>
Alternatively, you can set the pdx-read-serialized
attribute to true
for a Apache Geode cache client application,
as follows:
<gfe:client-cache pdx-read-serialized="true"/>
Doing so causes all values read from the cache (that is, Regions) as well as information passed between client and servers
(or peers) to remain in serialized form, including, but not limited to, Function arguments.
Apache Geode serializes only application domain object types that you have specifically configured (registered)
either by using Apache Geode’s
ReflectionBasedAutoSerializer
,
or specifically (and recommended) by using a “custom” Apache Geode
PdxSerializer
.
If you use Spring Data for Apache Geode’s Repository extension, you might even want to consider using Spring Data for Apache Geode’s
MappingPdxSerializer
,
which uses an entity’s mapping metadata to determine data from the application domain object that is serialized
to the PDX instance.
What is less than apparent, though, is that Apache Geode automatically handles Java Enum
types regardless
of whether they are explicitly configured (that is, registered with a ReflectionBasedAutoSerializer
,
using a regex pattern and the classes
parameter or are handled by a “custom” Apache Geode PdxSerializer
),
despite the fact that Java enumerations implement java.io.Serializable
.
So, when you set pdx-read-serialized
to true
on Apache Geode servers where the Apache Geode Functions
(including Spring Data for Apache Geode Function-annotated POJO classes) are registered, then you may encounter surprising behavior
when invoking the Function Execution
.
You might pass the following arguments when invoking the Function:
orderProcessingFunctions.process(new Order(123, customer, LocalDateTime.now(), items), OrderSource.ONLINE, 400);
However, the Apache Geode Function on the server gets the following:
process(regionData, order:PdxInstance, :PdxInstanceEnum, 400);
The Order
and OrderSource
have been passed to the Function as
PDX instances.
Again, this all happens because pdx-read-serialized
is set to true
, which may be necessary in cases where
the Apache Geode servers interact with multiple different clients (for example, a combination of Java clients
and native clients, such as C/C++, C#, and others).
This flies in the face of Spring Data for Apache Geode’s strongly-typed Function-annotated POJO class method signatures, where you would
reasonably expect application domain object types instead, not PDX serialized instances.
Consequently, Spring Data for Apache Geode includes enhanced Function support to automatically convert PDX typed method arguments
to the desired application domain object types defined by the Function method’s signature (parameter types).
However, this also requires you to explicitly register a Apache Geode PdxSerializer
on Apache Geode servers
where Spring Data for Apache Geode Function-annotated POJOs are registered and used, as the following example shows:
<bean id="customPdxSerializer" class="x.y.z.gemfire.serialization.pdx.MyCustomPdxSerializer"/>
<gfe:cache pdx-serializer-ref="customPdxSerializeer" pdx-read-serialized="true"/>
Alternatively, you can use Apache Geode’s
ReflectionBasedAutoSerializer
for convenience. Of course, we recommend that, where possible, you use a custom PdxSerializer
to maintain
finer-grained control over your serialization strategy.
Finally, Spring Data for Apache Geode is careful not to convert your Function arguments if you treat your Function arguments generically
or as one of Apache Geode’s PDX types, as follows:
@GemfireFunction
public Object genericFunction(String value, Object domainObject, PdxInstanceEnum pdxEnum) {
// ...
}
Spring Data for Apache Geode converts PDX typed data to the corresponding application domain types if and only if the corresponding
application domain types are on the classpath and the Function-annotated POJO method expects it.
For a good example of custom, composed application-specific Apache Geode PdxSerializers
as well as appropriate
POJO Function parameter type handling based on the method signatures, see Spring Data for Apache Geode’s
ClientCacheFunctionExecutionWithPdxIntegrationTest
class.
12. Apache Lucene Integration
Apache Geode integrates with Apache Lucene to let you
index and search on data stored in Apache Geode by using Lucene queries. Search-based queries also include
the ability to page through query results.
Additionally, Spring Data for Apache Geode adds support for query projections based on the Spring Data Commons projection infrastructure.
This feature lets the query results be projected into first-class application domain types as needed by the application.
A Lucene Index
must be created before any Lucene search-based query can be run. A LuceneIndex
can be created in Spring (Data for Apache Geode) XML config as follows:
<gfe:lucene-index id="IndexOne" fields="fieldOne, fieldTwo" region-path="/Example"/>
Additionally, Apache Lucene allows the specification of
analyzers
per field and can be configured as shown in the following example:
<gfe:lucene-index id="IndexTwo" lucene-service-ref="luceneService" region-path="/AnotherExample">
<gfe:field-analyzers>
<map>
<entry key="fieldOne">
<bean class="example.AnalyzerOne"/>
</entry>
<entry key="fieldTwo">
<bean class="example.AnalyzerTwo"/>
</entry>
</map>
</gfe:field-analyzers>
</gfe:lucene-index>
The Map
can be specified as a top-level bean definition and referenced by using the ref
attribute
in the nested <gfe:field-analyzers>
element, as follows:
<gfe-field-analyzers ref="refToTopLevelMapBeanDefinition"/>
.
Spring Data for Apache Geode’s LuceneIndexFactoryBean
API and SDG’s XML namespace also lets a
org.apache.geode.cache.lucene.LuceneSerializer
be specified when you create the LuceneIndex
. The LuceneSerializer
lets you configure the way objects are converted
to Lucene documents for the index when the object is indexed.
The following example shows how to add an LuceneSerializer
to the LuceneIndex
:
<bean id="MyLuceneSerializer" class="example.CustomLuceneSerializer"/>
<gfe:lucene-index id="IndexThree" lucene-service-ref="luceneService" region-path="/YetAnotherExample">
<gfe:lucene-serializer ref="MyLuceneSerializer">
</gfe:lucene-index>
You can specify the LuceneSerializer
as an anonymous, nested bean definition as well, as follows:
<gfe:lucene-index id="IndexThree" lucene-service-ref="luceneService" region-path="/YetAnotherExample">
<gfe:lucene-serializer>
<bean class="example.CustomLuceneSerializer"/>
</gfe:lucene-serializer>
</gfe:lucene-index>
Alternatively, you can declare or define a LuceneIndex
in Spring Java config, inside a @Configuration
class,
as the following example shows:
@Bean(name = "Books")
@DependsOn("bookTitleIndex")
PartitionedRegionFactoryBean<Long, Book> booksRegion(GemFireCache gemfireCache) {
PartitionedRegionFactoryBean<Long, Book> peopleRegion =
new PartitionedRegionFactoryBean<>();
peopleRegion.setCache(gemfireCache);
peopleRegion.setClose(false);
peopleRegion.setPersistent(false);
return peopleRegion;
}
@Bean
LuceneIndexFactoryBean bookTitleIndex(GemFireCache gemFireCache,
LuceneSerializer luceneSerializer) {
LuceneIndexFactoryBean luceneIndex = new LuceneIndexFactoryBean();
luceneIndex.setCache(gemFireCache);
luceneIndex.setFields("title");
luceneIndex.setLuceneSerializer(luceneSerializer);
luceneIndex.setRegionPath("/Books");
return luceneIndex;
}
@Bean
CustomLuceneSerializer myLuceneSerialier() {
return new CustomeLuceneSerializer();
}
There are a few limitations of Apache Geode’s, Apache Lucene integration and support.
First, a LuceneIndex
can only be created on a Apache Geode PARTITION
Region.
Second, all LuceneIndexes
must be created before the Region to which the LuceneIndex
applies.
To help ensure that all declared LuceneIndexes
defined in a Spring container are created before the Regions
on which they apply, SDG includes the org.springframework.data.gemfire.config.support.LuceneIndexRegionBeanFactoryPostProcessor
.
You may register this Spring BeanFactoryPostProcessor
in XML config by using <bean class="org.springframework.data.gemfire.config.support.LuceneIndexRegionBeanFactoryPostProcessor"/>
.
The o.s.d.g.config.support.LuceneIndexRegionBeanFactoryPostProcessor
may only be used when using SDG XML config.
More details about Spring’s BeanFactoryPostProcessors
can be found here.
It is possible that these Apache Geode restrictions will not apply in a future release which is why
the SDG LuceneIndexFactoryBean
API takes a reference to the Region directly as well,
rather than just the Region path.
This is more ideal when you want to define a LuceneIndex
on an existing Region with data at a later point
during the application’s lifecycle and as requirements demand. Where possible, SDG strives to adhere to
strongly-typed objects. However, for the time being, you must use the regionPath
property to specify the Region
to which the LuceneIndex
is applied.
Additionally, in the preceding example, note the presence of Spring’s @DependsOn
annotation
on the Books
Region bean definition. This creates a dependency from the Books
Region bean to the bookTitleIndex
LuceneIndex
bean definition, ensuring that the LuceneIndex
is created before the Region on which it applies.
Now that once we have a LuceneIndex
, we can perform Lucene-based data access operations, such as queries.
12.1. Lucene Template Data Accessors
Spring Data for Apache Geode provides two primary templates for Lucene data access operations, depending on how low of a level
your application is prepared to deal with.
The LuceneOperations
interface defines query operations by using Apache Geode
Lucene types,
which are defined in the following interface definition:
public interface LuceneOperations {
<K, V> List<LuceneResultStruct<K, V>> query(String query, String defaultField [, int resultLimit]
, String... projectionFields);
<K, V> PageableLuceneQueryResults<K, V> query(String query, String defaultField,
int resultLimit, int pageSize, String... projectionFields);
<K, V> List<LuceneResultStruct<K, V>> query(LuceneQueryProvider queryProvider [, int resultLimit]
, String... projectionFields);
<K, V> PageableLuceneQueryResults<K, V> query(LuceneQueryProvider queryProvider,
int resultLimit, int pageSize, String... projectionFields);
<K> Collection<K> queryForKeys(String query, String defaultField [, int resultLimit]);
<K> Collection<K> queryForKeys(LuceneQueryProvider queryProvider [, int resultLimit]);
<V> Collection<V> queryForValues(String query, String defaultField [, int resultLimit]);
<V> Collection<V> queryForValues(LuceneQueryProvider queryProvider [, int resultLimit]);
}
The [, int resultLimit]
indicates that the resultLimit
parameter is optional.
The operations in the LuceneOperations
interface match the operations provided by the Apache Geode’s
LuceneQuery interface.
However, SDG has the added value of translating proprietary Apache Geode or Apache Lucene Exceptions
into Spring’s highly consistent and expressive DAO
exception hierarchy,
particularly as many modern data access operations involve more than one store or repository.
Additionally, SDG’s LuceneOperations
interface can shield your application from interface-breaking changes
introduced by the underlying Apache Geode or Apache Lucene APIs when they occur.
However, it would be sad to offer a Lucene Data Access Object (DAO) that only uses Apache Geode and Apache Lucene
data types (such as Apache Geode’s LuceneResultStruct
). Therefore, SDG gives you the
ProjectingLuceneOperations
interface to remedy these important application concerns. The following listing shows
the ProjectingLuceneOperations
interface definition:
public interface ProjectingLuceneOperations {
<T> List<T> query(String query, String defaultField [, int resultLimit], Class<T> projectionType);
<T> Page<T> query(String query, String defaultField, int resultLimit, int pageSize, Class<T> projectionType);
<T> List<T> query(LuceneQueryProvider queryProvider [, int resultLimit], Class<T> projectionType);
<T> Page<T> query(LuceneQueryProvider queryProvider, int resultLimit, int pageSize, Class<T> projectionType);
}
The ProjectingLuceneOperations
interface primarily uses application domain object types that let you work with
your application data. The query
method variants accept a projection type, and the template applies the query results
to instances of the given projection type by using the Spring Data Commons Projection infrastructure.
Additionally, the template wraps the paged Lucene query results in an instance of the Spring Data Commons
Page
abstraction. The same projection logic can still be applied to the results in the page and are lazily projected
as each page in the collection is accessed.
By way of example, suppose you have a class representing a Person
, as follows:
class Person {
Gender gender;
LocalDate birthDate;
String firstName;
String lastName;
...
String getName() {
return String.format("%1$s %2$s", getFirstName(), getLastName());
}
}
Additionally, you might have a single interface to represent people as Customers
, depending on your application view,
as follows:
interface Customer {
String getName()
}
If I define the following LuceneIndex
…
@Bean
LuceneIndexFactoryBean personLastNameIndex(GemFireCache gemfireCache) {
LuceneIndexFactoryBean personLastNameIndex =
new LuceneIndexFactoryBean();
personLastNameIndex.setCache(gemfireCache);
personLastNameIndex.setFields("lastName");
personLastNameIndex.setRegionPath("/People");
return personLastNameIndex;
}
Then you could query for people as Person
objects, as follows:
List<Person> people = luceneTemplate.query("lastName: D*", "lastName", Person.class);
Alternatively, you could query for a Page
of type Customer
, as follows:
Page<Customer> customers = luceneTemplate.query("lastName: D*", "lastName", 100, 20, Customer.class);
The Page
can then be used to fetch individual pages of the results, as follows:
List<Customer> firstPage = customers.getContent();
Conveniently, the Spring Data Commons Page
interface also implements java.lang.Iterable<T>
, making it easy
to iterate over the contents.
The only restriction to the Spring Data Commons Projection infrastructure is that the projection type must be
an interface. However, it is possible to extend the provided SDC Projection infrastructure and provide a custom
ProjectionFactory
that uses CGLIB to generate proxy classes as the projected entity.
You can use setProjectionFactory(:ProjectionFactory)
to set a custom ProjectionFactory
on a Lucene template.
12.2. Annotation Configuration Support
Finally, Spring Data for Apache Geode provides annotation configuration support for LuceneIndexes
.
Eventually, the SDG Lucene support will finds its way into the Repository infrastructure extension for
Apache Geode so that Lucene queries can be expressed as methods on an application Repository
interface,
in much the same way as the OQL support works today.
However, in the meantime, if you want to conveniently express LuceneIndexes
, you can do so directly on
your application domain objects, as the following example shows:
@PartitionRegion("People")
class Person {
Gender gender;
@Index
LocalDate birthDate;
String firstName;
@LuceneIndex;
String lastName;
...
}
To enable this feature, you must use SDG’s annotation configuration support specifically with the
@EnableEntityDefineRegions
and @EnableIndexing
annotations, as follows:
@PeerCacheApplication
@EnableEntityDefinedRegions
@EnableIndexing
class ApplicationConfiguration {
...
}
LuceneIndexes
can only be created on Apache Geode servers since LuceneIndexes
only apply
to PARTITION
Regions.
Given our earlier definition of the Person
class, the SDG annotation configuration support finds
the Person
entity class definition and determines that people are stored in a PARTITION
Region called “People”
and that the Person
has an OQL Index
on birthDate
along with a LuceneIndex
on lastName
.
13. Bootstrapping a Spring ApplicationContext in Apache Geode
Normally, a Spring-based application bootstraps Apache Geode by using Spring Data for Apache Geode’s features.
By specifying a <gfe:cache/>
element that uses the Spring Data for Apache Geode XML namespace, a single embedded Apache Geode
peer Cache
instance is created and initialized with default settings in the same JVM process as your application.
However, it is sometimes necessary (perhaps as a requirement imposed by your IT organization) that Apache Geode
be fully managed and operated by the provided Apache Geode tool suite, perhaps using
Gfsh. By using Gfsh, Apache Geode bootstraps
your Spring ApplicationContext
rather than the other way around. Instead of an application server or a Java main class
that uses Spring Boot, Apache Geode does the bootstrapping and hosts your application.
Apache Geode is not an application server. In addition, there are limitations to using this approach
where the Apache Geode cache configuration is concerned.
13.1. Using Apache Geode to Bootstrap a Spring Context Started with Gfsh
In order to bootstrap a Spring ApplicationContext
in Apache Geode when starting a Apache Geode server
using Gfsh, you must use Apache Geode’s
initalizer capability.
An initializer block can declare a application callback that is launched after the cache is initialized
by Apache Geode.
An initializer is declared within an initializer element
by using a minimal snippet of Apache Geode’s native cache.xml
. To bootstrap the Spring ApplicationContext
,
a cache.xml
file is required, in much the same way as a minimal snippet of Spring XML config is needed to bootstrap
a Spring ApplicationContext
configured with component scanning
(for example <context:component-scan base-packages="…"/>
).
Fortunately, such an initializer is already conveniently provided by the framework: the
SpringContextBootstrappingInitializer
.
The following example shows a typical, yet minimal, configuration for this class inside Apache Geode’s
cache.xml
file:
<?xml version="1.0" encoding="UTF-8"?>
<cache xmlns="http://geode.apache.org/schema/cache"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://geode.apache.org/schema/cache https://geode.apache.org/schema/cache/cache-1.0.xsd"
version="1.0">
<initializer>
<class-name>org.springframework.data.gemfire.support.SpringContextBootstrappingInitializer</class-name>
<parameter name="contextConfigLocations">
<string>classpath:application-context.xml</string>
</parameter>
</initializer>
</cache>
The SpringContextBootstrappingInitializer
class follows conventions similar to Spring’s ContextLoaderListener
class, which is used to bootstrap a Spring ApplicationContext
inside a web application, where ApplicationContext
configuration files are specified with the contextConfigLocations
Servlet context parameter.
In addition, the SpringContextBootstrappingInitializer
class can also be used with a basePackages
parameter
to specify a comma-separated list of base packages that contain appropriately annotated application components.
The Spring container searches these components to find and create Spring beans and other application components
in the classpath, as the following example shows:
<?xml version="1.0" encoding="UTF-8"?>
<cache xmlns="http://geode.apache.org/schema/cache"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://geode.apache.org/schema/cache https://geode.apache.org/schema/cache/cache-1.0.xsd"
version="1.0">
<initializer>
<class-name>org.springframework.data.gemfire.support.SpringContextBootstrappingInitializer</class-name>
<parameter name="basePackages">
<string>org.mycompany.myapp.services,org.mycompany.myapp.dao,...</string>
</parameter>
</initializer>
</cache>
Then, with a properly configured and constructed CLASSPATH
and cache.xml
file (shown earlier) specified as
a command-line option when starting a Apache Geode server in Gfsh, the command-line would be as follows:
gfsh>start server --name=ExampleServer --log-level=config ...
--classpath="/path/to/application/classes.jar:/path/to/spring-data-geode-<major>.<minor>.<maint>.RELEASE.jar"
--cache-xml-file="/path/to/geode/cache.xml"
The application-context.xml
can be any valid Spring configuration metadata, including all of the SDG
XML namespace elements. The only limitation with this approach is that a Apache Geode cache cannot be configured
by using the SDG XML namespace. In other words, none of the <gfe:cache/>
element attributes
(such as cache-xml-location
, properties-ref
, critical-heap-percentage
, pdx-serializer-ref
, lock-lease
,
and others) can be specified. If used, these attributes are ignored.
The reason for this is that Apache Geode itself has already created and initialized the cache before the initializer
gets invoked. As a result, the cache already exists and, since it is a “singleton”, it cannot be re-initialized
or have any of its configuration augmented.
13.2. Lazy-wiring Apache Geode Components
Spring Data for Apache Geode already provides support for auto-wiring Apache Geode components (such as CacheListeners
,
CacheLoaders
, CacheWriters
and so on) that are declared and created by Apache Geode in cache.xml
by using
SDG’s WiringDeclarableSupport
class, as described in Configuration using auto-wiring and annotations. However, this works
only when Spring is the one doing the bootstrapping (that is, when Spring bootstraps Apache Geode).
When your Spring ApplicationContext
is bootstrapped by Apache Geode, these Apache Geode application components
go unnoticed, because the Spring ApplicationContext
does not exist yet. The Spring ApplicationContext
does not get
created until Apache Geode calls the initializer block, which only occurs after all the other Apache Geode
components (cache, Regions, and others) have already been created and initialized.
To solve this problem, a new LazyWiringDeclarableSupport
class was introduced. This new class is aware of the
Spring ApplicationContext
. The intention behind this abstract base class is that any implementing class registers
itself to be configured by the Spring container that is eventually created by Apache Geode once the initializer
is called. In essence, this gives your Apache Geode application components a chance to be configured and auto-wired
with Spring beans defined in the Spring container.
In order for your Apache Geode application components to be auto-wired by the Spring container, you should create
an application class that extends the LazyWiringDeclarableSupport
and annotate any class member that needs to be
provided as a Spring bean dependency, similar to the following example:
public class UserDataSourceCacheLoader extends LazyWiringDeclarableSupport
implements CacheLoader<String, User> {
@Autowired
private DataSource userDataSource;
...
}
As implied in the CacheLoader
example above, you might necessarily (though rarely) have defined both a Region
and a CacheListener
component in Apache Geode cache.xml
. The CacheLoader
may need access to an application
Repository (or perhaps a JDBC DataSource
defined in the Spring ApplicationContext
) for loading Users
into a
Apache Geode REPLICATE
Region on startup.
CAUTION
Be careful when mixing the different life-cycles of Apache Geode and the Spring container together in this manner.
Not all use cases and scenarios are supported. The Apache Geode cache.xml
configuration would be similar to
the following (which comes from SDG’s test suite):
<?xml version="1.0" encoding="UTF-8"?>
<cache xmlns="http://geode.apache.org/schema/cache"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://geode.apache.org/schema/cache https://geode.apache.org/schema/cache/cache-1.0.xsd"
version="1.0">
<region name="Users" refid="REPLICATE">
<region-attributes initial-capacity="101" load-factor="0.85">
<key-constraint>java.lang.String</key-constraint>
<value-constraint>org.springframework.data.gemfire.repository.sample.User</value-constraint>
<cache-loader>
<class-name>
org.springframework.data.gemfire.support.SpringContextBootstrappingInitializerIntegrationTests$UserDataStoreCacheLoader
</class-name>
</cache-loader>
</region-attributes>
</region>
<initializer>
<class-name>org.springframework.data.gemfire.support.SpringContextBootstrappingInitializer</class-name>
<parameter name="basePackages">
<string>org.springframework.data.gemfire.support.sample</string>
</parameter>
</initializer>
</cache>
14. Sample Applications
Sample applications are now maintained in the
Spring Apache Geode Examples repository.
The Spring Data for Apache Geode project also includes one sample application. Named “Hello World”, the sample application
demonstrates how to configure and use Apache Geode inside a Spring application. At run time, the sample offers
a shell that lets you run various commands against the data grid. It provides an excellent
starting point for developers who are unfamiliar with the essential components or with Spring and Apache Geode concepts.
The sample is bundled with the distribution and is Maven-based. You can import it into any
Maven-aware IDE (such as the Spring Tool Suite) or run them from the command-line.
14.1. Hello World
The “Hello World” sample application demonstrates the core functionality of the Spring Data for Apache Geode project.
It bootstraps Apache Geode, configures it, executes arbitrary commands against the cache, and shuts it down
when the application exits. Multiple instances of the application can be started at the same time
and work together, sharing data without any user intervention.
Running under Linux
If you experience networking problems when starting Apache Geode or the samples, try adding the following
system property java.net.preferIPv4Stack=true
to the command line (for example, -Djava.net.preferIPv4Stack=true
).
For an alternative (global) fix (especially on Ubuntu), see SGF-28.
14.1.1. Starting and Stopping the Sample
The “Hello World” sample application is designed as a stand-alone Java application. It features a main
class that can be started
either from your IDE (in Eclipse or STS, through Run As/Java Application
) or from the command line
through Maven with mvn exec:java
. If the classpath is properly set, you can also use java
directly on the resulting artifact.
To stop the sample, type exit
at the command line or press Ctrl+C
to stop the JVM and shutdown
the Spring container.
14.1.2. Using the Sample
Once started, the sample creates a shared data grid and lets you issue commands against it.
The output should resemble the following:
INFO: Created {data-store-name} Cache [Spring {data-store-name} World] v. X.Y.Z
INFO: Created new cache region [myWorld]
INFO: Member xxxxxx:50694/51611 connecting to region [myWorld]
Hello World!
Want to interact with the world ? ...
Supported commands are:
get <key> - retrieves an entry (by key) from the grid
put <key> <value> - puts a new entry into the grid
remove <key> - removes an entry (by key) from the grid
...
For example, to add new items to the grid, you can use the following commands:
-> Bold Section qName:emphasis level:5, chunks:[put 1 unu] attrs:[role:bold]
INFO: Added [1=unu] to the cache
null
-> Bold Section qName:emphasis level:5, chunks:[put 1 one] attrs:[role:bold]
INFO: Updated [1] from [unu] to [one]
unu
-> Bold Section qName:emphasis level:5, chunks:[size] attrs:[role:bold]
1
-> Bold Section qName:emphasis level:5, chunks:[put 2 two] attrs:[role:bold]
INFO: Added [2=two] to the cache
null
-> Bold Section qName:emphasis level:5, chunks:[size] attrs:[role:bold]
2
Multiple instances can be ran at the same time. Once started, the new VMs automatically see the existing region
and its information, as the following example shows:
INFO: Connected to Distributed System ['Spring {data-store-name} World'=xxxx:56218/49320@yyyyy]
Hello World!
...
-> Bold Section qName:emphasis level:5, chunks:[size] attrs:[role:bold]
2
-> Bold Section qName:emphasis level:5, chunks:[map] attrs:[role:bold]
[2=two] [1=one]
-> Bold Section qName:emphasis level:5, chunks:[query length = 3] attrs:[role:bold]
[one, two]
We encourage you to experiment with the example, start (and stop) as many instances as you want, and run various commands in one instance
and see how the others react. To preserve data, at least one instance needs to be alive all times. If all instances
are shutdown, the grid data is completely destroyed.
14.1.3. Hello World Sample Explained
The “Hello World” sample uses both Spring XML and annotations for its configuration. The initial bootstrapping configuration is
app-context.xml
, which includes the cache configuration defined in the cache-context.xml
file
and performs classpath
component scanning
for Spring
components.
The cache configuration defines the Apache Geode cache, a region, and for illustrative purposes, a CacheListener
that acts as a logger.
The main beans are HelloWorld
and CommandProcessor
, which rely on the GemfireTemplate
to interact with
the distributed fabric. Both classes use annotations to define their dependency and life-cycle callbacks.