您的位置:首页 > 运维架构 > Tomcat

apache分别基于三种方案实现tomcat的代理、负载均衡及会话绑定

2017-12-28 21:21 781 查看
apache分别基于mod_proxy_ajp, mod_proxy_http, mod_jk三种方案实现代理、负载均衡、会话绑定及Tomcat session cluster

1、nginx, haproxy, apache(mod_proxy_ajp, mod_proxy_http, mod_jk)反代用户请求至tomcat;
2、nginx, haproxy, apache(mod_proxy_ajp, mod_proxy_http, mod_jk)负载用户请求至tomcat servers;额外实现session sticky;
3、tomcat session cluster;
4、tomcat session server(msm);

规划:
172.16.1.2 apache反代服务器
172.16.1.3 tomcat A
172.16.1.10 tomcat B

说明:
如题,可以选择使用一个去反代用户请求至tomcat,因为apache和Tomcat都是Apache 软件基金会的产品,两者有更好的兼容性,所以选择用apache,并且会以mod_proxy_ajp, mod_proxy_http, mod_jk三种方式分别实现题目要求。

1.方案一:mod_proxy_http
vim /etc/httpd/conf.d/lb_tomcat.conf

Header add Set-Cookie "ROUTEID=.%{BALANCER_WORKER_ROUTE}e; path=/" env=BALANCER_ROUTE_CHANGED

<Proxy balancer://tcsrvs>
BalancerMember http://172.16.1.10:8080/testapp route=TomcatA loadfactor=1
BalancerMember http://172.16.1.3:8080/testapp route=TomcatB loadfactor=2
ProxySet lbmethod=byrequests
ProxySet stickysession=ROUTEID
</Proxy>

<VirtualHost *:80>
ServerName      lb.zrs.com
ProxyVia On
ProxyRequests Off
<Proxy *>
Require all granted
</Proxy>
ProxyPass / balancer://tcsrvs/
<Location />
Require all granted
</Location>
</VirtualHost>


分别为两个Tomcat的server.xml配置文件中的引擎中添加jvmRoute

第一个tomcat配置为TomcatA:
<Engine name="Catalina" defaultHost="localhost" jvmRoute="TomcatA">

第二个tomcat配置为TomcatB:
<Engine name="Catalina" defaultHost="localhost" jvmRoute="TomcatB">


分别为两个Tomcat Server的testapp提供测试页面:
tomcatA:
# mkdir -pv /usr/share/tomcat/webapps/testapp/{WEB-INF,classes,lib}
# vim /usr/share/tomcat/webapps/testapp/index.jsp
添加如下内容:

<%@ page language="java" %>
<html>
<head><title>TomcatA</title></head>
<body>
<h1><font color="red">TomcatA.zrs.com</font></h1>
<table align="centre" border="1">
<tr>
<td>Session ID</td>
<% session.setAttribute("zrs.com","zrs.com"); %>
<td><%= session.getId() %></td>
</tr>
<tr>
<td>Created on</td>
<td><%= session.getCreationTime() %></td>
</tr>
</table>
</body>
</html>


tomcatB:
# mkdir -pv /usr/share/tomcat/webapps/testapp/{WEB-INF,classes,lib}
# vim /usr/share/tomcat/webapps/testapp/index.jsp
添加如下内容:

<%@ page language="java" %>
<html>
<head><title>TomcatB</title></head>
<body>
<h1><font color="blue">TomcatB.zrs.com</font></h1>
<table align="centre" border="1">
<tr>
<td>Session ID</td>
<% session.setAttribute("zrs.com","zrs.com"); %>
<td><%= session.getId() %></td>
</tr>
<tr>
<td>Created on</td>
<td><%= session.getCreationTime() %></td>
</tr>
</table>
</body>
</html>


客户端测试,因为做了session sticky,会一直调度到后端一台服务器上,如我的一直是TomcatA



关掉TomcatA,便会调度到TomcatB上。



2.方案二:mod_proxy_ajp
使用mod_proxy_ajp,仅需要把mod_proxy_http的http改为ajp,端口8080改为8009即可,需要注意的是前端用了ajp,可以把后端的tomcat的8080端口注释掉,不让其监听,从而使安全性更高。

vim /etc/httpd/conf.d/lb_tomcat.conf

Header add Set-Cookie "ROUTEID=.%{BALANCER_WORKER_ROUTE}e; path=/" env=BALANCER_ROUTE_CHANGED

<Proxy balancer://tcsrvs>
BalancerMember ajp://172.16.1.10:8009/testapp route=TomcatA loadfactor=1
BalancerMember ajp://172.16.1.3:8009/testapp route=TomcatB loadfactor=2
ProxySet lbmethod=byrequests
ProxySet stickysession=ROUTEID
</Proxy>

<VirtualHost *:80>
ServerName      lb.qhdlink.com
ProxyVia On
ProxyRequests Off
<Proxy *>
Require all granted
</Proxy>
ProxyPass / balancer://tcsrvs/
<Location />
Require all granted
</Location>
</VirtualHost>
关闭session sticky,客户端查看





3.方案三:mod_jk

此方案需要mod_jk模块
需要安装这个文件tomcat-connectors-VERSION.tar.gz

配置文件:
vim /etc/httpd/conf.d/mod_jk_lb.conf
LoadModule jk_module modules/mod_jk.so
JkWorkersFile /etc/httpd/conf.d/workers.properties
JkLogFile logs/mod_jk.log
JkLogLevel debug
JkMount /* tcsrvs
JkMount /jk-status Stat
vim /etc/httpd/conf.d/workers.properties
worker.list=tcsrvs,Stat
worker.TomcatA.host=172.16.1.10
worker.TomcatA.port=8009
worker.TomcatA.type=ajp13
worker.TomcatA.lbfactor=1
worker.TomcatB.host=172.16.1.3
worker.TomcatB.port=8009
worker.TomcatB.type=ajp13
worker.TomcatB.lbfactor=2
worker.tcsrvs.type=lb
worker.tcsrvs.balance_workers=TomcatA,TomcatB
worker.tcsrvs.sticky_session=1
worker.Stat.type=status


4.tomcat session cluster

在两个tomcat的server.xml配置启用集群
添加如下配置内容,并且注释掉8080端口
需要注意的是下面的内容中,将address="auto"的auto改为该主机的IP地址

<Cluster className="org.apache.catalina.ha.tcp.SimpleTcpCluster"
channelSendOptions="8">
<Manager className="org.apache.catalina.ha.session.DeltaManager"
expireSessionsOnShutdown="false"
notifyListenersOnReplication="true"/>

<Channel className="org.apache.catalina.tribes.group.GroupChannel">
<Membership className="org.apache.catalina.tribes.membership.McastService"
address="228.100.0.4"
port="45564"
frequency="500"
dropTime="3000"/>
<Receiver className="org.apache.catalina.tribes.transport.nio.NioReceiver"
address="auto"
port="4000"
autoBind="100"
selectorTimeout="5000"
maxThreads="6"/>

<Sender className="org.apache.catalina.tribes.transport.ReplicationTransmitter">
<Transport className="org.apache.catalina.tribes.transport.nio.PooledParallelSender"/>
</Sender>

<Interceptor className="org.apache.catalina.tribes.group.interceptors.TcpFailureDetector"/>
<Interceptor className="org.apache.catalina.tribes.group.interceptors.MessageDispatch15Interceptor"/>
</Channel>

<Valve className="org.apache.catalina.ha.tcp.ReplicationValve"
filter=""/>
<Valve className="org.apache.catalina.ha.session.JvmRouteBinderValve"/>

<Deployer className="org.apache.catalina.ha.deploy.FarmWarDeployer"
tempDir="/tmp/war-temp/"
deployDir="/tmp/war-deploy/"
watchDir="/tmp/war-listen/"
watchEnabled="false"/>
<ClusterListener className="org.apache.catalina.ha.session.JvmRouteSessionIDBinderListener"/>
<ClusterListener className="org.apache.catalina.ha.session.ClusterSessionListener"/>
</Cluster>

并且在Host中添加一个Context组件
<Host name="localhost" appBase="webapps"
unpackWARs="true" autoDeploy="true">
<Context path="/test" docBase="/myapps/webapps" reloadable="true"/>

创建目录
[root@zrs tomcat]# mkdir -pv /myapps/webapps/{classes,lib,WEB-INF}

创建测试页面
跟上面的一样,创建AB两个不同的页面
[root@zrs tomcat]# vim /myapps/webapps/index.jsp

复制web.xml到指定目录
[root@zrs tomcat]# cp web.xml /myapps/webapps/WEB-INF/

编辑WEB-INF/web.xml
在任意非段落外添加<distributable/>
[root@zrs tomcat]# vim /myapps/webapps/WEB-INF/web.xml

反代服务器配置
[root@director conf.d]# vim mod_jk_proxy.conf

LoadModule jk_module modules/mod_jk.so
JkWorkersFile /etc/httpd/conf.d/workers.properties
JkLogFile logs/mod_jk.log
JkLogLevel debug
JkMount /* tcsrvs
JkMount /jk-status Stat


[root@director conf.d]# vim workers.properties

worker.list=tcsrvs,Stat
worker.TomcatA.host=172.16.1.10
worker.TomcatA.port=8009
worker.TomcatA.type=ajp13
worker.TomcatA.lbfactor=1
worker.TomcatB.host=172.16.1.3
worker.TomcatB.port=8009
worker.TomcatB.type=ajp13
worker.TomcatB.lbfactor=2
worker.tcsrvs.type=lb
worker.tcsrvs.balance_workers=TomcatA,TomcatB
worker.tcsrvs.sticky_session=0
worker.Stat.type=status

分别开启三台主机的httpd服务和tomcat服务

客户端测试
可以看到反代服务器以ABB的次序轮询后端两台tomcat服务器,但是ID不变,从而实现session cluster。





5.tomcat session server(msm)
接着刚才的配置环境,这个构建方案的原理是将所有会话放在同一台后端tomcat服务器上,当后端这台tomcat服务器宕机时,整个搭建便无法正常提供服务,所以也要给tomcat服务器做高可用。

这个构建方案需要memcached作为旁挂式缓存服务器,所以需要提前下载并启动memcached服务。

停止两台tomcat服务,还原server.xml配置文件
注释掉8080端口,在Engine中添加jvmRoute,如下

<Engine name="Catalina" defaultHost="localhost" jvmRoute="TomcatA" >

<Engine name="Catalina" defaultHost="localhost" jvmRoute="TomcatB" >

在Host段添加配置内容

<Context path="/test" docBase="/myapps/webapps" reloadable="true">
<Manager className="de.javakaffee.web.msm.MemcachedBackupSessionManager"
memcachedNodes="n1:172.16.1.10:11211,n2:172.16.1.3:11211"
failoverNodes="n1"
requestUriIgnorePattern=".*\.(ico|png|gif|jpg|css|js|html|htm)$"
transcoderFactoryClass="de.javakaffee.web.msm.serializer.javolution.JavolutionTranscoderFactory"
/>
</Context>

导入类文件
memcached-session-manager项目地址,http://code.google.com/p/memcached-session-manager/

下载如下jar文件至各tomcat节点的tomcat安装目录下的lib目录中,其中的${version}要换成你所需要的版本号,tc${6,7,8}要换成与tomcat版本相同的版本号。
memcached-session-manager-${version}.jar
memcached-session-manager-tc${6,7,8}-${version}.jar
spymemcached-${version}.jar
msm-javolution-serializer-${version}.jar
javolution-${version}.jar

将所有的类放在/usr/share/java/tomcat目录中。

启动tomcat服务即可看到反代服务器以AB的次序轮询后端两台tomcat服务器,但是session ID不变。
内容来自用户分享和网络整理,不保证内容的准确性,如有侵权内容,可联系管理员处理 点击这里给我发消息
标签:  tomcat apache