failover

High availability computing: How to deal with a non-returning system call, without risking false positives?

女生的网名这么多〃 提交于 2019-12-02 04:31:25
I have a process that's running on a Linux computer as part of a high-availability system. The process has a main thread that receives requests from the other computers on the network and responds to them. There is also a heartbeat thread that sends out multicast heartbeat packets periodically, to let the other processes on the network know that this process is still alive and available -- if they don't heart any heartbeat packets from it for a while, one of them will assume this process has died and will take over its duties, so that the system as a whole can continue to work. This all works

NHibernate and database connection failover?

这一生的挚爱 提交于 2019-12-01 23:07:19
I am using NHibernate to connect to a legacy rdbms system. Under high production load the rdbms service fails. To maintain the availability we have a failover rdbms service. Is there a way to configure NHibernate to use the FailOver Connection String when the primary connection is down? Additional Info: I am using Castle over NHibernate. If Castle provides handling of failover connections then that will also do it for me. You can build your own NHibernate.Connection.IConnectionProvider which provides failover support. This should be a subclass of ConnectionProvider which overrides

Kafka consumer fails to consume if first broker is down

六眼飞鱼酱① 提交于 2019-12-01 18:23:04
I'm using latest version of kafka(kafka_2.12-1.0.0.tgz). I have setup simple cluster with 3 brokers(just changed broker.id=1 and listeners=PLAINTEXT://:9092 in properties file for each instance).After cluster is up I created topic with the following command ./kafka-topics.sh --create --zookeeper localhost:2181 --replication-factor 3 --partitions 13 --topic demo then start kafka consumer and producers with the following commands ./kafka-console-producer.sh --topic demo --broker-list localhost:9094,localhost:9093,localhost:9092 ./kafka-console-consumer.sh --group test --bootstrap-server

What does Spark recover the data from a failed node?

南笙酒味 提交于 2019-12-01 11:27:34
Suppose we have an RDD, which is being used multiple times. So to save the computations again and again, we persisted this RDD using the rdd.persist() method. So when we are persisting this RDD, the nodes computing the RDD will be storing their partitions. So now suppose, the node containing this persisted partition of RDD fails, then what will happen? How will spark recover the lost data? Is there any replication mechanism? Or some other mechanism? When you do rdd.persist, rdd doesn't materialize the content. It does when you perform an action on the rdd. It follows the same lazy evaluation

What does Spark recover the data from a failed node?

吃可爱长大的小学妹 提交于 2019-12-01 08:59:03
问题 Suppose we have an RDD, which is being used multiple times. So to save the computations again and again, we persisted this RDD using the rdd.persist() method. So when we are persisting this RDD, the nodes computing the RDD will be storing their partitions. So now suppose, the node containing this persisted partition of RDD fails, then what will happen? How will spark recover the lost data? Is there any replication mechanism? Or some other mechanism? 回答1: When you do rdd.persist, rdd doesn't

Twisted: ReconnectingClientFactory connection to different servers

半城伤御伤魂 提交于 2019-12-01 06:51:51
I have a twisted ReconnectingClientFactory and i can successfully connect to given ip and port couple with this factory. And it works well. reactor.connectTCP(ip, port, myHandsomeReconnectingClientFactory) In this situation, when the server is gone, myHandsomeReconnectingClientFactory tries to connect same ip and port (as expected). My goal is, when the server which serves on given ip and port couple is gone, connecting to a backup server (which have different ip and port). Any ideas/comments on how to achieve this goal will be appreciated. Id try something like: class

Basic Apache Camel LoadBalancer Failover Example

情到浓时终转凉″ 提交于 2019-12-01 04:24:15
问题 To start I just want to let you know I am new to Camel and very recently I grasped its main concepts. I am trying to create a basic working example using Apache-Camel with ActiveMQ as a broker and using jms-component as a client of a loadbalancer using the failover construct. All this is done using the Java DSL only (if possible). The example consists of 4 main apps, called MyApp-A, MyApp-B, MyApp-C and MyApp-D. In a normal scenario MyApp-A reads a file from my computer and then transforms it

Configure GlassFish JDBC connection pool to handle Amazon RDS Multi-AZ failover

本小妞迷上赌 提交于 2019-11-30 14:19:39
问题 I have a Java EE application running in GlassFish on EC2, with a MySQL database on Amazon RDS. I am trying to configure the JDBC connection pool to in order to minimize downtime in case of database failover. My current configuration isn't working correctly during a Multi-AZ failover, as the standby database instance appears to be available in a couple of minutes (according to the AWS console) while my GlassFish instance remains stuck for a long time (about 15 minutes) before resuming work.

How to enforce message queue sequence with multiple WCF service instances

爱⌒轻易说出口 提交于 2019-11-30 13:56:14
I want to create a WCF service which uses an MSMQ binding as I have a high volume of notifications the service is to process. It is important that clients are not held up by the service and that the notifications are processed in the order they are raised, hence the queue implementation. Another consideration is resilience. I know I could cluster MSMQ itself to make the queue more robust, but I want to be able to run an instance of my service on different servers, so if a server crashes notifications do not build up in the queue but another server carries on processing. I have experimented

Configure GlassFish JDBC connection pool to handle Amazon RDS Multi-AZ failover

半世苍凉 提交于 2019-11-30 10:15:34
I have a Java EE application running in GlassFish on EC2, with a MySQL database on Amazon RDS. I am trying to configure the JDBC connection pool to in order to minimize downtime in case of database failover. My current configuration isn't working correctly during a Multi-AZ failover, as the standby database instance appears to be available in a couple of minutes (according to the AWS console) while my GlassFish instance remains stuck for a long time (about 15 minutes) before resuming work. The connection pool is configured like this: asadmin create-jdbc-connection-pool --restype javax.sql