kafka-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: Kafka-0.8 #32
Date Sat, 25 Aug 2012 07:25:20 GMT
See <https://builds.apache.org/job/Kafka-0.8/32/changes>

Changes:

[junrao] Log errors for unrecognized config options; patched by Jun Rao; reviewed by Jay Kreps;
KAFKA-181

------------------------------------------
[...truncated 4430 lines...]
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[2012-08-25 07:24:26,446] ERROR Failed to fetch broker info for broker id 0 (kafka.admin.AdminUtils$:96)
[2012-08-25 07:24:26,449] ERROR Failed to fetch broker info for broker id 0 (kafka.admin.AdminUtils$:96)
[2012-08-25 07:24:26,452] ERROR Failed to fetch broker info for broker id 0 (kafka.admin.AdminUtils$:96)
[2012-08-25 07:24:26,455] ERROR Failed to fetch broker info for broker id 0 (kafka.admin.AdminUtils$:96)
[2012-08-25 07:24:26,456] ERROR Failed to send the following requests: ListBuffer(ProducerData(new-topic,test,ArrayBuffer(message(magic
= 1, attributes = 0, crc = 2326977762, payload = java.nio.HeapByteBuffer[pos=0 lim=5 cap=5]))))
(kafka.producer.async.DefaultEventHandler:96)
[2012-08-25 07:24:26,780] WARN EndOfStreamException: Unable to read additional data from client
sessionid 0x1395caa850f0002, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[info] Test Passed: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Starting: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[2012-08-25 07:24:28,000] WARN EndOfStreamException: Unable to read additional data from client
sessionid 0x1395caa850f0002, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[2012-08-25 07:24:29,270] WARN EndOfStreamException: Unable to read additional data from client
sessionid 0x1395caa850f0002, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[2012-08-25 07:24:30,470] WARN EndOfStreamException: Unable to read additional data from client
sessionid 0x1395caa850f0002, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[2012-08-25 07:24:32,175] ERROR Failed to send the following requests: ListBuffer(ProducerData(new-topic,test,ArrayBuffer(message(magic
= 1, attributes = 0, crc = 3632233996, payload = java.nio.HeapByteBuffer[pos=0 lim=4 cap=4]))))
(kafka.producer.async.DefaultEventHandler:96)
[2012-08-25 07:24:32,280] WARN EndOfStreamException: Unable to read additional data from client
sessionid 0x1395caa850f0002, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[info] Test Passed: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] 
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] Test Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[info] Test Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Starting: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Passed: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] 
[info] == core-kafka / kafka.server.ISRExpirationTest ==
[info] Test Starting: testISRExpirationForStuckFollowers(kafka.server.ISRExpirationTest)
[info] Test Passed: testISRExpirationForStuckFollowers(kafka.server.ISRExpirationTest)
[info] Test Starting: testISRExpirationForSlowFollowers(kafka.server.ISRExpirationTest)
[info] Test Passed: testISRExpirationForSlowFollowers(kafka.server.ISRExpirationTest)
[info] == core-kafka / kafka.server.ISRExpirationTest ==
[info] 
[info] == core-kafka / kafka.log.SegmentListTest ==
[info] Test Starting: testAppend
[info] Test Passed: testAppend
[info] Test Starting: testTrunc
[info] Test Passed: testTrunc
[info] Test Starting: testTruncLast
[info] Test Passed: testTruncLast
[info] Test Starting: testTruncBeyondList
[info] Test Passed: testTruncBeyondList
[info] == core-kafka / kafka.log.SegmentListTest ==
[info] 
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] Test Starting: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Starting: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Passed: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
Offsets = 240,216,108,0
[info] Test Passed: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] 
[info] == core-kafka / kafka.server.ServerShutdownTest ==
[info] Test Starting: testCleanShutdown(kafka.server.ServerShutdownTest)
[info] Test Passed: testCleanShutdown(kafka.server.ServerShutdownTest)
[info] == core-kafka / kafka.server.ServerShutdownTest ==
[info] 
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] Test Starting: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] 
[info] == core-kafka / unit.kafka.metrics.KafkaTimerTest ==
[info] Test Starting: testKafkaTimer(unit.kafka.metrics.KafkaTimerTest)
[info] Test Passed: testKafkaTimer(unit.kafka.metrics.KafkaTimerTest)
[info] == core-kafka / unit.kafka.metrics.KafkaTimerTest ==
[info] 
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] Test Starting: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] 
[info] == core-kafka / kafka.log.LogTest ==
[info] Test Starting: testTimeBasedLogRoll
[info] Test Passed: testTimeBasedLogRoll
[info] Test Starting: testSizeBasedLogRoll
[info] Test Passed: testSizeBasedLogRoll
[info] Test Starting: testLoadEmptyLog
[info] Test Passed: testLoadEmptyLog
[info] Test Starting: testLoadInvalidLogsFails
[info] Test Passed: testLoadInvalidLogsFails
[info] Test Starting: testAppendAndRead
[info] Test Passed: testAppendAndRead
[info] Test Starting: testReadOutOfRange
[info] Test Passed: testReadOutOfRange
[info] Test Starting: testLogRolls
[info] Test Passed: testLogRolls
[info] Test Starting: testFindSegment
[info] Test Passed: testFindSegment
[info] Test Starting: testEdgeLogRolls
[info] Test Passed: testEdgeLogRolls
[info] == core-kafka / kafka.log.LogTest ==
[info] 
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.server.HighwatermarkPersistenceTest ==
[info] Test Starting: testHighWatermarkPersistenceSinglePartition(kafka.server.HighwatermarkPersistenceTest)
[info] Test Passed: testHighWatermarkPersistenceSinglePartition(kafka.server.HighwatermarkPersistenceTest)
[info] Test Starting: testHighWatermarkPersistenceMultiplePartitions(kafka.server.HighwatermarkPersistenceTest)
[info] Test Passed: testHighWatermarkPersistenceMultiplePartitions(kafka.server.HighwatermarkPersistenceTest)
[info] == core-kafka / kafka.server.HighwatermarkPersistenceTest ==
[info] 
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] Test Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] 
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testValidBytes
[info] Test Passed: testValidBytes
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testWriteTo
[info] Test Passed: testWriteTo
[info] Test Starting: testSmallFetchSize
[info] Test Passed: testSmallFetchSize
[info] Test Starting: testIterator
[info] Test Passed: testIterator
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] Test Starting: simpleRequest
[info] Test Passed: simpleRequest
[info] Test Starting: tooBigRequestIsRejected
[info] Test Passed: tooBigRequestIsRejected
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] 
[info] == core-kafka / kafka.integration.PrimitiveApiTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testMultiProduce(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testMultiProduce(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testFetchRequestEnforcesUniqueTopicsForOffsetDetails(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testFetchRequestEnforcesUniqueTopicsForOffsetDetails(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testEmptyFetchRequest(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testEmptyFetchRequest(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest)
[info] == core-kafka / kafka.integration.PrimitiveApiTest ==
[info] 
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] Test Starting: testFetcher(kafka.consumer.FetcherTest)
[info] Test Passed: testFetcher(kafka.consumer.FetcherTest)
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] 
[info] == core-kafka / kafka.utils.UtilsTest ==
[info] Test Starting: testSwallow
[info] Test Passed: testSwallow
[info] Test Starting: testCircularIterator
[info] Test Passed: testCircularIterator
[info] == core-kafka / kafka.utils.UtilsTest ==
[info] 
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] Test Starting: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] Test Passed: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] 
[info] == core-kafka / kafka.message.MessageTest ==
[info] Test Starting: testFieldValues
[info] Test Passed: testFieldValues
[info] Test Starting: testChecksum
[info] Test Passed: testChecksum
[info] Test Starting: testEquality
[info] Test Passed: testEquality
[info] Test Starting: testIsHashable
[info] Test Passed: testIsHashable
[info] == core-kafka / kafka.message.MessageTest ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 137, Failed 1, Errors 0, Passed 136, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /var/tmp/sbt_3091e321
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[error] Error running kafka.message.CompressionUtilTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 209 s, completed Aug 25, 2012 7:25:42 AM
[info] 
[info] Total session time: 210 s, completed Aug 25, 2012 7:25:42 AM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Mime
View raw message