kafka-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: Kafka-0.8 #103
Date Fri, 09 Nov 2012 23:19:19 GMT
See <https://builds.apache.org/job/Kafka-0.8/103/changes>

Changes:

[junrao] Add test cases to test log size retention and more; patched by John Fung; reviewed
by Jun Rao; KAFKA-591

------------------------------------------
[...truncated 3014 lines...]
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map1.foreach(Map.scala:105)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map1.map(Map.scala:93)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[info] Test Passed: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[2012-11-09 23:18:53,243] ERROR [KafkaApi-0] Error processing ProducerRequest on test:0 (kafka.server.KafkaApis:102)
kafka.common.MessageSizeTooLargeException: Message size is 1000023 bytes which exceeds the
maximum configured message size of 1000000.
	at kafka.log.Log$$anonfun$analyzeAndValidateMessageSet$1.apply(Log.scala:327)
	at kafka.log.Log$$anonfun$analyzeAndValidateMessageSet$1.apply(Log.scala:313)
	at scala.collection.Iterator$class.foreach(Iterator.scala:631)
	at kafka.utils.IteratorTemplate.foreach(IteratorTemplate.scala:32)
	at kafka.log.Log.analyzeAndValidateMessageSet(Log.scala:313)
	at kafka.log.Log.append(Log.scala:241)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:250)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map1.foreach(Map.scala:105)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map1.map(Map.scala:93)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[info] Test Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[2012-11-09 23:18:53,502] ERROR [KafkaApi-0] Error processing ProducerRequest on topic1:0
(kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic1 partition 0 doesn't exist on 0
	at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[2012-11-09 23:18:53,503] ERROR [KafkaApi-0] Error processing ProducerRequest on topic3:0
(kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic3 partition 0 doesn't exist on 0
	at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[2012-11-09 23:18:53,504] ERROR [KafkaApi-0] Error processing ProducerRequest on topic2:0
(kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic2 partition 0 doesn't exist on 0
	at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[2012-11-09 23:18:54,657] ERROR [KafkaApi-0] Error processing ProducerRequest on topic2:0
(kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic2 partition 0 doesn't exist on 0
	at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[info] Test Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] Test Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] 
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] Test Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[info] Test Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Starting: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Passed: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] 
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] Test Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] 
[info] == core-kafka / kafka.log.OffsetIndexTest ==
[info] Test Starting: truncate
[info] Test Passed: truncate
[info] Test Starting: randomLookupTest
[info] Test Passed: randomLookupTest
[info] Test Starting: lookupExtremeCases
[info] Test Passed: lookupExtremeCases
[info] Test Starting: appendTooMany
[info] Test Passed: appendTooMany
[info] Test Starting: appendOutOfOrder
[info] Test Passed: appendOutOfOrder
[info] Test Starting: testReopen
[info] Test Passed: testReopen
[info] == core-kafka / kafka.log.OffsetIndexTest ==
[info] 
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] 
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testWriteTo
[info] Test Passed: testWriteTo
[info] Test Starting: testValidBytes
[info] Test Passed: testValidBytes
[info] Test Starting: testValidBytesWithCompression
[info] Test Passed: testValidBytesWithCompression
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testIterator
[info] Test Passed: testIterator
[info] Test Starting: testOffsetAssignment
[info] Test Passed: testOffsetAssignment
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] Test Starting: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Starting: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Passed: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] 
[info] == core-kafka / kafka.consumer.TopicFilterTest ==
[info] Test Starting: testWhitelists
[info] Test Passed: testWhitelists
[info] Test Starting: testBlacklists
[info] Test Passed: testBlacklists
[info] == core-kafka / kafka.consumer.TopicFilterTest ==
[info] 
[info] == core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testIteratorIsConsistentWithCompression
[info] Test Passed: testIteratorIsConsistentWithCompression
[info] Test Starting: testSizeInBytesWithCompression
[info] Test Passed: testSizeInBytesWithCompression
[info] Test Starting: testEqualsWithCompression
[info] Test Passed: testEqualsWithCompression
[info] == core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / kafka.integration.FetcherTest ==
[info] Test Starting: testFetcher(kafka.integration.FetcherTest)
[info] Test Passed: testFetcher(kafka.integration.FetcherTest)
[info] == core-kafka / kafka.integration.FetcherTest ==
[info] 
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest
==
[info] Test Starting: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest
==
[info] 
[info] == core-kafka / kafka.metrics.KafkaTimerTest ==
[info] Test Starting: testKafkaTimer(kafka.metrics.KafkaTimerTest)
[info] Test Passed: testKafkaTimer(kafka.metrics.KafkaTimerTest)
[info] == core-kafka / kafka.metrics.KafkaTimerTest ==
[info] 
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] Test Starting: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Starting: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /tmp/sbt_8262927a
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 165, Failed 2, Errors 0, Passed 163, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[info] 
[info] == hadoop consumer / copy-test-resources ==
[info] == hadoop consumer / copy-test-resources ==
[error] Error running kafka.log.LogTest: Test FAILED
[error] Error running kafka.server.LogRecoveryTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 173 s, completed Nov 9, 2012 11:19:19 PM
[info] 
[info] Total session time: 174 s, completed Nov 9, 2012 11:19:19 PM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Mime
View raw message