This is an automated email from the ASF dual-hosted git repository.
guozhang pushed a commit to branch trunk
in repository https://gitbox.apache.org/repos/asf/kafka.git
The following commit(s) were added to refs/heads/trunk by this push:
new 5c549b2 MINOR: Replace tbd with the actual link for out-of-ordering data (#6035)
5c549b2 is described below
commit 5c549b2a89c2364eb966bd56f136a7891f0ffc9f
Author: Guozhang Wang <wangguoz@gmail.com>
AuthorDate: Fri Dec 14 09:37:43 2018 -0800
MINOR: Replace tbd with the actual link for out-of-ordering data (#6035)
Reviewers: Jason Gustafson <jason@confluent.io>
---
docs/streams/core-concepts.html | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/docs/streams/core-concepts.html b/docs/streams/core-concepts.html
index bffaa8a..c925c2e 100644
--- a/docs/streams/core-concepts.html
+++ b/docs/streams/core-concepts.html
@@ -224,7 +224,7 @@
<p>
Besides the guarantee that each record will be processed exactly-once, another issue
that many stream processing application will face is how to
- handle <a href="tbd">out-of-order data</a> that may impact their business
logic. In Kafka Streams, there are two causes that could potentially
+ handle <a href="https://www.confluent.io/wp-content/uploads/streams-tables-two-sides-same-coin.pdf">out-of-order
data</a> that may impact their business logic. In Kafka Streams, there are two causes
that could potentially
result in out-of-order data arrivals with respect to their timestamps:
</p>
|