Flink (38): API call of Flink TableAPI and FlinkSQL (in)
catalogue
0. Links to related articles
1. Table query
1.1. Call of table API
1.2. SQL query
2. Convert DataStream to table
2.1. Code expression
2.2. Correspondence between data type and Table schema
3. Convert DataSet to table
4. Create a Temporary View
0. Links to related articles
1. Table query
Using the connector of the extern ...
Posted by svguerin3 on Sat, 18 Dec 2021 19:29:05 +0100
flink sql knows why | flink sql dating call (read this article is enough)
Thank you for your little love (attention + praise + look again). Your affirmation of the blogger will urge the blogger to continuously output more high-quality practical content!!!
1. Preface - structure of this paper
Big data sheep said
Using data to improve the probability of good things~
32 original content
official account
Thi ...
Posted by pococool on Thu, 16 Dec 2021 17:36:46 +0100
Flink_ Malicious login monitoring (using CEP)
For websites,
User login is not a frequent business operation
. If
If a user fails to log in frequently in a short time, there may be a malicious attack on the program, such as password brute force cracking
. Therefore, we consider that the login failure actions of users should be counted. Specifically, if
If the same user (ca ...
Posted by Xo_ on Wed, 15 Dec 2021 19:18:33 +0100
FlinkCEP introduction and entry case (based on the source code of flink 1.13.x)
1. What is FlinkCEP?
1.1. CEP
CEP [composite event processing] finds out relevant event combinations (occurrence events) from different event streams through rules, and further processes the found (processing occurrence events).
CEP: first capture various subtle events (basic events or simple events), then analyze and sort out [event pattern ...
Posted by tgavin on Thu, 09 Dec 2021 06:47:45 +0100
FlinkCDC+Hudi+Hive big data real-time basic combat into the lake
catalogue
The new architecture is integrated with the lake warehouse
1, Version Description
2, Compile and package Hudi version 0.10.0
1. Use git to clone the latest master on github
2. Compilation and packaging
3, Create a flick project
1. Main contents of POM document
2.checkpoint
3.flinkcdc code
4.hudi code (refer to the official ...
Posted by WebbDawg on Fri, 03 Dec 2021 03:42:39 +0100
How does Flink analyze and handle back pressure?
1. Concept
backpressure is a very common problem in flow calculation. backpressure means that a node in the data pipeline becomes a bottleneck, and the processing rate can not keep up with the data sent by the upstream, so it is necessary to speed limit the upstream. Since real-time computing applications usually use message queues to decouple ...
Posted by blankextacy on Tue, 23 Nov 2021 14:12:55 +0100
The core abstraction of flink -- a comprehensive and detailed explanation of Window and Window operation
1. Windows
1.1. Basic concepts
Window is the core of processing infinite flow. The window divides the flow into "buckets" of fixed size, which is convenient for programmers to apply various calculations on it.
Window operation is a very core abstraction of streaming data processing. It divides an infinite stream data set into bound ...
Posted by pliant on Sat, 20 Nov 2021 03:55:28 +0100
Flink Project Series 3 - real time traffic statistics
1, Project overview
1.1 module creation and data preparation
create a new package for networkflow analysis.
copy the log file apache.log of the Apache server to the resource file directory src/main/resources Next, we will read the data from here.
of course, we can still use UserBehavior.csv as the data source. At th ...
Posted by mwd2005 on Fri, 12 Nov 2021 16:33:13 +0100
[calculate] interpretation of Apache calculate verification process source code
1. Core structure and conceptThe Validator process provided by calculate is extremely complex, but to sum up, it mainly does one thing: verify whether the semantics of each SqlNode is correct in combination with metadata. These semantics include:Verify whether the table name exists;Whether the selected column exists in the corresponding table a ...
Posted by presence on Tue, 09 Nov 2021 05:45:10 +0100
Flink practice tutorial: getting started 5 - writing ClickHouse
Introduction to flow computing OceanusStream computing Oceanus is a powerful tool for real-time analysis of big data product ecosystem. It is an enterprise level real-time big data analysis platform based on Apache Flink with the characteristics of one-stop development, seamless connection, sub second delay, low cost, security and stability. St ...
Posted by szym9341 on Sat, 06 Nov 2021 21:44:06 +0100