Skip to content

Linkis helps easily connect to various back-end computation/storage engines(Spark, Python, TiDB...), exposes various interfaces(REST, JDBC, Java ...), with multi-tenancy, high performance, and resource control.

License

Notifications You must be signed in to change notification settings

binbinCheng/incubator-linkis

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Apache Linkis(Incubating)

Linkis builds a computation middleware layer to facilitate connection, governance and orchestration between the upper applications and the underlying data engines.

https://linkis.apache.org/

EN docs 简体中文文档

github forks github stars github contributors



English | 中文

Introduction

Linkis builds a layer of computation middleware between upper applications and underlying engines. By using standard interfaces such as REST/WS/JDBC provided by Linkis, the upper applications can easily access the underlying engines such as MySQL/Spark/Hive/Presto/Flink, etc., and achieve the intercommunication of user resources like unified variables, scripts, UDFs, functions and resource files at the same time.

As a computation middleware, Linkis provides powerful connectivity, reuse, orchestration, expansion, and governance capabilities. By decoupling the application layer and the engine layer, it simplifies the complex network call relationship, and thus reduces the overall complexity and saves the development and maintenance costs as well.

Since the first release of Linkis in 2019, it has accumulated more than 700 trial companies and 1000+ sandbox trial users, which involving diverse industries, from finance, banking, tele-communication, to manufactory, internet companies and so on. Lots of companies have already used Linkis as a unified entrance for the underlying computation and storage engines of the big data platform.

linkis-intro-01

linkis-intro-03

Features

  • Support for diverse underlying computation storage engines

    • Currently supported computation/storage engines: Spark、Hive、Flink、Python、Pipeline、Sqoop、openLooKeng、Presto、ElasticSearch、JDBC, Shell, etc
    • Computation/storage engines to be supported: Trino (planned 1.3.1), SeaTunnel (planned 1.3.1), etc
    • Supported scripting languages: SparkSQL、HiveQL、Python、Shell、Pyspark、R、Scala and JDBC, etc
  • Powerful task/request governance capabilities With services such as Orchestrator, Label Manager and customized Spring Cloud Gateway, Linkis is able to provide multi-level labels based, cross-cluster/cross-IDC fine-grained routing, load balance, multi-tenancy, traffic control, resource control, and orchestration strategies like dual-active, active-standby, etc

  • Support full stack computation/storage engine As a computation middleware, it will receive, execute and manage tasks and requests for various computation storage engines, including batch tasks, interactive query tasks, real-time streaming tasks and storage tasks

  • Resource management capabilities ResourceManager is not only capable of managing resources for Yarn and Linkis EngineManger, but also able to provide label-based multi-level resource allocation and recycling, allowing itself to have powerful resource management capabilities across mutiple Yarn clusters and mutiple computation resource types

  • Unified Context Service Generate Context ID for each task/request, associate and manage user and system resource files (JAR, ZIP, Properties, etc.), result set, parameter variable, function, etc., across user, system, and computing engine. Set in one place, automatic reference everywhere

  • Unified materials System and user-level unified material management, which can be shared and transferred across users and systems

Supported Engine Types

Engine Name Suppor Component Version
(Default Dependent Version)
Linkis Version Requirements Included in Release Package
By Default
Description
Spark Apache 2.0.0~2.4.7,
CDH >= 5.4.0,
(default Apache Spark 2.4.3)
>=1.0.3 Yes Spark EngineConn, supports SQL , Scala, Pyspark and R code
Hive Apache >= 1.0.0,
CDH >= 5.4.0,
(default Apache Hive 2.3.3)
>=1.0.3 Yes Hive EngineConn, supports HiveQL code
Python Python >= 2.6,
(default Python2*)
>=1.0.3 Yes Python EngineConn, supports python code
Shell Bash >= 2.0 >=1.0.3 Yes Shell EngineConn, supports Bash shell code
JDBC MySQL >= 5.0, Hive >=1.2.1,
(default Hive-jdbc 2.3.4)
>=1.0.3 No JDBC EngineConn, already supports MySQL and HiveQL, can be extended quickly Support other engines with JDBC Driver package, such as Oracle
Flink Flink >= 1.12.2,
(default Apache Flink 1.12.2)
>=1.0.3 No Flink EngineConn, supports FlinkSQL code, also supports starting a new Yarn in the form of Flink Jar Application
Pipeline - >=1.0.3 No Pipeline EngineConn, supports file import and export
openLooKeng openLooKeng >= 1.5.0,
(default openLookEng 1.5.0)
>=1.1.1 No openLooKeng EngineConn, supports querying data virtualization engine with Sql openLooKeng
Sqoop Sqoop >= 1.4.6,
(default Apache Sqoop 1.4.6)
>=1.1.2 No Sqoop EngineConn, support data migration tool Sqoop engine
Presto Presto >= 0.180,
(default Presto 0.234)
>=1.2.0 - Presto EngineConn, supports Presto SQL code
ElasticSearch ElasticSearch >=6.0,
(default ElasticSearch 7.6.2)
>=1.2.0 - ElasticSearch EngineConn, supports SQL and DSL code
Impala Impala >= 3.2.0, CDH >=6.3.0 ongoing - Impala EngineConn, supports Impala SQL code
MLSQL MLSQL >=1.1.0 ongoing - MLSQL EngineConn, supports MLSQL code.
Hadoop Apache >=2.6.0,
CDH >=5.4.0
ongoing - Hadoop EngineConn, supports Hadoop MR/YARN application
TiSpark 1.1 ongoing - TiSpark EngineConn, supports querying TiDB with SparkSQL

Download

Please go to the Linkis Releases Page to download a compiled distribution or a source code package of Linkis.

Compile and Deploy

For more detailed guidance see:

Note: If you want use `-Dlinkis.build.web=true` to build  linkis-web image, you need to compile linkis-web first.

## compile backend
### Mac OS/Linux

# 1. When compiling for the first time, execute the following command first
./mvnw -N install

# 2. make the linkis distribution package
# - Option 1: make the linkis distribution package only
./mvnw clean install -Dmaven.javadoc.skip=true -Dmaven.test.skip=true

# - Option 2: make the linkis distribution package and docker image
#   - Option 2.1: image without mysql jdbc jars
./mvnw clean install -Pdocker -Dmaven.javadoc.skip=true -Dmaven.test.skip=true
#   - Option 2.2: image with mysql jdbc jars
./mvnw clean install -Pdocker -Dmaven.javadoc.skip=true -Dmaven.test.skip=true -Dlinkis.build.with.jdbc=true

# - Option 3: linkis distribution package and docker image (included web)
./mvnw clean install -Pdocker -Dmaven.javadoc.skip=true -Dmaven.test.skip=true -Dlinkis.build.web=true

# - Option 4: linkis distribution package and docker image (included web and ldh (hadoop all in one for test))
./mvnw clean install -Pdocker -Dmaven.javadoc.skip=true -Dmaven.test.skip=true -Dlinkis.build.web=true -Dlinkis.build.ldh=true -Dlinkis.build.with.jdbc=true

### Windows
mvnw.cmd -N install
mvnw.cmd clean install -Dmaven.javadoc.skip=true -Dmaven.test.skip=true

## compile web
cd incubator-linkis/linkis-web
npm install
npm run build

Bundled with MySQL JDBC Driver

Due to the MySQL licensing restrictions, the MySQL Java Database Connectivity (JDBC) driver is not bundled with the official released linkis image by default. However, at current stage, linkis still relies on this library to work properly. To solve this problem, we provide a script which can help to creating an custom image with mysql jdbc from the official linkis image by yourself, the image created by this tool will be tagged as linkis:with-jdbc by default.

$> LINKIS_IMAGE=linkis:1.3.1 
$> ./linkis-dist/docker/scripts/make-linikis-image-with-mysql-jdbc.sh

Please refer to Quick Deployment to do the deployment.

Examples and Guidance

Documentation & Vedio

Architecture

Linkis services could be divided into three categories: computation governance services, public enhancement services and microservice governance services

  • The computation governance services, support the 3 major stages of processing a task/request: submission -> preparation -> execution
  • The public enhancement services, including the material library service, context service, and data source service
  • The microservice governance services, including Spring Cloud Gateway, Eureka and Open Feign

Below is the Linkis architecture diagram. You can find more detailed architecture docs in Linkis-Doc/Architecture. architecture

Contributing

Contributions are always welcomed, we need more contributors to build Linkis together. either code, or doc, or other supports that could help the community.
For code and documentation contributions, please follow the contribution guide.

Contact Us

  • Any questions or suggestions please kindly submit an issue.
  • By mail dev@linkis.apache.org
  • You can scan the QR code below to join our WeChat group to get more immediate response

wechatgroup

Who is Using Linkis

We opened an issue [Who is Using Linkis] for users to feedback and record who is using Linkis.
Since the first release of Linkis in 2019, it has accumulated more than 700 trial companies and 1000+ sandbox trial users, which involving diverse industries, from finance, banking, tele-communication, to manufactory, internet companies and so on.

About

Linkis helps easily connect to various back-end computation/storage engines(Spark, Python, TiDB...), exposes various interfaces(REST, JDBC, Java ...), with multi-tenancy, high performance, and resource control.

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Java 45.6%
  • Scala 40.8%
  • Vue 6.8%
  • JavaScript 3.9%
  • Shell 1.5%
  • SCSS 1.1%
  • Other 0.3%