change.1.0.1.source-code.application.conf Maven / Gradle / Ivy
The newest version!
{
# Spark relation config
spark: {
app: {
name: Spark Writer
}
driver: {
cores: 1
maxResultSize: 1G
}
cores {
max: 16
}
}
# Nebula Graph relation config
nebula: {
addresses: ["127.0.0.1:3699"]
user: user
pswd: password
space: test
connection {
timeout: 3000
retry: 3
}
execution {
retry: 3
}
error: {
max: 32
output: /tmp/errors
}
rate: {
limit: 64M
timeout: 1000
}
}
# Processing tags
tags: [
# Loading tag from HDFS and data type is parquet
{
name: tag-name-0
type: parquet
path: hdfs tag path 0
fields: {
hive-field-0: nebula-field-0,
hive-field-1: nebula-field-1,
hive-field-2: nebula-field-2
}
vertex: hive-field-0
batch: 2
separator: ","
header: true
}
# Loading from Hive
{
name: tag-name-1
type: hive
exec: "select hive-field0, hive-field1, hive-field2 from database.table"
fields: {
hive-field-0: nebula-field-0,
hive-field-1: nebula-field-1,
hive-field-2: nebula-field-2
}
vertex: {
field: hive-field-0
policy: "hash"
}
vertex: hive-field-0
partition: 32
}
]
# Processing edges
edges: [
# Loading tag from HDFS and data type is parquet
{
name: edge-name-0
type: json
path: hdfs edge path 0
fields: {
hive-field-0: nebula-field-0,
hive-field-1: nebula-field-1,
hive-field-2: nebula-field-2
}
source: {
field: hive-field-0
policy: "hash"
}
target: {
field: hive-field-1
policy: "uuid"
}
ranking: hive-field-2
partition: 32
}
# Loading from Hive
{
name: edge-name-1
type: hive
exec: "select hive-field0, hive-field1, hive-field2 from database.table"
fields: {
hive-field-0: nebula-field-0,
hive-field-1: nebula-field-1,
hive-field-2: nebula-field-2
}
source: hive-field-0
target: hive-field-1
partition: 32
}
]
}