Atlas高可用

Atlas高可用

主节点
0000006的atlas-application.properties分发到各/etc/hive/conf/

Licensed to the Apache Software Foundation (ASF) under one

or more contributor license agreements. See the NOTICE file

distributed with this work for additional information

regarding copyright ownership. The ASF licenses this file

to you under the Apache License, Version 2.0 (the

“License”); you may not use this file except in compliance

with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software

distributed under the License is distributed on an “AS IS” BASIS,

WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.

See the License for the specific language governing permissions and

limitations under the License.

######### Graph Database Configs #########

Graph Database

#Configures the graph database to use. Defaults to JanusGraph
#atlas.graphdb.backend=org.apache.atlas.repository.graphdb.janus.AtlasJanusGraphDatabase

Graph Storage

Set atlas.graph.storage.backend to the correct value for your desired storage

backend. Possible values:

hbase

cassandra

embeddedcassandra - Should only be set by building Atlas with -Pdist,embedded-cassandra-solr

berkeleyje

See the configuration documentation for more information about configuring the various storage backends.

atlas.graph.storage.backend=hbase2
atlas.graph.storage.hbase.table=apache_atlas_janus

#Hbase
#For standalone mode , specify localhost
#for distributed mode, specify zookeeper quorum here
atlas.graph.storage.hostname=bd0000006,bd0000008,bd00000030
atlas.graph.storage.hbase.regions-per-server=1
atlas.graph.storage.lock.wait-time=10000

#In order to use Cassandra as a backend, comment out the hbase specific properties above, and uncomment the
#the following properties
#atlas.graph.storage.clustername=
#atlas.graph.storage.port=

Gremlin Query Optimizer

Enables rewriting gremlin queries to maximize performance. This flag is provided as

a possible way to work around any defects that are found in the optimizer until they

are resolved.

#atlas.query.gremlinOptimizerEnabled=true

Delete handler

This allows the default behavior of doing “soft” deletes to be changed.

Allowed Values:

org.apache.atlas.repository.store.graph.v1.SoftDeleteHandlerV1 - all deletes are “soft” deletes

org.apache.atlas.repository.store.graph.v1.HardDeleteHandlerV1 - all deletes are “hard” deletes

#atlas.DeleteHandlerV1.impl=org.apache.atlas.repository.store.graph.v1.SoftDeleteHandlerV1

Entity audit repository

This allows the default behavior of logging entity changes to hbase to be changed.

Allowed Values:

org.apache.atlas.repository.audit.HBaseBasedAuditRepository - log entity changes to hbase

org.apache.atlas.repository.audit.CassandraBasedAuditRepository - log entity changes to cassandra

org.apache.atlas.repository.audit.NoopEntityAuditRepository - disable the audit repository

atlas.EntityAuditRepository.impl=org.apache.atlas.repository.audit.HBaseBasedAuditRepository

if Cassandra is used as a backend for audit from the above property, uncomment and set the following

properties appropriately. If using the embedded cassandra profile, these properties can remain

commented out.

atlas.EntityAuditRepository.keyspace=atlas_audit

atlas.EntityAuditRepository.replicationFactor=1

Graph Search Index

atlas.graph.index.search.backend=elasticsearch

#Solr
#Solr cloud mode properties
#atlas.graph.index.search.solr.mode=cloud
#atlas.graph.index.search.solr.zookeeper-url=
#atlas.graph.index.search.solr.zookeeper-connect-timeout=60000
#atlas.graph.index.search.solr.zookeeper-session-timeout=60000
#atlas.graph.index.search.solr.wait-searcher=true

#Solr http mode properties
#atlas.graph.index.search.solr.mode=http
#atlas.graph.index.search.solr.http-urls=http://localhost:8983/solr

ElasticSearch support (Tech Preview)

Comment out above solr configuration, and uncomment the following two lines. Additionally, make sure the

hostname field is set to a comma delimited set of elasticsearch master nodes, or an ELB that fronts the masters.

Elasticsearch does not provide authentication out of the box, but does provide an option with the X-Pack product

https://www.elastic.co/products/x-pack/security

Alternatively, the JanusGraph documentation provides some tips on how to secure Elasticsearch without additional

plugins: http://docs.janusgraph.org/latest/elasticsearch.html

atlas.graph.index.hostname=ip.0000006:9200,ip.0000008:9200
atlas.graph.index.search.elasticsearch.client-only=true

Solr-specific configuration property

#atlas.graph.index.search.max-result-set-size=150

######### Notification Configs #########
atlas.notification.embedded=false
atlas.kafka.data=${sys:atlas.home}/data/kafka
#atlas.kafka.zookeeper.connect=localhost:9026
#atlas.kafka.bootstrap.servers=localhost:9027
atlas.kafka.zookeeper.connect=bd0000006:2181,bd0000008:2181,bd00000030:2181
atlas.kafka.bootstrap.servers=bd0000006:9092,bd0000007:9092,bd0000008:9092,bd0000009:9092,bd00000030:9092
atlas.kafka.zookeeper.session.timeout.ms=60000
atlas.kafka.zookeeper.connection.timeout.ms=30000
atlas.kafka.zookeeper.sync.time.ms=20
atlas.kafka.auto.commit.interval.ms=1000
atlas.kafka.hook.group.id=atlas

atlas.kafka.enable.auto.commit=true
atlas.kafka.auto.offset.reset=earliest
atlas.kafka.session.timeout.ms=30000
atlas.kafka.offsets.topic.replication.factor=1
atlas.kafka.poll.timeout.ms=1000
atlas.kafka.max.poll.interval.ms=300000

#Hive
atlas.hook.hive.synchronous=false
atlas.hook.hive.numRetries=3
atlas.hook.hive.queueSize=10000
atlas.cluster.name=primary
hive.atlas.hook=true
hive.exec.post.hooks=org.apache.atlas.hive.hook.HiveHook

atlas.notification.create.topics=true
atlas.notification.replicas=1
atlas.notification.topics=ATLAS_HOOK,ATLAS_ENTITIES
atlas.notification.log.failed.messages=true
atlas.notification.consumer.retry.interval=500
atlas.notification.hook.retry.interval=1000

Enable for Kerberized Kafka clusters

#atlas.notification.kafka.service.principal=kafka/_HOST@EXAMPLE.COM
#atlas.notification.kafka.keytab.location=/etc/security/keytabs/kafka.service.keytab

Server port configuration

atlas.client.readTimeoutMSecs=60000
atlas.client.connectTimeoutMSecs=60000
#atlas.server.http.port=21000
#atlas.server.https.port=21443

######### Security Properties #########

SSL config

atlas.enableTLS=false

#truststore.file=/path/to/truststore.jks
#cert.stores.credential.provider.path=jceks://file/path/to/credentialstore.jceks

#following only required for 2-way SSL
#keystore.file=/path/to/keystore.jks

Authentication config

atlas.authentication.method.kerberos=false
atlas.authentication.method.file=true

ldap.type= LDAP or AD

atlas.authentication.method.ldap.type=none

user credentials file

atlas.authentication.method.file.filename=${sys:atlas.home}/conf/users-credentials.properties

groups from UGI

#atlas.authentication.method.ldap.ugi-groups=true

######## LDAP properties #########
#atlas.authentication.method.ldap.url=ldap://:389
#atlas.authentication.method.ldap.userDNpattern=uid={0},ou=People,dc=example,dc=com
#atlas.authentication.method.ldap.groupSearchBase=dc=example,dc=com
#atlas.authentication.method.ldap.groupSearchFilter=(member=uid={0},ou=Users,dc=example,dc=com)
#atlas.authentication.method.ldap.groupRoleAttribute=cn
#atlas.authentication.method.ldap.base.dn=dc=example,dc=com
#atlas.authentication.method.ldap.bind.dn=cn=Manager,dc=example,dc=com
#atlas.authentication.method.ldap.bind.password=
#atlas.authentication.method.ldap.referral=ignore
#atlas.authentication.method.ldap.user.searchfilter=(uid={0})
#atlas.authentication.method.ldap.default.role=

######### Active directory properties #######
#atlas.authentication.method.ldap.ad.domain=example.com
#atlas.authentication.method.ldap.ad.url=ldap://:389
#atlas.authentication.method.ldap.ad.base.dn=(sAMAccountName={0})
#atlas.authentication.method.ldap.ad.bind.dn=CN=team,CN=Users,DC=example,DC=com
#atlas.authentication.method.ldap.ad.bind.password=
#atlas.authentication.method.ldap.ad.referral=ignore
#atlas.authentication.method.ldap.ad.user.searchfilter=(sAMAccountName={0})
#atlas.authentication.method.ldap.ad.default.role=

######### JAAS Configuration ########

#atlas.jaas.KafkaClient.loginModuleName = com.sun.security.auth.module.Krb5LoginModule
#atlas.jaas.KafkaClient.loginModuleControlFlag = required
#atlas.jaas.KafkaClient.option.useKeyTab = true
#atlas.jaas.KafkaClient.option.storeKey = true
#atlas.jaas.KafkaClient.option.serviceName = kafka
#atlas.jaas.KafkaClient.option.keyTab = /etc/security/keytabs/atlas.service.keytab
#atlas.jaas.KafkaClient.option.principal = atlas/_HOST@EXAMPLE.COM

######### Server Properties #########
atlas.rest.address=http://bd0000006:21000

If enabled and set to true, this will run setup steps when the server starts

#atlas.server.run.setup.on.start=false

######### Entity Audit Configs #########
atlas.audit.hbase.tablename=apache_atlas_entity_audit
atlas.audit.zookeeper.session.timeout.ms=3000
atlas.audit.hbase.zookeeper.quorum=bd0000006:2181,bd0000008:2181,bd00000030:2181

######### High Availability Configuration ########
atlas.server.ha.enabled=true

Enabled the configs below as per need if HA is enabled

atlas.server.ids=id1,id2
atlas.server.address.id1=bd0000006:21000
atlas.server.address.id2=bd0000008:21001
atlas.server.ha.zookeeper.connect=bd0000006:2181,bd0000008:2181,bd00000030:2181
atlas.server.ha.zookeeper.retry.sleeptime.ms=1000
atlas.server.ha.zookeeper.num.retries=3
atlas.server.ha.zookeeper.session.timeout.ms=20000

if ACLs need to be set on the created nodes, uncomment these lines and set the values

#atlas.server.ha.zookeeper.acl=:
#atlas.server.ha.zookeeper.auth=:

######### Atlas Authorization #########
atlas.authorizer.impl=simple
atlas.authorizer.simple.authz.policy.file=atlas-simple-authz-policy.json

######### Type Cache Implementation ########

A type cache class which implements

org.apache.atlas.typesystem.types.cache.TypeCache.

The default implementation is org.apache.atlas.typesystem.types.cache.DefaultTypeCache which is a local in-memory type cache.

#atlas.TypeCache.impl=

######### Performance Configs #########
#atlas.graph.storage.lock.retries=10
#atlas.graph.storage.cache.db-cache-time=120000

######### CSRF Configs #########
atlas.rest-csrf.enabled=true
atlas.rest-csrf.browser-useragents-regex=Mozilla.*,Opera.,^Chrome.
atlas.rest-csrf.methods-to-ignore=GET,OPTIONS,HEAD,TRACE
atlas.rest-csrf.custom-header=X-XSRF-HEADER

############ KNOX Configs ################
#atlas.sso.knox.browser.useragent=Mozilla,Chrome,Opera
#atlas.sso.knox.enabled=true
#atlas.sso.knox.providerurl=https://:8443/gateway/knoxsso/api/v1/websso
#atlas.sso.knox.publicKey=

############ Atlas Metric/Stats configs ################

Format: atlas.metric.query..

atlas.metric.query.cache.ttlInSecs=900
#atlas.metric.query.general.typeCount=
#atlas.metric.query.general.typeUnusedCount=
#atlas.metric.query.general.entityCount=
#atlas.metric.query.general.tagCount=
#atlas.metric.query.general.entityDeleted=

#atlas.metric.query.entity.typeEntities=
#atlas.metric.query.entity.entityTagged=

#atlas.metric.query.tags.entityTags=

######### Compiled Query Cache Configuration #########

The size of the compiled query cache. Older queries will be evicted from the cache

when we reach the capacity.

#atlas.CompiledQueryCache.capacity=1000

Allows notifications when items are evicted from the compiled query

cache because it has become full. A warning will be issued when

the specified number of evictions have occurred. If the eviction

warning threshold <= 0, no eviction warnings will be issued.

#atlas.CompiledQueryCache.evictionWarningThrottle=0

######### Full Text Search Configuration #########

#Set to false to disable full text search.
#atlas.search.fulltext.enable=true

######### Gremlin Search Configuration #########

#Set to false to disable gremlin search.
atlas.search.gremlin.enable=false

########## Add http headers ###########

#atlas.headers.Access-Control-Allow-Origin=*
#atlas.headers.Access-Control-Allow-Methods=GET,OPTIONS,HEAD,PUT,POST
#atlas.headers.=

第二节点
0000008的atlas-application.properties

Licensed to the Apache Software Foundation (ASF) under one

or more contributor license agreements. See the NOTICE file

distributed with this work for additional information

regarding copyright ownership. The ASF licenses this file

to you under the Apache License, Version 2.0 (the

“License”); you may not use this file except in compliance

with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software

distributed under the License is distributed on an “AS IS” BASIS,

WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.

See the License for the specific language governing permissions and

limitations under the License.

######### Graph Database Configs #########

Graph Database

#Configures the graph database to use. Defaults to JanusGraph
#atlas.graphdb.backend=org.apache.atlas.repository.graphdb.janus.AtlasJanusGraphDatabase

Graph Storage

Set atlas.graph.storage.backend to the correct value for your desired storage

backend. Possible values:

hbase

cassandra

embeddedcassandra - Should only be set by building Atlas with -Pdist,embedded-cassandra-solr

berkeleyje

See the configuration documentation for more information about configuring the various storage backends.

atlas.graph.storage.backend=hbase2
atlas.graph.storage.hbase.table=apache_atlas_janus

#Hbase
#For standalone mode , specify localhost
#for distributed mode, specify zookeeper quorum here
atlas.graph.storage.hostname=bd0000006,bd0000008,bd00000030
atlas.graph.storage.hbase.regions-per-server=1
atlas.graph.storage.lock.wait-time=10000

#In order to use Cassandra as a backend, comment out the hbase specific properties above, and uncomment the
#the following properties
#atlas.graph.storage.clustername=
#atlas.graph.storage.port=

Gremlin Query Optimizer

Enables rewriting gremlin queries to maximize performance. This flag is provided as

a possible way to work around any defects that are found in the optimizer until they

are resolved.

#atlas.query.gremlinOptimizerEnabled=true

Delete handler

This allows the default behavior of doing “soft” deletes to be changed.

Allowed Values:

org.apache.atlas.repository.store.graph.v1.SoftDeleteHandlerV1 - all deletes are “soft” deletes

org.apache.atlas.repository.store.graph.v1.HardDeleteHandlerV1 - all deletes are “hard” deletes

#atlas.DeleteHandlerV1.impl=org.apache.atlas.repository.store.graph.v1.SoftDeleteHandlerV1

Entity audit repository

This allows the default behavior of logging entity changes to hbase to be changed.

Allowed Values:

org.apache.atlas.repository.audit.HBaseBasedAuditRepository - log entity changes to hbase

org.apache.atlas.repository.audit.CassandraBasedAuditRepository - log entity changes to cassandra

org.apache.atlas.repository.audit.NoopEntityAuditRepository - disable the audit repository

atlas.EntityAuditRepository.impl=org.apache.atlas.repository.audit.HBaseBasedAuditRepository

if Cassandra is used as a backend for audit from the above property, uncomment and set the following

properties appropriately. If using the embedded cassandra profile, these properties can remain

commented out.

atlas.EntityAuditRepository.keyspace=atlas_audit

atlas.EntityAuditRepository.replicationFactor=1

Graph Search Index

atlas.graph.index.search.backend=elasticsearch

#Solr
#Solr cloud mode properties
#atlas.graph.index.search.solr.mode=cloud
#atlas.graph.index.search.solr.zookeeper-url=
#atlas.graph.index.search.solr.zookeeper-connect-timeout=60000
#atlas.graph.index.search.solr.zookeeper-session-timeout=60000
#atlas.graph.index.search.solr.wait-searcher=true

#Solr http mode properties
#atlas.graph.index.search.solr.mode=http
#atlas.graph.index.search.solr.http-urls=http://localhost:8983/solr

ElasticSearch support (Tech Preview)

Comment out above solr configuration, and uncomment the following two lines. Additionally, make sure the

hostname field is set to a comma delimited set of elasticsearch master nodes, or an ELB that fronts the masters.

Elasticsearch does not provide authentication out of the box, but does provide an option with the X-Pack product

https://www.elastic.co/products/x-pack/security

Alternatively, the JanusGraph documentation provides some tips on how to secure Elasticsearch without additional

plugins: http://docs.janusgraph.org/latest/elasticsearch.html

atlas.graph.index.hostname=ip.0000006:9200,ip.0000008:9200
atlas.graph.index.search.elasticsearch.client-only=true

Solr-specific configuration property

#atlas.graph.index.search.max-result-set-size=150

######### Notification Configs #########
atlas.notification.embedded=false
atlas.kafka.data=${sys:atlas.home}/data/kafka
#atlas.kafka.zookeeper.connect=localhost:9026
#atlas.kafka.bootstrap.servers=localhost:9027
atlas.kafka.zookeeper.connect=bd0000006:2181,bd0000008:2181,bd00000030:2181
atlas.kafka.bootstrap.servers=bd0000006:9092,bd000007:9092,bd0000008:9092,bd000009:9092,bd00000030:9092
atlas.kafka.zookeeper.session.timeout.ms=60000
atlas.kafka.zookeeper.connection.timeout.ms=30000
atlas.kafka.zookeeper.sync.time.ms=20
atlas.kafka.auto.commit.interval.ms=1000
atlas.kafka.hook.group.id=atlas

atlas.kafka.enable.auto.commit=true
atlas.kafka.auto.offset.reset=earliest
atlas.kafka.session.timeout.ms=30000
atlas.kafka.offsets.topic.replication.factor=1
atlas.kafka.poll.timeout.ms=1000
atlas.kafka.max.poll.interval.ms=300000

#Hive
atlas.hook.hive.synchronous=false
atlas.hook.hive.numRetries=3
atlas.hook.hive.queueSize=10000
atlas.cluster.name=primary
hive.atlas.hook=true
hive.exec.post.hooks=org.apache.atlas.hive.hook.HiveHook

atlas.notification.create.topics=true
atlas.notification.replicas=1
atlas.notification.topics=ATLAS_HOOK,ATLAS_ENTITIES
atlas.notification.log.failed.messages=true
atlas.notification.consumer.retry.interval=500
atlas.notification.hook.retry.interval=1000

Enable for Kerberized Kafka clusters

#atlas.notification.kafka.service.principal=kafka/_HOST@EXAMPLE.COM
#atlas.notification.kafka.keytab.location=/etc/security/keytabs/kafka.service.keytab

Server port configuration

atlas.client.readTimeoutMSecs=60000
atlas.client.connectTimeoutMSecs=60000
atlas.server.http.port=21001
#atlas.server.https.port=21443

######### Security Properties #########

SSL config

atlas.enableTLS=false

#truststore.file=/path/to/truststore.jks
#cert.stores.credential.provider.path=jceks://file/path/to/credentialstore.jceks

#following only required for 2-way SSL
#keystore.file=/path/to/keystore.jks

Authentication config

atlas.authentication.method.kerberos=false
atlas.authentication.method.file=true

ldap.type= LDAP or AD

atlas.authentication.method.ldap.type=none

user credentials file

atlas.authentication.method.file.filename=${sys:atlas.home}/conf/users-credentials.properties

groups from UGI

#atlas.authentication.method.ldap.ugi-groups=true

######## LDAP properties #########
#atlas.authentication.method.ldap.url=ldap://:389
#atlas.authentication.method.ldap.userDNpattern=uid={0},ou=People,dc=example,dc=com
#atlas.authentication.method.ldap.groupSearchBase=dc=example,dc=com
#atlas.authentication.method.ldap.groupSearchFilter=(member=uid={0},ou=Users,dc=example,dc=com)
#atlas.authentication.method.ldap.groupRoleAttribute=cn
#atlas.authentication.method.ldap.base.dn=dc=example,dc=com
#atlas.authentication.method.ldap.bind.dn=cn=Manager,dc=example,dc=com
#atlas.authentication.method.ldap.bind.password=
#atlas.authentication.method.ldap.referral=ignore
#atlas.authentication.method.ldap.user.searchfilter=(uid={0})
#atlas.authentication.method.ldap.default.role=

######### Active directory properties #######
#atlas.authentication.method.ldap.ad.domain=example.com
#atlas.authentication.method.ldap.ad.url=ldap://:389
#atlas.authentication.method.ldap.ad.base.dn=(sAMAccountName={0})
#atlas.authentication.method.ldap.ad.bind.dn=CN=team,CN=Users,DC=example,DC=com
#atlas.authentication.method.ldap.ad.bind.password=
#atlas.authentication.method.ldap.ad.referral=ignore
#atlas.authentication.method.ldap.ad.user.searchfilter=(sAMAccountName={0})
#atlas.authentication.method.ldap.ad.default.role=

######### JAAS Configuration ########

#atlas.jaas.KafkaClient.loginModuleName = com.sun.security.auth.module.Krb5LoginModule
#atlas.jaas.KafkaClient.loginModuleControlFlag = required
#atlas.jaas.KafkaClient.option.useKeyTab = true
#atlas.jaas.KafkaClient.option.storeKey = true
#atlas.jaas.KafkaClient.option.serviceName = kafka
#atlas.jaas.KafkaClient.option.keyTab = /etc/security/keytabs/atlas.service.keytab
#atlas.jaas.KafkaClient.option.principal = atlas/_HOST@EXAMPLE.COM

######### Server Properties #########
atlas.rest.address=http://bd0000006:21000,http://bd0000008:21001

If enabled and set to true, this will run setup steps when the server starts

#atlas.server.run.setup.on.start=false

######### Entity Audit Configs #########
atlas.audit.hbase.tablename=apache_atlas_entity_audit
atlas.audit.zookeeper.session.timeout.ms=3000
atlas.audit.hbase.zookeeper.quorum=bd0000006:2181,bd0000008:2181,bd00000030:2181

######### High Availability Configuration ########
atlas.server.ha.enabled=true

Enabled the configs below as per need if HA is enabled

atlas.server.ids=id1,id2
atlas.server.address.id1=bd0000006:21000
atlas.server.address.id2=bd0000008:21001
atlas.server.ha.zookeeper.connect=bd0000006:2181,bd0000008:2181,bd00000030:2181
atlas.server.ha.zookeeper.retry.sleeptime.ms=1000
atlas.server.ha.zookeeper.num.retries=3
atlas.server.ha.zookeeper.session.timeout.ms=20000

if ACLs need to be set on the created nodes, uncomment these lines and set the values

#atlas.server.ha.zookeeper.acl=:
#atlas.server.ha.zookeeper.auth=:

######### Atlas Authorization #########
atlas.authorizer.impl=simple
atlas.authorizer.simple.authz.policy.file=atlas-simple-authz-policy.json

######### Type Cache Implementation ########

A type cache class which implements

org.apache.atlas.typesystem.types.cache.TypeCache.

The default implementation is org.apache.atlas.typesystem.types.cache.DefaultTypeCache which is a local in-memory type cache.

#atlas.TypeCache.impl=

######### Performance Configs #########
#atlas.graph.storage.lock.retries=10
#atlas.graph.storage.cache.db-cache-time=120000

######### CSRF Configs #########
atlas.rest-csrf.enabled=true
atlas.rest-csrf.browser-useragents-regex=Mozilla.*,Opera.,^Chrome.
atlas.rest-csrf.methods-to-ignore=GET,OPTIONS,HEAD,TRACE
atlas.rest-csrf.custom-header=X-XSRF-HEADER

############ KNOX Configs ################
#atlas.sso.knox.browser.useragent=Mozilla,Chrome,Opera
#atlas.sso.knox.enabled=true
#atlas.sso.knox.providerurl=https://:8443/gateway/knoxsso/api/v1/websso
#atlas.sso.knox.publicKey=

############ Atlas Metric/Stats configs ################

Format: atlas.metric.query..

atlas.metric.query.cache.ttlInSecs=900
#atlas.metric.query.general.typeCount=
#atlas.metric.query.general.typeUnusedCount=
#atlas.metric.query.general.entityCount=
#atlas.metric.query.general.tagCount=
#atlas.metric.query.general.entityDeleted=

#atlas.metric.query.entity.typeEntities=
#atlas.metric.query.entity.entityTagged=

#atlas.metric.query.tags.entityTags=

######### Compiled Query Cache Configuration #########

The size of the compiled query cache. Older queries will be evicted from the cache

when we reach the capacity.

#atlas.CompiledQueryCache.capacity=1000

Allows notifications when items are evicted from the compiled query

cache because it has become full. A warning will be issued when

the specified number of evictions have occurred. If the eviction

warning threshold <= 0, no eviction warnings will be issued.

#atlas.CompiledQueryCache.evictionWarningThrottle=0

######### Full Text Search Configuration #########

#Set to false to disable full text search.
#atlas.search.fulltext.enable=true

######### Gremlin Search Configuration #########

#Set to false to disable gremlin search.
atlas.search.gremlin.enable=false

########## Add http headers ###########

#atlas.headers.Access-Control-Allow-Origin=*
#atlas.headers.Access-Control-Allow-Methods=GET,OPTIONS,HEAD,PUT,POST
#atlas.headers.=

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值