-
Notifications
You must be signed in to change notification settings - Fork 1
/
config.example.yml
168 lines (160 loc) · 7.33 KB
/
config.example.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
# Sample configuration for BaseConfig.
# This can be modified and included as part of an extractor built on the utils.
# The values here are defaults and can be left out, except if specified otherwise.
# Version of the config file, accepted versions are specified in each version of the extractor
# Needs to be set
version:
# Configuration for the pusher to CDF.
cognite:
# The project to connect to in the API
project:
# Cognite service url
host: "https://api.cognitedata.com"
# Config for authentication if a bearer access token has to be used for authentication.
# Leave empty to disable.
idp-authentication:
# URL to fetch tokens from, either this, or tenant must be present.
token-url:
# Identity provider authority endpoint (optional, only used in combination with tenant)
authority: "https://login.microsoftonline.com/"
# Directory tenant
tenant:
# Application Id
client-id:
# Client secret
secret:
# List of resource scopes, ex:
# scopes:
# - scopeA
# - scopeB
scopes:
# Audience
audience:
# Minimum time-to-live in seconds for the token (optional)
min-ttl: 30
# Configuration for certificate based authentication
certificate:
# Path to a .pfx or .pem certificate file (.pem only for .NET 7 extractors)
path:
# Optional certificate key password
password:
# Optional authority URL. If this is not specified, authority + tenant is used.
authority-url:
# Configure automatic retries on requests to CDF. Can be left out to keep default values.
cdf-retries:
# Timeout in ms for each individual try
timeout: 80000
# Maximum number of retries, less than 0 retries forever
max-retries: 5
# Max delay in ms between each try. Base delay is calculated according to 125*2^retry ms.
# If less than 0, there is no maximum. (0 means that there is never any delay)
max-delay: 5000
# Configure chunking of data on requests to CDF. Note that increasing these may cause requests to fail due to limits in CDF.
cdf-chunking:
# Maximum number of timeseries per get/create timeseries request
time-series: 1000
# Maximum number of assets per get/create asset request
assets: 1000
# Maximum number of timeseries per datapoint create request
data-point-time-series: 10000
# Maximum number of datapoints per datapoint create request
data-points: 100000
# Minimum number of datapoints in request to switch to using gzip.
# Set to -1 to disable, and 0 to always enable (not recommended)
# The minimum HTTP packet size is generally 1500 bytes, so this should never be
# set below 100, at least for numeric datapoints. Even for larger packages gzip is efficient enough
# that packages are compressed below 1500 bytes which is costly. At 5000 it is always a performance gain.
# It can be set lower if bandwith is a major issue.
data-points-gzip-limit: 5000
# Maximum number of ranges per delete datapoints request
data-point-delete: 10000
# Maximum number of timeseries per datapoint read request, used when getting the first point in a timeseries
data-point-list: 100
# Maximum number of timeseries per datapoint read latest request, used when getting last point in a timeseries
data-point-latest: 100
# Maximum number of rows per request to cdf raw. Used with raw state-store.
raw-rows: 10000
# Maximum number of events per get/create events request
events: 1000
# Maximum number of sequences per get/create sequences request
sequences: 1000
# Maximum number of sequences per create sequence rows request
sequence-row-sequences: 1000
# Maximum number of rows per sequence when creating rows
sequence-rows: 10000
# Configure how requests to CDF should be throttled
cdf-throttling:
# Maximum number of parallel requests per timeseries operation
time-series: 20
# Maximum number of parallel requests per assets operation
assets: 20
# Maximum number of parallel requests per datapoints operation
data-points: 10
# Maximum number of parallel requests per raw operation
raw: 10
# Maximum number of parallel requests per get first/last datapoint operation
ranges: 20
# Maximum number of parallel requests per events operation
events: 20
# Maximum number of parallel requests per sequences operation
sequences: 10
# Configure if the SDK should do logging of requests
sdk-logging:
# True to disable logging from the SDK
disable: false
# Minimum level of logging, one of trace, debug, information, warning, error, critical, none
level: debug
# Format of the log message
format: "CDF ({Message}): {HttpMethod} {Url} {ResponseHeader[X-Request-ID]} - {Elapsed} ms"
# Configure an extraction pipeline manager
extraction-pipeline:
# ExternalId of extraction pipeline
# external-id:
# Frequency to report "Seen", in seconds. Less than or equal to zero will not report automatically.
# frequency: 600
# Optional configuration for special handling of SSL certificates.
# This should never be considered a permanent solution to certificate problems.
certificates:
# True to accept all certificates.
# This poses a severe security risk.
accept-all: false
# List of thumbprints of certificates to allow.
# This is a smaller risk compared to accepting all certificates.
allow-list:
# - 99E92D8447AEF30483B1D7527812C9B7B3A915A7
# Store state in a local database or in CDF raw to speed up starting, by not having to read state from destinations
state-store:
# Path to .db file used by the state storage, or name of table in raw. Default is none, so this needs to be specified.
location: "buffer.db"
# Which type of database to use. One of "None", "Raw", "LiteDb"
database: "None"
# Configure logger to console or file
logger:
# Writes log events at this level to the Console. One of verbose, debug, information, warning, error, fatal.
# If not present, or if the level is invalid, Console is not used.
console:
level:
# Log events at or above this level are redirected to standard error.
stderrLevel:
# Writes log events at this level to a file. Logs will roll over to new files daily.
# If not present, or if the level is invalid, logging to file is disabled.
file:
level:
# Path for logging output. If not present, logging to file is disabled.
path: # "logs/log.txt"
# Maximum number of logs files that are kept in the log folder.
retention-limit: 31
# Rolling interval for log files. Either "day" or "hour".
rolling-interval: "day"
# Configure destinations for prometheus metrics
metrics:
# Start a metrics server in the extractor for Prometheus scrape
server:
host:
port: 0
# Multiple Prometheus PushGateway destinations:
push-gateways:
- host:
job:
username:
password: