FirstPhone
Some checks failed
K8S Fission Deployment / Deployment fission functions (push) Failing after 21s
Some checks failed
K8S Fission Deployment / Deployment fission functions (push) Failing after 21s
This commit is contained in:
81
.devcontainer/helm/fission-values.yaml
Normal file
81
.devcontainer/helm/fission-values.yaml
Normal file
@@ -0,0 +1,81 @@
|
||||
### helm show values fission-charts/fission-all > .devcontainer/fission-values.yaml
|
||||
|
||||
serviceMonitor:
|
||||
enabled: true
|
||||
##namespace in which you want to deploy servicemonitor
|
||||
##
|
||||
namespace: "monitoring"
|
||||
## Map of additional labels to add to the ServiceMonitor resources
|
||||
# to allow selecting specific ServiceMonitors
|
||||
# in case of multiple prometheus deployments
|
||||
additionalServiceMonitorLabels:
|
||||
release: "prometheus"
|
||||
# key: "value"
|
||||
|
||||
##The following components expose Prometheus metrics and have podmonitors in this chart (disabled by default)
|
||||
##
|
||||
podMonitor:
|
||||
enabled: true
|
||||
##namespace in which you want to deploy podmonitor
|
||||
##
|
||||
namespace: "monitoring"
|
||||
## Map of additional labels to add to the PodMonitor resources
|
||||
# to allow selecting specific PodMonitor
|
||||
# in case of multiple prometheus deployments
|
||||
additionalPodMonitorLabels:
|
||||
release: "monitoring"
|
||||
# key: "value"
|
||||
|
||||
## Enable Grafana Dashboard configmaps for auto dashboard provisioning
|
||||
## If you use kube-prometheus stack for monitoring, these will get imported into grafana
|
||||
grafana:
|
||||
## The namespace in which grafana pod is present
|
||||
namespace: monitoring
|
||||
dashboards:
|
||||
## Disabled by default. switch to true to deploy them
|
||||
enable: true
|
||||
|
||||
# OpenTelemetry is a set of tools for collecting, analyzing, and visualizing
|
||||
# distributed tracing data across function calls.
|
||||
openTelemetry:
|
||||
## Use this flag to set the collector endpoint for OpenTelemetry.
|
||||
## The variable is endpoint of the collector in the format shown below.
|
||||
## otlpCollectorEndpoint: "otel-collector.observability.svc:4317"
|
||||
##
|
||||
otlpCollectorEndpoint: "otel-collector.opentelemetry-operator-system.svc.cluster.local:4317"
|
||||
## Set this flag to false if you are using secure endpoint for the collector.
|
||||
##
|
||||
otlpInsecure: true
|
||||
## Key-value pairs to be used as headers associated with gRPC or HTTP requests to the collector.
|
||||
## Eg. otlpHeaders: "key1=value1,key2=value2"
|
||||
##
|
||||
# otlpHeaders: ""
|
||||
## Supported samplers:
|
||||
## always_on - Sampler that always samples spans, regardless of the parent span's sampling decision.
|
||||
## always_off - Sampler that never samples spans, regardless of the parent span's sampling decision.
|
||||
## traceidratio - Sampler that samples probabalistically based on rate.
|
||||
## parentbased_always_on - (default if empty) Sampler that respects its parent span's sampling decision, but otherwise always samples.
|
||||
## parentbased_always_off - Sampler that respects its parent span's sampling decision, but otherwise never samples.
|
||||
## parentbased_traceidratio - Sampler that respects its parent span's sampling decision, but otherwise samples probabalistically based on rate.
|
||||
## See https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/sdk-environment-variables.md#general-sdk-configuration
|
||||
##
|
||||
tracesSampler: "parentbased_traceidratio"
|
||||
## Each Sampler type defines its own expected input, if any.
|
||||
## Currently we get trace ratio for the case of,
|
||||
## 1. traceidratio
|
||||
## 2. parentbased_traceidratio
|
||||
## Sampling probability, a number in the [0..1] range, e.g. "0.1". Default is 0.1.
|
||||
##
|
||||
tracesSamplingRate: "1"
|
||||
## Supported providers:
|
||||
## tracecontext - W3C Trace Context
|
||||
## baggage - W3C Baggage
|
||||
## b3 - B3 Single
|
||||
## b3multi - B3 Multi
|
||||
## jaeger - Jaeger uber-trace-id header
|
||||
## xray - AWS X-Ray (third party)
|
||||
## ottrace - OpenTracing Trace (third party)
|
||||
## none - No tracing
|
||||
## See https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/sdk-environment-variables.md#general-sdk-configuration
|
||||
##
|
||||
# propagators: "tracecontext,baggage"
|
||||
132
.devcontainer/helm/opentelemetry-collector.yml
Normal file
132
.devcontainer/helm/opentelemetry-collector.yml
Normal file
@@ -0,0 +1,132 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: otel-collector-conf
|
||||
namespace: opentelemetry-operator-system
|
||||
labels:
|
||||
app: opentelemetry
|
||||
component: otel-collector-conf
|
||||
data:
|
||||
otel-collector-config: |
|
||||
receivers:
|
||||
# Make sure to add the otlp receiver.
|
||||
# This will open up the receiver on port 4317
|
||||
otlp:
|
||||
protocols:
|
||||
grpc:
|
||||
endpoint: "0.0.0.0:4317"
|
||||
processors:
|
||||
extensions:
|
||||
health_check: {}
|
||||
exporters:
|
||||
jaeger:
|
||||
# <service-name>.<namespace>.svc.cluster.local:<service-port>
|
||||
endpoint: "jaeger-collector.jaeger.svc.cluster.local:14250"
|
||||
insecure: true
|
||||
prometheus:
|
||||
endpoint: 0.0.0.0:8889
|
||||
namespace: "testapp"
|
||||
logging:
|
||||
|
||||
service:
|
||||
extensions: [ health_check ]
|
||||
pipelines:
|
||||
traces:
|
||||
receivers: [ otlp ]
|
||||
processors: []
|
||||
exporters: [ jaeger ]
|
||||
metrics:
|
||||
receivers: [ otlp ]
|
||||
processors: []
|
||||
exporters: [ prometheus, logging ]
|
||||
---
|
||||
apiVersion: v1
|
||||
kind: Service
|
||||
metadata:
|
||||
name: otel-collector
|
||||
namespace: opentelemetry-operator-system
|
||||
labels:
|
||||
app: opentelemetry
|
||||
component: otel-collector
|
||||
spec:
|
||||
ports:
|
||||
- name: otlp # Default endpoint for otlp receiver.
|
||||
port: 4317
|
||||
protocol: TCP
|
||||
targetPort: 4317
|
||||
nodePort: 30080
|
||||
- name: metrics # Default endpoint for metrics.
|
||||
port: 8889
|
||||
protocol: TCP
|
||||
targetPort: 8889
|
||||
selector:
|
||||
component: otel-collector
|
||||
type: NodePort
|
||||
---
|
||||
apiVersion: apps/v1
|
||||
kind: Deployment
|
||||
metadata:
|
||||
name: otel-collector
|
||||
namespace: opentelemetry-operator-system
|
||||
labels:
|
||||
app: opentelemetry
|
||||
component: otel-collector
|
||||
spec:
|
||||
selector:
|
||||
matchLabels:
|
||||
app: opentelemetry
|
||||
component: otel-collector
|
||||
minReadySeconds: 5
|
||||
progressDeadlineSeconds: 120
|
||||
replicas: 1 #TODO - adjust this to your own requirements
|
||||
template:
|
||||
metadata:
|
||||
annotations:
|
||||
prometheus.io/path: "/metrics"
|
||||
prometheus.io/port: "8889"
|
||||
prometheus.io/scrape: "true"
|
||||
labels:
|
||||
app: opentelemetry
|
||||
component: otel-collector
|
||||
spec:
|
||||
containers:
|
||||
- command:
|
||||
- "/otelcol"
|
||||
- "--config=/conf/otel-collector-config.yaml"
|
||||
# Memory Ballast size should be max 1/3 to 1/2 of memory.
|
||||
- "--mem-ballast-size-mib=683"
|
||||
env:
|
||||
- name: GOGC
|
||||
value: "80"
|
||||
image: otel/opentelemetry-collector:0.6.0
|
||||
name: otel-collector
|
||||
resources:
|
||||
limits:
|
||||
cpu: 1
|
||||
memory: 2Gi
|
||||
requests:
|
||||
cpu: 200m
|
||||
memory: 400Mi
|
||||
ports:
|
||||
- containerPort: 4317 # Default endpoint for otlp receiver.
|
||||
- containerPort: 8889 # Default endpoint for querying metrics.
|
||||
volumeMounts:
|
||||
- name: otel-collector-config-vol
|
||||
mountPath: /conf
|
||||
# - name: otel-collector-secrets
|
||||
# mountPath: /secrets
|
||||
livenessProbe:
|
||||
httpGet:
|
||||
path: /
|
||||
port: 13133 # Health Check extension default port.
|
||||
readinessProbe:
|
||||
httpGet:
|
||||
path: /
|
||||
port: 13133 # Health Check extension default port.
|
||||
volumes:
|
||||
- configMap:
|
||||
name: otel-collector-conf
|
||||
items:
|
||||
- key: otel-collector-config
|
||||
path: otel-collector-config.yaml
|
||||
name: otel-collector-config-vol
|
||||
Reference in New Issue
Block a user