mirror of
https://github.com/ZwareBear/JFrog-Cloud-Installers.git
synced 2026-01-21 02:06:56 -06:00
@@ -7,4 +7,4 @@ scorecard:
|
||||
- olm:
|
||||
cr-manifest:
|
||||
- "deploy/crds/charts.helm.k8s.io_v1alpha1_openshiftartifactoryha_cr.yaml"
|
||||
csv-path: "deploy/olm-catalog/artifactory-ha-operator/1.0.0/artifactory-ha-operator.v1.0.0.clusterserviceversion.yaml"
|
||||
csv-path: "deploy/olm-catalog/artifactory-ha-operator/1.0.2/artifactory-ha-operator.v1.0.2.clusterserviceversion.yaml"
|
||||
|
||||
@@ -1,6 +1,12 @@
|
||||
# JFrog Openshift Artifactory-ha Chart Changelog
|
||||
All changes to this chart will be documented in this file.
|
||||
|
||||
## [3.0.5] - July 16, 2020
|
||||
* Updating to latest jfrog/artifactory-ha helm chart version 3.0.5 artifactory version 7.6.3
|
||||
|
||||
## [2.6.0] - June 29, 2020
|
||||
* Updating to latest jfrog/artifactory-ha helm chart version 2.6.0 artifactory version 7.6.1
|
||||
|
||||
## [2.4.6] - May 12, 2020
|
||||
* Updating to latest jfrog/artifactory-ha helm chart version 2.4.6 artifactory version 7.4.3
|
||||
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
@@ -1,4 +1,4 @@
|
||||
packageName: openshiftartifactoryha-operator
|
||||
channels:
|
||||
- name: alpha
|
||||
currentCSV: artifactory-ha-operator.v1.0.0
|
||||
currentCSV: artifactory-ha-operator.v1.0.2
|
||||
@@ -20,7 +20,7 @@ spec:
|
||||
name: volume
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/artifactory-pro
|
||||
version: 7.4.3
|
||||
version: 7.6.1
|
||||
joinKey: EEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEE
|
||||
masterKey: FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF
|
||||
node:
|
||||
|
||||
@@ -13,7 +13,6 @@ metadata:
|
||||
"spec": {
|
||||
"artifactory-ha": {
|
||||
"artifactory": {
|
||||
"customInitContainersBegin": "- name: \"redhat-custom-setup\"\n image: {{ index .Values \"initContainerImage\" }}\n imagePullPolicy: \"{{ .Values.artifactory.image.pullPolicy }}\"\n command:\n - 'sh'\n - '-c'\n - 'chown -R 1030:1030 {{ .Values.artifactory.persistence.mountPath }}'\n securityContext:\n runAsUser: 0\n volumeMounts:\n - mountPath: \"{{ .Values.artifactory.persistence.mountPath }}\"\n name: volume\n",
|
||||
"image": {
|
||||
"repository": "registry.connect.redhat.com/jfrog/artifactory-pro",
|
||||
"version": "7.4.3"
|
||||
@@ -35,7 +34,6 @@ metadata:
|
||||
"user": "OVERRIDE"
|
||||
},
|
||||
"initContainerImage": "registry.redhat.io/ubi8-minimal",
|
||||
"installerInfo": "{ \"productId\": \"Openshift_artifactory-ha/{{ .Chart.Version }}\", \"features\": [ { \"featureId\": \"ArtifactoryVersion/{{ default .Chart.AppVersion .Values.artifactory.image.version }}\" }, { \"featureId\": \"{{ if .Values.postgresql.enabled }}postgresql{{ else }}{{ .Values.database.type }}{{ end }}/0.0.0\" }, { \"featureId\": \"Platform/Openshift\" }, { \"featureId\": \"Partner/ACC-006983\" }, { \"featureId\": \"Channel/Openshift\" } ] }",
|
||||
"nginx": {
|
||||
"http": {
|
||||
"externalPort": 80,
|
||||
@@ -49,7 +47,6 @@ metadata:
|
||||
"repository": "registry.redhat.io/rhel8/nginx-116",
|
||||
"version": "latest"
|
||||
},
|
||||
"mainConf": "# Main Nginx configuration file\nworker_processes 4;\nerror_log {{ .Values.nginx.persistence.mountPath }}/logs//error.log warn;\npid /tmp/nginx.pid;\nevents {\n worker_connections 1024;\n}\nhttp {\n include /etc/nginx/mime.types;\n default_type application/octet-stream;\n variables_hash_max_size 1024;\n variables_hash_bucket_size 64;\n server_names_hash_max_size 4096;\n server_names_hash_bucket_size 128;\n types_hash_max_size 2048;\n types_hash_bucket_size 64;\n proxy_read_timeout 2400s;\n client_header_timeout 2400s;\n client_body_timeout 2400s;\n proxy_connect_timeout 75s;\n proxy_send_timeout 2400s;\n proxy_buffer_size 32k;\n proxy_buffers 40 32k;\n proxy_busy_buffers_size 64k;\n proxy_temp_file_write_size 250m;\n proxy_http_version 1.1;\n client_body_buffer_size 128k;\n log_format main '$remote_addr - $remote_user [$time_local] \"$request\" '\n '$status $body_bytes_sent \"$http_referer\" '\n '\"$http_user_agent\" \"$http_x_forwarded_for\"';\n log_format timing 'ip = $remote_addr '\n 'user = \\\"$remote_user\\\" '\n 'local_time = \\\"$time_local\\\" '\n 'host = $host '\n 'request = \\\"$request\\\" '\n 'status = $status '\n 'bytes = $body_bytes_sent '\n 'upstream = \\\"$upstream_addr\\\" '\n 'upstream_time = $upstream_response_time '\n 'request_time = $request_time '\n 'referer = \\\"$http_referer\\\" '\n 'UA = \\\"$http_user_agent\\\"';\n access_log {{ .Values.nginx.persistence.mountPath }}/logs/access.log timing;\n sendfile on;\n #tcp_nopush on;\n keepalive_timeout 65;\n #gzip on;\n include {{ .Values.nginx.persistence.mountPath }}/conf.d/*.conf;\n}\n",
|
||||
"service": {
|
||||
"ssloffload": false
|
||||
},
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
@@ -1,6 +1,12 @@
|
||||
# JFrog Openshift Artifactory-ha Chart Changelog
|
||||
All changes to this chart will be documented in this file.
|
||||
|
||||
## [3.0.5] - Jul 16, 2020
|
||||
* Updating to latest jfrog/artifactory helm chart version 3.0.5 artifactory version 7.6.3
|
||||
|
||||
## [2.6.0] - June 29, 2020
|
||||
* Updating to latest jfrog/artifactory helm chart version 2.6.0 artifactory version 7.6.1
|
||||
|
||||
## [2.4.6] - May 12, 2020
|
||||
* Updating to latest jfrog/artifactory-ha helm chart version 2.4.6 artifactory version 7.4.3
|
||||
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
apiVersion: v1
|
||||
appVersion: 7.4.3
|
||||
appVersion: 7.6.3
|
||||
description: Openshift JFrog Artifactory HA subcharting Artifactory HA to work in Openshift environment
|
||||
home: https://www.jfrog.com/artifactory/
|
||||
icon: https://raw.githubusercontent.com/jfrog/charts/master/stable/artifactory-ha/logo/artifactory-logo.png
|
||||
@@ -16,4 +16,4 @@ name: openshift-artifactory-ha
|
||||
sources:
|
||||
- https://bintray.com/jfrog/product/JFrog-Artifactory-Pro/view
|
||||
- https://github.com/jfrog/charts
|
||||
version: 2.4.6
|
||||
version: 3.0.5
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
dependencies:
|
||||
- name: artifactory-ha
|
||||
repository: https://charts.jfrog.io/
|
||||
version: 2.4.6
|
||||
digest: sha256:e0c6b67c9745748aba555b2383d832fee3a977fcde31c5a4f3a5f73f4a357a92
|
||||
generated: "2020-05-12T11:37:46.61737-07:00"
|
||||
version: 3.0.5
|
||||
digest: sha256:59deb56ee27e8a629a22f48cc051453e774999228ece09c77584d95c8c54ce6d
|
||||
generated: "2020-07-16T14:29:16.129919-07:00"
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
dependencies:
|
||||
- name: artifactory-ha
|
||||
version: 2.4.6
|
||||
version: 3.0.5
|
||||
repository: https://charts.jfrog.io/
|
||||
|
||||
@@ -34,7 +34,7 @@ artifactory-ha:
|
||||
## Change to use RH UBI images
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/artifactory-pro
|
||||
version: 7.4.3
|
||||
version: 7.6.3
|
||||
node:
|
||||
replicaCount: 2
|
||||
waitForPrimaryStartup:
|
||||
|
||||
@@ -1,17 +1,8 @@
|
||||
# JFrog Openshift Artifactory-Xray Chart Changelog
|
||||
All changes to this chart will be documented in this file.
|
||||
|
||||
## [2.4.0] - April 14, 2020
|
||||
* Updating to latest jfrog/artifactory-ha helm chart version 2.4.0 adding new requirements.yaml entry for xray helm charts to combine together into one umbrella chart
|
||||
## [4.1.2] July 28, 2020
|
||||
* Updating to Xray chart version 4.1.2 and Xray app version 3.6.2
|
||||
|
||||
## [2.3.0] - April 13, 2020
|
||||
* Updating to latest jfrog/artifactory-ha helm chart version 2.3.0
|
||||
|
||||
## [2.2.9] - April 11, 2020
|
||||
* Fixed issues with master key
|
||||
|
||||
## [2.1.9] - March 17, 2020
|
||||
* Updated Artifactory version to 7.3.2
|
||||
|
||||
## [2.0.35] - March 09, 2020
|
||||
* Updated Artifactory version to 7.2.1
|
||||
## [3.5.1] June 29, 2020
|
||||
* Updating to Xray chart version 3.5.1 and Xray app version 3.5.2
|
||||
@@ -1,5 +1,5 @@
|
||||
apiVersion: v1
|
||||
appVersion: 3.3.0
|
||||
appVersion: 3.6.2
|
||||
description: Universal component scan for security and license inventory and impact analysis
|
||||
sources:
|
||||
- https://bintray.com/jfrog/product/xray/view
|
||||
@@ -13,4 +13,4 @@ maintainers:
|
||||
- email: johnp@jfrog.com
|
||||
name: John Peterson
|
||||
name: openshift-xray
|
||||
version: 3.3.1
|
||||
version: 4.1.2
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
dependencies:
|
||||
- name: xray
|
||||
repository: https://charts.jfrog.io/
|
||||
version: 3.3.1
|
||||
digest: sha256:22010f573f0dfaf95a05835e6b712ef74438aa7c5f39674cd8fd27390bc99d7e
|
||||
generated: "2020-05-21T13:54:18.60088-07:00"
|
||||
version: 4.1.2
|
||||
digest: sha256:79e535f41be683f61d7f181a094d91f2688df43b7c3511be0c5c3216a6ce342b
|
||||
generated: "2020-07-28T11:11:46.534466-07:00"
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
dependencies:
|
||||
- name: xray
|
||||
version: 3.3.1
|
||||
version: 4.1.2
|
||||
repository: https://charts.jfrog.io/
|
||||
|
||||
@@ -32,7 +32,7 @@ xray:
|
||||
name: xray-analysis
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/xray-analysis
|
||||
version: 3.3.0
|
||||
version: 3.6.2
|
||||
updateStrategy: RollingUpdate
|
||||
podManagementPolicy: Parallel
|
||||
preStartCommand:
|
||||
@@ -40,14 +40,14 @@ xray:
|
||||
name: xray-indexer
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/xray-indexer
|
||||
version: 3.3.0
|
||||
version: 3.6.2
|
||||
updateStrategy: RollingUpdate
|
||||
podManagementPolicy: Parallel
|
||||
persist:
|
||||
name: xray-persist
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/xray-persist
|
||||
version: 3.3.0
|
||||
version: 3.6.2
|
||||
updateStrategy: RollingUpdate
|
||||
podManagementPolicy: Parallel
|
||||
persistence:
|
||||
@@ -57,7 +57,7 @@ xray:
|
||||
name: xray-server
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/xray-server
|
||||
version: 3.3.0
|
||||
version: 3.6.2
|
||||
updateStrategy: RollingUpdate
|
||||
podManagementPolicy: Parallel
|
||||
replicaCount: 1
|
||||
@@ -65,7 +65,7 @@ xray:
|
||||
name: router
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/xray-router
|
||||
version: 1.2.1
|
||||
version: 1.4.2
|
||||
imagePullPolicy: IfNotPresent
|
||||
rabbitmq-ha:
|
||||
enabled: true
|
||||
|
||||
@@ -1,6 +1,12 @@
|
||||
# JFrog Openshift Xray Chart Changelog
|
||||
All changes to this chart will be documented in this file.
|
||||
|
||||
## [3.6.2] - July 28, 2020
|
||||
* Deploying JFrog Xray 3.6.2 as an Operator into Openshift
|
||||
|
||||
## [3.5.2] - June 29, 2020
|
||||
* Deploying JFrog Xray 3.5.2 as an Operator into Openshift
|
||||
|
||||
## [3.3.0] - May 22, 2020
|
||||
* Deploying JFrog Xray 3.3.0 as an Operator initial version of Jfrog Xray supported
|
||||
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
packageName: openshiftxray-operator
|
||||
channels:
|
||||
- name: alpha
|
||||
currentCSV: xray-operator.v1.0.0
|
||||
currentCSV: xray-operator.v1.0.2
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
@@ -1,17 +1,8 @@
|
||||
# JFrog Openshift Artifactory-Xray Chart Changelog
|
||||
All changes to this chart will be documented in this file.
|
||||
|
||||
## [2.4.0] - April 14, 2020
|
||||
* Updating to latest jfrog/artifactory-ha helm chart version 2.4.0 adding new requirements.yaml entry for xray helm charts to combine together into one umbrella chart
|
||||
## [4.1.2] July 28, 2020
|
||||
* Updating to Xray chart version 4.1.2 and Xray app version 3.6.2
|
||||
|
||||
## [2.3.0] - April 13, 2020
|
||||
* Updating to latest jfrog/artifactory-ha helm chart version 2.3.0
|
||||
|
||||
## [2.2.9] - April 11, 2020
|
||||
* Fixed issues with master key
|
||||
|
||||
## [2.1.9] - March 17, 2020
|
||||
* Updated Artifactory version to 7.3.2
|
||||
|
||||
## [2.0.35] - March 09, 2020
|
||||
* Updated Artifactory version to 7.2.1
|
||||
## [3.5.1] June 29, 2020
|
||||
* Updating to Xray chart version 3.5.1 and Xray app version 3.5.2
|
||||
@@ -1,5 +1,5 @@
|
||||
apiVersion: v1
|
||||
appVersion: 3.3.0
|
||||
appVersion: 3.6.2
|
||||
description: Universal component scan for security and license inventory and impact
|
||||
analysis
|
||||
keywords:
|
||||
@@ -14,4 +14,4 @@ name: openshift-xray
|
||||
sources:
|
||||
- https://bintray.com/jfrog/product/xray/view
|
||||
- https://github.com/jfrog/charts
|
||||
version: 3.3.1
|
||||
version: 4.1.2
|
||||
|
||||
509
Openshift4/xray-operator/helm-charts/openshift-xray/README.md
Normal file
509
Openshift4/xray-operator/helm-charts/openshift-xray/README.md
Normal file
@@ -0,0 +1,509 @@
|
||||
# JFrog Xray HA on Kubernetes Helm Chart
|
||||
|
||||
## Openshift
|
||||
The Xray chart has been made a subchart of this chart.
|
||||
|
||||
Note due to this change we now reference values through the subchart name as shown below:
|
||||
|
||||
original:
|
||||
```
|
||||
xray.jfrogUrl
|
||||
```
|
||||
|
||||
now:
|
||||
```
|
||||
xray.xray.jfrogUrl
|
||||
```
|
||||
|
||||
This is due to helm referencing the value through the subchart named xray now.
|
||||
|
||||
## Prerequisites Details
|
||||
|
||||
* Kubernetes 1.12+
|
||||
|
||||
## Chart Details
|
||||
|
||||
This chart will do the following:
|
||||
|
||||
* Optionally deploy PostgreSQL
|
||||
* Deploy RabbitMQ (optionally as an HA cluster)
|
||||
* Deploy JFrog Xray micro-services
|
||||
|
||||
## Requirements
|
||||
|
||||
- A running Kubernetes cluster
|
||||
- Dynamic storage provisioning enabled
|
||||
- Default StorageClass set to allow services using the default StorageClass for persistent storage
|
||||
- A running Artifactory
|
||||
- [Kubectl](https://kubernetes.io/docs/tasks/tools/install-kubectl/) installed and setup to use the cluster
|
||||
- [Helm](https://helm.sh/) v2 or v3 installed
|
||||
|
||||
|
||||
## Install JFrog Xray
|
||||
|
||||
### Add JFrog Helm repository
|
||||
|
||||
Before installing JFrog helm charts, you need to add the [JFrog helm repository](https://charts.jfrog.io/) to your helm client
|
||||
|
||||
```bash
|
||||
helm repo add jfrog https://charts.jfrog.io
|
||||
```
|
||||
|
||||
### Install Chart
|
||||
|
||||
#### Artifactory Connection Details
|
||||
|
||||
In order to connect Xray to your Artifactory installation, you have to use a Join Key, hence it is *MANDATORY* to provide a Join Key and Jfrog Url to your Xray installation. Here's how you do that:
|
||||
|
||||
Retrieve the connection details of your Artifactory installation, from the UI - https://www.jfrog.com/confluence/display/JFROG/General+Security+Settings#GeneralSecuritySettings-ViewingtheJoinKey.
|
||||
|
||||
#### Initiate Installation
|
||||
|
||||
Provide join key and jfrog url as a parameter to the Xray chart installation:
|
||||
|
||||
```bash
|
||||
helm upgrade --install --set xray.joinKey=<YOUR_PREVIOUSLY_RETIREVED_JOIN_KEY> \
|
||||
--set xray.jfrogUrl=<YOUR_PREVIOUSLY_RETIREVED_BASE_URL> --namespace xray jfrog/xray
|
||||
```
|
||||
|
||||
Alternatively, you can create a secret containing the join key manually and pass it to the template at install/upgrade time.
|
||||
```bash
|
||||
|
||||
# Create a secret containing the key. The key in the secret must be named join-key
|
||||
kubectl create secret generic my-secret --from-literal=join-key=<YOUR_PREVIOUSLY_RETIREVED_JOIN_KEY>
|
||||
|
||||
# Pass the created secret to helm
|
||||
helm upgrade --install --set xray.joinKeySecretName=my-secret --namespace xray jfrog/xray
|
||||
```
|
||||
**NOTE:** In either case, make sure to pass the same join key on all future calls to `helm install` and `helm upgrade`! This means always passing `--set xray.joinKey=<YOUR_PREVIOUSLY_RETIREVED_JOIN_KEY>`. In the second, this means always passing `--set xray.joinKeySecretName=my-secret` and ensuring the contents of the secret remain unchanged.
|
||||
|
||||
|
||||
### System Configuration
|
||||
|
||||
Xray uses a common system configuration file - `system.yaml`. See [official documentation](https://www.jfrog.com/confluence/display/JFROG/System+YAML+Configuration+File) on its usage.
|
||||
|
||||
## Status
|
||||
|
||||
See the status of your deployed **helm** releases
|
||||
|
||||
```bash
|
||||
helm status xray
|
||||
```
|
||||
|
||||
## Upgrade
|
||||
To upgrade an existing Xray, you still use **helm**
|
||||
|
||||
```bash
|
||||
# Update existing deployed version to 2.1.2
|
||||
helm upgrade --set common.xrayVersion=2.1.2 jfrog/xray
|
||||
```
|
||||
|
||||
If Xray was installed without providing a value to postgresql.postgresqlPassword (a password was autogenerated), follow these instructions:
|
||||
1. Get the current password by running:
|
||||
|
||||
```bash
|
||||
POSTGRES_PASSWORD=$(kubectl get secret -n <namespace> <myrelease>-postgresql -o jsonpath="{.data.postgresql-password}" | base64 --decode)
|
||||
```
|
||||
|
||||
2. Upgrade the release by passing the previously auto-generated secret:
|
||||
|
||||
```bash
|
||||
helm upgrade <myrelease> jfrog/xray --set postgresql.postgresqlPassword=${POSTGRES_PASSWORD}
|
||||
```
|
||||
|
||||
If Xray was installed without providing a value to rabbitmq.rabbitmqPassword/rabbitmq-ha.rabbitmqPassword (a password was autogenerated), follow these instructions:
|
||||
1. Get the current password by running:
|
||||
|
||||
```bash
|
||||
RABBITMQ_PASSWORD=$(kubectl get secret -n <namespace> <myrelease>-rabbitmq -o jsonpath="{.data.rabbitmq-password}" | base64 --decode)
|
||||
```
|
||||
|
||||
2. Upgrade the release by passing the previously auto-generated secret:
|
||||
|
||||
```bash
|
||||
helm upgrade <myrelease> jfrog/xray --set rabbitmq.rabbitmqPassword=${RABBITMQ_PASSWORD}/rabbitmq-ha.rabbitmqPassword=${RABBITMQ_PASSWORD}
|
||||
```
|
||||
|
||||
If Xray was installed with all of the default values (e.g. with no user-provided values for rabbit/postgres), follow these steps:
|
||||
1. Retrieve all current passwords (rabbitmq/postgresql) as explained in the above section.
|
||||
2. Upgrade the release by passing the previously auto-generated secrets:
|
||||
|
||||
```bash
|
||||
helm upgrade --install xray --namespace xray jfrog/xray --set rabbitmq-ha.rabbitmqPassword=<rabbit-password> --set postgresql.postgresqlPassword=<postgresql-password>
|
||||
```
|
||||
|
||||
## Remove
|
||||
|
||||
Removing a **helm** release is done with
|
||||
|
||||
```bash
|
||||
# Remove the Xray services and data tools
|
||||
|
||||
#On helm v2:
|
||||
helm delete --purge xray
|
||||
|
||||
#On helm v3:
|
||||
helm delete xray --namespace xray
|
||||
|
||||
# Remove the data disks
|
||||
kubectl delete pvc -l release=xray
|
||||
```
|
||||
|
||||
### Deploying Xray for small/medium/large instllations
|
||||
In the chart directory, we have added three values files, one for each installation type - small/medium/large. These values files are recommendations for setting resources requests and limits for your installation. The values are derived from the following [documentation](https://www.jfrog.com/confluence/display/EP/Installing+on+Kubernetes#InstallingonKubernetes-Systemrequirements). You can find them in the corresponding chart directory - values-small.yaml, values-medium.yaml and values-large.yaml
|
||||
|
||||
### Create a unique Master Key
|
||||
|
||||
JFrog Xray requires a unique master key to be used by all micro-services in the same cluster. By default the chart has one set in values.yaml (`xray.masterKey`).
|
||||
|
||||
**This key is for demo purpose and should not be used in a production environment!**
|
||||
|
||||
You should generate a unique one and pass it to the template at install/upgrade time.
|
||||
|
||||
```bash
|
||||
# Create a key
|
||||
export MASTER_KEY=$(openssl rand -hex 32)
|
||||
echo ${MASTER_KEY}
|
||||
|
||||
# Pass the created master key to helm
|
||||
helm upgrade --install --set xray.masterKey=${MASTER_KEY} --namespace xray jfrog/xray
|
||||
|
||||
```
|
||||
|
||||
Alternatively, you can create a secret containing the master key manually and pass it to the template at install/upgrade time.
|
||||
```bash
|
||||
# Create a key
|
||||
export MASTER_KEY=$(openssl rand -hex 32)
|
||||
echo ${MASTER_KEY}
|
||||
|
||||
# Create a secret containing the key. The key in the secret must be named master-key
|
||||
kubectl create secret generic my-secret --from-literal=master-key=${MASTER_KEY}
|
||||
|
||||
# Pass the created secret to helm
|
||||
helm upgrade --install xray --set xray.masterKeySecretName=my-secret --namespace xray jfrog/xray
|
||||
```
|
||||
**NOTE:** In either case, make sure to pass the same master key on all future calls to `helm install` and `helm upgrade`! In the first case, this means always passing `--set xray.masterKey=${MASTER_KEY}`. In the second, this means always passing `--set xray.masterKeySecretName=my-secret` and ensuring the contents of the secret remain unchanged.
|
||||
|
||||
|
||||
## Special deployments
|
||||
This is a list of special use cases for non-standard deployments
|
||||
|
||||
### High Availability
|
||||
|
||||
For **high availability** of Xray, set the replica count to be equal or higher than **2**. Recommended is **3**.
|
||||
> It is highly recommended to also set **RabbitMQ** to run as an HA cluster.
|
||||
|
||||
```bash
|
||||
# Start Xray with 3 replicas per service and 3 replicas for RabbitMQ
|
||||
helm upgarde --install xray --namespace xray --set server.replicaCount=3 jfrog/xray
|
||||
```
|
||||
|
||||
### External Databases
|
||||
There is an option to use external PostgreSQL database for your Xray.
|
||||
|
||||
#### PostgreSQL
|
||||
|
||||
##### PostgreSQL without TLS
|
||||
|
||||
To use an external **PostgreSQL**, you need to disable the use of the bundled **PostgreSQL** and set a custom **PostgreSQL** connection URL.
|
||||
|
||||
For this, pass the parameters: `postgresql.enabled=false` and `database.url=${XRAY_POSTGRESQL_CONN_URL}`.
|
||||
|
||||
**IMPORTANT:** Make sure the DB is already created before deploying Xray services
|
||||
|
||||
```bash
|
||||
# Passing a custom PostgreSQL to Xray
|
||||
|
||||
# Example
|
||||
export POSTGRESQL_HOST=custom-postgresql-host
|
||||
export POSTGRESQL_PORT=5432
|
||||
export POSTGRESQL_USER=xray
|
||||
export POSTGRESQL_PASSWORD=password2_X
|
||||
export POSTGRESQL_DATABASE=xraydb
|
||||
|
||||
export XRAY_POSTGRESQL_CONN_URL="postgres://${POSTGRESQL_HOST}:${POSTGRESQL_PORT}/${POSTGRESQL_DATABASE}?sslmode=disable"
|
||||
helm upgrade --install xray --namespace xray \
|
||||
--set postgresql.enabled=false \
|
||||
--set database.url="${XRAY_POSTGRESQL_CONN_URL}" \
|
||||
--set database.user="${POSTGRESQL_USER}" \
|
||||
--set database.password="${POSTGRESQL_PASSWORD}" \
|
||||
jfrog/xray
|
||||
```
|
||||
|
||||
##### PostgreSQL with TLS
|
||||
If external **PostgreSQL** is set with TLS, you need to disable the use of the bundled **PostgreSQL**, set a custom **PostgreSQL** connection URL and provide a secret with **PostgreSQL** TLS certificates.
|
||||
|
||||
Create the Kubernetes secret (assuming the local files are `client-cert.pem client-key.pem server-ca.pem`)
|
||||
|
||||
```bash
|
||||
kubectl create secret generic postgres-tls --from-file=client-key.pem --from-file=client-cert.pem --from-file=server-ca.pem
|
||||
|
||||
```
|
||||
|
||||
**IMPORTANT:** `PostgreSQL` connection URL needs to have listed TLS files with the path `/var/opt/jfrog/xray/data/tls/`
|
||||
and `sslmode==verify-ca` otherwise Xray will fail to connect to Postgres.
|
||||
|
||||
```bash
|
||||
# Passing a custom PostgreSQL with TLS to Xray
|
||||
|
||||
# Example
|
||||
export POSTGRESQL_HOST=custom-postgresql-host
|
||||
export POSTGRESQL_PORT=5432
|
||||
export POSTGRESQL_USER=xray
|
||||
export POSTGRESQL_PASSWORD=password2_X
|
||||
export POSTGRESQL_DATABASE=xraydb
|
||||
export POSTGRESQL_SERVER_CA=server-ca.pem
|
||||
export POSTGRESQL_CLIENT_CERT=client-key.pem
|
||||
export POSTGRESQL_CLIENT_KEY=client-cert.pem
|
||||
export POSTGRESQL_TLS_SECRET=postgres-tls
|
||||
|
||||
export XRAY_POSTGRESQL_CONN_URL="postgres://${POSTGRESQL_HOST}:${POSTGRESQL_PORT}/${POSTGRESQL_DATABASE}?sslrootcert=/var/opt/jfrog/xray/data/tls/${POSTGRESQL_SERVER_CA}&sslkey=/var/opt/jfrog/xray/data/tls/${POSTGRESQL_CLIENT_KEY}&sslcert=/var/opt/jfrog/xray/data/tls/${POSTGRESQL_CLIENT_CERT}&sslmode=verify-ca"
|
||||
helm upgrade --install xray --namespace xray \
|
||||
--set postgresql.enabled=false \
|
||||
--set database.url="${XRAY_POSTGRESQL_CONN_URL}" \
|
||||
--set database.user="${POSTGRESQL_USER}" \
|
||||
--set database.password="${POSTGRESQL_PASSWORD}" \
|
||||
jfrog/xray
|
||||
```
|
||||
|
||||
### Custom init containers
|
||||
|
||||
There are cases where a special, unsupported init processes is needed like checking something on the file system or testing something before spinning up the main container.
|
||||
|
||||
For this, there is a section for writing custom init containers before and after the predefined init containers in the [values.yaml](values.yaml) . By default it's commented out
|
||||
|
||||
```yaml
|
||||
common:
|
||||
## Add custom init containers executed before predefined init containers
|
||||
customInitContainersBegin: |
|
||||
## Init containers template goes here ##
|
||||
|
||||
## Add custom init containers executed after predefined init containers
|
||||
customInitContainers: |
|
||||
## Init containers template goes here ##
|
||||
```
|
||||
|
||||
## Configuration
|
||||
|
||||
The following table lists the configurable parameters of the xray chart and their default values.
|
||||
|
||||
| Parameter | Description | Default |
|
||||
|------------------------------|--------------------------------------------------|------------------------------------|
|
||||
| `imagePullSecrets` | Docker registry pull secret | |
|
||||
| `imagePullPolicy` | Container pull policy | `IfNotPresent` |
|
||||
| `initContainerImage` | Init container image | `alpine:3.6` |
|
||||
| `xray.jfrogUrl` | Main Artifactory URL, without the `/artifactory` prefix .Mandatory | |
|
||||
| `xray.persistence.mountPath` | Xray persistence mount path | `/var/opt/jfrog/xray` |
|
||||
| `xray.masterKey` | Xray Master Key (Can be generated with `openssl rand -hex 32`) | `` |
|
||||
| `xray.masterKeySecretName` | Xray Master Key secret name | |
|
||||
| `xray.joinKey` | Xray Join Key to connect to Artifactory . Mandatory | `` |
|
||||
| `xray.joinKeySecretName` | Xray Join Key secret name | |
|
||||
| `xray.systemYaml` | Xray system configuration (`system.yaml`) as described here - https://www.jfrog.com/confluence/display/JFROG/Xray+System+YAML | |
|
||||
| `xray.autoscaling.enabled` | Enable Xray Pods autoscaling using `HorizontalPodAutoscaler` | `false` |
|
||||
| `xray.autoscaling.minReplicas` | Minimum number of Xray replicas | `1` |
|
||||
| `xray.autoscaling.maxReplicas` | Maximum number of Xray replicas | `1` |
|
||||
| `xray.autoscaling.targetCPUUtilizationPercentage` | CPU usage percentage that will trigger autoscaling | `50` |
|
||||
| `xray.autoscaling.targetMemoryUtilizationPercentage` | Memory usage percentage that will trigger autoscaling | `75` |
|
||||
| `serviceAccount.create` | Specifies whether a ServiceAccount should be created| `true` |
|
||||
| `serviceAccount.name` | The name of the ServiceAccount to create | Generated using the fullname template |
|
||||
| `rbac.create` | Specifies whether RBAC resources should be created | `true` |
|
||||
| `rbac.role.rules` | Rules to create | `[]` |
|
||||
| `postgresql.enabled` | Use enclosed PostgreSQL as database | `true` |
|
||||
| `postgresql.image.registry` | PostgreSQL Docker image registry | `docker.bintray.io` |
|
||||
| `postgresql.image.repository` | PostgreSQL Docker image repository | `bitnami/postgresql` |
|
||||
| `postgresql.image.tag` | PostgreSQL Docker image tag | `9.6.15-debian-9-r91` |
|
||||
| `postgresql.postgresqlUsername` | PostgreSQL database user | `xray` |
|
||||
| `postgresql.postgresqlPassword` | PostgreSQL database password | ` ` |
|
||||
| `postgresql.postgresqlDatabase` | PostgreSQL database name | `xraydb` |
|
||||
| `postgresql.postgresqlExtendedConf.listenAddresses` | PostgreSQL listen address | `"'*'"` |
|
||||
| `postgresql.postgresqlExtendedConf.maxConnections` | PostgreSQL max_connections parameter | `500` |
|
||||
| `postgresql.service.port` | PostgreSQL database port | `5432` |
|
||||
| `postgresql.persistence.enabled` | PostgreSQL use persistent storage | `true` |
|
||||
| `postgresql.persistence.size` | PostgreSQL persistent storage size | `50Gi` |
|
||||
| `postgresql.persistence.existingClaim` | PostgreSQL name of existing Persistent Volume Claim to use | ` ` |
|
||||
| `postgresql.resources.requests.memory` | PostgreSQL initial memory request | |
|
||||
| `postgresql.resources.requests.cpu` | PostgreSQL initial cpu request | |
|
||||
| `postgresql.resources.limits.memory` | PostgreSQL memory limit | |
|
||||
| `postgresql.resources.limits.cpu` | PostgreSQL cpu limit | |
|
||||
| `postgresql.nodeSelector` | PostgreSQL node selector | `{}` |
|
||||
| `postgresql.affinity` | PostgreSQL node affinity | `{}` |
|
||||
| `postgresql.tolerations` | PostgreSQL node tolerations | `[]` |
|
||||
| `database.url` | External database connection URL | |
|
||||
| `database.user` | External database username | |
|
||||
| `database.password` | External database password | |
|
||||
| `database.secrets.user.name` | External database username `Secret` name | |
|
||||
| `database.secrets.user.key` | External database username `Secret` key | |
|
||||
| `database.secrets.password.name` | External database password `Secret` name | |
|
||||
| `database.secrets.password.key` | External database password `Secret` key | |
|
||||
| `database.secrets.url.name` | External database url `Secret` name | |
|
||||
| `database.secrets.url.key` | External database url `Secret` key | |
|
||||
| `rabbitmq.enabled` | RabbitMQ enabled uses rabbitmq | `false` |
|
||||
| `rabbitmq.replicas` | RabbitMQ replica count | `1` |
|
||||
| `rabbitmq.rbacEnabled` | If true, create & use RBAC resources | `true` |
|
||||
| `rabbitmq.rabbitmq.username` | RabbitMQ application username | `guest` |
|
||||
| `rabbitmq.rabbitmq.password` | RabbitMQ application password | |
|
||||
| `rabbitmq.rabbitmq.existingPasswordSecret` | RabbitMQ existingPasswordSecret | |
|
||||
| `rabbitmq.rabbitmq.erlangCookie` | RabbitMQ Erlang cookie | `XRAYRABBITMQCLUSTER`|
|
||||
| `rabbitmq.service.nodePort` | RabbitMQ node port | `5672` |
|
||||
| `rabbitmq.persistence.enabled` | If `true`, persistent volume claims are created | `true` |
|
||||
| `rabbitmq.persistence.accessMode` | RabbitMQ persistent volume claims access mode | `ReadWriteOnce` |
|
||||
| `rabbitmq.persistence.size` | RabbitMQ Persistent volume size | `20Gi` |
|
||||
| `rabbitmq-ha.enabled` | RabbitMQ enabled uses rabbitmq-ha | `true` |
|
||||
| `rabbitmq-ha.replicaCount` | RabbitMQ Number of replica | `1` |
|
||||
| `rabbitmq-ha.rabbitmqUsername` | RabbitMQ application username | `guest` |
|
||||
| `rabbitmq-ha.rabbitmqPassword` | RabbitMQ application password | ` ` |
|
||||
| `rabbitmq-ha.existingSecret` | RabbitMQ existingSecret | ` ` |
|
||||
| `rabbitmq-ha.rabbitmqErlangCookie` | RabbitMQ Erlang cookie | `XRAYRABBITMQCLUSTER`|
|
||||
| `rabbitmq-ha.rabbitmqMemoryHighWatermark` | RabbitMQ Memory high watermark | `500MB` |
|
||||
| `rabbitmq-ha.persistentVolume.enabled` | If `true`, persistent volume claims are created | `true` |
|
||||
| `rabbitmq-ha.persistentVolume.size` | RabbitMQ Persistent volume size | `20Gi` |
|
||||
| `rabbitmq-ha.rbac.create` | If true, create & use RBAC resources | `true` |
|
||||
| `rabbitmq-ha.nodeSelector` | RabbitMQ node selector | `{}` |
|
||||
| `rabbitmq-ha.tolerations` | RabbitMQ node tolerations | `[]` |
|
||||
| `common.xrayVersion` | Xray image tag | `.Chart.AppVersion` |
|
||||
| `common.preStartCommand` | Xray Custom command to run before startup. Runs BEFORE any microservice-specific preStartCommand | |
|
||||
| `common.xrayUserId` | Xray User Id | `1035` |
|
||||
| `common.xrayGroupId` | Xray Group Id | `1035` |
|
||||
| `common.persistence.enabled` | Xray common persistence volume enabled | `false` |
|
||||
| `common.persistence.existingClaim` | Provide an existing PersistentVolumeClaim | `nil` |
|
||||
| `common.persistence.storageClass` | Storage class of backing PVC | `nil (uses default storage class annotation)` |
|
||||
| `common.persistence.accessMode` | Xray common persistence volume access mode | `ReadWriteOnce` |
|
||||
| `common.persistence.size` | Xray common persistence volume size | `50Gi` |
|
||||
| `xray.systemYaml` | Xray system configuration (`system.yaml`) | `see values.yaml` |
|
||||
| `common.customInitContainersBegin` | Custom init containers to run before existing init containers | ` ` |
|
||||
| `common.customInitContainers` | Custom init containers to run after existing init containers | ` ` |
|
||||
| `common.xrayConfig` | Additional xray yaml configuration to be written to xray_config.yaml file | See [values.yaml](stable/xray/values.yaml) |
|
||||
| `database.url` | Xray external PostgreSQL URL | ` ` |
|
||||
| `global.postgresqlTlsSecret` | Xray external PostgreSQL TLS files secret | ` ` |
|
||||
| `analysis.name` | Xray Analysis name | `xray-analysis` |
|
||||
| `analysis.image` | Xray Analysis container image | `docker.bintray.io/jfrog/xray-analysis` |
|
||||
| `analysis.updateStrategy` | Xray Analysis update strategy | `RollingUpdate` |
|
||||
| `analysis.podManagementPolicy` | Xray Analysis pod management policy | `Parallel` |
|
||||
| `analysis.internalPort` | Xray Analysis internal port | `7000` |
|
||||
| `analysis.externalPort` | Xray Analysis external port | `7000` |
|
||||
| `analysis.livenessProbe` | Xray Analysis livenessProbe | See `values.yaml` |
|
||||
| `analysis.readinessProbe` | Xray Analysis readinessProbe | See `values.yaml` |
|
||||
| `analysis.persistence.size` | Xray Analysis storage size limit | `10Gi` |
|
||||
| `analysis.resources` | Xray Analysis resources | `{}` |
|
||||
| `analysis.preStartCommand` | Xray Analysis Custom command to run before startup. Runs AFTER the `common.preStartCommand` | |
|
||||
| `analysis.nodeSelector` | Xray Analysis node selector | `{}` |
|
||||
| `analysis.affinity` | Xray Analysis node affinity | `{}` |
|
||||
| `analysis.tolerations` | Xray Analysis node tolerations | `[]` |
|
||||
| `analysis.annotations` | Xray Analysis annotations | `{}` |
|
||||
| `indexer.name` | Xray Indexer name | `xray-indexer` |
|
||||
| `indexer.image` | Xray Indexer container image | `docker.bintray.io/jfrog/xray-indexer` |
|
||||
| `indexer.annotations` | Xray Indexer annotations | `{}` |
|
||||
| `indexer.updateStrategy` | Xray Indexer update strategy | `RollingUpdate` |
|
||||
| `indexer.podManagementPolicy` | Xray Indexer pod management policy | `Parallel` |
|
||||
| `indexer.internalPort` | Xray Indexer internal port | `7002` |
|
||||
| `indexer.externalPort` | Xray Indexer external port | `7002` |
|
||||
| `indexer.livenessProbe` | Xray Indexer livenessProbe | See `values.yaml` |
|
||||
| `indexer.readinessProbe` | Xray Indexer readinessProbe | See `values.yaml` |
|
||||
| `indexer.customVolumes` | Custom volumes | |
|
||||
| `indexer.customVolumeMounts` | Custom Server volumeMounts | |
|
||||
| `indexer.persistence.existingClaim` | Provide an existing PersistentVolumeClaim | `nil` |
|
||||
| `indexer.persistence.storageClass` | Storage class of backing PVC | `nil (uses default storage class annotation)` |
|
||||
| `indexer.persistence.enabled` | Xray Indexer persistence volume enabled | `false` |
|
||||
| `indexer.persistence.accessMode` | Xray Indexer persistence volume access mode | `ReadWriteOnce` |
|
||||
| `indexer.persistence.size` | Xray Indexer persistence volume size | `50Gi` |
|
||||
| `indexer.resources` | Xray Indexer resources | `{}` |
|
||||
| `indexer.preStartCommand` | Xray Indexer Custom command to run before startup. Runs AFTER the `common.preStartCommand` | |
|
||||
| `indexer.nodeSelector` | Xray Indexer node selector | `{}` |
|
||||
| `indexer.affinity` | Xray Indexer node affinity | `{}` |
|
||||
| `indexer.tolerations` | Xray Indexer node tolerations | `[]` |
|
||||
| `persist.name` | Xray Persist name | `xray-persist` |
|
||||
| `persist.image` | Xray Persist container image | `docker.bintray.io/jfrog/xray-persist` |
|
||||
| `persist.annotations` | Xray Persist annotations | `{}` |
|
||||
| `persist.updateStrategy` | Xray Persist update strategy | `RollingUpdate` |
|
||||
| `persist.podManagementPolicy` | Xray Persist pod management policy | `Parallel` |
|
||||
| `persist.internalPort` | Xray Persist internal port | `7003` |
|
||||
| `persist.externalPort` | Xray Persist external port | `7003` |
|
||||
| `persist.livenessProbe` | Xray Persist livenessProbe | See `values.yaml` |
|
||||
| `persist.readinessProbe` | Xray Persist readinessProbe | See `values.yaml` |
|
||||
| `persist.persistence.size` | Xray Persist storage size limit | `10Gi` |
|
||||
| `persist.preStartCommand` | Xray Persist Custom command to run before startup. Runs AFTER the `common.preStartCommand` | |
|
||||
| `persist.resources` | Xray Persist resources | `{}` |
|
||||
| `persist.nodeSelector` | Xray Persist node selector | `{}` |
|
||||
| `persist.affinity` | Xray Persist node affinity | `{}` |
|
||||
| `persist.tolerations` | Xray Persist node tolerations | `[]` |
|
||||
| `server.name` | Xray server name | `xray-server` |
|
||||
| `server.image` | Xray server container image | `docker.bintray.io/jfrog/xray-server` |
|
||||
| `server.annotations` | Xray server annotations | `{}` |
|
||||
| `server.customVolumes` | Custom volumes | |
|
||||
| `server.customVolumeMounts` | Custom Server volumeMounts | |
|
||||
| `server.replicaCount` | Xray services replica count | `1` |
|
||||
| `server.updateStrategy` | Xray server update strategy | `RollingUpdate` |
|
||||
| `server.podManagementPolicy` | Xray server pod management policy | `Parallel` |
|
||||
| `server.internalPort` | Xray server internal port | `8000` |
|
||||
| `server.externalPort` | Xray server external port | `80` |
|
||||
| `server.service.name` | Xray server service name | `xray` |
|
||||
| `server.service.type` | Xray server service type | `ClusterIP` |
|
||||
| `server.service.annotations` | Xray server service annotations | `{}` |
|
||||
| `server.livenessProbe` | Xray server livenessProbe | See `values.yaml` |
|
||||
| `server.readinessProbe` | Xray server readinessProbe | See `values.yaml` |
|
||||
| `server.preStartCommand` | Xray server Custom command to run before startup. Runs AFTER the `common.preStartCommand` | |
|
||||
| `server.resources` | Xray server resources | `{}` |
|
||||
| `server.nodeSelector` | Xray server node selector | `{}` |
|
||||
| `server.affinity` | Xray server node affinity | `{}` |
|
||||
| `server.tolerations` | Xray server node tolerations | `[]` |
|
||||
| `router.name` | Router name | `router` |
|
||||
| `router.image.repository` | Container image | `docker.bintray.io/jfrog/router` |
|
||||
| `router.image.version` | Container image tag | `.Chart.AppVersion` |
|
||||
| `router.image.pullPolicy` | Container pull policy | `IfNotPresent` |
|
||||
| `router.internalPort` | Router internal port | `8082` |
|
||||
| `router.externalPort` | Router external port | `8082` |
|
||||
| `router.resources.requests.memory` | Router initial memory request | |
|
||||
| `router.resources.requests.cpu` | Router initial cpu request | |
|
||||
| `router.resources.limits.memory` | Router memory limit | |
|
||||
| `router.resources.limits.cpu` | Router cpu limit | |
|
||||
| `router.livenessProbe.enabled` | Enable Router livenessProbe | `true` |
|
||||
| `router.livenessProbe.config` | Router livenessProbe configuration | See `values.yaml` |
|
||||
| `router.readinessProbe.enabled` | Enable Router readinessProbe | `true` |
|
||||
| `router.readinessProbe.config` | Router readinessProbe configuration | See `values.yaml` |
|
||||
| `router.persistence.accessMode` | Router persistence access mode | `ReadWriteOnce` |
|
||||
| `router.persistence.mountPath` | Router persistence mount path | `/var/opt/jfrog/router` |
|
||||
| `router.persistence.size` | Router persistence size | `5Gi` |
|
||||
| `router.readinessProbe.config` | Router readinessProbe configuration | See `values.yaml` |
|
||||
| `router.readinessProbe.config` | Router readinessProbe configuration | See `values.yaml` |
|
||||
| `router.nodeSelector` | Router node selector | `{}` |
|
||||
| `router.affinity` | Router node affinity | `{}` |
|
||||
| `router.tolerations` | Router node tolerations | `[]` |
|
||||
| `filebeat.enabled` | Enable a filebeat container to send your logs to a log management solution like ELK | `false` |
|
||||
| `filebeat.name` | filebeat container name | `xray-filebeat` |
|
||||
| `filebeat.image.repository` | filebeat Docker image repository | `docker.elastic.co/beats/filebeat` |
|
||||
| `filebeat.image.version` | filebeat Docker image version | `7.5.1` |
|
||||
| `filebeat.logstashUrl` | The URL to the central Logstash service, if you have one | `logstash:5044` |
|
||||
| `filebeat.livenessProbe.exec.command` | liveness probe exec command | see [values.yaml](stable/xray/values.yaml) |
|
||||
| `filebeat.livenessProbe.failureThreshold` | Minimum consecutive failures for the probe to be considered failed after having succeeded. | 10 |
|
||||
| `filebeat.livenessProbe.initialDelaySeconds` | Delay before liveness probe is initiated | 180 |
|
||||
| `filebeat.livenessProbe.periodSeconds` | How often to perform the probe | 10 |
|
||||
| `filebeat.readinessProbe.exec.command` | readiness probe exec command | see [values.yaml](stable/xray/values.yaml) |
|
||||
| `filebeat.readinessProbe.failureThreshold` | Minimum consecutive failures for the probe to be considered failed after having succeeded. | 10 |
|
||||
| `filebeat.readinessProbe.initialDelaySeconds` | Delay before readiness probe is initiated | 180 |
|
||||
| `filebeat.readinessProbe.periodSeconds` | How often to perform the probe | 10 |
|
||||
| `filebeat.resources.requests.memory` | Filebeat initial memory request | |
|
||||
| `filebeat.resources.requests.cpu` | Filebeat initial cpu request | |
|
||||
| `filebeat.resources.limits.memory` | Filebeat memory limit | |
|
||||
| `filebeat.resources.limits.cpu` | Filebeat cpu limit | |
|
||||
| `filebeat.filebeatYml` | Filebeat yaml configuration file | see [values.yaml](stable/xray/values.yaml) |
|
||||
|
||||
|
||||
Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`.
|
||||
|
||||
### Custom volumes
|
||||
|
||||
If you need to use a custom volume in a custom init or sidecar container, you can use this option.
|
||||
|
||||
For this, there is a section for defining custom volumes in the [values.yaml](values.yaml). By default it's commented out
|
||||
|
||||
```yaml
|
||||
server:
|
||||
## Add custom volumes
|
||||
customVolumes: |
|
||||
## Custom volume comes here ##
|
||||
```
|
||||
|
||||
## Useful links
|
||||
- https://www.jfrog.com/confluence/display/XRAY/Xray+High+Availability
|
||||
- https://www.jfrog.com/confluence/display/EP/Getting+Started
|
||||
- https://www.jfrog.com/confluence/
|
||||
@@ -1,6 +1,6 @@
|
||||
dependencies:
|
||||
- name: xray
|
||||
repository: https://charts.jfrog.io/
|
||||
version: 3.3.1
|
||||
digest: sha256:22010f573f0dfaf95a05835e6b712ef74438aa7c5f39674cd8fd27390bc99d7e
|
||||
generated: "2020-05-21T13:54:18.60088-07:00"
|
||||
version: 4.1.2
|
||||
digest: sha256:79e535f41be683f61d7f181a094d91f2688df43b7c3511be0c5c3216a6ce342b
|
||||
generated: "2020-07-28T11:11:46.534466-07:00"
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
dependencies:
|
||||
- name: xray
|
||||
version: 3.3.1
|
||||
version: 4.1.2
|
||||
repository: https://charts.jfrog.io/
|
||||
|
||||
@@ -2,7 +2,7 @@ xray:
|
||||
analysis:
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/xray-analysis
|
||||
version: 3.3.0
|
||||
version: 3.6.2
|
||||
name: xray-analysis
|
||||
podManagementPolicy: Parallel
|
||||
preStartCommand: null
|
||||
@@ -16,14 +16,14 @@ xray:
|
||||
indexer:
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/xray-indexer
|
||||
version: 3.3.0
|
||||
version: 3.6.2
|
||||
name: xray-indexer
|
||||
podManagementPolicy: Parallel
|
||||
updateStrategy: RollingUpdate
|
||||
persist:
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/xray-persist
|
||||
version: 3.3.0
|
||||
version: 3.6.2
|
||||
name: xray-persist
|
||||
persistence:
|
||||
size: 10Gi
|
||||
@@ -46,16 +46,17 @@ xray:
|
||||
image:
|
||||
imagePullPolicy: IfNotPresent
|
||||
repository: registry.connect.redhat.com/jfrog/xray-router
|
||||
version: 1.2.1
|
||||
version: 1.4.2
|
||||
name: router
|
||||
server:
|
||||
image:
|
||||
repository: registry.connect.redhat.com/jfrog/xray-server
|
||||
version: 3.3.0
|
||||
version: 3.6.2
|
||||
name: xray-server
|
||||
podManagementPolicy: Parallel
|
||||
replicaCount: 1
|
||||
updateStrategy: RollingUpdate
|
||||
unifiedUpgradeAllowed: true
|
||||
xray:
|
||||
consoleLog: false
|
||||
jfrogUrl: OVERRIDE
|
||||
|
||||
Reference in New Issue
Block a user