inital copy from github repo
This commit is contained in:
30
README.md
Normal file
30
README.md
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
# DockerSetup
|
||||||
|
This repo contains files I've used to setup my personal docker setup, used for my home media library.
|
||||||
|
|
||||||
|
Containers used:
|
||||||
|
|
||||||
|
Tautulli - Plexstats page
|
||||||
|
|
||||||
|
Ombi - used for requesting new media
|
||||||
|
|
||||||
|
nzbget - downloader, usenet
|
||||||
|
|
||||||
|
sonarr - manager, monitors tv shows
|
||||||
|
|
||||||
|
radarr - manager, monitors movies
|
||||||
|
|
||||||
|
jackett - indexer, used to add sites for sonarr/ radarr
|
||||||
|
|
||||||
|
arch-deluge - downloader, uses deluge behind an openvpn connection
|
||||||
|
|
||||||
|
portainer - GUI for managing docker containers
|
||||||
|
|
||||||
|
traefik - reverse proxy, used to host multiple services on one domain
|
||||||
|
|
||||||
|
fail2ban - ban ips after so many failed attempts
|
||||||
|
|
||||||
|
influxdb - data analytics, real-time database for system information
|
||||||
|
|
||||||
|
grafana - data analytics, dashboard for viewing data from influxdb
|
||||||
|
|
||||||
|
telegraf - data analytics, used for feeding docker info into influxdb
|
||||||
623
appdata/grafana/grafana.ini
Normal file
623
appdata/grafana/grafana.ini
Normal file
@@ -0,0 +1,623 @@
|
|||||||
|
##################### Grafana Configuration Example #####################
|
||||||
|
#
|
||||||
|
# Everything has defaults so you only need to uncomment things you want to
|
||||||
|
# change
|
||||||
|
|
||||||
|
# possible values : production, development
|
||||||
|
;app_mode = production
|
||||||
|
|
||||||
|
# instance name, defaults to HOSTNAME environment variable value or hostname if HOSTNAME var is empty
|
||||||
|
;instance_name = ${HOSTNAME}
|
||||||
|
|
||||||
|
#################################### Paths ####################################
|
||||||
|
[paths]
|
||||||
|
# Path to where grafana can store temp files, sessions, and the sqlite3 db (if that is used)
|
||||||
|
;data = /var/lib/grafana
|
||||||
|
|
||||||
|
# Temporary files in `data` directory older than given duration will be removed
|
||||||
|
;temp_data_lifetime = 24h
|
||||||
|
|
||||||
|
# Directory where grafana can store logs
|
||||||
|
;logs = /var/log/grafana
|
||||||
|
|
||||||
|
# Directory where grafana will automatically scan and look for plugins
|
||||||
|
;plugins = /var/lib/grafana/plugins
|
||||||
|
|
||||||
|
# folder that contains provisioning config files that grafana will apply on startup and while running.
|
||||||
|
;provisioning = conf/provisioning
|
||||||
|
|
||||||
|
#################################### Server ####################################
|
||||||
|
[server]
|
||||||
|
# Protocol (http, https, h2, socket)
|
||||||
|
;protocol = http
|
||||||
|
|
||||||
|
# The ip address to bind to, empty will bind to all interfaces
|
||||||
|
;http_addr =
|
||||||
|
|
||||||
|
# The http port to use
|
||||||
|
;http_port = 3000
|
||||||
|
|
||||||
|
# The public facing domain name used to access grafana from a browser
|
||||||
|
;domain = localhost
|
||||||
|
|
||||||
|
# Redirect to correct domain if host header does not match domain
|
||||||
|
# Prevents DNS rebinding attacks
|
||||||
|
;enforce_domain = false
|
||||||
|
|
||||||
|
# The full public facing url you use in browser, used for redirects and emails
|
||||||
|
# If you use reverse proxy and sub path specify full url (with sub path)
|
||||||
|
;root_url = %(protocol)s://%(domain)s:%(http_port)s/
|
||||||
|
|
||||||
|
# Serve Grafana from subpath specified in `root_url` setting. By default it is set to `false` for compatibility reasons.
|
||||||
|
;serve_from_sub_path = false
|
||||||
|
|
||||||
|
# Log web requests
|
||||||
|
;router_logging = false
|
||||||
|
|
||||||
|
# the path relative working path
|
||||||
|
;static_root_path = public
|
||||||
|
|
||||||
|
# enable gzip
|
||||||
|
;enable_gzip = false
|
||||||
|
|
||||||
|
# https certs & key file
|
||||||
|
;cert_file =
|
||||||
|
;cert_key =
|
||||||
|
|
||||||
|
# Unix socket path
|
||||||
|
;socket =
|
||||||
|
|
||||||
|
#################################### Database ####################################
|
||||||
|
[database]
|
||||||
|
# You can configure the database connection by specifying type, host, name, user and password
|
||||||
|
# as separate properties or as on string using the url properties.
|
||||||
|
|
||||||
|
# Either "mysql", "postgres" or "sqlite3", it's your choice
|
||||||
|
;type = sqlite3
|
||||||
|
;host = 127.0.0.1:3306
|
||||||
|
;name = grafana
|
||||||
|
;user = root
|
||||||
|
# If the password contains # or ; you have to wrap it with triple quotes. Ex """#password;"""
|
||||||
|
;password =
|
||||||
|
|
||||||
|
# Use either URL or the previous fields to configure the database
|
||||||
|
# Example: mysql://user:secret@host:port/database
|
||||||
|
;url =
|
||||||
|
|
||||||
|
# For "postgres" only, either "disable", "require" or "verify-full"
|
||||||
|
;ssl_mode = disable
|
||||||
|
|
||||||
|
# For "sqlite3" only, path relative to data_path setting
|
||||||
|
;path = grafana.db
|
||||||
|
|
||||||
|
# Max idle conn setting default is 2
|
||||||
|
;max_idle_conn = 2
|
||||||
|
|
||||||
|
# Max conn setting default is 0 (mean not set)
|
||||||
|
;max_open_conn =
|
||||||
|
|
||||||
|
# Connection Max Lifetime default is 14400 (means 14400 seconds or 4 hours)
|
||||||
|
;conn_max_lifetime = 14400
|
||||||
|
|
||||||
|
# Set to true to log the sql calls and execution times.
|
||||||
|
;log_queries =
|
||||||
|
|
||||||
|
# For "sqlite3" only. cache mode setting used for connecting to the database. (private, shared)
|
||||||
|
;cache_mode = private
|
||||||
|
|
||||||
|
#################################### Cache server #############################
|
||||||
|
[remote_cache]
|
||||||
|
# Either "redis", "memcached" or "database" default is "database"
|
||||||
|
;type = database
|
||||||
|
|
||||||
|
# cache connectionstring options
|
||||||
|
# database: will use Grafana primary database.
|
||||||
|
# redis: config like redis server e.g. `addr=127.0.0.1:6379,pool_size=100,db=0,ssl=false`. Only addr is required. ssl may be 'true', 'false', or 'insecure'.
|
||||||
|
# memcache: 127.0.0.1:11211
|
||||||
|
;connstr =
|
||||||
|
|
||||||
|
#################################### Data proxy ###########################
|
||||||
|
[dataproxy]
|
||||||
|
|
||||||
|
# This enables data proxy logging, default is false
|
||||||
|
;logging = false
|
||||||
|
|
||||||
|
# How long the data proxy should wait before timing out default is 30 (seconds)
|
||||||
|
;timeout = 30
|
||||||
|
|
||||||
|
# If enabled and user is not anonymous, data proxy will add X-Grafana-User header with username into the request, default is false.
|
||||||
|
;send_user_header = false
|
||||||
|
|
||||||
|
#################################### Analytics ####################################
|
||||||
|
[analytics]
|
||||||
|
# Server reporting, sends usage counters to stats.grafana.org every 24 hours.
|
||||||
|
# No ip addresses are being tracked, only simple counters to track
|
||||||
|
# running instances, dashboard and error counts. It is very helpful to us.
|
||||||
|
# Change this option to false to disable reporting.
|
||||||
|
;reporting_enabled = true
|
||||||
|
|
||||||
|
# Set to false to disable all checks to https://grafana.net
|
||||||
|
# for new vesions (grafana itself and plugins), check is used
|
||||||
|
# in some UI views to notify that grafana or plugin update exists
|
||||||
|
# This option does not cause any auto updates, nor send any information
|
||||||
|
# only a GET request to http://grafana.com to get latest versions
|
||||||
|
;check_for_updates = true
|
||||||
|
|
||||||
|
# Google Analytics universal tracking code, only enabled if you specify an id here
|
||||||
|
;google_analytics_ua_id =
|
||||||
|
|
||||||
|
# Google Tag Manager ID, only enabled if you specify an id here
|
||||||
|
;google_tag_manager_id =
|
||||||
|
|
||||||
|
#################################### Security ####################################
|
||||||
|
[security]
|
||||||
|
# disable creation of admin user on first start of grafana
|
||||||
|
;disable_initial_admin_creation = false
|
||||||
|
|
||||||
|
# default admin user, created on startup
|
||||||
|
;admin_user = admin
|
||||||
|
|
||||||
|
# default admin password, can be changed before first start of grafana, or in profile settings
|
||||||
|
;admin_password = admin
|
||||||
|
|
||||||
|
# used for signing
|
||||||
|
;secret_key = SW2YcwTIb9zpOOhoPsMm
|
||||||
|
|
||||||
|
# disable gravatar profile images
|
||||||
|
;disable_gravatar = false
|
||||||
|
|
||||||
|
# data source proxy whitelist (ip_or_domain:port separated by spaces)
|
||||||
|
;data_source_proxy_whitelist =
|
||||||
|
|
||||||
|
# disable protection against brute force login attempts
|
||||||
|
;disable_brute_force_login_protection = false
|
||||||
|
|
||||||
|
# set to true if you host Grafana behind HTTPS. default is false.
|
||||||
|
;cookie_secure = false
|
||||||
|
|
||||||
|
# set cookie SameSite attribute. defaults to `lax`. can be set to "lax", "strict" and "none"
|
||||||
|
;cookie_samesite = lax
|
||||||
|
|
||||||
|
# set to true if you want to allow browsers to render Grafana in a <frame>, <iframe>, <embed> or <object>. default is false.
|
||||||
|
;allow_embedding = false
|
||||||
|
|
||||||
|
# Set to true if you want to enable http strict transport security (HSTS) response header.
|
||||||
|
# This is only sent when HTTPS is enabled in this configuration.
|
||||||
|
# HSTS tells browsers that the site should only be accessed using HTTPS.
|
||||||
|
# The default version will change to true in the next minor release, 6.3.
|
||||||
|
;strict_transport_security = false
|
||||||
|
|
||||||
|
# Sets how long a browser should cache HSTS. Only applied if strict_transport_security is enabled.
|
||||||
|
;strict_transport_security_max_age_seconds = 86400
|
||||||
|
|
||||||
|
# Set to true if to enable HSTS preloading option. Only applied if strict_transport_security is enabled.
|
||||||
|
;strict_transport_security_preload = false
|
||||||
|
|
||||||
|
# Set to true if to enable the HSTS includeSubDomains option. Only applied if strict_transport_security is enabled.
|
||||||
|
;strict_transport_security_subdomains = false
|
||||||
|
|
||||||
|
# Set to true to enable the X-Content-Type-Options response header.
|
||||||
|
# The X-Content-Type-Options response HTTP header is a marker used by the server to indicate that the MIME types advertised
|
||||||
|
# in the Content-Type headers should not be changed and be followed. The default will change to true in the next minor release, 6.3.
|
||||||
|
;x_content_type_options = false
|
||||||
|
|
||||||
|
# Set to true to enable the X-XSS-Protection header, which tells browsers to stop pages from loading
|
||||||
|
# when they detect reflected cross-site scripting (XSS) attacks. The default will change to true in the next minor release, 6.3.
|
||||||
|
;x_xss_protection = false
|
||||||
|
|
||||||
|
#################################### Snapshots ###########################
|
||||||
|
[snapshots]
|
||||||
|
# snapshot sharing options
|
||||||
|
;external_enabled = true
|
||||||
|
;external_snapshot_url = https://snapshots-origin.raintank.io
|
||||||
|
;external_snapshot_name = Publish to snapshot.raintank.io
|
||||||
|
|
||||||
|
# Set to true to enable this Grafana instance act as an external snapshot server and allow unauthenticated requests for
|
||||||
|
# creating and deleting snapshots.
|
||||||
|
;public_mode = false
|
||||||
|
|
||||||
|
# remove expired snapshot
|
||||||
|
;snapshot_remove_expired = true
|
||||||
|
|
||||||
|
#################################### Dashboards History ##################
|
||||||
|
[dashboards]
|
||||||
|
# Number dashboard versions to keep (per dashboard). Default: 20, Minimum: 1
|
||||||
|
;versions_to_keep = 20
|
||||||
|
|
||||||
|
#################################### Users ###############################
|
||||||
|
[users]
|
||||||
|
# disable user signup / registration
|
||||||
|
;allow_sign_up = true
|
||||||
|
|
||||||
|
# Allow non admin users to create organizations
|
||||||
|
;allow_org_create = true
|
||||||
|
|
||||||
|
# Set to true to automatically assign new users to the default organization (id 1)
|
||||||
|
;auto_assign_org = true
|
||||||
|
|
||||||
|
# Default role new users will be automatically assigned (if disabled above is set to true)
|
||||||
|
;auto_assign_org_role = Viewer
|
||||||
|
|
||||||
|
# Background text for the user field on the login page
|
||||||
|
;login_hint = email or username
|
||||||
|
;password_hint = password
|
||||||
|
|
||||||
|
# Default UI theme ("dark" or "light")
|
||||||
|
;default_theme = dark
|
||||||
|
|
||||||
|
# External user management, these options affect the organization users view
|
||||||
|
;external_manage_link_url =
|
||||||
|
;external_manage_link_name =
|
||||||
|
;external_manage_info =
|
||||||
|
|
||||||
|
# Viewers can edit/inspect dashboard settings in the browser. But not save the dashboard.
|
||||||
|
;viewers_can_edit = false
|
||||||
|
|
||||||
|
# Editors can administrate dashboard, folders and teams they create
|
||||||
|
;editors_can_admin = false
|
||||||
|
|
||||||
|
[auth]
|
||||||
|
# Login cookie name
|
||||||
|
;login_cookie_name = grafana_session
|
||||||
|
|
||||||
|
# The lifetime (days) an authenticated user can be inactive before being required to login at next visit. Default is 7 days,
|
||||||
|
;login_maximum_inactive_lifetime_days = 7
|
||||||
|
|
||||||
|
# The maximum lifetime (days) an authenticated user can be logged in since login time before being required to login. Default is 30 days.
|
||||||
|
;login_maximum_lifetime_days = 30
|
||||||
|
|
||||||
|
# How often should auth tokens be rotated for authenticated users when being active. The default is each 10 minutes.
|
||||||
|
;token_rotation_interval_minutes = 10
|
||||||
|
|
||||||
|
# Set to true to disable (hide) the login form, useful if you use OAuth, defaults to false
|
||||||
|
;disable_login_form = false
|
||||||
|
|
||||||
|
# Set to true to disable the signout link in the side menu. useful if you use auth.proxy, defaults to false
|
||||||
|
;disable_signout_menu = false
|
||||||
|
|
||||||
|
# URL to redirect the user to after sign out
|
||||||
|
;signout_redirect_url =
|
||||||
|
|
||||||
|
# Set to true to attempt login with OAuth automatically, skipping the login screen.
|
||||||
|
# This setting is ignored if multiple OAuth providers are configured.
|
||||||
|
;oauth_auto_login = false
|
||||||
|
|
||||||
|
#################################### Anonymous Auth ######################
|
||||||
|
[auth.anonymous]
|
||||||
|
# enable anonymous access
|
||||||
|
enabled = true
|
||||||
|
|
||||||
|
# specify organization name that should be used for unauthenticated users
|
||||||
|
org_name = "Zware's HomeLab"
|
||||||
|
|
||||||
|
# specify role for unauthenticated users
|
||||||
|
;org_role = Viewer
|
||||||
|
|
||||||
|
#################################### Github Auth ##########################
|
||||||
|
[auth.github]
|
||||||
|
;enabled = false
|
||||||
|
;allow_sign_up = true
|
||||||
|
;client_id = some_id
|
||||||
|
;client_secret = some_secret
|
||||||
|
;scopes = user:email,read:org
|
||||||
|
;auth_url = https://github.com/login/oauth/authorize
|
||||||
|
;token_url = https://github.com/login/oauth/access_token
|
||||||
|
;api_url = https://api.github.com/user
|
||||||
|
;team_ids =
|
||||||
|
;allowed_organizations =
|
||||||
|
|
||||||
|
#################################### Google Auth ##########################
|
||||||
|
[auth.google]
|
||||||
|
;enabled = false
|
||||||
|
;allow_sign_up = true
|
||||||
|
;client_id = some_client_id
|
||||||
|
;client_secret = some_client_secret
|
||||||
|
;scopes = https://www.googleapis.com/auth/userinfo.profile https://www.googleapis.com/auth/userinfo.email
|
||||||
|
;auth_url = https://accounts.google.com/o/oauth2/auth
|
||||||
|
;token_url = https://accounts.google.com/o/oauth2/token
|
||||||
|
;api_url = https://www.googleapis.com/oauth2/v1/userinfo
|
||||||
|
;allowed_domains =
|
||||||
|
|
||||||
|
#################################### Generic OAuth ##########################
|
||||||
|
[auth.generic_oauth]
|
||||||
|
;enabled = false
|
||||||
|
;name = OAuth
|
||||||
|
;allow_sign_up = true
|
||||||
|
;client_id = some_id
|
||||||
|
;client_secret = some_secret
|
||||||
|
;scopes = user:email,read:org
|
||||||
|
;email_attribute_name = email:primary
|
||||||
|
;email_attribute_path =
|
||||||
|
;auth_url = https://foo.bar/login/oauth/authorize
|
||||||
|
;token_url = https://foo.bar/login/oauth/access_token
|
||||||
|
;api_url = https://foo.bar/user
|
||||||
|
;team_ids =
|
||||||
|
;allowed_organizations =
|
||||||
|
;role_attribute_path =
|
||||||
|
;tls_skip_verify_insecure = false
|
||||||
|
;tls_client_cert =
|
||||||
|
;tls_client_key =
|
||||||
|
;tls_client_ca =
|
||||||
|
|
||||||
|
; Set to true to enable sending client_id and client_secret via POST body instead of Basic authentication HTTP header
|
||||||
|
; This might be required if the OAuth provider is not RFC6749 compliant, only supporting credentials passed via POST payload
|
||||||
|
;send_client_credentials_via_post = false
|
||||||
|
|
||||||
|
#################################### SAML Auth ###########################
|
||||||
|
[auth.saml] # Enterprise only
|
||||||
|
# Defaults to false. If true, the feature is enabled.
|
||||||
|
;enabled = false
|
||||||
|
|
||||||
|
# Base64-encoded public X.509 certificate. Used to sign requests to the IdP
|
||||||
|
;certificate =
|
||||||
|
|
||||||
|
# Path to the public X.509 certificate. Used to sign requests to the IdP
|
||||||
|
;certificate_path =
|
||||||
|
|
||||||
|
# Base64-encoded private key. Used to decrypt assertions from the IdP
|
||||||
|
;private_key =
|
||||||
|
|
||||||
|
;# Path to the private key. Used to decrypt assertions from the IdP
|
||||||
|
;private_key_path =
|
||||||
|
|
||||||
|
# Base64-encoded IdP SAML metadata XML. Used to verify and obtain binding locations from the IdP
|
||||||
|
;idp_metadata =
|
||||||
|
|
||||||
|
# Path to the SAML metadata XML. Used to verify and obtain binding locations from the IdP
|
||||||
|
;idp_metadata_path =
|
||||||
|
|
||||||
|
# URL to fetch SAML IdP metadata. Used to verify and obtain binding locations from the IdP
|
||||||
|
;idp_metadata_url =
|
||||||
|
|
||||||
|
# Duration, since the IdP issued a response and the SP is allowed to process it. Defaults to 90 seconds.
|
||||||
|
;max_issue_delay = 90s
|
||||||
|
|
||||||
|
# Duration, for how long the SP's metadata should be valid. Defaults to 48 hours.
|
||||||
|
;metadata_valid_duration = 48h
|
||||||
|
|
||||||
|
# Friendly name or name of the attribute within the SAML assertion to use as the user's name
|
||||||
|
;assertion_attribute_name = displayName
|
||||||
|
|
||||||
|
# Friendly name or name of the attribute within the SAML assertion to use as the user's login handle
|
||||||
|
;assertion_attribute_login = mail
|
||||||
|
|
||||||
|
# Friendly name or name of the attribute within the SAML assertion to use as the user's email
|
||||||
|
;assertion_attribute_email = mail
|
||||||
|
|
||||||
|
#################################### Grafana.com Auth ####################
|
||||||
|
[auth.grafana_com]
|
||||||
|
;enabled = false
|
||||||
|
;allow_sign_up = true
|
||||||
|
;client_id = some_id
|
||||||
|
;client_secret = some_secret
|
||||||
|
;scopes = user:email
|
||||||
|
;allowed_organizations =
|
||||||
|
|
||||||
|
#################################### Auth Proxy ##########################
|
||||||
|
[auth.proxy]
|
||||||
|
;enabled = false
|
||||||
|
;header_name = X-WEBAUTH-USER
|
||||||
|
;header_property = username
|
||||||
|
;auto_sign_up = true
|
||||||
|
;sync_ttl = 60
|
||||||
|
;whitelist = 192.168.1.1, 192.168.2.1
|
||||||
|
;headers = Email:X-User-Email, Name:X-User-Name
|
||||||
|
# Read the auth proxy docs for details on what the setting below enables
|
||||||
|
;enable_login_token = false
|
||||||
|
|
||||||
|
#################################### Basic Auth ##########################
|
||||||
|
[auth.basic]
|
||||||
|
;enabled = true
|
||||||
|
|
||||||
|
#################################### Auth LDAP ##########################
|
||||||
|
[auth.ldap]
|
||||||
|
;enabled = false
|
||||||
|
;config_file = /etc/grafana/ldap.toml
|
||||||
|
;allow_sign_up = true
|
||||||
|
|
||||||
|
# LDAP backround sync (Enterprise only)
|
||||||
|
# At 1 am every day
|
||||||
|
;sync_cron = "0 0 1 * * *"
|
||||||
|
;active_sync_enabled = true
|
||||||
|
|
||||||
|
#################################### SMTP / Emailing ##########################
|
||||||
|
[smtp]
|
||||||
|
enabled = true
|
||||||
|
host = localhost:25
|
||||||
|
;user =
|
||||||
|
# If the password contains # or ; you have to wrap it with triple quotes. Ex """#password;"""
|
||||||
|
;password =
|
||||||
|
;cert_file =
|
||||||
|
;key_file =
|
||||||
|
;skip_verify = false
|
||||||
|
from_address = admin@grafana.localhost
|
||||||
|
from_name = Grafana
|
||||||
|
# EHLO identity in SMTP dialog (defaults to instance_name)
|
||||||
|
;ehlo_identity = dashboard.example.com
|
||||||
|
|
||||||
|
[emails]
|
||||||
|
;welcome_email_on_sign_up = false
|
||||||
|
|
||||||
|
#################################### Logging ##########################
|
||||||
|
[log]
|
||||||
|
# Either "console", "file", "syslog". Default is console and file
|
||||||
|
# Use space to separate multiple modes, e.g. "console file"
|
||||||
|
;mode = console file
|
||||||
|
|
||||||
|
# Either "debug", "info", "warn", "error", "critical", default is "info"
|
||||||
|
;level = info
|
||||||
|
|
||||||
|
# optional settings to set different levels for specific loggers. Ex filters = sqlstore:debug
|
||||||
|
;filters =
|
||||||
|
|
||||||
|
# For "console" mode only
|
||||||
|
[log.console]
|
||||||
|
;level =
|
||||||
|
|
||||||
|
# log line format, valid options are text, console and json
|
||||||
|
;format = console
|
||||||
|
|
||||||
|
# For "file" mode only
|
||||||
|
[log.file]
|
||||||
|
;level =
|
||||||
|
|
||||||
|
# log line format, valid options are text, console and json
|
||||||
|
;format = text
|
||||||
|
|
||||||
|
# This enables automated log rotate(switch of following options), default is true
|
||||||
|
;log_rotate = true
|
||||||
|
|
||||||
|
# Max line number of single file, default is 1000000
|
||||||
|
;max_lines = 1000000
|
||||||
|
|
||||||
|
# Max size shift of single file, default is 28 means 1 << 28, 256MB
|
||||||
|
;max_size_shift = 28
|
||||||
|
|
||||||
|
# Segment log daily, default is true
|
||||||
|
;daily_rotate = true
|
||||||
|
|
||||||
|
# Expired days of log file(delete after max days), default is 7
|
||||||
|
;max_days = 7
|
||||||
|
|
||||||
|
[log.syslog]
|
||||||
|
;level =
|
||||||
|
|
||||||
|
# log line format, valid options are text, console and json
|
||||||
|
;format = text
|
||||||
|
|
||||||
|
# Syslog network type and address. This can be udp, tcp, or unix. If left blank, the default unix endpoints will be used.
|
||||||
|
;network =
|
||||||
|
;address =
|
||||||
|
|
||||||
|
# Syslog facility. user, daemon and local0 through local7 are valid.
|
||||||
|
;facility =
|
||||||
|
|
||||||
|
# Syslog tag. By default, the process' argv[0] is used.
|
||||||
|
;tag =
|
||||||
|
|
||||||
|
#################################### Alerting ############################
|
||||||
|
[alerting]
|
||||||
|
# Disable alerting engine & UI features
|
||||||
|
;enabled = true
|
||||||
|
# Makes it possible to turn off alert rule execution but alerting UI is visible
|
||||||
|
;execute_alerts = true
|
||||||
|
|
||||||
|
# Default setting for new alert rules. Defaults to categorize error and timeouts as alerting. (alerting, keep_state)
|
||||||
|
;error_or_timeout = alerting
|
||||||
|
|
||||||
|
# Default setting for how Grafana handles nodata or null values in alerting. (alerting, no_data, keep_state, ok)
|
||||||
|
;nodata_or_nullvalues = no_data
|
||||||
|
|
||||||
|
# Alert notifications can include images, but rendering many images at the same time can overload the server
|
||||||
|
# This limit will protect the server from render overloading and make sure notifications are sent out quickly
|
||||||
|
;concurrent_render_limit = 5
|
||||||
|
|
||||||
|
|
||||||
|
# Default setting for alert calculation timeout. Default value is 30
|
||||||
|
;evaluation_timeout_seconds = 30
|
||||||
|
|
||||||
|
# Default setting for alert notification timeout. Default value is 30
|
||||||
|
;notification_timeout_seconds = 30
|
||||||
|
|
||||||
|
# Default setting for max attempts to sending alert notifications. Default value is 3
|
||||||
|
;max_attempts = 3
|
||||||
|
|
||||||
|
#################################### Explore #############################
|
||||||
|
[explore]
|
||||||
|
# Enable the Explore section
|
||||||
|
;enabled = true
|
||||||
|
|
||||||
|
#################################### Internal Grafana Metrics ##########################
|
||||||
|
# Metrics available at HTTP API Url /metrics
|
||||||
|
[metrics]
|
||||||
|
# Disable / Enable internal metrics
|
||||||
|
;enabled = true
|
||||||
|
# Disable total stats (stat_totals_*) metrics to be generated
|
||||||
|
;disable_total_stats = false
|
||||||
|
|
||||||
|
# Publish interval
|
||||||
|
;interval_seconds = 10
|
||||||
|
|
||||||
|
# Send internal metrics to Graphite
|
||||||
|
[metrics.graphite]
|
||||||
|
# Enable by setting the address setting (ex localhost:2003)
|
||||||
|
;address =
|
||||||
|
;prefix = prod.grafana.%(instance_name)s.
|
||||||
|
|
||||||
|
#################################### Distributed tracing ############
|
||||||
|
[tracing.jaeger]
|
||||||
|
# Enable by setting the address sending traces to jaeger (ex localhost:6831)
|
||||||
|
;address = localhost:6831
|
||||||
|
# Tag that will always be included in when creating new spans. ex (tag1:value1,tag2:value2)
|
||||||
|
;always_included_tag = tag1:value1
|
||||||
|
# Type specifies the type of the sampler: const, probabilistic, rateLimiting, or remote
|
||||||
|
;sampler_type = const
|
||||||
|
# jaeger samplerconfig param
|
||||||
|
# for "const" sampler, 0 or 1 for always false/true respectively
|
||||||
|
# for "probabilistic" sampler, a probability between 0 and 1
|
||||||
|
# for "rateLimiting" sampler, the number of spans per second
|
||||||
|
# for "remote" sampler, param is the same as for "probabilistic"
|
||||||
|
# and indicates the initial sampling rate before the actual one
|
||||||
|
# is received from the mothership
|
||||||
|
;sampler_param = 1
|
||||||
|
# Whether or not to use Zipkin propagation (x-b3- HTTP headers).
|
||||||
|
;zipkin_propagation = false
|
||||||
|
# Setting this to true disables shared RPC spans.
|
||||||
|
# Not disabling is the most common setting when using Zipkin elsewhere in your infrastructure.
|
||||||
|
;disable_shared_zipkin_spans = false
|
||||||
|
|
||||||
|
#################################### Grafana.com integration ##########################
|
||||||
|
# Url used to import dashboards directly from Grafana.com
|
||||||
|
[grafana_com]
|
||||||
|
;url = https://grafana.com
|
||||||
|
|
||||||
|
#################################### External image storage ##########################
|
||||||
|
[external_image_storage]
|
||||||
|
# Used for uploading images to public servers so they can be included in slack/email messages.
|
||||||
|
# you can choose between (s3, webdav, gcs, azure_blob, local)
|
||||||
|
;provider =
|
||||||
|
|
||||||
|
[external_image_storage.s3]
|
||||||
|
;bucket =
|
||||||
|
;region =
|
||||||
|
;path =
|
||||||
|
;access_key =
|
||||||
|
;secret_key =
|
||||||
|
|
||||||
|
[external_image_storage.webdav]
|
||||||
|
;url =
|
||||||
|
;public_url =
|
||||||
|
;username =
|
||||||
|
;password =
|
||||||
|
|
||||||
|
[external_image_storage.gcs]
|
||||||
|
;key_file =
|
||||||
|
;bucket =
|
||||||
|
;path =
|
||||||
|
|
||||||
|
[external_image_storage.azure_blob]
|
||||||
|
;account_name =
|
||||||
|
;account_key =
|
||||||
|
;container_name =
|
||||||
|
|
||||||
|
[external_image_storage.local]
|
||||||
|
# does not require any configuration
|
||||||
|
|
||||||
|
[rendering]
|
||||||
|
# Options to configure a remote HTTP image rendering service, e.g. using https://github.com/grafana/grafana-image-renderer.
|
||||||
|
# URL to a remote HTTP image renderer service, e.g. http://localhost:8081/render, will enable Grafana to render panels and dashboards to PNG-images using HTTP requests to an external service.
|
||||||
|
;server_url =
|
||||||
|
# If the remote HTTP image renderer service runs on a different server than the Grafana server you may have to configure this to a URL where Grafana is reachable, e.g. http://grafana.domain/.
|
||||||
|
;callback_url =
|
||||||
|
|
||||||
|
[enterprise]
|
||||||
|
# Path to a valid Grafana Enterprise license.jwt file
|
||||||
|
;license_path =
|
||||||
|
|
||||||
|
[panels]
|
||||||
|
# If set to true Grafana will allow script tags in text panels. Not recommended as it enable XSS vulnerabilities.
|
||||||
|
;disable_sanitize_html = false
|
||||||
|
|
||||||
|
[plugins]
|
||||||
|
;enable_alpha = false
|
||||||
|
;app_tls_skip_verify_insecure = false
|
||||||
158
appdata/influxdb/influxdb.conf
Normal file
158
appdata/influxdb/influxdb.conf
Normal file
@@ -0,0 +1,158 @@
|
|||||||
|
[admin]
|
||||||
|
enabled = true
|
||||||
|
reporting-disabled = false
|
||||||
|
bind-address = "localhost:8083"
|
||||||
|
https-enabled = false
|
||||||
|
#fucking hell
|
||||||
|
[meta]
|
||||||
|
dir = "/var/lib/influxdb/meta"
|
||||||
|
retention-autocreate = true
|
||||||
|
logging-enabled = true
|
||||||
|
|
||||||
|
[data]
|
||||||
|
dir = "/var/lib/influxdb/data"
|
||||||
|
index-version = "inmem"
|
||||||
|
wal-dir = "/var/lib/influxdb/wal"
|
||||||
|
wal-fsync-delay = "0s"
|
||||||
|
validate-keys = false
|
||||||
|
query-log-enabled = true
|
||||||
|
cache-max-memory-size = 1073741824
|
||||||
|
cache-snapshot-memory-size = 26214400
|
||||||
|
cache-snapshot-write-cold-duration = "10m0s"
|
||||||
|
compact-full-write-cold-duration = "4h0m0s"
|
||||||
|
compact-throughput = 50331648
|
||||||
|
compact-throughput-burst = 50331648
|
||||||
|
max-series-per-database = 1000000
|
||||||
|
max-values-per-tag = 100000
|
||||||
|
max-concurrent-compactions = 0
|
||||||
|
max-index-log-file-size = 1048576
|
||||||
|
series-id-set-cache-size = 100
|
||||||
|
trace-logging-enabled = false
|
||||||
|
tsm-use-madv-willneed = false
|
||||||
|
|
||||||
|
[coordinator]
|
||||||
|
write-timeout = "10s"
|
||||||
|
max-concurrent-queries = 0
|
||||||
|
query-timeout = "0s"
|
||||||
|
log-queries-after = "0s"
|
||||||
|
max-select-point = 0
|
||||||
|
max-select-series = 0
|
||||||
|
max-select-buckets = 0
|
||||||
|
|
||||||
|
[retention]
|
||||||
|
enabled = true
|
||||||
|
check-interval = "30m0s"
|
||||||
|
|
||||||
|
[shard-precreation]
|
||||||
|
enabled = true
|
||||||
|
check-interval = "10m0s"
|
||||||
|
advance-period = "30m0s"
|
||||||
|
|
||||||
|
[monitor]
|
||||||
|
store-enabled = true
|
||||||
|
store-database = "_internal"
|
||||||
|
store-interval = "10s"
|
||||||
|
|
||||||
|
[subscriber]
|
||||||
|
enabled = true
|
||||||
|
http-timeout = "30s"
|
||||||
|
insecure-skip-verify = false
|
||||||
|
ca-certs = ""
|
||||||
|
write-concurrency = 40
|
||||||
|
write-buffer-size = 1000
|
||||||
|
|
||||||
|
[http]
|
||||||
|
enabled = true
|
||||||
|
bind-address = ":8086"
|
||||||
|
auth-enabled = false
|
||||||
|
log-enabled = true
|
||||||
|
suppress-write-log = false
|
||||||
|
write-tracing = false
|
||||||
|
flux-enabled = false
|
||||||
|
flux-log-enabled = false
|
||||||
|
pprof-enabled = true
|
||||||
|
pprof-auth-enabled = false
|
||||||
|
debug-pprof-enabled = false
|
||||||
|
ping-auth-enabled = false
|
||||||
|
https-enabled = false
|
||||||
|
https-certificate = "/etc/ssl/influxdb.pem"
|
||||||
|
https-private-key = ""
|
||||||
|
max-row-limit = 0
|
||||||
|
max-connection-limit = 0
|
||||||
|
shared-secret = ""
|
||||||
|
realm = "InfluxDB"
|
||||||
|
unix-socket-enabled = false
|
||||||
|
unix-socket-permissions = "0777"
|
||||||
|
bind-socket = "/var/run/influxdb.sock"
|
||||||
|
max-body-size = 25000000
|
||||||
|
access-log-path = ""
|
||||||
|
max-concurrent-write-limit = 0
|
||||||
|
max-enqueued-write-limit = 0
|
||||||
|
enqueued-write-timeout = 30000000000
|
||||||
|
|
||||||
|
[logging]
|
||||||
|
format = "auto"
|
||||||
|
level = "info"
|
||||||
|
suppress-logo = false
|
||||||
|
|
||||||
|
[[graphite]]
|
||||||
|
enabled = false
|
||||||
|
bind-address = ":2003"
|
||||||
|
database = "graphite"
|
||||||
|
retention-policy = ""
|
||||||
|
protocol = "tcp"
|
||||||
|
batch-size = 5000
|
||||||
|
batch-pending = 10
|
||||||
|
batch-timeout = "1s"
|
||||||
|
consistency-level = "one"
|
||||||
|
separator = "."
|
||||||
|
udp-read-buffer = 0
|
||||||
|
|
||||||
|
[[collectd]]
|
||||||
|
enabled = false
|
||||||
|
bind-address = ":25826"
|
||||||
|
database = "collectd"
|
||||||
|
retention-policy = ""
|
||||||
|
batch-size = 5000
|
||||||
|
batch-pending = 10
|
||||||
|
batch-timeout = "10s"
|
||||||
|
read-buffer = 0
|
||||||
|
typesdb = "/usr/share/collectd/types.db"
|
||||||
|
security-level = "none"
|
||||||
|
auth-file = "/etc/collectd/auth_file"
|
||||||
|
parse-multivalue-plugin = "split"
|
||||||
|
|
||||||
|
[[opentsdb]]
|
||||||
|
enabled = false
|
||||||
|
bind-address = ":4242"
|
||||||
|
database = "opentsdb"
|
||||||
|
retention-policy = ""
|
||||||
|
consistency-level = "one"
|
||||||
|
tls-enabled = false
|
||||||
|
certificate = "/etc/ssl/influxdb.pem"
|
||||||
|
batch-size = 1000
|
||||||
|
batch-pending = 5
|
||||||
|
batch-timeout = "1s"
|
||||||
|
log-point-errors = true
|
||||||
|
|
||||||
|
[[udp]]
|
||||||
|
enabled = true
|
||||||
|
bind-address = ":8090"
|
||||||
|
database = "proxmox"
|
||||||
|
retention-policy = ""
|
||||||
|
batch-size = 1000
|
||||||
|
batch-pending = 10
|
||||||
|
read-buffer = 0
|
||||||
|
batch-timeout = "1s"
|
||||||
|
precision = ""
|
||||||
|
|
||||||
|
[continuous_queries]
|
||||||
|
log-enabled = true
|
||||||
|
enabled = true
|
||||||
|
query-stats-enabled = false
|
||||||
|
run-interval = "1s"
|
||||||
|
|
||||||
|
[tls]
|
||||||
|
min-version = ""
|
||||||
|
max-version = ""
|
||||||
|
|
||||||
28
appdata/telegraf/telegraf.conf
Normal file
28
appdata/telegraf/telegraf.conf
Normal file
@@ -0,0 +1,28 @@
|
|||||||
|
[global_tags]
|
||||||
|
user = "$USER"
|
||||||
|
[agent]
|
||||||
|
interval = "10s"
|
||||||
|
round_interval = true
|
||||||
|
metric_batch_size = 1000
|
||||||
|
metric_buffer_limit = 10000
|
||||||
|
collection_jitter = "0s"
|
||||||
|
flush_interval = "10s"
|
||||||
|
flush_jitter = "0s"
|
||||||
|
precision = ""
|
||||||
|
debug = false
|
||||||
|
quiet = false
|
||||||
|
hostname = ""
|
||||||
|
omit_hostname = false
|
||||||
|
[[outputs.influxdb]]
|
||||||
|
urls = ["http://192.168.0.115:8086"]
|
||||||
|
database = "telegraf" # required
|
||||||
|
retention_policy = ""
|
||||||
|
write_consistency = "any"
|
||||||
|
timeout = "5s"
|
||||||
|
[[inputs.cpu]]
|
||||||
|
percpu = true
|
||||||
|
totalcpu = true
|
||||||
|
fielddrop = ["time_*"]
|
||||||
|
[[inputs.disk]]
|
||||||
|
ignore_fs = ["tmpfs", "devtmpfs"]
|
||||||
|
[[inputs.diskio]]
|
||||||
486
docker-compose.yml
Normal file
486
docker-compose.yml
Normal file
@@ -0,0 +1,486 @@
|
|||||||
|
# /opt/docker-compose.yml
|
||||||
|
---
|
||||||
|
version: '2'
|
||||||
|
services:
|
||||||
|
#Plex server stats page
|
||||||
|
tautulli:
|
||||||
|
image: linuxserver/tautulli:latest
|
||||||
|
container_name: tautulli
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/tautulli:/config
|
||||||
|
- /opt/appdata/tautulli/Library/Application\ Support/Plex\ Media\ Server/Logs:/logs
|
||||||
|
ports:
|
||||||
|
- 8181:8181
|
||||||
|
networks:
|
||||||
|
- traefik_proxy1
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=true"
|
||||||
|
- "traefik.backend=tautulli"
|
||||||
|
- "traefik.frontend.rule=Host:plexstats.${DOMAINNAME}"
|
||||||
|
- "traefik.port=8181"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "docker.group=external"
|
||||||
|
restart: unless-stopped
|
||||||
|
#Grabber, manage TV shows
|
||||||
|
sonarr:
|
||||||
|
image: linuxserver/sonarr:latest
|
||||||
|
container_name: sonarr
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/sonarr:/config
|
||||||
|
- /mnt/storage/tv:/tv
|
||||||
|
- /downloads/:/downloads
|
||||||
|
- /downloads/complete/tv_import:/tv_import
|
||||||
|
ports:
|
||||||
|
- 27021:8989
|
||||||
|
networks:
|
||||||
|
- traefik_proxy1
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=true"
|
||||||
|
- "traefik.backend=sonarr"
|
||||||
|
- "traefik.frontend.rule=Host:sonarr.${DOMAINNAME}" #
|
||||||
|
- "traefik.port=8989"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "docker.group=external"
|
||||||
|
restart: always
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
#Grabber, manage movies
|
||||||
|
radarr:
|
||||||
|
image: linuxserver/radarr:latest
|
||||||
|
container_name: radarr
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/radarr:/config
|
||||||
|
- /downloads:/downloads
|
||||||
|
- /mnt/storage/movies:/movies
|
||||||
|
- /downloads/watch:/watch
|
||||||
|
ports:
|
||||||
|
- 7878:7878
|
||||||
|
networks:
|
||||||
|
- traefik_proxy1
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=true"
|
||||||
|
- "traefik.backend=radarr"
|
||||||
|
- "traefik.frontend.rule=Host:radarr.${DOMAINNAME}" #
|
||||||
|
- "traefik.port=7878"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "docker.group=external"
|
||||||
|
restart: always
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
#Manager, Used to maintain docker containers
|
||||||
|
portainer:
|
||||||
|
image: portainer/portainer:latest
|
||||||
|
container_name: portainer
|
||||||
|
ports:
|
||||||
|
- 9000:9000
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/portainer:/data
|
||||||
|
- /var/run/docker.sock:/var/run/docker.sock
|
||||||
|
- /opt/appdata/shared:/shared
|
||||||
|
environment:
|
||||||
|
- TZ=${TZ}
|
||||||
|
networks:
|
||||||
|
- traefik_proxy1
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=false"
|
||||||
|
- "traefik.backend=portainer"
|
||||||
|
- "traefik.frontend.rule=Host:portainer.${DOMAINNAME}"
|
||||||
|
- "traefik.port=9000"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "docker.group=internal"
|
||||||
|
restart:
|
||||||
|
always
|
||||||
|
#Indexer, Used by managers to query sites for content
|
||||||
|
jackett:
|
||||||
|
image: linuxserver/jackett:latest
|
||||||
|
container_name: jackett
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/jackett:/config
|
||||||
|
- /downloads:/downloads
|
||||||
|
ports:
|
||||||
|
- 9117:9117
|
||||||
|
networks:
|
||||||
|
- traefik_proxy1
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=true"
|
||||||
|
- "traefik.backend=jackett"
|
||||||
|
- "traefik.frontend.rule=Host:jackett.${DOMAINNAME}" #
|
||||||
|
- "traefik.port=9117"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "docker.group=internal"
|
||||||
|
restart: always
|
||||||
|
#Reverse Proxy, Used to route applications to domain
|
||||||
|
traefik:
|
||||||
|
hostname: traefik
|
||||||
|
image: traefik:v1.7
|
||||||
|
container_name: traefik
|
||||||
|
restart: always
|
||||||
|
domainname: ${DOMAINNAME}
|
||||||
|
networks:
|
||||||
|
- default
|
||||||
|
- traefik_proxy1
|
||||||
|
ports:
|
||||||
|
- "80:80"
|
||||||
|
- "443:443"
|
||||||
|
- "8080:8080"
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=true"
|
||||||
|
- "traefik.backend=traefik"
|
||||||
|
- "traefik.frontend.rule=Host:traefik.${DOMAINNAME}"
|
||||||
|
- "traefik.port=80"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "traefik.frontend.auth.basic.users=${HTTP_USERNAME}:${HTTP_PASSWORD}"
|
||||||
|
- "docker.group=internal"
|
||||||
|
volumes:
|
||||||
|
- /var/run/docker.sock:/var/run/docker.sock:ro
|
||||||
|
- /home/dockeruser/traefik:/etc/traefik
|
||||||
|
- /opt/appdata/shared:/shared
|
||||||
|
- /opt/appdata/traefik:/var/log
|
||||||
|
- /opt/appdata/traefik/tmp:/tmp
|
||||||
|
#fail2ban - ban IPs after so many failed attempts
|
||||||
|
fail2ban:
|
||||||
|
hostname: fail2ban
|
||||||
|
image: crazymax/fail2ban:latest
|
||||||
|
container_name: fail2ban
|
||||||
|
network_mode: "host"
|
||||||
|
cap_add:
|
||||||
|
- NET_ADMIN
|
||||||
|
- NET_RAW
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
volumes:
|
||||||
|
- /var/log:/var/log:ro
|
||||||
|
- /opt/appdata/fail2ban/data:/data
|
||||||
|
- /opt/appdata/shared:/shared
|
||||||
|
restart: always
|
||||||
|
#Downloader, pull in content from torrents & hide traffic behind VPN
|
||||||
|
arch-deluge:
|
||||||
|
image: binhex/arch-delugevpn:latest
|
||||||
|
container_name: arch-deluge
|
||||||
|
ports:
|
||||||
|
- 8112:8112
|
||||||
|
- 8118:8118
|
||||||
|
- 58846:58846
|
||||||
|
- 58946:58946
|
||||||
|
cap_add:
|
||||||
|
- NET_ADMIN
|
||||||
|
networks:
|
||||||
|
- traefik_proxy1
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/arch-deluge:/config
|
||||||
|
- /downloads:/downloads
|
||||||
|
- /mnt/storage:/mnt/storage
|
||||||
|
- /opt/appdata/shared:/shared
|
||||||
|
- /opt/appdata/arch-deluge/data:/data
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
- VPN_ENABLED=yes
|
||||||
|
- VPN_USER=${VPN_USER}
|
||||||
|
- VPN_PASS=${VPN_PASS}
|
||||||
|
- VPN_PROV=custom
|
||||||
|
#- VPN_OPTIONS=<additional openvpn cli options>
|
||||||
|
- STRICT_PORT_FORWARD=yes
|
||||||
|
- ENABLE_PRIVOXY=yes
|
||||||
|
- LAN_NETWORK=192.168.0.100/24
|
||||||
|
- NAME_SERVERS=103.86.96.100, 103.86.99.100
|
||||||
|
- DELUGE_DAEMON_LOG_LEVEL=error
|
||||||
|
- DELUGE_WEB_LOG_LEVEL=error
|
||||||
|
- DEBUG=true
|
||||||
|
- UMASK=000
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=true"
|
||||||
|
- "traefik.backend=arch-deluge"
|
||||||
|
- "traefik.frontend.rule=Host:deluge.${DOMAINNAME}"
|
||||||
|
- "traefik.port=8112"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "docker.group=external"
|
||||||
|
privileged: true
|
||||||
|
restart: unless-stopped
|
||||||
|
#updated version of plex requests
|
||||||
|
ombi:
|
||||||
|
image: linuxserver/ombi
|
||||||
|
container_name: ombi
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
- BASE_URL=/ombi #optional
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/ombi:/config
|
||||||
|
- /opt/appdata/shared:/shared
|
||||||
|
networks:
|
||||||
|
- traefik_proxy1
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=true"
|
||||||
|
- "traefik.backend=ombi"
|
||||||
|
- "traefik.frontend.rule=Host:requests.${DOMAINNAME}" #
|
||||||
|
- "traefik.port=3579"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "docker.group=external"
|
||||||
|
ports:
|
||||||
|
- 3579:3579
|
||||||
|
restart: unless-stopped
|
||||||
|
#NZBGet - usenet downloader
|
||||||
|
nzbget:
|
||||||
|
image: linuxserver/nzbget:latest
|
||||||
|
container_name: nzbget
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/nzbget/config:/config
|
||||||
|
- /opt/appdata/nzbget/data:/data
|
||||||
|
- /downloads:/downloads
|
||||||
|
- /opt/appdata/shared:/shared
|
||||||
|
ports:
|
||||||
|
- 27020:6789
|
||||||
|
networks:
|
||||||
|
- traefik_proxy1
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=true"
|
||||||
|
- "traefik.backend=nzbget"
|
||||||
|
- "traefik.frontend.rule=Host:nzbget.${DOMAINNAME}" #
|
||||||
|
- "traefik.port=6789"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "docker.group=external"
|
||||||
|
restart: always
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
#Data Metrics: https://docs.docker.com/samples/library/influxdb/
|
||||||
|
influxdb:
|
||||||
|
image: influxdb:latest
|
||||||
|
container_name: influxdb
|
||||||
|
restart: always
|
||||||
|
ports:
|
||||||
|
- 8086:8086
|
||||||
|
- 8083:8083
|
||||||
|
- 8090:8090/udp
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/influxdb/lib:/var/lib/influxdb
|
||||||
|
- /opt/appdata/influxdb/:/etc/influxdb/
|
||||||
|
- /opt/appdata/shared:/shared
|
||||||
|
restart:
|
||||||
|
always
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
#Server Metrics: https://github.com/grafana/grafana/tree/master/packaging/docker
|
||||||
|
grafana:
|
||||||
|
image: grafana/grafana:latest
|
||||||
|
container_name: grafana
|
||||||
|
restart: always
|
||||||
|
user: "104"
|
||||||
|
links:
|
||||||
|
- influxdb
|
||||||
|
ports:
|
||||||
|
- 3001:3000
|
||||||
|
networks:
|
||||||
|
- traefik_proxy1
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=true"
|
||||||
|
- "traefik.backend=grafana"
|
||||||
|
- "traefik.frontend.rule=Host:grafana.${DOMAINNAME}" #
|
||||||
|
- "traefik.port=3000"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "docker.group=external"
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/grafana/var:/var/lib/grafana
|
||||||
|
- /opt/appdata/grafana:/etc/grafana
|
||||||
|
- /opt/appdata/shared:/shared
|
||||||
|
restart:
|
||||||
|
always
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
- GF_INSTALL_PLUGINS=${GF_INSTALL_PLUGINS}
|
||||||
|
#Server Metrics: feeds to influxdb
|
||||||
|
telegraf:
|
||||||
|
image: telegraf:latest
|
||||||
|
container_name: telegraf
|
||||||
|
links:
|
||||||
|
- influxdb
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/telegraf:/var/lib/telegraf
|
||||||
|
- /opt/appdata/telegraf/:/etc/telegraf/
|
||||||
|
- /var/run/docker.sock:/var/run/docker.sock
|
||||||
|
restart:
|
||||||
|
always
|
||||||
|
#tdarr - convert / transcode video files
|
||||||
|
tdarr_aio:
|
||||||
|
image: haveagitgat/tdarr_aio
|
||||||
|
container_name: tdarr_aio
|
||||||
|
ports:
|
||||||
|
- 8265:8265
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/tdarr:/home/Tdarr/Documents/Tdarr
|
||||||
|
- /downloads/tdarrDB:/var/lib/mongodb
|
||||||
|
- /mnt/storage:/home/Tdarr/media
|
||||||
|
- /mnt/storage/movies:/home/Tdarr/media/movies
|
||||||
|
- /mnt/storage/tv:/home/Tdarr/media/tv
|
||||||
|
- /downloads/tdarrDB/cache:/home/Tdarr/cache
|
||||||
|
- /etc/localtime:/etc/localtime:ro
|
||||||
|
restart:
|
||||||
|
always
|
||||||
|
lidarr:
|
||||||
|
image: linuxserver/lidarr
|
||||||
|
container_name: lidarr
|
||||||
|
ports:
|
||||||
|
- 8686:8686
|
||||||
|
environment:
|
||||||
|
- PUID=${PUID}
|
||||||
|
- PGID=${PGID}
|
||||||
|
- TZ=${TZ}
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/lidarr:/config
|
||||||
|
- /mnt/storage/music:/music
|
||||||
|
- /downloads:/downloads
|
||||||
|
restart:
|
||||||
|
always
|
||||||
|
gitea:
|
||||||
|
image: gitea/gitea:1.13.0
|
||||||
|
container_name: gitea
|
||||||
|
environment:
|
||||||
|
- USER_UID=${PUID}
|
||||||
|
- USER_GID=${PGID}
|
||||||
|
restart: always
|
||||||
|
networks:
|
||||||
|
- traefik_proxy1
|
||||||
|
labels:
|
||||||
|
- "traefik.enable=true"
|
||||||
|
- "traefik.backend=gitea"
|
||||||
|
- "traefik.frontend.rule=Host:gitea.${DOMAINNAME}" #
|
||||||
|
- "traefik.port=3000"
|
||||||
|
- "traefik.docker.network=traefik_proxy1"
|
||||||
|
- "traefik.frontend.headers.SSLRedirect=true"
|
||||||
|
- "traefik.frontend.headers.STSSeconds=315360000"
|
||||||
|
- "traefik.frontend.headers.browserXSSFilter=true"
|
||||||
|
- "traefik.frontend.headers.contentTypeNosniff=true"
|
||||||
|
- "traefik.frontend.headers.forceSTSHeader=true"
|
||||||
|
- "traefik.frontend.headers.SSLHost=example.com"
|
||||||
|
- "traefik.frontend.headers.STSIncludeSubdomains=true"
|
||||||
|
- "traefik.frontend.headers.STSPreload=true"
|
||||||
|
- "traefik.frontend.headers.frameDeny=true"
|
||||||
|
- "docker.group=external"
|
||||||
|
volumes:
|
||||||
|
- /opt/appdata/gitea:/data
|
||||||
|
- /etc/timezone:/etc/timezone:ro
|
||||||
|
- /etc/localtime:/etc/localtime:ro
|
||||||
|
ports:
|
||||||
|
- "3000:3000"
|
||||||
|
- "222:22"
|
||||||
|
#Networks, Needed for Traefik
|
||||||
|
networks:
|
||||||
|
traefik_proxy1:
|
||||||
|
external: true
|
||||||
|
default:
|
||||||
|
driver: bridge
|
||||||
1
github.git/COMMIT_EDITMSG
Normal file
1
github.git/COMMIT_EDITMSG
Normal file
@@ -0,0 +1 @@
|
|||||||
|
fixed indention
|
||||||
2
github.git/FETCH_HEAD
Normal file
2
github.git/FETCH_HEAD
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
03be7b969b298f884054e5e1e48804578c1d1f36 branch 'master' of https://github.com/ZwareBear/DockerSetup
|
||||||
|
bf9ebf9bd7f0445d9c4446e2e6c7ac23fca88dcb not-for-merge branch 'test' of https://github.com/ZwareBear/DockerSetup
|
||||||
1
github.git/HEAD
Normal file
1
github.git/HEAD
Normal file
@@ -0,0 +1 @@
|
|||||||
|
ref: refs/heads/master
|
||||||
1
github.git/ORIG_HEAD
Normal file
1
github.git/ORIG_HEAD
Normal file
@@ -0,0 +1 @@
|
|||||||
|
dfe7b57536860175ea9c4ab10902bad328786869
|
||||||
11
github.git/config
Normal file
11
github.git/config
Normal file
@@ -0,0 +1,11 @@
|
|||||||
|
[core]
|
||||||
|
repositoryformatversion = 0
|
||||||
|
filemode = true
|
||||||
|
bare = false
|
||||||
|
logallrefupdates = true
|
||||||
|
[remote "origin"]
|
||||||
|
url = https://github.com/ZwareBear/DockerSetup.git
|
||||||
|
fetch = +refs/heads/*:refs/remotes/origin/*
|
||||||
|
[branch "master"]
|
||||||
|
remote = origin
|
||||||
|
merge = refs/heads/master
|
||||||
1
github.git/description
Normal file
1
github.git/description
Normal file
@@ -0,0 +1 @@
|
|||||||
|
Unnamed repository; edit this file 'description' to name the repository.
|
||||||
15
github.git/hooks/applypatch-msg.sample
Executable file
15
github.git/hooks/applypatch-msg.sample
Executable file
@@ -0,0 +1,15 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
#
|
||||||
|
# An example hook script to check the commit log message taken by
|
||||||
|
# applypatch from an e-mail message.
|
||||||
|
#
|
||||||
|
# The hook should exit with non-zero status after issuing an
|
||||||
|
# appropriate message if it wants to stop the commit. The hook is
|
||||||
|
# allowed to edit the commit message file.
|
||||||
|
#
|
||||||
|
# To enable this hook, rename this file to "applypatch-msg".
|
||||||
|
|
||||||
|
. git-sh-setup
|
||||||
|
commitmsg="$(git rev-parse --git-path hooks/commit-msg)"
|
||||||
|
test -x "$commitmsg" && exec "$commitmsg" ${1+"$@"}
|
||||||
|
:
|
||||||
24
github.git/hooks/commit-msg.sample
Executable file
24
github.git/hooks/commit-msg.sample
Executable file
@@ -0,0 +1,24 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
#
|
||||||
|
# An example hook script to check the commit log message.
|
||||||
|
# Called by "git commit" with one argument, the name of the file
|
||||||
|
# that has the commit message. The hook should exit with non-zero
|
||||||
|
# status after issuing an appropriate message if it wants to stop the
|
||||||
|
# commit. The hook is allowed to edit the commit message file.
|
||||||
|
#
|
||||||
|
# To enable this hook, rename this file to "commit-msg".
|
||||||
|
|
||||||
|
# Uncomment the below to add a Signed-off-by line to the message.
|
||||||
|
# Doing this in a hook is a bad idea in general, but the prepare-commit-msg
|
||||||
|
# hook is more suited to it.
|
||||||
|
#
|
||||||
|
# SOB=$(git var GIT_AUTHOR_IDENT | sed -n 's/^\(.*>\).*$/Signed-off-by: \1/p')
|
||||||
|
# grep -qs "^$SOB" "$1" || echo "$SOB" >> "$1"
|
||||||
|
|
||||||
|
# This example catches duplicate Signed-off-by lines.
|
||||||
|
|
||||||
|
test "" = "$(grep '^Signed-off-by: ' "$1" |
|
||||||
|
sort | uniq -c | sed -e '/^[ ]*1[ ]/d')" || {
|
||||||
|
echo >&2 Duplicate Signed-off-by lines.
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
173
github.git/hooks/fsmonitor-watchman.sample
Executable file
173
github.git/hooks/fsmonitor-watchman.sample
Executable file
@@ -0,0 +1,173 @@
|
|||||||
|
#!/usr/bin/perl
|
||||||
|
|
||||||
|
use strict;
|
||||||
|
use warnings;
|
||||||
|
use IPC::Open2;
|
||||||
|
|
||||||
|
# An example hook script to integrate Watchman
|
||||||
|
# (https://facebook.github.io/watchman/) with git to speed up detecting
|
||||||
|
# new and modified files.
|
||||||
|
#
|
||||||
|
# The hook is passed a version (currently 2) and last update token
|
||||||
|
# formatted as a string and outputs to stdout a new update token and
|
||||||
|
# all files that have been modified since the update token. Paths must
|
||||||
|
# be relative to the root of the working tree and separated by a single NUL.
|
||||||
|
#
|
||||||
|
# To enable this hook, rename this file to "query-watchman" and set
|
||||||
|
# 'git config core.fsmonitor .git/hooks/query-watchman'
|
||||||
|
#
|
||||||
|
my ($version, $last_update_token) = @ARGV;
|
||||||
|
|
||||||
|
# Uncomment for debugging
|
||||||
|
# print STDERR "$0 $version $last_update_token\n";
|
||||||
|
|
||||||
|
# Check the hook interface version
|
||||||
|
if ($version ne 2) {
|
||||||
|
die "Unsupported query-fsmonitor hook version '$version'.\n" .
|
||||||
|
"Falling back to scanning...\n";
|
||||||
|
}
|
||||||
|
|
||||||
|
my $git_work_tree = get_working_dir();
|
||||||
|
|
||||||
|
my $retry = 1;
|
||||||
|
|
||||||
|
my $json_pkg;
|
||||||
|
eval {
|
||||||
|
require JSON::XS;
|
||||||
|
$json_pkg = "JSON::XS";
|
||||||
|
1;
|
||||||
|
} or do {
|
||||||
|
require JSON::PP;
|
||||||
|
$json_pkg = "JSON::PP";
|
||||||
|
};
|
||||||
|
|
||||||
|
launch_watchman();
|
||||||
|
|
||||||
|
sub launch_watchman {
|
||||||
|
my $o = watchman_query();
|
||||||
|
if (is_work_tree_watched($o)) {
|
||||||
|
output_result($o->{clock}, @{$o->{files}});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
sub output_result {
|
||||||
|
my ($clockid, @files) = @_;
|
||||||
|
|
||||||
|
# Uncomment for debugging watchman output
|
||||||
|
# open (my $fh, ">", ".git/watchman-output.out");
|
||||||
|
# binmode $fh, ":utf8";
|
||||||
|
# print $fh "$clockid\n@files\n";
|
||||||
|
# close $fh;
|
||||||
|
|
||||||
|
binmode STDOUT, ":utf8";
|
||||||
|
print $clockid;
|
||||||
|
print "\0";
|
||||||
|
local $, = "\0";
|
||||||
|
print @files;
|
||||||
|
}
|
||||||
|
|
||||||
|
sub watchman_clock {
|
||||||
|
my $response = qx/watchman clock "$git_work_tree"/;
|
||||||
|
die "Failed to get clock id on '$git_work_tree'.\n" .
|
||||||
|
"Falling back to scanning...\n" if $? != 0;
|
||||||
|
|
||||||
|
return $json_pkg->new->utf8->decode($response);
|
||||||
|
}
|
||||||
|
|
||||||
|
sub watchman_query {
|
||||||
|
my $pid = open2(\*CHLD_OUT, \*CHLD_IN, 'watchman -j --no-pretty')
|
||||||
|
or die "open2() failed: $!\n" .
|
||||||
|
"Falling back to scanning...\n";
|
||||||
|
|
||||||
|
# In the query expression below we're asking for names of files that
|
||||||
|
# changed since $last_update_token but not from the .git folder.
|
||||||
|
#
|
||||||
|
# To accomplish this, we're using the "since" generator to use the
|
||||||
|
# recency index to select candidate nodes and "fields" to limit the
|
||||||
|
# output to file names only. Then we're using the "expression" term to
|
||||||
|
# further constrain the results.
|
||||||
|
if (substr($last_update_token, 0, 1) eq "c") {
|
||||||
|
$last_update_token = "\"$last_update_token\"";
|
||||||
|
}
|
||||||
|
my $query = <<" END";
|
||||||
|
["query", "$git_work_tree", {
|
||||||
|
"since": $last_update_token,
|
||||||
|
"fields": ["name"],
|
||||||
|
"expression": ["not", ["dirname", ".git"]]
|
||||||
|
}]
|
||||||
|
END
|
||||||
|
|
||||||
|
# Uncomment for debugging the watchman query
|
||||||
|
# open (my $fh, ">", ".git/watchman-query.json");
|
||||||
|
# print $fh $query;
|
||||||
|
# close $fh;
|
||||||
|
|
||||||
|
print CHLD_IN $query;
|
||||||
|
close CHLD_IN;
|
||||||
|
my $response = do {local $/; <CHLD_OUT>};
|
||||||
|
|
||||||
|
# Uncomment for debugging the watch response
|
||||||
|
# open ($fh, ">", ".git/watchman-response.json");
|
||||||
|
# print $fh $response;
|
||||||
|
# close $fh;
|
||||||
|
|
||||||
|
die "Watchman: command returned no output.\n" .
|
||||||
|
"Falling back to scanning...\n" if $response eq "";
|
||||||
|
die "Watchman: command returned invalid output: $response\n" .
|
||||||
|
"Falling back to scanning...\n" unless $response =~ /^\{/;
|
||||||
|
|
||||||
|
return $json_pkg->new->utf8->decode($response);
|
||||||
|
}
|
||||||
|
|
||||||
|
sub is_work_tree_watched {
|
||||||
|
my ($output) = @_;
|
||||||
|
my $error = $output->{error};
|
||||||
|
if ($retry > 0 and $error and $error =~ m/unable to resolve root .* directory (.*) is not watched/) {
|
||||||
|
$retry--;
|
||||||
|
my $response = qx/watchman watch "$git_work_tree"/;
|
||||||
|
die "Failed to make watchman watch '$git_work_tree'.\n" .
|
||||||
|
"Falling back to scanning...\n" if $? != 0;
|
||||||
|
$output = $json_pkg->new->utf8->decode($response);
|
||||||
|
$error = $output->{error};
|
||||||
|
die "Watchman: $error.\n" .
|
||||||
|
"Falling back to scanning...\n" if $error;
|
||||||
|
|
||||||
|
# Uncomment for debugging watchman output
|
||||||
|
# open (my $fh, ">", ".git/watchman-output.out");
|
||||||
|
# close $fh;
|
||||||
|
|
||||||
|
# Watchman will always return all files on the first query so
|
||||||
|
# return the fast "everything is dirty" flag to git and do the
|
||||||
|
# Watchman query just to get it over with now so we won't pay
|
||||||
|
# the cost in git to look up each individual file.
|
||||||
|
my $o = watchman_clock();
|
||||||
|
$error = $output->{error};
|
||||||
|
|
||||||
|
die "Watchman: $error.\n" .
|
||||||
|
"Falling back to scanning...\n" if $error;
|
||||||
|
|
||||||
|
output_result($o->{clock}, ("/"));
|
||||||
|
$last_update_token = $o->{clock};
|
||||||
|
|
||||||
|
eval { launch_watchman() };
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
die "Watchman: $error.\n" .
|
||||||
|
"Falling back to scanning...\n" if $error;
|
||||||
|
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
sub get_working_dir {
|
||||||
|
my $working_dir;
|
||||||
|
if ($^O =~ 'msys' || $^O =~ 'cygwin') {
|
||||||
|
$working_dir = Win32::GetCwd();
|
||||||
|
$working_dir =~ tr/\\/\//;
|
||||||
|
} else {
|
||||||
|
require Cwd;
|
||||||
|
$working_dir = Cwd::cwd();
|
||||||
|
}
|
||||||
|
|
||||||
|
return $working_dir;
|
||||||
|
}
|
||||||
8
github.git/hooks/post-update.sample
Executable file
8
github.git/hooks/post-update.sample
Executable file
@@ -0,0 +1,8 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
#
|
||||||
|
# An example hook script to prepare a packed repository for use over
|
||||||
|
# dumb transports.
|
||||||
|
#
|
||||||
|
# To enable this hook, rename this file to "post-update".
|
||||||
|
|
||||||
|
exec git update-server-info
|
||||||
14
github.git/hooks/pre-applypatch.sample
Executable file
14
github.git/hooks/pre-applypatch.sample
Executable file
@@ -0,0 +1,14 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
#
|
||||||
|
# An example hook script to verify what is about to be committed
|
||||||
|
# by applypatch from an e-mail message.
|
||||||
|
#
|
||||||
|
# The hook should exit with non-zero status after issuing an
|
||||||
|
# appropriate message if it wants to stop the commit.
|
||||||
|
#
|
||||||
|
# To enable this hook, rename this file to "pre-applypatch".
|
||||||
|
|
||||||
|
. git-sh-setup
|
||||||
|
precommit="$(git rev-parse --git-path hooks/pre-commit)"
|
||||||
|
test -x "$precommit" && exec "$precommit" ${1+"$@"}
|
||||||
|
:
|
||||||
49
github.git/hooks/pre-commit.sample
Executable file
49
github.git/hooks/pre-commit.sample
Executable file
@@ -0,0 +1,49 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
#
|
||||||
|
# An example hook script to verify what is about to be committed.
|
||||||
|
# Called by "git commit" with no arguments. The hook should
|
||||||
|
# exit with non-zero status after issuing an appropriate message if
|
||||||
|
# it wants to stop the commit.
|
||||||
|
#
|
||||||
|
# To enable this hook, rename this file to "pre-commit".
|
||||||
|
|
||||||
|
if git rev-parse --verify HEAD >/dev/null 2>&1
|
||||||
|
then
|
||||||
|
against=HEAD
|
||||||
|
else
|
||||||
|
# Initial commit: diff against an empty tree object
|
||||||
|
against=$(git hash-object -t tree /dev/null)
|
||||||
|
fi
|
||||||
|
|
||||||
|
# If you want to allow non-ASCII filenames set this variable to true.
|
||||||
|
allownonascii=$(git config --type=bool hooks.allownonascii)
|
||||||
|
|
||||||
|
# Redirect output to stderr.
|
||||||
|
exec 1>&2
|
||||||
|
|
||||||
|
# Cross platform projects tend to avoid non-ASCII filenames; prevent
|
||||||
|
# them from being added to the repository. We exploit the fact that the
|
||||||
|
# printable range starts at the space character and ends with tilde.
|
||||||
|
if [ "$allownonascii" != "true" ] &&
|
||||||
|
# Note that the use of brackets around a tr range is ok here, (it's
|
||||||
|
# even required, for portability to Solaris 10's /usr/bin/tr), since
|
||||||
|
# the square bracket bytes happen to fall in the designated range.
|
||||||
|
test $(git diff --cached --name-only --diff-filter=A -z $against |
|
||||||
|
LC_ALL=C tr -d '[ -~]\0' | wc -c) != 0
|
||||||
|
then
|
||||||
|
cat <<\EOF
|
||||||
|
Error: Attempt to add a non-ASCII file name.
|
||||||
|
|
||||||
|
This can cause problems if you want to work with people on other platforms.
|
||||||
|
|
||||||
|
To be portable it is advisable to rename the file.
|
||||||
|
|
||||||
|
If you know what you are doing you can disable this check using:
|
||||||
|
|
||||||
|
git config hooks.allownonascii true
|
||||||
|
EOF
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
# If there are whitespace errors, print the offending file names and fail.
|
||||||
|
exec git diff-index --check --cached $against --
|
||||||
13
github.git/hooks/pre-merge-commit.sample
Executable file
13
github.git/hooks/pre-merge-commit.sample
Executable file
@@ -0,0 +1,13 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
#
|
||||||
|
# An example hook script to verify what is about to be committed.
|
||||||
|
# Called by "git merge" with no arguments. The hook should
|
||||||
|
# exit with non-zero status after issuing an appropriate message to
|
||||||
|
# stderr if it wants to stop the merge commit.
|
||||||
|
#
|
||||||
|
# To enable this hook, rename this file to "pre-merge-commit".
|
||||||
|
|
||||||
|
. git-sh-setup
|
||||||
|
test -x "$GIT_DIR/hooks/pre-commit" &&
|
||||||
|
exec "$GIT_DIR/hooks/pre-commit"
|
||||||
|
:
|
||||||
53
github.git/hooks/pre-push.sample
Executable file
53
github.git/hooks/pre-push.sample
Executable file
@@ -0,0 +1,53 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
# An example hook script to verify what is about to be pushed. Called by "git
|
||||||
|
# push" after it has checked the remote status, but before anything has been
|
||||||
|
# pushed. If this script exits with a non-zero status nothing will be pushed.
|
||||||
|
#
|
||||||
|
# This hook is called with the following parameters:
|
||||||
|
#
|
||||||
|
# $1 -- Name of the remote to which the push is being done
|
||||||
|
# $2 -- URL to which the push is being done
|
||||||
|
#
|
||||||
|
# If pushing without using a named remote those arguments will be equal.
|
||||||
|
#
|
||||||
|
# Information about the commits which are being pushed is supplied as lines to
|
||||||
|
# the standard input in the form:
|
||||||
|
#
|
||||||
|
# <local ref> <local sha1> <remote ref> <remote sha1>
|
||||||
|
#
|
||||||
|
# This sample shows how to prevent push of commits where the log message starts
|
||||||
|
# with "WIP" (work in progress).
|
||||||
|
|
||||||
|
remote="$1"
|
||||||
|
url="$2"
|
||||||
|
|
||||||
|
z40=0000000000000000000000000000000000000000
|
||||||
|
|
||||||
|
while read local_ref local_sha remote_ref remote_sha
|
||||||
|
do
|
||||||
|
if [ "$local_sha" = $z40 ]
|
||||||
|
then
|
||||||
|
# Handle delete
|
||||||
|
:
|
||||||
|
else
|
||||||
|
if [ "$remote_sha" = $z40 ]
|
||||||
|
then
|
||||||
|
# New branch, examine all commits
|
||||||
|
range="$local_sha"
|
||||||
|
else
|
||||||
|
# Update to existing branch, examine new commits
|
||||||
|
range="$remote_sha..$local_sha"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check for WIP commit
|
||||||
|
commit=`git rev-list -n 1 --grep '^WIP' "$range"`
|
||||||
|
if [ -n "$commit" ]
|
||||||
|
then
|
||||||
|
echo >&2 "Found WIP commit in $local_ref, not pushing"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
exit 0
|
||||||
169
github.git/hooks/pre-rebase.sample
Executable file
169
github.git/hooks/pre-rebase.sample
Executable file
@@ -0,0 +1,169 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
#
|
||||||
|
# Copyright (c) 2006, 2008 Junio C Hamano
|
||||||
|
#
|
||||||
|
# The "pre-rebase" hook is run just before "git rebase" starts doing
|
||||||
|
# its job, and can prevent the command from running by exiting with
|
||||||
|
# non-zero status.
|
||||||
|
#
|
||||||
|
# The hook is called with the following parameters:
|
||||||
|
#
|
||||||
|
# $1 -- the upstream the series was forked from.
|
||||||
|
# $2 -- the branch being rebased (or empty when rebasing the current branch).
|
||||||
|
#
|
||||||
|
# This sample shows how to prevent topic branches that are already
|
||||||
|
# merged to 'next' branch from getting rebased, because allowing it
|
||||||
|
# would result in rebasing already published history.
|
||||||
|
|
||||||
|
publish=next
|
||||||
|
basebranch="$1"
|
||||||
|
if test "$#" = 2
|
||||||
|
then
|
||||||
|
topic="refs/heads/$2"
|
||||||
|
else
|
||||||
|
topic=`git symbolic-ref HEAD` ||
|
||||||
|
exit 0 ;# we do not interrupt rebasing detached HEAD
|
||||||
|
fi
|
||||||
|
|
||||||
|
case "$topic" in
|
||||||
|
refs/heads/??/*)
|
||||||
|
;;
|
||||||
|
*)
|
||||||
|
exit 0 ;# we do not interrupt others.
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
|
||||||
|
# Now we are dealing with a topic branch being rebased
|
||||||
|
# on top of master. Is it OK to rebase it?
|
||||||
|
|
||||||
|
# Does the topic really exist?
|
||||||
|
git show-ref -q "$topic" || {
|
||||||
|
echo >&2 "No such branch $topic"
|
||||||
|
exit 1
|
||||||
|
}
|
||||||
|
|
||||||
|
# Is topic fully merged to master?
|
||||||
|
not_in_master=`git rev-list --pretty=oneline ^master "$topic"`
|
||||||
|
if test -z "$not_in_master"
|
||||||
|
then
|
||||||
|
echo >&2 "$topic is fully merged to master; better remove it."
|
||||||
|
exit 1 ;# we could allow it, but there is no point.
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Is topic ever merged to next? If so you should not be rebasing it.
|
||||||
|
only_next_1=`git rev-list ^master "^$topic" ${publish} | sort`
|
||||||
|
only_next_2=`git rev-list ^master ${publish} | sort`
|
||||||
|
if test "$only_next_1" = "$only_next_2"
|
||||||
|
then
|
||||||
|
not_in_topic=`git rev-list "^$topic" master`
|
||||||
|
if test -z "$not_in_topic"
|
||||||
|
then
|
||||||
|
echo >&2 "$topic is already up to date with master"
|
||||||
|
exit 1 ;# we could allow it, but there is no point.
|
||||||
|
else
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
not_in_next=`git rev-list --pretty=oneline ^${publish} "$topic"`
|
||||||
|
/usr/bin/perl -e '
|
||||||
|
my $topic = $ARGV[0];
|
||||||
|
my $msg = "* $topic has commits already merged to public branch:\n";
|
||||||
|
my (%not_in_next) = map {
|
||||||
|
/^([0-9a-f]+) /;
|
||||||
|
($1 => 1);
|
||||||
|
} split(/\n/, $ARGV[1]);
|
||||||
|
for my $elem (map {
|
||||||
|
/^([0-9a-f]+) (.*)$/;
|
||||||
|
[$1 => $2];
|
||||||
|
} split(/\n/, $ARGV[2])) {
|
||||||
|
if (!exists $not_in_next{$elem->[0]}) {
|
||||||
|
if ($msg) {
|
||||||
|
print STDERR $msg;
|
||||||
|
undef $msg;
|
||||||
|
}
|
||||||
|
print STDERR " $elem->[1]\n";
|
||||||
|
}
|
||||||
|
}
|
||||||
|
' "$topic" "$not_in_next" "$not_in_master"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
<<\DOC_END
|
||||||
|
|
||||||
|
This sample hook safeguards topic branches that have been
|
||||||
|
published from being rewound.
|
||||||
|
|
||||||
|
The workflow assumed here is:
|
||||||
|
|
||||||
|
* Once a topic branch forks from "master", "master" is never
|
||||||
|
merged into it again (either directly or indirectly).
|
||||||
|
|
||||||
|
* Once a topic branch is fully cooked and merged into "master",
|
||||||
|
it is deleted. If you need to build on top of it to correct
|
||||||
|
earlier mistakes, a new topic branch is created by forking at
|
||||||
|
the tip of the "master". This is not strictly necessary, but
|
||||||
|
it makes it easier to keep your history simple.
|
||||||
|
|
||||||
|
* Whenever you need to test or publish your changes to topic
|
||||||
|
branches, merge them into "next" branch.
|
||||||
|
|
||||||
|
The script, being an example, hardcodes the publish branch name
|
||||||
|
to be "next", but it is trivial to make it configurable via
|
||||||
|
$GIT_DIR/config mechanism.
|
||||||
|
|
||||||
|
With this workflow, you would want to know:
|
||||||
|
|
||||||
|
(1) ... if a topic branch has ever been merged to "next". Young
|
||||||
|
topic branches can have stupid mistakes you would rather
|
||||||
|
clean up before publishing, and things that have not been
|
||||||
|
merged into other branches can be easily rebased without
|
||||||
|
affecting other people. But once it is published, you would
|
||||||
|
not want to rewind it.
|
||||||
|
|
||||||
|
(2) ... if a topic branch has been fully merged to "master".
|
||||||
|
Then you can delete it. More importantly, you should not
|
||||||
|
build on top of it -- other people may already want to
|
||||||
|
change things related to the topic as patches against your
|
||||||
|
"master", so if you need further changes, it is better to
|
||||||
|
fork the topic (perhaps with the same name) afresh from the
|
||||||
|
tip of "master".
|
||||||
|
|
||||||
|
Let's look at this example:
|
||||||
|
|
||||||
|
o---o---o---o---o---o---o---o---o---o "next"
|
||||||
|
/ / / /
|
||||||
|
/ a---a---b A / /
|
||||||
|
/ / / /
|
||||||
|
/ / c---c---c---c B /
|
||||||
|
/ / / \ /
|
||||||
|
/ / / b---b C \ /
|
||||||
|
/ / / / \ /
|
||||||
|
---o---o---o---o---o---o---o---o---o---o---o "master"
|
||||||
|
|
||||||
|
|
||||||
|
A, B and C are topic branches.
|
||||||
|
|
||||||
|
* A has one fix since it was merged up to "next".
|
||||||
|
|
||||||
|
* B has finished. It has been fully merged up to "master" and "next",
|
||||||
|
and is ready to be deleted.
|
||||||
|
|
||||||
|
* C has not merged to "next" at all.
|
||||||
|
|
||||||
|
We would want to allow C to be rebased, refuse A, and encourage
|
||||||
|
B to be deleted.
|
||||||
|
|
||||||
|
To compute (1):
|
||||||
|
|
||||||
|
git rev-list ^master ^topic next
|
||||||
|
git rev-list ^master next
|
||||||
|
|
||||||
|
if these match, topic has not merged in next at all.
|
||||||
|
|
||||||
|
To compute (2):
|
||||||
|
|
||||||
|
git rev-list master..topic
|
||||||
|
|
||||||
|
if this is empty, it is fully merged to "master".
|
||||||
|
|
||||||
|
DOC_END
|
||||||
24
github.git/hooks/pre-receive.sample
Executable file
24
github.git/hooks/pre-receive.sample
Executable file
@@ -0,0 +1,24 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
#
|
||||||
|
# An example hook script to make use of push options.
|
||||||
|
# The example simply echoes all push options that start with 'echoback='
|
||||||
|
# and rejects all pushes when the "reject" push option is used.
|
||||||
|
#
|
||||||
|
# To enable this hook, rename this file to "pre-receive".
|
||||||
|
|
||||||
|
if test -n "$GIT_PUSH_OPTION_COUNT"
|
||||||
|
then
|
||||||
|
i=0
|
||||||
|
while test "$i" -lt "$GIT_PUSH_OPTION_COUNT"
|
||||||
|
do
|
||||||
|
eval "value=\$GIT_PUSH_OPTION_$i"
|
||||||
|
case "$value" in
|
||||||
|
echoback=*)
|
||||||
|
echo "echo from the pre-receive-hook: ${value#*=}" >&2
|
||||||
|
;;
|
||||||
|
reject)
|
||||||
|
exit 1
|
||||||
|
esac
|
||||||
|
i=$((i + 1))
|
||||||
|
done
|
||||||
|
fi
|
||||||
42
github.git/hooks/prepare-commit-msg.sample
Executable file
42
github.git/hooks/prepare-commit-msg.sample
Executable file
@@ -0,0 +1,42 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
#
|
||||||
|
# An example hook script to prepare the commit log message.
|
||||||
|
# Called by "git commit" with the name of the file that has the
|
||||||
|
# commit message, followed by the description of the commit
|
||||||
|
# message's source. The hook's purpose is to edit the commit
|
||||||
|
# message file. If the hook fails with a non-zero status,
|
||||||
|
# the commit is aborted.
|
||||||
|
#
|
||||||
|
# To enable this hook, rename this file to "prepare-commit-msg".
|
||||||
|
|
||||||
|
# This hook includes three examples. The first one removes the
|
||||||
|
# "# Please enter the commit message..." help message.
|
||||||
|
#
|
||||||
|
# The second includes the output of "git diff --name-status -r"
|
||||||
|
# into the message, just before the "git status" output. It is
|
||||||
|
# commented because it doesn't cope with --amend or with squashed
|
||||||
|
# commits.
|
||||||
|
#
|
||||||
|
# The third example adds a Signed-off-by line to the message, that can
|
||||||
|
# still be edited. This is rarely a good idea.
|
||||||
|
|
||||||
|
COMMIT_MSG_FILE=$1
|
||||||
|
COMMIT_SOURCE=$2
|
||||||
|
SHA1=$3
|
||||||
|
|
||||||
|
/usr/bin/perl -i.bak -ne 'print unless(m/^. Please enter the commit message/..m/^#$/)' "$COMMIT_MSG_FILE"
|
||||||
|
|
||||||
|
# case "$COMMIT_SOURCE,$SHA1" in
|
||||||
|
# ,|template,)
|
||||||
|
# /usr/bin/perl -i.bak -pe '
|
||||||
|
# print "\n" . `git diff --cached --name-status -r`
|
||||||
|
# if /^#/ && $first++ == 0' "$COMMIT_MSG_FILE" ;;
|
||||||
|
# *) ;;
|
||||||
|
# esac
|
||||||
|
|
||||||
|
# SOB=$(git var GIT_COMMITTER_IDENT | sed -n 's/^\(.*>\).*$/Signed-off-by: \1/p')
|
||||||
|
# git interpret-trailers --in-place --trailer "$SOB" "$COMMIT_MSG_FILE"
|
||||||
|
# if test -z "$COMMIT_SOURCE"
|
||||||
|
# then
|
||||||
|
# /usr/bin/perl -i.bak -pe 'print "\n" if !$first_line++' "$COMMIT_MSG_FILE"
|
||||||
|
# fi
|
||||||
128
github.git/hooks/update.sample
Executable file
128
github.git/hooks/update.sample
Executable file
@@ -0,0 +1,128 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
#
|
||||||
|
# An example hook script to block unannotated tags from entering.
|
||||||
|
# Called by "git receive-pack" with arguments: refname sha1-old sha1-new
|
||||||
|
#
|
||||||
|
# To enable this hook, rename this file to "update".
|
||||||
|
#
|
||||||
|
# Config
|
||||||
|
# ------
|
||||||
|
# hooks.allowunannotated
|
||||||
|
# This boolean sets whether unannotated tags will be allowed into the
|
||||||
|
# repository. By default they won't be.
|
||||||
|
# hooks.allowdeletetag
|
||||||
|
# This boolean sets whether deleting tags will be allowed in the
|
||||||
|
# repository. By default they won't be.
|
||||||
|
# hooks.allowmodifytag
|
||||||
|
# This boolean sets whether a tag may be modified after creation. By default
|
||||||
|
# it won't be.
|
||||||
|
# hooks.allowdeletebranch
|
||||||
|
# This boolean sets whether deleting branches will be allowed in the
|
||||||
|
# repository. By default they won't be.
|
||||||
|
# hooks.denycreatebranch
|
||||||
|
# This boolean sets whether remotely creating branches will be denied
|
||||||
|
# in the repository. By default this is allowed.
|
||||||
|
#
|
||||||
|
|
||||||
|
# --- Command line
|
||||||
|
refname="$1"
|
||||||
|
oldrev="$2"
|
||||||
|
newrev="$3"
|
||||||
|
|
||||||
|
# --- Safety check
|
||||||
|
if [ -z "$GIT_DIR" ]; then
|
||||||
|
echo "Don't run this script from the command line." >&2
|
||||||
|
echo " (if you want, you could supply GIT_DIR then run" >&2
|
||||||
|
echo " $0 <ref> <oldrev> <newrev>)" >&2
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -z "$refname" -o -z "$oldrev" -o -z "$newrev" ]; then
|
||||||
|
echo "usage: $0 <ref> <oldrev> <newrev>" >&2
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
# --- Config
|
||||||
|
allowunannotated=$(git config --type=bool hooks.allowunannotated)
|
||||||
|
allowdeletebranch=$(git config --type=bool hooks.allowdeletebranch)
|
||||||
|
denycreatebranch=$(git config --type=bool hooks.denycreatebranch)
|
||||||
|
allowdeletetag=$(git config --type=bool hooks.allowdeletetag)
|
||||||
|
allowmodifytag=$(git config --type=bool hooks.allowmodifytag)
|
||||||
|
|
||||||
|
# check for no description
|
||||||
|
projectdesc=$(sed -e '1q' "$GIT_DIR/description")
|
||||||
|
case "$projectdesc" in
|
||||||
|
"Unnamed repository"* | "")
|
||||||
|
echo "*** Project description file hasn't been set" >&2
|
||||||
|
exit 1
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
|
||||||
|
# --- Check types
|
||||||
|
# if $newrev is 0000...0000, it's a commit to delete a ref.
|
||||||
|
zero="0000000000000000000000000000000000000000"
|
||||||
|
if [ "$newrev" = "$zero" ]; then
|
||||||
|
newrev_type=delete
|
||||||
|
else
|
||||||
|
newrev_type=$(git cat-file -t $newrev)
|
||||||
|
fi
|
||||||
|
|
||||||
|
case "$refname","$newrev_type" in
|
||||||
|
refs/tags/*,commit)
|
||||||
|
# un-annotated tag
|
||||||
|
short_refname=${refname##refs/tags/}
|
||||||
|
if [ "$allowunannotated" != "true" ]; then
|
||||||
|
echo "*** The un-annotated tag, $short_refname, is not allowed in this repository" >&2
|
||||||
|
echo "*** Use 'git tag [ -a | -s ]' for tags you want to propagate." >&2
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
refs/tags/*,delete)
|
||||||
|
# delete tag
|
||||||
|
if [ "$allowdeletetag" != "true" ]; then
|
||||||
|
echo "*** Deleting a tag is not allowed in this repository" >&2
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
refs/tags/*,tag)
|
||||||
|
# annotated tag
|
||||||
|
if [ "$allowmodifytag" != "true" ] && git rev-parse $refname > /dev/null 2>&1
|
||||||
|
then
|
||||||
|
echo "*** Tag '$refname' already exists." >&2
|
||||||
|
echo "*** Modifying a tag is not allowed in this repository." >&2
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
refs/heads/*,commit)
|
||||||
|
# branch
|
||||||
|
if [ "$oldrev" = "$zero" -a "$denycreatebranch" = "true" ]; then
|
||||||
|
echo "*** Creating a branch is not allowed in this repository" >&2
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
refs/heads/*,delete)
|
||||||
|
# delete branch
|
||||||
|
if [ "$allowdeletebranch" != "true" ]; then
|
||||||
|
echo "*** Deleting a branch is not allowed in this repository" >&2
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
refs/remotes/*,commit)
|
||||||
|
# tracking branch
|
||||||
|
;;
|
||||||
|
refs/remotes/*,delete)
|
||||||
|
# delete tracking branch
|
||||||
|
if [ "$allowdeletebranch" != "true" ]; then
|
||||||
|
echo "*** Deleting a tracking branch is not allowed in this repository" >&2
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
;;
|
||||||
|
*)
|
||||||
|
# Anything else (is there anything else?)
|
||||||
|
echo "*** Update hook: unknown type of update to ref $refname of type $newrev_type" >&2
|
||||||
|
exit 1
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
|
||||||
|
# --- Finished
|
||||||
|
exit 0
|
||||||
BIN
github.git/index
Normal file
BIN
github.git/index
Normal file
Binary file not shown.
6
github.git/info/exclude
Normal file
6
github.git/info/exclude
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
# git ls-files --others --exclude-from=.git/info/exclude
|
||||||
|
# Lines that start with '#' are comments.
|
||||||
|
# For a project mostly in C, the following would be a good set of
|
||||||
|
# exclude patterns (uncomment them if you want to use them):
|
||||||
|
# *.[oa]
|
||||||
|
# *~
|
||||||
5
github.git/logs/HEAD
Normal file
5
github.git/logs/HEAD
Normal file
@@ -0,0 +1,5 @@
|
|||||||
|
0000000000000000000000000000000000000000 0c67d5918530687fe1ddd7950b2ab61cb71febb3 zware <zware@dell-lat.(none)> 1596226970 -0500 clone: from https://github.com/ZwareBear/DockerSetup.git
|
||||||
|
0c67d5918530687fe1ddd7950b2ab61cb71febb3 dfe7b57536860175ea9c4ab10902bad328786869 ZwareBear <github@zware.33mail.com> 1596227334 -0500 commit: added lidarr to docker-compose
|
||||||
|
dfe7b57536860175ea9c4ab10902bad328786869 03be7b969b298f884054e5e1e48804578c1d1f36 ZwareBear <github@zware.33mail.com> 1607276841 -0600 pull: Fast-forward
|
||||||
|
03be7b969b298f884054e5e1e48804578c1d1f36 6ef15396d607fa6cc862e9fb4a50be8d38982788 ZwareBear <github@zware.33mail.com> 1607278308 -0600 commit: adding gitea config
|
||||||
|
6ef15396d607fa6cc862e9fb4a50be8d38982788 1991645d49c5af38a4144f93494e4ce773bfee7a ZwareBear <github@zware.33mail.com> 1607278836 -0600 commit: fixed indention
|
||||||
5
github.git/logs/refs/heads/master
Normal file
5
github.git/logs/refs/heads/master
Normal file
@@ -0,0 +1,5 @@
|
|||||||
|
0000000000000000000000000000000000000000 0c67d5918530687fe1ddd7950b2ab61cb71febb3 zware <zware@dell-lat.(none)> 1596226970 -0500 clone: from https://github.com/ZwareBear/DockerSetup.git
|
||||||
|
0c67d5918530687fe1ddd7950b2ab61cb71febb3 dfe7b57536860175ea9c4ab10902bad328786869 ZwareBear <github@zware.33mail.com> 1596227334 -0500 commit: added lidarr to docker-compose
|
||||||
|
dfe7b57536860175ea9c4ab10902bad328786869 03be7b969b298f884054e5e1e48804578c1d1f36 ZwareBear <github@zware.33mail.com> 1607276841 -0600 pull: Fast-forward
|
||||||
|
03be7b969b298f884054e5e1e48804578c1d1f36 6ef15396d607fa6cc862e9fb4a50be8d38982788 ZwareBear <github@zware.33mail.com> 1607278308 -0600 commit: adding gitea config
|
||||||
|
6ef15396d607fa6cc862e9fb4a50be8d38982788 1991645d49c5af38a4144f93494e4ce773bfee7a ZwareBear <github@zware.33mail.com> 1607278836 -0600 commit: fixed indention
|
||||||
1
github.git/logs/refs/remotes/origin/HEAD
Normal file
1
github.git/logs/refs/remotes/origin/HEAD
Normal file
@@ -0,0 +1 @@
|
|||||||
|
0000000000000000000000000000000000000000 0c67d5918530687fe1ddd7950b2ab61cb71febb3 zware <zware@dell-lat.(none)> 1596226970 -0500 clone: from https://github.com/ZwareBear/DockerSetup.git
|
||||||
4
github.git/logs/refs/remotes/origin/master
Normal file
4
github.git/logs/refs/remotes/origin/master
Normal file
@@ -0,0 +1,4 @@
|
|||||||
|
0c67d5918530687fe1ddd7950b2ab61cb71febb3 dfe7b57536860175ea9c4ab10902bad328786869 ZwareBear <github@zware.33mail.com> 1596227402 -0500 update by push
|
||||||
|
dfe7b57536860175ea9c4ab10902bad328786869 03be7b969b298f884054e5e1e48804578c1d1f36 ZwareBear <github@zware.33mail.com> 1607276841 -0600 pull: fast-forward
|
||||||
|
03be7b969b298f884054e5e1e48804578c1d1f36 6ef15396d607fa6cc862e9fb4a50be8d38982788 ZwareBear <github@zware.33mail.com> 1607278321 -0600 update by push
|
||||||
|
6ef15396d607fa6cc862e9fb4a50be8d38982788 1991645d49c5af38a4144f93494e4ce773bfee7a ZwareBear <github@zware.33mail.com> 1607278890 -0600 update by push
|
||||||
BIN
github.git/objects/02/9d5372d6be2ef9d1b7ee4e7688f0e8a6cb9178
Normal file
BIN
github.git/objects/02/9d5372d6be2ef9d1b7ee4e7688f0e8a6cb9178
Normal file
Binary file not shown.
@@ -0,0 +1,5 @@
|
|||||||
|
xu<>Ms<4D>@<18><>̯<EFBFBD>{*<01><1A>J<EFBFBD>%D<>
|
||||||
|
<EFBFBD><1F>
|
||||||
|
<EFBFBD><EFBFBD>5<EFBFBD><EFBFBD>n<1F><>~<7E><><EFBFBD><EFBFBD>M]@A<><41>:B<><42>"Y<><59><EFBFBD>SQNUD<55>D<EFBFBD>T<EFBFBD>iD<69>%39%T<>Zԑ3)%0Q<30>"<22><>*̠B<CCA0><42>e<EFBFBD><65>]<13>J<EFBFBD><06><><EFBFBD>s<EFBFBD>gyӁ<79><D381><EFBFBD>3 <09><>kV<6B><56>Oާo<DEA7>"I5*<2A><17>Կ<EFBFBD>L<EFBFBD>UI<55>P<EFBFBD><50><EFBFBD><EFBFBD><02><><EFBFBD>'<19><>S0<53>O<EFBFBD><4F><EFBFBD><EFBFBD>H[<5B><>?7<><13><><EFBFBD>Zd<5A><64>[<5B><>xk<78>u<EFBFBD>`<60>9k#<23><><EFBFBD>98p<38><70><EFBFBD>4<0C>2<EFBFBD><32>t1<74><31><EFBFBD>3<EFBFBD>0/<2F><>M꼻a<EABCBB><61><EFBFBD><19>yZ_?T{%<25><><EFBFBD><EFBFBD>K<EFBFBD><4B><EFBFBD><EFBFBD>ās<C481><73><EFBFBD><EFBFBD><EFBFBD>&e<><65><EFBFBD>/
|
||||||
|
U<EFBFBD>f{}8<>C<><43><EFBFBD>,<2C><>d<EFBFBD><64>F<EFBFBD>$<24>|<7C><>S<EFBFBD>R<<3C>g<EFBFBD><67><EFBFBD>-<o`<60><><EFBFBD>W<<1F>\<5C>+7<><37><EFBFBD><EFBFBD><EFBFBD><EFBFBD>6!<21>Ixv<78>ѼK<17>MӰ<4D><D3B0>$Ԉ<><D488><EFBFBD>&<26>l<EFBFBD><07>jS<><53><EFBFBD>8<EFBFBD><38><EFBFBD><EFBFBD>9<EFBFBD><39><EFBFBD>i<>έ<EFBFBD><CEAD><0C><>ր<EFBFBD><D680><EFBFBD>c<EFBFBD>P<EFBFBD>]
|
||||||
|
<EFBFBD><EFBFBD>~<7E><>Z5r<>8<10>}} o<>(^
|
||||||
@@ -0,0 +1,4 @@
|
|||||||
|
x<01><>[
|
||||||
|
!@<40>vn<><6E><EFBFBD>\"<22>e<EFBFBD>w<EFBFBD>k <09><18>P<EFBFBD><50><EFBFBD>-<2D>{<7B>N<><4E>R:W<>z#<23><><11><>VG)<29><><EFBFBD>B<EFBFBD>h <09>
|
||||||
|
<EFBFBD>*+ef
|
||||||
|
셍<EFBFBD><EFBFBD><EFBFBD><EFBFBD><EFBFBD><04>f<EFBFBD>(<28>9<18>E <20><><EFBFBD>NY<4E>n<>Q<1B><><EFBFBD>w%l<>t/<2F><><EFBFBD><EFBFBD>w'<27><><0B><><14>r<EFBFBD>r<EFBFBD>vG?
|
||||||
@@ -0,0 +1,2 @@
|
|||||||
|
x<01><>Aj<41>0<10><><EFBFBD>u<EFBFBD><75>@<40>X#K#<08><>ݍ<><DD8D>#<23><>`d9}<7D>+t<><74><EFBFBD>{y_<79><5F><EFBFBD><>C<12>M90ېdv<1A>;A<><41>f/<2F><>Hl<48>1<EFBFBD><31><1C>u@JR<>1<EFBFBD>ȕ<EFBFBD><C895>@<40>N<EFBFBD><4E><EFBFBD>́<EFBFBD>T<EFBFBD>J<EFBFBD><4A><EFBFBD><1F>߯<><DFAF>*ܖ<>g<>z<EFBFBD><7A>+<2B>*<2A><><EFBFBD><EFBFBD><EFBFBD>&<26><>&d<><64>G4<47><34><EFBFBD>]<5D>ɍ<EFBFBD>ҶƠ
|
||||||
|
<EFBFBD>}<7D>m1<6D><31>8J<38>
|
||||||
BIN
github.git/objects/7c/5dbce056dc3c5caeba7085cae25fc4d209f7b0
Normal file
BIN
github.git/objects/7c/5dbce056dc3c5caeba7085cae25fc4d209f7b0
Normal file
Binary file not shown.
BIN
github.git/objects/80/77b52f8a2bf92ab427d4ab1e1f9fae1914f2b9
Normal file
BIN
github.git/objects/80/77b52f8a2bf92ab427d4ab1e1f9fae1914f2b9
Normal file
Binary file not shown.
BIN
github.git/objects/86/a46783cd222acd5776af9b2ff618e41b14def5
Normal file
BIN
github.git/objects/86/a46783cd222acd5776af9b2ff618e41b14def5
Normal file
Binary file not shown.
BIN
github.git/objects/b8/32bc78827ba54e9064a03f4fc6a80d33a2d999
Normal file
BIN
github.git/objects/b8/32bc78827ba54e9064a03f4fc6a80d33a2d999
Normal file
Binary file not shown.
BIN
github.git/objects/df/e7b57536860175ea9c4ab10902bad328786869
Normal file
BIN
github.git/objects/df/e7b57536860175ea9c4ab10902bad328786869
Normal file
Binary file not shown.
BIN
github.git/objects/e9/a602c73c118094ebd7a7e6d2b4756a2f224536
Normal file
BIN
github.git/objects/e9/a602c73c118094ebd7a7e6d2b4756a2f224536
Normal file
Binary file not shown.
BIN
github.git/objects/ec/52e19b4ad3be5db46233d20526bb82d9a568f0
Normal file
BIN
github.git/objects/ec/52e19b4ad3be5db46233d20526bb82d9a568f0
Normal file
Binary file not shown.
BIN
github.git/objects/fe/ab86cfab95bf3e4ce5302d6ad3447621f82a72
Normal file
BIN
github.git/objects/fe/ab86cfab95bf3e4ce5302d6ad3447621f82a72
Normal file
Binary file not shown.
Binary file not shown.
Binary file not shown.
3
github.git/packed-refs
Normal file
3
github.git/packed-refs
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
# pack-refs with: peeled fully-peeled sorted
|
||||||
|
0c67d5918530687fe1ddd7950b2ab61cb71febb3 refs/remotes/origin/master
|
||||||
|
bf9ebf9bd7f0445d9c4446e2e6c7ac23fca88dcb refs/remotes/origin/test
|
||||||
1
github.git/refs/heads/master
Normal file
1
github.git/refs/heads/master
Normal file
@@ -0,0 +1 @@
|
|||||||
|
1991645d49c5af38a4144f93494e4ce773bfee7a
|
||||||
1
github.git/refs/remotes/origin/HEAD
Normal file
1
github.git/refs/remotes/origin/HEAD
Normal file
@@ -0,0 +1 @@
|
|||||||
|
ref: refs/remotes/origin/master
|
||||||
1
github.git/refs/remotes/origin/master
Normal file
1
github.git/refs/remotes/origin/master
Normal file
@@ -0,0 +1 @@
|
|||||||
|
1991645d49c5af38a4144f93494e4ce773bfee7a
|
||||||
Reference in New Issue
Block a user