##// END OF EJS Templates
pick_port: unified code for testing/hooks
pick_port: unified code for testing/hooks

File last commit:

r4859:169b0860 default
r4866:6b029be8 default
Show More
development.ini
923 lines | 32.0 KiB | text/x-ini | IniLexer
## -*- coding: utf-8 -*-
; #########################################
; RHODECODE COMMUNITY EDITION CONFIGURATION
; #########################################
[DEFAULT]
; Debug flag sets all loggers to debug, and enables request tracking
debug = true
; ########################################################################
; EMAIL CONFIGURATION
; These settings will be used by the RhodeCode mailing system
; ########################################################################
; prefix all emails subjects with given prefix, helps filtering out emails
#email_prefix = [RhodeCode]
; email FROM address all mails will be sent
#app_email_from = rhodecode-noreply@localhost
#smtp_server = mail.server.com
#smtp_username =
#smtp_password =
#smtp_port =
#smtp_use_tls = false
#smtp_use_ssl = true
[server:main]
; COMMON HOST/IP CONFIG
host = 127.0.0.1
port = 5000
; ##################################################
; WAITRESS WSGI SERVER - Recommended for Development
; ##################################################
; use server type
use = egg:waitress#main
; number of worker threads
threads = 5
; MAX BODY SIZE 100GB
max_request_body_size = 107374182400
; Use poll instead of select, fixes file descriptors limits problems.
; May not work on old windows systems.
asyncore_use_poll = true
; ###########################
; GUNICORN APPLICATION SERVER
; ###########################
; run with gunicorn --log-config rhodecode.ini --paste rhodecode.ini
; Module to use, this setting shouldn't be changed
#use = egg:gunicorn#main
; Sets the number of process workers. More workers means more concurrent connections
; RhodeCode can handle at the same time. Each additional worker also it increases
; memory usage as each has it's own set of caches.
; Recommended value is (2 * NUMBER_OF_CPUS + 1), eg 2CPU = 5 workers, but no more
; than 8-10 unless for really big deployments .e.g 700-1000 users.
; `instance_id = *` must be set in the [app:main] section below (which is the default)
; when using more than 1 worker.
#workers = 2
; Gunicorn access log level
#loglevel = info
; Process name visible in process list
#proc_name = rhodecode
; Type of worker class, one of `sync`, `gevent`
; Recommended type is `gevent`
#worker_class = gevent
; The maximum number of simultaneous clients. Valid only for gevent
#worker_connections = 10
; Max number of requests that worker will handle before being gracefully restarted.
; Prevents memory leaks, jitter adds variability so not all workers are restarted at once.
#max_requests = 1000
#max_requests_jitter = 30
; Amount of time a worker can spend with handling a request before it
; gets killed and restarted. By default set to 21600 (6hrs)
; Examples: 1800 (30min), 3600 (1hr), 7200 (2hr), 43200 (12h)
#timeout = 21600
; The maximum size of HTTP request line in bytes.
; 0 for unlimited
#limit_request_line = 0
; Limit the number of HTTP headers fields in a request.
; By default this value is 100 and can't be larger than 32768.
#limit_request_fields = 32768
; Limit the allowed size of an HTTP request header field.
; Value is a positive number or 0.
; Setting it to 0 will allow unlimited header field sizes.
#limit_request_field_size = 0
; Timeout for graceful workers restart.
; After receiving a restart signal, workers have this much time to finish
; serving requests. Workers still alive after the timeout (starting from the
; receipt of the restart signal) are force killed.
; Examples: 1800 (30min), 3600 (1hr), 7200 (2hr), 43200 (12h)
#graceful_timeout = 3600
# The number of seconds to wait for requests on a Keep-Alive connection.
# Generally set in the 1-5 seconds range.
#keepalive = 2
; Maximum memory usage that each worker can use before it will receive a
; graceful restart signal 0 = memory monitoring is disabled
; Examples: 268435456 (256MB), 536870912 (512MB)
; 1073741824 (1GB), 2147483648 (2GB), 4294967296 (4GB)
#memory_max_usage = 0
; How often in seconds to check for memory usage for each gunicorn worker
#memory_usage_check_interval = 60
; Threshold value for which we don't recycle worker if GarbageCollection
; frees up enough resources. Before each restart we try to run GC on worker
; in case we get enough free memory after that, restart will not happen.
#memory_usage_recovery_threshold = 0.8
; Prefix middleware for RhodeCode.
; recommended when using proxy setup.
; allows to set RhodeCode under a prefix in server.
; eg https://server.com/custom_prefix. Enable `filter-with =` option below as well.
; And set your prefix like: `prefix = /custom_prefix`
; be sure to also set beaker.session.cookie_path = /custom_prefix if you need
; to make your cookies only work on prefix url
[filter:proxy-prefix]
use = egg:PasteDeploy#prefix
prefix = /
[app:main]
; The %(here)s variable will be replaced with the absolute path of parent directory
; of this file
; Each option in the app:main can be override by an environmental variable
;
;To override an option:
;
;RC_<KeyName>
;Everything should be uppercase, . and - should be replaced by _.
;For example, if you have these configuration settings:
;rc_cache.repo_object.backend = foo
;can be overridden by
;export RC_CACHE_REPO_OBJECT_BACKEND=foo
use = egg:rhodecode-enterprise-ce
; enable proxy prefix middleware, defined above
#filter-with = proxy-prefix
; #############
; DEBUG OPTIONS
; #############
pyramid.reload_templates = true
# During development the we want to have the debug toolbar enabled
pyramid.includes =
pyramid_debugtoolbar
debugtoolbar.hosts = 0.0.0.0/0
debugtoolbar.exclude_prefixes =
/css
/fonts
/images
/js
## RHODECODE PLUGINS ##
rhodecode.includes =
rhodecode.api
# api prefix url
rhodecode.api.url = /_admin/api
; enable debug style page
debug_style = true
; #################
; END DEBUG OPTIONS
; #################
; encryption key used to encrypt social plugin tokens,
; remote_urls with credentials etc, if not set it defaults to
; `beaker.session.secret`
#rhodecode.encrypted_values.secret =
; decryption strict mode (enabled by default). It controls if decryption raises
; `SignatureVerificationError` in case of wrong key, or damaged encryption data.
#rhodecode.encrypted_values.strict = false
; Pick algorithm for encryption. Either fernet (more secure) or aes (default)
; fernet is safer, and we strongly recommend switching to it.
; Due to backward compatibility aes is used as default.
#rhodecode.encrypted_values.algorithm = fernet
; Return gzipped responses from RhodeCode (static files/application)
gzip_responses = false
; Auto-generate javascript routes file on startup
generate_js_files = false
; System global default language.
; All available languages: en (default), be, de, es, fr, it, ja, pl, pt, ru, zh
lang = en
; Perform a full repository scan and import on each server start.
; Settings this to true could lead to very long startup time.
startup.import_repos = false
; Uncomment and set this path to use archive download cache.
; Once enabled, generated archives will be cached at this location
; and served from the cache during subsequent requests for the same archive of
; the repository.
#archive_cache_dir = /tmp/tarballcache
; URL at which the application is running. This is used for Bootstrapping
; requests in context when no web request is available. Used in ishell, or
; SSH calls. Set this for events to receive proper url for SSH calls.
app.base_url = http://rhodecode.local
; Unique application ID. Should be a random unique string for security.
app_instance_uuid = rc-production
; Cut off limit for large diffs (size in bytes). If overall diff size on
; commit, or pull request exceeds this limit this diff will be displayed
; partially. E.g 512000 == 512Kb
cut_off_limit_diff = 512000
; Cut off limit for large files inside diffs (size in bytes). Each individual
; file inside diff which exceeds this limit will be displayed partially.
; E.g 128000 == 128Kb
cut_off_limit_file = 128000
; Use cached version of vcs repositories everywhere. Recommended to be `true`
vcs_full_cache = true
; Force https in RhodeCode, fixes https redirects, assumes it's always https.
; Normally this is controlled by proper flags sent from http server such as Nginx or Apache
force_https = false
; use Strict-Transport-Security headers
use_htsts = false
; Set to true if your repos are exposed using the dumb protocol
git_update_server_info = false
; RSS/ATOM feed options
rss_cut_off_limit = 256000
rss_items_per_page = 10
rss_include_diff = false
; gist URL alias, used to create nicer urls for gist. This should be an
; url that does rewrites to _admin/gists/{gistid}.
; example: http://gist.rhodecode.org/{gistid}. Empty means use the internal
; RhodeCode url, ie. http[s]://rhodecode.server/_admin/gists/{gistid}
gist_alias_url =
; List of views (using glob pattern syntax) that AUTH TOKENS could be
; used for access.
; Adding ?auth_token=TOKEN_HASH to the url authenticates this request as if it
; came from the the logged in user who own this authentication token.
; Additionally @TOKEN syntax can be used to bound the view to specific
; authentication token. Such view would be only accessible when used together
; with this authentication token
; list of all views can be found under `/_admin/permissions/auth_token_access`
; The list should be "," separated and on a single line.
; Most common views to enable:
# RepoCommitsView:repo_commit_download
# RepoCommitsView:repo_commit_patch
# RepoCommitsView:repo_commit_raw
# RepoCommitsView:repo_commit_raw@TOKEN
# RepoFilesView:repo_files_diff
# RepoFilesView:repo_archivefile
# RepoFilesView:repo_file_raw
# GistView:*
api_access_controllers_whitelist =
; Default encoding used to convert from and to unicode
; can be also a comma separated list of encoding in case of mixed encodings
default_encoding = UTF-8
; instance-id prefix
; a prefix key for this instance used for cache invalidation when running
; multiple instances of RhodeCode, make sure it's globally unique for
; all running RhodeCode instances. Leave empty if you don't use it
instance_id =
; Fallback authentication plugin. Set this to a plugin ID to force the usage
; of an authentication plugin also if it is disabled by it's settings.
; This could be useful if you are unable to log in to the system due to broken
; authentication settings. Then you can enable e.g. the internal RhodeCode auth
; module to log in again and fix the settings.
; Available builtin plugin IDs (hash is part of the ID):
; egg:rhodecode-enterprise-ce#rhodecode
; egg:rhodecode-enterprise-ce#pam
; egg:rhodecode-enterprise-ce#ldap
; egg:rhodecode-enterprise-ce#jasig_cas
; egg:rhodecode-enterprise-ce#headers
; egg:rhodecode-enterprise-ce#crowd
#rhodecode.auth_plugin_fallback = egg:rhodecode-enterprise-ce#rhodecode
; Flag to control loading of legacy plugins in py:/path format
auth_plugin.import_legacy_plugins = true
; alternative return HTTP header for failed authentication. Default HTTP
; response is 401 HTTPUnauthorized. Currently HG clients have troubles with
; handling that causing a series of failed authentication calls.
; Set this variable to 403 to return HTTPForbidden, or any other HTTP code
; This will be served instead of default 401 on bad authentication
auth_ret_code =
; use special detection method when serving auth_ret_code, instead of serving
; ret_code directly, use 401 initially (Which triggers credentials prompt)
; and then serve auth_ret_code to clients
auth_ret_code_detection = false
; locking return code. When repository is locked return this HTTP code. 2XX
; codes don't break the transactions while 4XX codes do
lock_ret_code = 423
; allows to change the repository location in settings page
allow_repo_location_change = true
; allows to setup custom hooks in settings page
allow_custom_hooks_settings = true
; Generated license token required for EE edition license.
; New generated token value can be found in Admin > settings > license page.
license_token =
; This flag hides sensitive information on the license page such as token, and license data
license.hide_license_info = false
; supervisor connection uri, for managing supervisor and logs.
supervisor.uri =
; supervisord group name/id we only want this RC instance to handle
supervisor.group_id = dev
; Display extended labs settings
labs_settings_active = true
; Custom exception store path, defaults to TMPDIR
; This is used to store exception from RhodeCode in shared directory
#exception_tracker.store_path =
; Send email with exception details when it happens
#exception_tracker.send_email = false
; Comma separated list of recipients for exception emails,
; e.g admin@rhodecode.com,devops@rhodecode.com
; Can be left empty, then emails will be sent to ALL super-admins
#exception_tracker.send_email_recipients =
; optional prefix to Add to email Subject
#exception_tracker.email_prefix = [RHODECODE ERROR]
; File store configuration. This is used to store and serve uploaded files
file_store.enabled = true
; Storage backend, available options are: local
file_store.backend = local
; path to store the uploaded binaries
file_store.storage_path = %(here)s/data/file_store
; #############
; CELERY CONFIG
; #############
; manually run celery: /path/to/celery worker -E --beat --app rhodecode.lib.celerylib.loader --scheduler rhodecode.lib.celerylib.scheduler.RcScheduler --loglevel DEBUG --ini /path/to/rhodecode.ini
use_celery = false
; path to store schedule database
#celerybeat-schedule.path =
; connection url to the message broker (default redis)
celery.broker_url = redis://localhost:6379/8
; rabbitmq example
#celery.broker_url = amqp://rabbitmq:qweqwe@localhost:5672/rabbitmqhost
; maximum tasks to execute before worker restart
celery.max_tasks_per_child = 100
; tasks will never be sent to the queue, but executed locally instead.
celery.task_always_eager = false
; #############
; DOGPILE CACHE
; #############
; Default cache dir for caches. Putting this into a ramdisk can boost performance.
; eg. /tmpfs/data_ramdisk, however this directory might require large amount of space
cache_dir = %(here)s/data
; *********************************************
; `sql_cache_short` cache for heavy SQL queries
; Only supported backend is `memory_lru`
; *********************************************
rc_cache.sql_cache_short.backend = dogpile.cache.rc.memory_lru
rc_cache.sql_cache_short.expiration_time = 30
; *****************************************************
; `cache_repo_longterm` cache for repo object instances
; Only supported backend is `memory_lru`
; *****************************************************
rc_cache.cache_repo_longterm.backend = dogpile.cache.rc.memory_lru
; by default we use 30 Days, cache is still invalidated on push
rc_cache.cache_repo_longterm.expiration_time = 2592000
; max items in LRU cache, set to smaller number to save memory, and expire last used caches
rc_cache.cache_repo_longterm.max_size = 10000
; *********************************************
; `cache_general` cache for general purpose use
; for simplicity use rc.file_namespace backend,
; for performance and scale use rc.redis
; *********************************************
rc_cache.cache_general.backend = dogpile.cache.rc.file_namespace
rc_cache.cache_general.expiration_time = 43200
; file cache store path. Defaults to `cache_dir =` value or tempdir if both values are not set
#rc_cache.cache_general.arguments.filename = /tmp/cache_general.db
; alternative `cache_general` redis backend with distributed lock
#rc_cache.cache_general.backend = dogpile.cache.rc.redis
#rc_cache.cache_general.expiration_time = 300
; redis_expiration_time needs to be greater then expiration_time
#rc_cache.cache_general.arguments.redis_expiration_time = 7200
#rc_cache.cache_general.arguments.host = localhost
#rc_cache.cache_general.arguments.port = 6379
#rc_cache.cache_general.arguments.db = 0
#rc_cache.cache_general.arguments.socket_timeout = 30
; more Redis options: https://dogpilecache.sqlalchemy.org/en/latest/api.html#redis-backends
#rc_cache.cache_general.arguments.distributed_lock = true
; auto-renew lock to prevent stale locks, slower but safer. Use only if problems happen
#rc_cache.cache_general.arguments.lock_auto_renewal = true
; *************************************************
; `cache_perms` cache for permission tree, auth TTL
; for simplicity use rc.file_namespace backend,
; for performance and scale use rc.redis
; *************************************************
rc_cache.cache_perms.backend = dogpile.cache.rc.file_namespace
rc_cache.cache_perms.expiration_time = 3600
; file cache store path. Defaults to `cache_dir =` value or tempdir if both values are not set
#rc_cache.cache_perms.arguments.filename = /tmp/cache_perms.db
; alternative `cache_perms` redis backend with distributed lock
#rc_cache.cache_perms.backend = dogpile.cache.rc.redis
#rc_cache.cache_perms.expiration_time = 300
; redis_expiration_time needs to be greater then expiration_time
#rc_cache.cache_perms.arguments.redis_expiration_time = 7200
#rc_cache.cache_perms.arguments.host = localhost
#rc_cache.cache_perms.arguments.port = 6379
#rc_cache.cache_perms.arguments.db = 0
#rc_cache.cache_perms.arguments.socket_timeout = 30
; more Redis options: https://dogpilecache.sqlalchemy.org/en/latest/api.html#redis-backends
#rc_cache.cache_perms.arguments.distributed_lock = true
; auto-renew lock to prevent stale locks, slower but safer. Use only if problems happen
#rc_cache.cache_perms.arguments.lock_auto_renewal = true
; ***************************************************
; `cache_repo` cache for file tree, Readme, RSS FEEDS
; for simplicity use rc.file_namespace backend,
; for performance and scale use rc.redis
; ***************************************************
rc_cache.cache_repo.backend = dogpile.cache.rc.file_namespace
rc_cache.cache_repo.expiration_time = 2592000
; file cache store path. Defaults to `cache_dir =` value or tempdir if both values are not set
#rc_cache.cache_repo.arguments.filename = /tmp/cache_repo.db
; alternative `cache_repo` redis backend with distributed lock
#rc_cache.cache_repo.backend = dogpile.cache.rc.redis
#rc_cache.cache_repo.expiration_time = 2592000
; redis_expiration_time needs to be greater then expiration_time
#rc_cache.cache_repo.arguments.redis_expiration_time = 2678400
#rc_cache.cache_repo.arguments.host = localhost
#rc_cache.cache_repo.arguments.port = 6379
#rc_cache.cache_repo.arguments.db = 1
#rc_cache.cache_repo.arguments.socket_timeout = 30
; more Redis options: https://dogpilecache.sqlalchemy.org/en/latest/api.html#redis-backends
#rc_cache.cache_repo.arguments.distributed_lock = true
; auto-renew lock to prevent stale locks, slower but safer. Use only if problems happen
#rc_cache.cache_repo.arguments.lock_auto_renewal = true
; ##############
; BEAKER SESSION
; ##############
; beaker.session.type is type of storage options for the logged users sessions. Current allowed
; types are file, ext:redis, ext:database, ext:memcached, and memory (default if not specified).
; Fastest ones are Redis and ext:database
beaker.session.type = file
beaker.session.data_dir = %(here)s/data/sessions
; Redis based sessions
#beaker.session.type = ext:redis
#beaker.session.url = redis://127.0.0.1:6379/2
; DB based session, fast, and allows easy management over logged in users
#beaker.session.type = ext:database
#beaker.session.table_name = db_session
#beaker.session.sa.url = postgresql://postgres:secret@localhost/rhodecode
#beaker.session.sa.url = mysql://root:secret@127.0.0.1/rhodecode
#beaker.session.sa.pool_recycle = 3600
#beaker.session.sa.echo = false
beaker.session.key = rhodecode
beaker.session.secret = develop-rc-uytcxaz
beaker.session.lock_dir = %(here)s/data/sessions/lock
; Secure encrypted cookie. Requires AES and AES python libraries
; you must disable beaker.session.secret to use this
#beaker.session.encrypt_key = key_for_encryption
#beaker.session.validate_key = validation_key
; Sets session as invalid (also logging out user) if it haven not been
; accessed for given amount of time in seconds
beaker.session.timeout = 2592000
beaker.session.httponly = true
; Path to use for the cookie. Set to prefix if you use prefix middleware
#beaker.session.cookie_path = /custom_prefix
; Set https secure cookie
beaker.session.secure = false
; default cookie expiration time in seconds, set to `true` to set expire
; at browser close
#beaker.session.cookie_expires = 3600
; #############################
; SEARCH INDEXING CONFIGURATION
; #############################
; Full text search indexer is available in rhodecode-tools under
; `rhodecode-tools index` command
; WHOOSH Backend, doesn't require additional services to run
; it works good with few dozen repos
search.module = rhodecode.lib.index.whoosh
search.location = %(here)s/data/index
; ####################
; CHANNELSTREAM CONFIG
; ####################
; channelstream enables persistent connections and live notification
; in the system. It's also used by the chat system
channelstream.enabled = false
; server address for channelstream server on the backend
channelstream.server = 127.0.0.1:9800
; location of the channelstream server from outside world
; use ws:// for http or wss:// for https. This address needs to be handled
; by external HTTP server such as Nginx or Apache
; see Nginx/Apache configuration examples in our docs
channelstream.ws_url = ws://rhodecode.yourserver.com/_channelstream
channelstream.secret = secret
channelstream.history.location = %(here)s/channelstream_history
; Internal application path that Javascript uses to connect into.
; If you use proxy-prefix the prefix should be added before /_channelstream
channelstream.proxy_path = /_channelstream
; ##############################
; MAIN RHODECODE DATABASE CONFIG
; ##############################
#sqlalchemy.db1.url = sqlite:///%(here)s/rhodecode.db?timeout=30
#sqlalchemy.db1.url = postgresql://postgres:qweqwe@localhost/rhodecode
#sqlalchemy.db1.url = mysql://root:qweqwe@localhost/rhodecode?charset=utf8
; pymysql is an alternative driver for MySQL, use in case of problems with default one
#sqlalchemy.db1.url = mysql+pymysql://root:qweqwe@localhost/rhodecode
sqlalchemy.db1.url = sqlite:///%(here)s/rhodecode.db?timeout=30
; see sqlalchemy docs for other advanced settings
; print the sql statements to output
sqlalchemy.db1.echo = false
; recycle the connections after this amount of seconds
sqlalchemy.db1.pool_recycle = 3600
sqlalchemy.db1.convert_unicode = true
; the number of connections to keep open inside the connection pool.
; 0 indicates no limit
#sqlalchemy.db1.pool_size = 5
; The number of connections to allow in connection pool "overflow", that is
; connections that can be opened above and beyond the pool_size setting,
; which defaults to five.
#sqlalchemy.db1.max_overflow = 10
; Connection check ping, used to detect broken database connections
; could be enabled to better handle cases if MySQL has gone away errors
#sqlalchemy.db1.ping_connection = true
; ##########
; VCS CONFIG
; ##########
vcs.server.enable = true
vcs.server = localhost:9900
; Web server connectivity protocol, responsible for web based VCS operations
; Available protocols are:
; `http` - use http-rpc backend (default)
vcs.server.protocol = http
; Push/Pull operations protocol, available options are:
; `http` - use http-rpc backend (default)
vcs.scm_app_implementation = http
; Push/Pull operations hooks protocol, available options are:
; `http` - use http-rpc backend (default)
vcs.hooks.protocol = http
; Host on which this instance is listening for hooks. vcsserver will call this host to pull/push hooks so it should be
; accessible via network.
; Use vcs.hooks.host = "*" to bind to current hostname (for Docker)
vcs.hooks.host = *
; Start VCSServer with this instance as a subprocess, useful for development
vcs.start_server = false
; List of enabled VCS backends, available options are:
; `hg` - mercurial
; `git` - git
; `svn` - subversion
vcs.backends = hg, git, svn
; Wait this number of seconds before killing connection to the vcsserver
vcs.connection_timeout = 3600
; Compatibility version when creating SVN repositories. Defaults to newest version when commented out.
; Set a numeric version for your current SVN e.g 1.8, or 1.12
; Legacy available options are: pre-1.4-compatible, pre-1.5-compatible, pre-1.6-compatible, pre-1.8-compatible, pre-1.9-compatible
#vcs.svn.compatible_version = 1.8
; Cache flag to cache vcsserver remote calls locally
; It uses cache_region `cache_repo`
vcs.methods.cache = true
; ####################################################
; Subversion proxy support (mod_dav_svn)
; Maps RhodeCode repo groups into SVN paths for Apache
; ####################################################
; Enable or disable the config file generation.
svn.proxy.generate_config = false
; Generate config file with `SVNListParentPath` set to `On`.
svn.proxy.list_parent_path = true
; Set location and file name of generated config file.
svn.proxy.config_file_path = %(here)s/mod_dav_svn.conf
; alternative mod_dav config template. This needs to be a valid mako template
; Example template can be found in the source code:
; rhodecode/apps/svn_support/templates/mod-dav-svn.conf.mako
#svn.proxy.config_template = ~/.rccontrol/enterprise-1/custom_svn_conf.mako
; Used as a prefix to the `Location` block in the generated config file.
; In most cases it should be set to `/`.
svn.proxy.location_root = /
; Command to reload the mod dav svn configuration on change.
; Example: `/etc/init.d/apache2 reload` or /home/USER/apache_reload.sh
; Make sure user who runs RhodeCode process is allowed to reload Apache
#svn.proxy.reload_cmd = /etc/init.d/apache2 reload
; If the timeout expires before the reload command finishes, the command will
; be killed. Setting it to zero means no timeout. Defaults to 10 seconds.
#svn.proxy.reload_timeout = 10
; ####################
; SSH Support Settings
; ####################
; Defines if a custom authorized_keys file should be created and written on
; any change user ssh keys. Setting this to false also disables possibility
; of adding SSH keys by users from web interface. Super admins can still
; manage SSH Keys.
ssh.generate_authorized_keyfile = false
; Options for ssh, default is `no-pty,no-port-forwarding,no-X11-forwarding,no-agent-forwarding`
# ssh.authorized_keys_ssh_opts =
; Path to the authorized_keys file where the generate entries are placed.
; It is possible to have multiple key files specified in `sshd_config` e.g.
; AuthorizedKeysFile %h/.ssh/authorized_keys %h/.ssh/authorized_keys_rhodecode
ssh.authorized_keys_file_path = ~/.ssh/authorized_keys_rhodecode
; Command to execute the SSH wrapper. The binary is available in the
; RhodeCode installation directory.
; e.g ~/.rccontrol/community-1/profile/bin/rc-ssh-wrapper
ssh.wrapper_cmd = ~/.rccontrol/community-1/rc-ssh-wrapper
; Allow shell when executing the ssh-wrapper command
ssh.wrapper_cmd_allow_shell = false
; Enables logging, and detailed output send back to the client during SSH
; operations. Useful for debugging, shouldn't be used in production.
ssh.enable_debug_logging = true
; Paths to binary executable, by default they are the names, but we can
; override them if we want to use a custom one
ssh.executable.hg = ~/.rccontrol/vcsserver-1/profile/bin/hg
ssh.executable.git = ~/.rccontrol/vcsserver-1/profile/bin/git
ssh.executable.svn = ~/.rccontrol/vcsserver-1/profile/bin/svnserve
; Enables SSH key generator web interface. Disabling this still allows users
; to add their own keys.
ssh.enable_ui_key_generator = true
; #################
; APPENLIGHT CONFIG
; #################
; Appenlight is tailored to work with RhodeCode, see
; http://appenlight.rhodecode.com for details how to obtain an account
; Appenlight integration enabled
#appenlight = false
#appenlight.server_url = https://api.appenlight.com
#appenlight.api_key = YOUR_API_KEY
#appenlight.transport_config = https://api.appenlight.com?threaded=1&timeout=5
; used for JS client
#appenlight.api_public_key = YOUR_API_PUBLIC_KEY
; TWEAK AMOUNT OF INFO SENT HERE
; enables 404 error logging (default False)
#appenlight.report_404 = false
; time in seconds after request is considered being slow (default 1)
#appenlight.slow_request_time = 1
; record slow requests in application
; (needs to be enabled for slow datastore recording and time tracking)
#appenlight.slow_requests = true
; enable hooking to application loggers
#appenlight.logging = true
; minimum log level for log capture
#ppenlight.logging.level = WARNING
; send logs only from erroneous/slow requests
; (saves API quota for intensive logging)
#appenlight.logging_on_error = false
; list of additional keywords that should be grabbed from environ object
; can be string with comma separated list of words in lowercase
; (by default client will always send following info:
; 'REMOTE_USER', 'REMOTE_ADDR', 'SERVER_NAME', 'CONTENT_TYPE' + all keys that
; start with HTTP* this list be extended with additional keywords here
#appenlight.environ_keys_whitelist =
; list of keywords that should be blanked from request object
; can be string with comma separated list of words in lowercase
; (by default client will always blank keys that contain following words
; 'password', 'passwd', 'pwd', 'auth_tkt', 'secret', 'csrf'
; this list be extended with additional keywords set here
#appenlight.request_keys_blacklist =
; list of namespaces that should be ignores when gathering log entries
; can be string with comma separated list of namespaces
; (by default the client ignores own entries: appenlight_client.client)
#appenlight.log_namespace_blacklist =
; Statsd client config, this is used to send metrics to statsd
; We recommend setting statsd_exported and scrape them using Promethues
#statsd.enabled = false
#statsd.statsd_host = 0.0.0.0
#statsd.statsd_port = 8125
#statsd.statsd_prefix =
#statsd.statsd_ipv6 = false
; configure logging automatically at server startup set to false
; to use the below custom logging config.
; RC_LOGGING_FORMATTER
; RC_LOGGING_LEVEL
; env variables can control the settings for logging in case of autoconfigure
#logging.autoconfigure = true
; specify your own custom logging config file to configure logging
#logging.logging_conf_file = /path/to/custom_logging.ini
; Dummy marker to add new entries after.
; Add any custom entries below. Please don't remove this marker.
custom.conf = 1
; #####################
; LOGGING CONFIGURATION
; #####################
[loggers]
keys = root, sqlalchemy, beaker, celery, rhodecode, ssh_wrapper
[handlers]
keys = console, console_sql
[formatters]
keys = generic, json, color_formatter, color_formatter_sql
; #######
; LOGGERS
; #######
[logger_root]
level = NOTSET
handlers = console
[logger_sqlalchemy]
level = INFO
handlers = console_sql
qualname = sqlalchemy.engine
propagate = 0
[logger_beaker]
level = DEBUG
handlers =
qualname = beaker.container
propagate = 1
[logger_rhodecode]
level = DEBUG
handlers =
qualname = rhodecode
propagate = 1
[logger_ssh_wrapper]
level = DEBUG
handlers =
qualname = ssh_wrapper
propagate = 1
[logger_celery]
level = DEBUG
handlers =
qualname = celery
; ########
; HANDLERS
; ########
[handler_console]
class = StreamHandler
args = (sys.stderr, )
level = DEBUG
; To enable JSON formatted logs replace 'generic/color_formatter' with 'json'
; This allows sending properly formatted logs to grafana loki or elasticsearch
formatter = color_formatter
[handler_console_sql]
; "level = DEBUG" logs SQL queries and results.
; "level = INFO" logs SQL queries.
; "level = WARN" logs neither. (Recommended for production systems.)
class = StreamHandler
args = (sys.stderr, )
level = WARN
; To enable JSON formatted logs replace 'generic/color_formatter_sql' with 'json'
; This allows sending properly formatted logs to grafana loki or elasticsearch
formatter = color_formatter_sql
; ##########
; FORMATTERS
; ##########
[formatter_generic]
class = rhodecode.lib.logging_formatter.ExceptionAwareFormatter
format = %(asctime)s.%(msecs)03d [%(process)d] %(levelname)-5.5s [%(name)s] %(message)s
datefmt = %Y-%m-%d %H:%M:%S
[formatter_color_formatter]
class = rhodecode.lib.logging_formatter.ColorFormatter
format = %(asctime)s.%(msecs)03d [%(process)d] %(levelname)-5.5s [%(name)s] %(message)s
datefmt = %Y-%m-%d %H:%M:%S
[formatter_color_formatter_sql]
class = rhodecode.lib.logging_formatter.ColorFormatterSql
format = %(asctime)s.%(msecs)03d [%(process)d] %(levelname)-5.5s [%(name)s] %(message)s
datefmt = %Y-%m-%d %H:%M:%S
[formatter_json]
format = %(timestamp)s %(levelname)s %(name)s %(message)s %(req_id)s
class = rhodecode.lib._vendor.jsonlogger.JsonFormatter