Varnish Administration Console

Api reference

GET

Backend name                   Admin      Probe                Last updated
boot.default                   probe      Healthy (no probe)   Sat, 25 May 2019 07:01:52 GMT


PUT

The following example will set the default backend to sick:

$  curl -is  http://localhost:6085/backend/default -u agent:pass -X PUT -d 'sick'

The following example will set the default backend to healthy:

$  curl -is  http://localhost:6085/backend/default -u agent:pass -X PUT -d 'healthy'


GET

Present bans:
1558767712.572449     0 C


GET

-----------------------------
Varnish Cache CLI 1.0
-----------------------------
Linux,5.1.2-arch1-1-ARCH,x86_64,-junix,-sdefault,-sdefault,-hcritbit
varnish-plus-6.0.3r3 revision 0aace84169eca8d72edab4678e4751ace953ac5b

Type 'help' for command list.
Type 'quit' to close CLI session.


GET

Child has not panicked or panic has been cleared.


GET


vsm_space
        Value is: 1M [bytes] (default)
        Minimum is: 1M
        Maximum is: 1G

        DEPRECATED: This parameter is ignored.
        There is no global limit on amount of shared memory now.


PUT

The following example will update the vsl_space param to 200M

$  curl -is  http://localhost:6085/param/vsl_space -u agent:pass -X PUT -d '200M'

Note the parameter value updates get into effect only after the child process is restarted.


GET


{
    "accept_filter": {
        "default": "",
        "description": "Not available This parameter depends on a feature which is not available on this platform. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": ""
    },
    "acceptor_sleep_decay": {
        "default": "0.9",
        "description": "If we run out of resources, such as file descriptors or worker threads, the acceptor will sleep between accepts. This parameter (multiplicatively) reduce the sleep duration for each successful accept. (ie: 0.9 = reduce by 10%) NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "1",
        "min": "0",
        "unit": "",
        "value": "0.9"
    },
    "acceptor_sleep_incr": {
        "default": "0.000",
        "description": "If we run out of resources, such as file descriptors or worker threads, the acceptor will sleep between accepts. This parameter control how much longer we sleep, each time we fail to accept a new connection. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "1.000",
        "min": "0.000",
        "unit": "seconds",
        "value": "0.000"
    },
    "acceptor_sleep_max": {
        "default": "0.050",
        "description": "If we run out of resources, such as file descriptors or worker threads, the acceptor will sleep between accepts. This parameter limits how long it can sleep between attempts to accept new connections. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "10.000",
        "min": "0.000",
        "unit": "seconds",
        "value": "0.050"
    },
    "auto_restart": {
        "default": "on",
        "description": "Automatically restart the child/worker process if it dies. ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "on"
    },
    "backend_idle_timeout": {
        "default": "60.000",
        "description": "Timeout before we close unused backend connections. ",
        "max": "",
        "min": "1.000",
        "unit": "seconds",
        "value": "60.000"
    },
    "backend_local_error_holddown": {
        "default": "10.000",
        "description": "When connecting to backends, certain error codes (EADDRNOTAVAIL, EACCESS, EPERM) signal a local resource shortage or configuration issue for which retrying connection attempts may worsen the situation due to the complexity of the operations involved in the kernel. This parameter prevents repeated connection attempts for the configured duration. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "10.000"
    },
    "backend_remote_error_holddown": {
        "default": "0.250",
        "description": "When connecting to backends, certain error codes (ECONNREFUSED, ENETUNREACH) signal fundamental connection issues such as the backend not accepting connections or routing problems for which repeated connection attempts are considered useless This parameter prevents repeated connection attempts for the configured duration. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "0.250"
    },
    "ban_cutoff": {
        "default": "0",
        "description": "Expurge long tail content from the cache to keep the number of bans below this value. 0 disables. When this parameter is set to a non-zero value, the ban lurker continues to work the ban list as usual top to bottom, but when it reaches the ban_cutoff-th ban, it treats all objects as if they matched a ban and expurges them from cache. As actively used objects get tested against the ban list at request time and thus are likely to be associated with bans near the top of the ban list, with ban_cutoff, least recently accessed objects (the \"long tail\") are removed. This parameter is a safety net to avoid bad response times due to bans being tested at lookup time. Setting a cutoff trades response time for cache efficiency. The recommended value is proportional to rate(bans_lurker_tests_tested) / n_objects while the ban lurker is working, which is the number of bans the system can sustain. The additional latency due to request ban testing is in the order of ban_cutoff / rate(bans_lurker_tests_tested). For example, for rate(bans_lurker_tests_tested) = 2M/s and a tolerable latency of 100ms, a good value for ban_cutoff may be 200K. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0",
        "unit": "bans",
        "value": "0"
    },
    "ban_dups": {
        "default": "on",
        "description": "Eliminate older identical bans when a new ban is added. This saves CPU cycles by not comparing objects to identical bans. This is a waste of time if you have many bans which are never identical. ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "on"
    },
    "ban_lurker_age": {
        "default": "60.000",
        "description": "The ban lurker will ignore bans until they are this old. When a ban is added, the active traffic will be tested against it as part of object lookup. Because many applications issue bans in bursts, this parameter holds the ban-lurker off until the rush is over. This should be set to the approximate time which a ban-burst takes. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "60.000"
    },
    "ban_lurker_batch": {
        "default": "1000",
        "description": "The ban lurker sleeps ${ban_lurker_sleep} after examining this many objects. Use this to pace the ban-lurker if it eats too many resources. ",
        "max": "",
        "min": "1",
        "unit": "",
        "value": "1000"
    },
    "ban_lurker_holdoff": {
        "default": "0.010",
        "description": "How long the ban lurker sleeps when giving way to lookup due to lock contention. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "0.010"
    },
    "ban_lurker_sleep": {
        "default": "0.010",
        "description": "How long the ban lurker sleeps after examining ${ban_lurker_batch} objects. Use this to pace the ban-lurker if it eats too many resources. A value of zero will disable the ban lurker entirely. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "0.010"
    },
    "between_bytes_timeout": {
        "default": "60.000",
        "description": "We only wait for this many seconds between bytes received from the backend before giving up the fetch. A value of zero means never give up. VCL values, per backend or per backend request take precedence. This parameter does not apply to pipe'ed requests. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "60.000"
    },
    "cc_command": {
        "default": "exec",
        "description": "Command used for compiling the C source code to a dlopen(3) loadable object. Any occurrence of %s in the string will be replaced with the source file name, and %o will be replaced with the output file name. NB: This parameter will not take any effect until the VCL programs have been reloaded. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "exec"
    },
    "cli_limit": {
        "default": "128k",
        "description": "Maximum size of CLI response. If the response exceeds this limit, the response code will be 201 instead of 200 and the last line will indicate the truncation. ",
        "max": "99999999b",
        "min": "128b",
        "unit": "bytes",
        "value": "128k"
    },
    "cli_timeout": {
        "default": "60.000",
        "description": "Timeout for the childs replies to CLI requests from the mgt_param. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "60.000"
    },
    "clock_skew": {
        "default": "10",
        "description": "How much clockskew we are willing to accept between the backend and our own clock. ",
        "max": "",
        "min": "0",
        "unit": "seconds",
        "value": "10"
    },
    "clock_step": {
        "default": "1.000",
        "description": "How much observed clock step we are willing to accept before we panic. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "1.000"
    },
    "connect_timeout": {
        "default": "3.500",
        "description": "Default connection timeout for backend connections. We only try to connect to the backend for this many seconds before giving up. VCL can override this default value for each backend and backend request. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "3.500"
    },
    "critbit_cooloff": {
        "default": "180.000",
        "description": "How long the critbit hasher keeps deleted objheads on the cooloff list. NB: Do not change this parameter, unless a developer tell you to do so. ",
        "max": "254.000",
        "min": "60.000",
        "unit": "seconds",
        "value": "180.000"
    },
    "crypto_buffer": {
        "default": "32k",
        "description": "Size of crypto buffer used for Total Encryption processing. These buffers are used for passing data to and from the kernel. If the buffers are too large, the kernel may block. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "2k",
        "unit": "bytes",
        "value": "32k"
    },
    "debug": {
        "default": "none",
        "description": "Enable/Disable various kinds of debugging. none Disable all debugging Use +/- prefix to set/reset individual bits: req_state VSL Request state engine workspace VSL Workspace operations waiter VSL Waiter internals waitinglist VSL Waitinglist events syncvsl Make VSL synchronous hashedge Edge cases in Hash vclrel Rapid VCL release lurker VSL Ban lurker esi_chop Chop ESI fetch to bits flush_head Flush after http1 head vtc_mode Varnishtest Mode witness Emit WITNESS lock records vsm_keep Keep the VSM file on restart drop_pools Drop thread pools (testing) slow_acceptor Slow down Acceptor h2_nocheck Disable various H2 checks vmod_so_keep Keep copied VMOD libraries processors Fetch/Deliver processors protocol Protocol debugging probe VSL health probe events cli CLI debug log to syslog slow_start Add 3 seconds to CLI start failresched Fail from waiting list vcl_keep Keep VCL C and so files ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "none"
    },
    "default_grace": {
        "default": "10.000",
        "description": "Default grace period. We will deliver an object this long after it has expired, provided another thread is attempting to get a new copy. NB: This parameter is evaluated only when objects are created. To change it for all objects, restart or ban everything. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "10.000"
    },
    "default_keep": {
        "default": "0.000",
        "description": "Default keep period. We will keep a useless object around this long, making it available for conditional backend fetches. That means that the object will be removed from the cache at the end of ttl+grace+keep. NB: This parameter is evaluated only when objects are created. To change it for all objects, restart or ban everything. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "0.000"
    },
    "default_ttl": {
        "default": "120.000",
        "description": "The TTL assigned to objects if neither the backend nor the VCL code assigns one. NB: This parameter is evaluated only when objects are created. To change it for all objects, restart or ban everything. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "120.000"
    },
    "esi_iovs": {
        "default": "10",
        "description": "Number of io vectors to allocate on the thread workspace for ESI requests. NB: Do not change this parameter, unless a developer tell you to do so. ",
        "max": "1024",
        "min": "3",
        "unit": "struct iovec",
        "value": "10"
    },
    "feature": {
        "default": "none",
        "description": "Enable/Disable various minor features. none Disable all features. Use +/- prefix to enable/disable individual feature: short_panic Short panic message. wait_silo Wait for persistent silo. no_coredump No coredumps. esi_ignore_https Treat HTTPS as HTTP in ESI:includes esi_disable_xml_check Don't check of body looks like XML esi_ignore_other_elements Ignore non-esi XML-elements esi_remove_bom Remove UTF-8 BOM https_scheme Also split https URIs http2 Support HTTP/2 protocol http_date_postel Relax parsing of timestamps in HTTP headers ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "none"
    },
    "fetch_chunksize": {
        "default": "16k",
        "description": "The default chunksize used by fetcher. This should be bigger than the majority of objects with short TTLs. Internal limits in the storage_file module makes increases above 128kb a dubious idea. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "4k",
        "unit": "bytes",
        "value": "16k"
    },
    "fetch_maxchunksize": {
        "default": "0.25G",
        "description": "The maximum chunksize we attempt to allocate from storage. Making this too large may cause delays and storage fragmentation. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "64k",
        "unit": "bytes",
        "value": "0.25G"
    },
    "first_byte_timeout": {
        "default": "60.000",
        "description": "Default timeout for receiving first byte from backend. We only wait for this many seconds for the first byte before giving up. A value of 0 means it will never time out. VCL can override this default value for each backend and backend request. This parameter does not apply to pipe. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "60.000"
    },
    "gzip_buffer": {
        "default": "32k",
        "description": "Size of malloc buffer used for gzip processing. These buffers are used for in-transit data, for instance gunzip'ed data being sent to a client.Making this space to small results in more overhead, writes to sockets etc, making it too big is probably just a waste of memory. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "2k",
        "unit": "bytes",
        "value": "32k"
    },
    "gzip_level": {
        "default": "6",
        "description": "Gzip compression level: 0=debug, 1=fast, 9=best ",
        "max": "9",
        "min": "0",
        "unit": "",
        "value": "6"
    },
    "gzip_memlevel": {
        "default": "8",
        "description": "Gzip memory level 1=slow/least, 9=fast/most compression. Memory impact is 1=1k, 2=2k, ... 9=256k. ",
        "max": "9",
        "min": "1",
        "unit": "",
        "value": "8"
    },
    "h2_header_table_size": {
        "default": "4k",
        "description": "HTTP2 header table size. This is the size that will be used for the HPACK dynamic decoding table. ",
        "max": "",
        "min": "0b",
        "unit": "bytes",
        "value": "4k"
    },
    "h2_initial_window_size": {
        "default": "65535b",
        "description": "HTTP2 initial flow control window size. ",
        "max": "2147483647b",
        "min": "0b",
        "unit": "bytes",
        "value": "65535b"
    },
    "h2_max_concurrent_streams": {
        "default": "100",
        "description": "HTTP2 Maximum number of concurrent streams. This is the number of requests that can be active at the same time for a single HTTP2 connection. ",
        "max": "",
        "min": "0",
        "unit": "streams",
        "value": "100"
    },
    "h2_max_frame_size": {
        "default": "16k",
        "description": "HTTP2 maximum per frame payload size we are willing to accept. ",
        "max": "16777215b",
        "min": "16k",
        "unit": "bytes",
        "value": "16k"
    },
    "h2_max_header_list_size": {
        "default": "2147483647b",
        "description": "HTTP2 maximum size of an uncompressed header list. ",
        "max": "",
        "min": "0b",
        "unit": "bytes",
        "value": "2147483647b"
    },
    "h2_rx_window_increment": {
        "default": "1M",
        "description": "HTTP2 Receive Window Increments. How big credits we send in WINDOW_UPDATE frames Only affects incoming request bodies (ie: POST, PUT etc.) NB: Do not change this parameter, unless a developer tell you to do so. ",
        "max": "1G",
        "min": "1M",
        "unit": "bytes",
        "value": "1M"
    },
    "h2_rx_window_low_water": {
        "default": "10M",
        "description": "HTTP2 Receive Window low water mark. We try to keep the window at least this big Only affects incoming request bodies (ie: POST, PUT etc.) NB: Do not change this parameter, unless a developer tell you to do so. ",
        "max": "1G",
        "min": "65535b",
        "unit": "bytes",
        "value": "10M"
    },
    "http_gzip_support": {
        "default": "on",
        "description": "Enable gzip support. When enabled Varnish request compressed objects from the backend and store them compressed. If a client does not support gzip encoding Varnish will uncompress compressed objects on demand. Varnish will also rewrite the Accept-Encoding header of clients indicating support for gzip to: Accept-Encoding: gzip Clients that do not support gzip will have their Accept-Encoding header removed. For more information on how gzip is implemented please see the chapter on gzip in the Varnish reference. When gzip support is disabled the variables beresp.do_gzip and beresp.do_gunzip have no effect in VCL. ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "on"
    },
    "http_max_hdr": {
        "default": "64",
        "description": "Maximum number of HTTP header lines we allow in {req|resp|bereq|beresp}.http (obj.http is autosized to the exact number of headers). Cheap, ~20 bytes, in terms of workspace memory. Note that the first line occupies five header lines. ",
        "max": "65535",
        "min": "32",
        "unit": "header lines",
        "value": "64"
    },
    "http_range_support": {
        "default": "on",
        "description": "Enable support for HTTP Range headers. ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "on"
    },
    "http_req_hdr_len": {
        "default": "8k",
        "description": "Maximum length of any HTTP client request header we will allow. The limit is inclusive its continuation lines. ",
        "max": "",
        "min": "40b",
        "unit": "bytes",
        "value": "8k"
    },
    "http_req_size": {
        "default": "32k",
        "description": "Maximum number of bytes of HTTP client request we will deal with. This is a limit on all bytes up to the double blank line which ends the HTTP request. The memory for the request is allocated from the client workspace (param: workspace_client) and this parameter limits how much of that the request is allowed to take up. ",
        "max": "",
        "min": "0.25k",
        "unit": "bytes",
        "value": "32k"
    },
    "http_resp_hdr_len": {
        "default": "8k",
        "description": "Maximum length of any HTTP backend response header we will allow. The limit is inclusive its continuation lines. ",
        "max": "",
        "min": "40b",
        "unit": "bytes",
        "value": "8k"
    },
    "http_resp_size": {
        "default": "32k",
        "description": "Maximum number of bytes of HTTP backend response we will deal with. This is a limit on all bytes up to the double blank line which ends the HTTP response. The memory for the response is allocated from the backend workspace (param: workspace_backend) and this parameter limits how much of that the response is allowed to take up. ",
        "max": "",
        "min": "0.25k",
        "unit": "bytes",
        "value": "32k"
    },
    "idle_send_timeout": {
        "default": "60.000",
        "description": "Send timeout for individual pieces of data on client connections. May get extended if 'send_timeout' applies. When this timeout is hit, the session is closed. See the man page for `setsockopt(2)` under ``SO_SNDTIMEO`` for more information. NB: This parameter may take quite some time to take (full) effect. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "60.000"
    },
    "listen_depth": {
        "default": "1024",
        "description": "Listen queue depth. NB: This parameter will not take any effect until the child process has been restarted. ",
        "max": "",
        "min": "0",
        "unit": "connections",
        "value": "1024"
    },
    "lru_interval": {
        "default": "2.000",
        "description": "Grace period before object moves on LRU list. Objects are only moved to the front of the LRU list if they have not been moved there already inside this timeout period. This reduces the amount of lock operations necessary for LRU list access. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "2.000"
    },
    "max_esi_depth": {
        "default": "5",
        "description": "Maximum depth of esi:include processing. ",
        "max": "",
        "min": "0",
        "unit": "levels",
        "value": "5"
    },
    "max_restarts": {
        "default": "4",
        "description": "Upper limit on how many times a request can restart. ",
        "max": "",
        "min": "0",
        "unit": "restarts",
        "value": "4"
    },
    "max_retries": {
        "default": "4",
        "description": "Upper limit on how many times a backend fetch can retry. ",
        "max": "",
        "min": "0",
        "unit": "retries",
        "value": "4"
    },
    "nuke_limit": {
        "default": "50",
        "description": "Maximum number of objects we attempt to nuke in order to make space for a object body. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0",
        "unit": "allocations",
        "value": "50"
    },
    "object_mutex_slots": {
        "default": "4096",
        "description": "Number of mutex and condvar slots for per object signalling. Objects are assigned randomly to one of these slots. Increasing this number may reduce mutex contention and spurious thread wake ups. NB: This parameter will not take any effect until the child process has been restarted. ",
        "max": "",
        "min": "1",
        "unit": "slots",
        "value": "4096"
    },
    "pcre_match_limit": {
        "default": "10000",
        "description": "The limit for the number of calls to the internal match() function in pcre_exec(). (See: PCRE_EXTRA_MATCH_LIMIT in pcre docs.) This parameter limits how much CPU time regular expression matching can soak up. ",
        "max": "",
        "min": "1",
        "unit": "",
        "value": "10000"
    },
    "pcre_match_limit_recursion": {
        "default": "20",
        "description": "The recursion depth-limit for the internal match() function in a pcre_exec(). (See: PCRE_EXTRA_MATCH_LIMIT_RECURSION in pcre docs.) This puts an upper limit on the amount of stack used by PCRE for certain classes of regular expressions. We have set the default value low in order to prevent crashes, at the cost of possible regexp matching failures. Matching failures will show up in the log as VCL_Error messages with regexp errors -27 or -21. Testcase r01576 can be useful when tuning this parameter. ",
        "max": "",
        "min": "1",
        "unit": "",
        "value": "20"
    },
    "ping_interval": {
        "default": "3",
        "description": "Interval between pings from parent to child. Zero will disable pinging entirely, which makes it possible to attach a debugger to the child. NB: This parameter will not take any effect until the child process has been restarted. ",
        "max": "",
        "min": "0",
        "unit": "seconds",
        "value": "3"
    },
    "pipe_timeout": {
        "default": "60.000",
        "description": "Idle timeout for PIPE sessions. If nothing have been received in either direction for this many seconds, the session is closed. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "60.000"
    },
    "pool_req": {
        "default": "10,100,10",
        "description": "Parameters for per worker pool request memory pool. The three numbers are: min_pool minimum size of free pool. max_pool maximum size of free pool. max_age max age of free element. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "10,100,10"
    },
    "pool_sess": {
        "default": "10,100,10",
        "description": "Parameters for per worker pool session memory pool. The three numbers are: min_pool minimum size of free pool. max_pool maximum size of free pool. max_age max age of free element. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "10,100,10"
    },
    "pool_sslbuffer": {
        "default": "10,100,10",
        "description": "Parameters for ssl buffer pool. The three numbers are: min_pool minimum size of free pool. max_pool maximum size of free pool. max_age max age of free element. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "10,100,10"
    },
    "pool_vbo": {
        "default": "10,100,10",
        "description": "Parameters for backend object fetch memory pool. The three numbers are: min_pool minimum size of free pool. max_pool maximum size of free pool. max_age max age of free element. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "10,100,10"
    },
    "prefer_ipv6": {
        "default": "off",
        "description": "Prefer IPv6 address when connecting to backends which have both IPv4 and IPv6 addresses. ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "off"
    },
    "rush_exponent": {
        "default": "3",
        "description": "How many parked request we start for each completed request on the object. NB: Even with the implict delay of delivery, this parameter controls an exponential increase in number of worker threads. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "2",
        "unit": "requests per request",
        "value": "3"
    },
    "send_timeout": {
        "default": "600.000",
        "description": "Total timeout for ordinary HTTP1 responses. Does not apply to some internally generated errors and pipe mode. When 'idle_send_timeout' is hit while sending an HTTP1 response, the timeout is extended unless the total time already taken for sending the response in its entirety exceeds this many seconds. When this timeout is hit, the session is closed NB: This parameter may take quite some time to take (full) effect. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "600.000"
    },
    "shm_reclen": {
        "default": "255b",
        "description": "Old name for vsl_reclen, use that instead. ",
        "max": "4084",
        "min": "16b",
        "unit": "bytes",
        "value": "255b"
    },
    "shortlived": {
        "default": "10.000",
        "description": "Objects created with (ttl+grace+keep) shorter than this are always put in transient storage. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "10.000"
    },
    "sigsegv_handler": {
        "default": "on",
        "description": "Install a signal handler which tries to dump debug information on segmentation faults, bus errors and abort signals. NB: This parameter will not take any effect until the child process has been restarted. ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "on"
    },
    "ssl_buffer": {
        "default": "16k",
        "description": "Size of SSL buffer. ",
        "max": "",
        "min": "0.25k",
        "unit": "bytes",
        "value": "16k"
    },
    "startup_timeout": {
        "default": "60.000",
        "description": "Timeout for CLI commands during the child's startup, including automatic restarts from the manager. This parameter only takes effect if it is larger than cli_timeout. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "60.000"
    },
    "syslog_cli_traffic": {
        "default": "on",
        "description": "Log all CLI traffic to syslog(LOG_INFO). ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "on"
    },
    "tcp_fastopen": {
        "default": "off",
        "description": "Enable TCP Fast Open extension. NB: This parameter will not take any effect until the child process has been restarted. ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "off"
    },
    "tcp_keepalive_intvl": {
        "default": "75.000",
        "description": "The number of seconds between TCP keep-alive probes. Ignored for Unix domain sockets. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "100.000",
        "min": "1.000",
        "unit": "seconds",
        "value": "75.000"
    },
    "tcp_keepalive_probes": {
        "default": "9",
        "description": "The maximum number of TCP keep-alive probes to send before giving up and killing the connection if no response is obtained from the other end. Ignored for Unix domain sockets. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "100",
        "min": "1",
        "unit": "probes",
        "value": "9"
    },
    "tcp_keepalive_time": {
        "default": "7200.000",
        "description": "The number of seconds a connection needs to be idle before TCP begins sending out keep-alive probes. Ignored for Unix domain sockets. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "7200.000",
        "min": "1.000",
        "unit": "seconds",
        "value": "7200.000"
    },
    "thread_pool_add_delay": {
        "default": "0.000",
        "description": "Wait at least this long after creating a thread. Some (buggy) systems may need a short (sub-second) delay between creating threads. Set this to a few milliseconds if you see the 'threads_failed' counter grow too much. Setting this too high results in insufficient worker threads. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "0.000"
    },
    "thread_pool_destroy_delay": {
        "default": "1.000",
        "description": "Wait this long after destroying a thread. This controls the decay of thread pools when idle(-ish). NB: This parameter may take quite some time to take (full) effect. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0.010",
        "unit": "seconds",
        "value": "1.000"
    },
    "thread_pool_fail_delay": {
        "default": "0.200",
        "description": "Wait at least this long after a failed thread creation before trying to create another thread. Failure to create a worker thread is often a sign that the end is near, because the process is running out of some resource. This delay tries to not rush the end on needlessly. If thread creation failures are a problem, check that thread_pool_max is not too high. It may also help to increase thread_pool_timeout and thread_pool_min, to reduce the rate at which treads are destroyed and later recreated. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0.010",
        "unit": "seconds",
        "value": "0.200"
    },
    "thread_pool_max": {
        "default": "5000",
        "description": "The maximum number of worker threads in each pool. The minimum value depends on thread_pool_min. Do not set this higher than you have to, since excess worker threads soak up RAM and CPU and generally just get in the way of getting work done. NB: This parameter may take quite some time to take (full) effect. ",
        "max": "",
        "min": "100",
        "unit": "threads",
        "value": "5000"
    },
    "thread_pool_min": {
        "default": "100",
        "description": "The minimum number of worker threads in each pool. The maximum value depends on thread_pool_max. Increasing this may help ramp up faster from low load situations or when threads have expired. Minimum is 10 threads. NB: This parameter may take quite some time to take (full) effect. ",
        "max": "5000",
        "min": "",
        "unit": "threads",
        "value": "100"
    },
    "thread_pool_reserve": {
        "default": "0",
        "description": "The number of worker threads reserved for vital tasks in each pool. Tasks may require other tasks to complete (for example, client requests may require backend requests). This reserve is to ensure that such tasks still get to run even under high load. Increasing the reserve may help setups with a high number of backend requests at the expense of client performance. Setting it too high will waste resources by keeping threads unused. Default is 0 to auto-tune (currently 5% of thread_pool_min). Minimum is 1 otherwise, maximum is 95% of thread_pool_min. NB: This parameter may take quite some time to take (full) effect. ",
        "max": "95",
        "min": "",
        "unit": "threads",
        "value": "0"
    },
    "thread_pool_stack": {
        "default": "48k",
        "description": "Worker thread stack size. This will likely be rounded up to a multiple of 4k (or whatever the page_size might be) by the kernel. The required stack size is primarily driven by the depth of the call-tree. The most common relevant determining factors in varnish core code are GZIP (un)compression, ESI processing and regular expression matches. VMODs may also require significant amounts of additional stack. The nesting depth of VCL subs is another factor, although typically not predominant. The stack size is per thread, so the maximum total memory required for worker thread stacks is in the order of size = thread_pools x thread_pool_max x thread_pool_stack. Thus, in particular for setups with many threads, keeping the stack size at a minimum helps reduce the amount of memory required by Varnish. On the other hand, thread_pool_stack must be large enough under all circumstances, otherwise varnish will crash due to a stack overflow. Usually, a stack overflow manifests itself as a segmentation fault (aka segfault / SIGSEGV) with the faulting address being near the stack pointer (sp). Unless stack usage can be reduced, thread_pool_stack must be increased when a stack overflow occurs. Setting it in 150%-200% increments is recommended until stack overflows cease to occur. NB: This parameter may take quite some time to take (full) effect. ",
        "max": "",
        "min": "16k",
        "unit": "bytes",
        "value": "48k"
    },
    "thread_pool_timeout": {
        "default": "300.000",
        "description": "Thread idle threshold. Threads in excess of thread_pool_min, which have been idle for at least this long, will be destroyed. NB: This parameter may take quite some time to take (full) effect. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "10.000",
        "unit": "seconds",
        "value": "300.000"
    },
    "thread_pool_watchdog": {
        "default": "60.000",
        "description": "Thread queue stuck watchdog. If no queued work have been released for this long, the worker process panics itself. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0.100",
        "unit": "seconds",
        "value": "60.000"
    },
    "thread_pools": {
        "default": "2",
        "description": "Number of worker thread pools. Increasing the number of worker pools decreases lock contention. Each worker pool also has a thread accepting new connections, so for very high rates of incoming new connections on systems with many cores, increasing the worker pools may be required. Too many pools waste CPU and RAM resources, and more than one pool for each CPU is most likely detrimental to performance. Can be increased on the fly, but decreases require a restart to take effect. NB: This parameter may take quite some time to take (full) effect. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "32",
        "min": "1",
        "unit": "pools",
        "value": "2"
    },
    "thread_queue_limit": {
        "default": "20",
        "description": "Permitted request queue length per thread-pool. This sets the number of requests we will queue, waiting for an available thread. Above this limit sessions will be dropped instead of queued. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0",
        "unit": "",
        "value": "20"
    },
    "thread_stats_rate": {
        "default": "10",
        "description": "Worker threads accumulate statistics, and dump these into the global stats counters if the lock is free when they finish a job (request/fetch etc.) This parameters defines the maximum number of jobs a worker thread may handle, before it is forced to dump its accumulated stats into the global counters. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0",
        "unit": "requests",
        "value": "10"
    },
    "timeout_idle": {
        "default": "5.000",
        "description": "Idle timeout for client connections. A connection is considered idle until we have received the full request headers. This parameter is particularly relevant for HTTP1 keepalive connections which are closed unless the next request is received before this timeout is reached. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "5.000"
    },
    "timeout_linger": {
        "default": "0.050",
        "description": "How long the worker thread lingers on an idle session before handing it over to the waiter. When sessions are reused, as much as half of all reuses happen within the first 100 msec of the previous request completing. Setting this too high results in worker threads not doing anything for their keep, setting it too low just means that more sessions take a detour around the waiter. NB: We do not know yet if it is a good idea to change this parameter, or if the default value is even sensible. Caution is advised, and feedback is most welcome. ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "0.050"
    },
    "vcc_allow_inline_c": {
        "default": "off",
        "description": "Allow inline C code in VCL. ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "off"
    },
    "vcc_err_unref": {
        "default": "on",
        "description": "Unreferenced VCL objects result in error. ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "on"
    },
    "vcc_unsafe_path": {
        "default": "on",
        "description": "Allow '/' in vmod & include paths. Allow 'import ... from ...'. ",
        "max": "",
        "min": "",
        "unit": "bool",
        "value": "on"
    },
    "vcl_cooldown": {
        "default": "600.000",
        "description": "How long a VCL is kept warm after being replaced as the active VCL (granularity approximately 30 seconds). ",
        "max": "",
        "min": "0.000",
        "unit": "seconds",
        "value": "600.000"
    },
    "vcl_dir": {
        "default": "/usr/etc/varnish:/usr/share/varnish-plus/vcl",
        "description": "Old name for vcl_path, use that instead. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "/usr/etc/varnish:/usr/share/varnish-plus/vcl"
    },
    "vcl_path": {
        "default": "/usr/etc/varnish:/usr/share/varnish-plus/vcl",
        "description": "Directory (or colon separated list of directories) from which relative VCL filenames (vcl.load and include) are to be found. By default Varnish searches VCL files in both the system configuration and shared data directories to allow packages to drop their VCL files in a standard location where relative includes would work. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "/usr/etc/varnish:/usr/share/varnish-plus/vcl"
    },
    "vmod_dir": {
        "default": "/usr/lib/varnish-plus/vmods",
        "description": "Old name for vmod_path, use that instead. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "/usr/lib/varnish-plus/vmods"
    },
    "vmod_path": {
        "default": "/usr/lib/varnish-plus/vmods",
        "description": "Directory (or colon separated list of directories) where VMODs are to be found. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "/usr/lib/varnish-plus/vmods"
    },
    "vsl_buffer": {
        "default": "4k",
        "description": "Bytes of (req-/backend-)workspace dedicated to buffering VSL records. When this parameter is adjusted, most likely workspace_client and workspace_backend will have to be adjusted by the same amount. Setting this too high costs memory, setting it too low will cause more VSL flushes and likely increase lock-contention on the VSL mutex. The minimum tracks the vsl_reclen parameter + 12 bytes. ",
        "max": "",
        "min": "267",
        "unit": "bytes",
        "value": "4k"
    },
    "vsl_mask": {
        "default": "-ObjProtocol,-ObjStatus,-ObjReason,-ObjHeader,-VCL_trace,-WorkThread,-Hash,-VfpAcct,-H2RxHdr,-H2RxBody,-H2TxHdr,-H2TxBody",
        "description": "Mask individual VSL messages from being logged. default Set default value Use +/- prefix in front of VSL tag name, to mask/unmask individual VSL messages. ",
        "max": "",
        "min": "",
        "unit": "",
        "value": "-ObjProtocol,-ObjStatus,-ObjReason,-ObjHeader,-VCL_trace,-WorkThread,-Hash,-VfpAcct,-H2RxHdr,-H2RxBody,-H2TxHdr,-H2TxBody"
    },
    "vsl_reclen": {
        "default": "255b",
        "description": "Maximum number of bytes in SHM log record. The maximum tracks the vsl_buffer parameter - 12 bytes. ",
        "max": "4084",
        "min": "16b",
        "unit": "bytes",
        "value": "255b"
    },
    "vsl_space": {
        "default": "80M",
        "description": "The amount of space to allocate for the VSL fifo buffer in the VSM memory segment. If you make this too small, varnish{ncsa|log} etc will not be able to keep up. Making it too large just costs memory resources. NB: This parameter will not take any effect until the child process has been restarted. ",
        "max": "4G",
        "min": "1M",
        "unit": "bytes",
        "value": "80M"
    },
    "vsm_free_cooldown": {
        "default": "60.000",
        "description": "How long VSM memory is kept warm after a deallocation (granularity approximately 2 seconds). ",
        "max": "600.000",
        "min": "10.000",
        "unit": "seconds",
        "value": "60.000"
    },
    "vsm_space": {
        "default": "1M",
        "description": "DEPRECATED: This parameter is ignored. There is no global limit on amount of shared memory now. ",
        "max": "1G",
        "min": "1M",
        "unit": "bytes",
        "value": "1M"
    },
    "workspace_backend": {
        "default": "64k",
        "description": "Bytes of HTTP protocol workspace for backend HTTP req/resp. If larger than 4k, use a multiple of 4k for VM efficiency. NB: This parameter may take quite some time to take (full) effect. ",
        "max": "",
        "min": "1k",
        "unit": "bytes",
        "value": "64k"
    },
    "workspace_client": {
        "default": "64k",
        "description": "Bytes of HTTP protocol workspace for clients HTTP req/resp. Use a multiple of 4k for VM efficiency. For HTTP/2 compliance this must be at least 20k, in order to receive fullsize (=16k) frames from the client. That usually happens only in POST/PUT bodies. For other traffic-patterns smaller values work just fine. NB: This parameter may take quite some time to take (full) effect. ",
        "max": "",
        "min": "9k",
        "unit": "bytes",
        "value": "64k"
    },
    "workspace_session": {
        "default": "0.50k",
        "description": "Allocation size for session structure and workspace. The workspace is primarily used for TCP connection addresses. If larger than 4k, use a multiple of 4k for VM efficiency. NB: This parameter may take quite some time to take (full) effect. ",
        "max": "",
        "min": "0.25k",
        "unit": "bytes",
        "value": "0.50k"
    },
    "workspace_thread": {
        "default": "2k",
        "description": "Bytes of auxiliary workspace per thread. This workspace is used for certain temporary data structures during the operation of a worker thread. One use is for the IO-vectors used during delivery. Setting this parameter too low may increase the number of writev() syscalls, setting it too high just wastes space. ~0.1k + UIO_MAXIOV * sizeof(struct iovec) (typically = ~16k for 64bit) is considered the maximum sensible value under any known circumstances (excluding exotic vmod use). NB: This parameter may take quite some time to take (full) effect. ",
        "max": "8k",
        "min": "0.25k",
        "unit": "bytes",
        "value": "2k"
    }
}



GET

PONG 1558780038


GET

[
    "DELETE /vcl/:name",
    "GET /backend",
    "GET /ban",
    "GET /banner",
    "GET /panic",
    "GET /param/:name",
    "GET /paramjson",
    "GET /ping",
    "GET /resources",
    "GET /stats",
    "GET /status",
    "GET /vcl",
    "GET /vcl/:name",
    "GET /vclactive",
    "GET /vcljson",
    "GET /version",
    "POST /ban",
    "POST /direct",
    "POST /start",
    "POST /stop",
    "POST /vcl",
    "PUT /backend/:name",
    "PUT /param/:name",
    "PUT /push/url/stats",
    "PUT /start",
    "PUT /stop",
    "PUT /vcl/:name",
    "PUT /vcldeploy/:name"
]


GET

{
    "LCK.backend.creat": {
        "value": 2
    },
    "LCK.backend.destroy": {
        "value": 0
    },
    "LCK.backend.locks": {
        "value": 1
    },
    "LCK.ban.creat": {
        "value": 1
    },
    "LCK.ban.destroy": {
        "value": 0
    },
    "LCK.ban.locks": {
        "value": 556
    },
    "LCK.busyobj.creat": {
        "value": 4096
    },
    "LCK.busyobj.destroy": {
        "value": 0
    },
    "LCK.busyobj.locks": {
        "value": 0
    },
    "LCK.cli.creat": {
        "value": 1
    },
    "LCK.cli.destroy": {
        "value": 0
    },
    "LCK.cli.locks": {
        "value": 4480
    },
    "LCK.exp.creat": {
        "value": 1
    },
    "LCK.exp.destroy": {
        "value": 0
    },
    "LCK.exp.locks": {
        "value": 4259
    },
    "LCK.hcb.creat": {
        "value": 1
    },
    "LCK.hcb.destroy": {
        "value": 0
    },
    "LCK.hcb.locks": {
        "value": 75
    },
    "LCK.lru.creat": {
        "value": 2
    },
    "LCK.lru.destroy": {
        "value": 0
    },
    "LCK.lru.locks": {
        "value": 0
    },
    "LCK.mempool.creat": {
        "value": 6
    },
    "LCK.mempool.destroy": {
        "value": 0
    },
    "LCK.mempool.locks": {
        "value": 51728
    },
    "LCK.mse.creat": {
        "value": 0
    },
    "LCK.mse.destroy": {
        "value": 0
    },
    "LCK.mse.locks": {
        "value": 0
    },
    "LCK.objhdr.creat": {
        "value": 1
    },
    "LCK.objhdr.destroy": {
        "value": 0
    },
    "LCK.objhdr.locks": {
        "value": 0
    },
    "LCK.pipestat.creat": {
        "value": 1
    },
    "LCK.pipestat.destroy": {
        "value": 0
    },
    "LCK.pipestat.locks": {
        "value": 0
    },
    "LCK.sess.creat": {
        "value": 0
    },
    "LCK.sess.destroy": {
        "value": 0
    },
    "LCK.sess.locks": {
        "value": 0
    },
    "LCK.sma.creat": {
        "value": 2
    },
    "LCK.sma.destroy": {
        "value": 0
    },
    "LCK.sma.locks": {
        "value": 0
    },
    "LCK.tcp_pool.creat": {
        "value": 2
    },
    "LCK.tcp_pool.destroy": {
        "value": 0
    },
    "LCK.tcp_pool.locks": {
        "value": 2
    },
    "LCK.vbe.creat": {
        "value": 1
    },
    "LCK.vbe.destroy": {
        "value": 0
    },
    "LCK.vbe.locks": {
        "value": 4470
    },
    "LCK.vcapace.creat": {
        "value": 1
    },
    "LCK.vcapace.destroy": {
        "value": 0
    },
    "LCK.vcapace.locks": {
        "value": 0
    },
    "LCK.vcl.creat": {
        "value": 1
    },
    "LCK.vcl.destroy": {
        "value": 0
    },
    "LCK.vcl.locks": {
        "value": 6
    },
    "LCK.vxid.creat": {
        "value": 1
    },
    "LCK.vxid.destroy": {
        "value": 0
    },
    "LCK.vxid.locks": {
        "value": 0
    },
    "LCK.waiter.creat": {
        "value": 2
    },
    "LCK.waiter.destroy": {
        "value": 0
    },
    "LCK.waiter.locks": {
        "value": 268
    },
    "LCK.wq.creat": {
        "value": 3
    },
    "LCK.wq.destroy": {
        "value": 0
    },
    "LCK.wq.locks": {
        "value": 13861
    },
    "LCK.wstat.creat": {
        "value": 1
    },
    "LCK.wstat.destroy": {
        "value": 0
    },
    "LCK.wstat.locks": {
        "value": 4609
    },
    "LCK.ykey.creat": {
        "value": 2
    },
    "LCK.ykey.destroy": {
        "value": 0
    },
    "LCK.ykey.locks": {
        "value": 0
    },
    "MAIN.backend_busy": {
        "value": 0
    },
    "MAIN.backend_conn": {
        "value": 0
    },
    "MAIN.backend_fail": {
        "value": 0
    },
    "MAIN.backend_recycle": {
        "value": 0
    },
    "MAIN.backend_req": {
        "value": 0
    },
    "MAIN.backend_retry": {
        "value": 0
    },
    "MAIN.backend_reuse": {
        "value": 0
    },
    "MAIN.backend_unhealthy": {
        "value": 0
    },
    "MAIN.bans": {
        "value": 1
    },
    "MAIN.bans_added": {
        "value": 1
    },
    "MAIN.bans_completed": {
        "value": 1
    },
    "MAIN.bans_deleted": {
        "value": 0
    },
    "MAIN.bans_dups": {
        "value": 0
    },
    "MAIN.bans_lurker_contention": {
        "value": 0
    },
    "MAIN.bans_lurker_obj_killed": {
        "value": 0
    },
    "MAIN.bans_lurker_obj_killed_cutoff": {
        "value": 0
    },
    "MAIN.bans_lurker_tested": {
        "value": 0
    },
    "MAIN.bans_lurker_tests_tested": {
        "value": 0
    },
    "MAIN.bans_obj": {
        "value": 0
    },
    "MAIN.bans_obj_killed": {
        "value": 0
    },
    "MAIN.bans_req": {
        "value": 0
    },
    "MAIN.bans_tested": {
        "value": 0
    },
    "MAIN.bans_tests_tested": {
        "value": 0
    },
    "MAIN.busy_killed": {
        "value": 0
    },
    "MAIN.busy_sleep": {
        "value": 0
    },
    "MAIN.busy_wakeup": {
        "value": 0
    },
    "MAIN.cache_hit": {
        "value": 0
    },
    "MAIN.cache_hit_grace": {
        "value": 0
    },
    "MAIN.cache_hitmiss": {
        "value": 0
    },
    "MAIN.cache_hitpass": {
        "value": 0
    },
    "MAIN.cache_miss": {
        "value": 0
    },
    "MAIN.client_req": {
        "value": 0
    },
    "MAIN.client_req_400": {
        "value": 0
    },
    "MAIN.client_req_417": {
        "value": 0
    },
    "MAIN.client_resp_500": {
        "value": 0
    },
    "MAIN.esi_errors": {
        "value": 0
    },
    "MAIN.esi_maxdepth": {
        "value": 0
    },
    "MAIN.esi_warnings": {
        "value": 0
    },
    "MAIN.exp_mailed": {
        "value": 0
    },
    "MAIN.exp_received": {
        "value": 0
    },
    "MAIN.fetch_1xx": {
        "value": 0
    },
    "MAIN.fetch_204": {
        "value": 0
    },
    "MAIN.fetch_304": {
        "value": 0
    },
    "MAIN.fetch_bad": {
        "value": 0
    },
    "MAIN.fetch_chunked": {
        "value": 0
    },
    "MAIN.fetch_eof": {
        "value": 0
    },
    "MAIN.fetch_failed": {
        "value": 0
    },
    "MAIN.fetch_head": {
        "value": 0
    },
    "MAIN.fetch_length": {
        "value": 0
    },
    "MAIN.fetch_no_thread": {
        "value": 0
    },
    "MAIN.fetch_none": {
        "value": 0
    },
    "MAIN.goto_dns_cache_hits": {
        "value": 0
    },
    "MAIN.goto_dns_lookup_fails": {
        "value": 0
    },
    "MAIN.goto_dns_lookups": {
        "value": 0
    },
    "MAIN.hcb_insert": {
        "value": 0
    },
    "MAIN.hcb_lock": {
        "value": 0
    },
    "MAIN.hcb_nolock": {
        "value": 0
    },
    "MAIN.losthdr": {
        "value": 0
    },
    "MAIN.n_backend": {
        "value": 1
    },
    "MAIN.n_expired": {
        "value": 0
    },
    "MAIN.n_gunzip": {
        "value": 0
    },
    "MAIN.n_gzip": {
        "value": 0
    },
    "MAIN.n_lru_limited": {
        "value": 0
    },
    "MAIN.n_lru_moved": {
        "value": 0
    },
    "MAIN.n_lru_nuked": {
        "value": 0
    },
    "MAIN.n_obj_purged": {
        "value": 0
    },
    "MAIN.n_object": {
        "value": 0
    },
    "MAIN.n_objectcore": {
        "value": 0
    },
    "MAIN.n_objecthead": {
        "value": 0
    },
    "MAIN.n_purges": {
        "value": 0
    },
    "MAIN.n_test_gunzip": {
        "value": 0
    },
    "MAIN.n_vampireobject": {
        "value": 0
    },
    "MAIN.n_vcl": {
        "value": 1
    },
    "MAIN.n_vcl_avail": {
        "value": 1
    },
    "MAIN.n_vcl_discard": {
        "value": 0
    },
    "MAIN.pools": {
        "value": 2
    },
    "MAIN.req_dropped": {
        "value": 0
    },
    "MAIN.s_fetch": {
        "value": 0
    },
    "MAIN.s_pass": {
        "value": 0
    },
    "MAIN.s_pipe": {
        "value": 0
    },
    "MAIN.s_pipe_hdrbytes": {
        "value": 0
    },
    "MAIN.s_pipe_in": {
        "value": 0
    },
    "MAIN.s_pipe_out": {
        "value": 0
    },
    "MAIN.s_req_bodybytes": {
        "value": 0
    },
    "MAIN.s_req_hdrbytes": {
        "value": 0
    },
    "MAIN.s_resp_bodybytes": {
        "value": 0
    },
    "MAIN.s_resp_hdrbytes": {
        "value": 0
    },
    "MAIN.s_sess": {
        "value": 0
    },
    "MAIN.s_synth": {
        "value": 0
    },
    "MAIN.sc_overload": {
        "value": 0
    },
    "MAIN.sc_pipe_overflow": {
        "value": 0
    },
    "MAIN.sc_range_short": {
        "value": 0
    },
    "MAIN.sc_rem_close": {
        "value": 0
    },
    "MAIN.sc_req_close": {
        "value": 0
    },
    "MAIN.sc_req_http10": {
        "value": 0
    },
    "MAIN.sc_req_http20": {
        "value": 0
    },
    "MAIN.sc_resp_close": {
        "value": 0
    },
    "MAIN.sc_rx_bad": {
        "value": 0
    },
    "MAIN.sc_rx_body": {
        "value": 0
    },
    "MAIN.sc_rx_junk": {
        "value": 0
    },
    "MAIN.sc_rx_overflow": {
        "value": 0
    },
    "MAIN.sc_rx_timeout": {
        "value": 0
    },
    "MAIN.sc_stream_failure": {
        "value": 0
    },
    "MAIN.sc_tx_eof": {
        "value": 0
    },
    "MAIN.sc_tx_error": {
        "value": 0
    },
    "MAIN.sc_tx_pipe": {
        "value": 0
    },
    "MAIN.sc_vcl_failure": {
        "value": 0
    },
    "MAIN.sc_vdp_error": {
        "value": 0
    },
    "MAIN.sess_closed": {
        "value": 0
    },
    "MAIN.sess_closed_err": {
        "value": 0
    },
    "MAIN.sess_conn": {
        "value": 0
    },
    "MAIN.sess_drop": {
        "value": 0
    },
    "MAIN.sess_dropped": {
        "value": 0
    },
    "MAIN.sess_fail": {
        "value": 4
    },
    "MAIN.sess_fail_ebadf": {
        "value": 0
    },
    "MAIN.sess_fail_econnaborted": {
        "value": 0
    },
    "MAIN.sess_fail_eintr": {
        "value": 4
    },
    "MAIN.sess_fail_emfile": {
        "value": 0
    },
    "MAIN.sess_fail_enomem": {
        "value": 0
    },
    "MAIN.sess_fail_other": {
        "value": 0
    },
    "MAIN.sess_herd": {
        "value": 0
    },
    "MAIN.sess_queued": {
        "value": 0
    },
    "MAIN.sess_readahead": {
        "value": 0
    },
    "MAIN.shm_cont": {
        "value": 0
    },
    "MAIN.shm_cycles": {
        "value": 0
    },
    "MAIN.shm_flushes": {
        "value": 0
    },
    "MAIN.shm_records": {
        "value": 8942
    },
    "MAIN.shm_writes": {
        "value": 8942
    },
    "MAIN.summs": {
        "value": 4
    },
    "MAIN.thread_queue_len": {
        "value": 0
    },
    "MAIN.threads": {
        "value": 200
    },
    "MAIN.threads_created": {
        "value": 200
    },
    "MAIN.threads_destroyed": {
        "value": 0
    },
    "MAIN.threads_failed": {
        "value": 0
    },
    "MAIN.threads_limited": {
        "value": 0
    },
    "MAIN.uptime": {
        "value": 13765
    },
    "MAIN.vcl_fail": {
        "value": 0
    },
    "MAIN.vmods": {
        "value": 0
    },
    "MAIN.ws_backend_overflow": {
        "value": 0
    },
    "MAIN.ws_client_overflow": {
        "value": 0
    },
    "MAIN.ws_session_overflow": {
        "value": 0
    },
    "MAIN.ws_thread_overflow": {
        "value": 0
    },
    "MEMPOOL.busyobj.allocs": {
        "value": 0
    },
    "MEMPOOL.busyobj.frees": {
        "value": 0
    },
    "MEMPOOL.busyobj.live": {
        "value": 0
    },
    "MEMPOOL.busyobj.pool": {
        "value": 10
    },
    "MEMPOOL.busyobj.randry": {
        "value": 0
    },
    "MEMPOOL.busyobj.recycle": {
        "value": 0
    },
    "MEMPOOL.busyobj.surplus": {
        "value": 0
    },
    "MEMPOOL.busyobj.sz_actual": {
        "value": 65504
    },
    "MEMPOOL.busyobj.sz_wanted": {
        "value": 65536
    },
    "MEMPOOL.busyobj.timeout": {
        "value": 0
    },
    "MEMPOOL.busyobj.toosmall": {
        "value": 0
    },
    "MEMPOOL.req0.allocs": {
        "value": 0
    },
    "MEMPOOL.req0.frees": {
        "value": 0
    },
    "MEMPOOL.req0.live": {
        "value": 0
    },
    "MEMPOOL.req0.pool": {
        "value": 10
    },
    "MEMPOOL.req0.randry": {
        "value": 0
    },
    "MEMPOOL.req0.recycle": {
        "value": 0
    },
    "MEMPOOL.req0.surplus": {
        "value": 0
    },
    "MEMPOOL.req0.sz_actual": {
        "value": 65504
    },
    "MEMPOOL.req0.sz_wanted": {
        "value": 65536
    },
    "MEMPOOL.req0.timeout": {
        "value": 0
    },
    "MEMPOOL.req0.toosmall": {
        "value": 0
    },
    "MEMPOOL.req1.allocs": {
        "value": 0
    },
    "MEMPOOL.req1.frees": {
        "value": 0
    },
    "MEMPOOL.req1.live": {
        "value": 0
    },
    "MEMPOOL.req1.pool": {
        "value": 10
    },
    "MEMPOOL.req1.randry": {
        "value": 0
    },
    "MEMPOOL.req1.recycle": {
        "value": 0
    },
    "MEMPOOL.req1.surplus": {
        "value": 0
    },
    "MEMPOOL.req1.sz_actual": {
        "value": 65504
    },
    "MEMPOOL.req1.sz_wanted": {
        "value": 65536
    },
    "MEMPOOL.req1.timeout": {
        "value": 0
    },
    "MEMPOOL.req1.toosmall": {
        "value": 0
    },
    "MEMPOOL.sess0.allocs": {
        "value": 0
    },
    "MEMPOOL.sess0.frees": {
        "value": 0
    },
    "MEMPOOL.sess0.live": {
        "value": 0
    },
    "MEMPOOL.sess0.pool": {
        "value": 10
    },
    "MEMPOOL.sess0.randry": {
        "value": 0
    },
    "MEMPOOL.sess0.recycle": {
        "value": 0
    },
    "MEMPOOL.sess0.surplus": {
        "value": 0
    },
    "MEMPOOL.sess0.sz_actual": {
        "value": 480
    },
    "MEMPOOL.sess0.sz_wanted": {
        "value": 512
    },
    "MEMPOOL.sess0.timeout": {
        "value": 0
    },
    "MEMPOOL.sess0.toosmall": {
        "value": 0
    },
    "MEMPOOL.sess1.allocs": {
        "value": 0
    },
    "MEMPOOL.sess1.frees": {
        "value": 0
    },
    "MEMPOOL.sess1.live": {
        "value": 0
    },
    "MEMPOOL.sess1.pool": {
        "value": 10
    },
    "MEMPOOL.sess1.randry": {
        "value": 0
    },
    "MEMPOOL.sess1.recycle": {
        "value": 0
    },
    "MEMPOOL.sess1.surplus": {
        "value": 0
    },
    "MEMPOOL.sess1.sz_actual": {
        "value": 480
    },
    "MEMPOOL.sess1.sz_wanted": {
        "value": 512
    },
    "MEMPOOL.sess1.timeout": {
        "value": 0
    },
    "MEMPOOL.sess1.toosmall": {
        "value": 0
    },
    "MEMPOOL.ssl_buf.allocs": {
        "value": 0
    },
    "MEMPOOL.ssl_buf.frees": {
        "value": 0
    },
    "MEMPOOL.ssl_buf.live": {
        "value": 0
    },
    "MEMPOOL.ssl_buf.pool": {
        "value": 10
    },
    "MEMPOOL.ssl_buf.randry": {
        "value": 0
    },
    "MEMPOOL.ssl_buf.recycle": {
        "value": 0
    },
    "MEMPOOL.ssl_buf.surplus": {
        "value": 0
    },
    "MEMPOOL.ssl_buf.sz_actual": {
        "value": 16352
    },
    "MEMPOOL.ssl_buf.sz_wanted": {
        "value": 16384
    },
    "MEMPOOL.ssl_buf.timeout": {
        "value": 0
    },
    "MEMPOOL.ssl_buf.toosmall": {
        "value": 0
    },
    "MGT.child_died": {
        "value": 0
    },
    "MGT.child_dump": {
        "value": 0
    },
    "MGT.child_exit": {
        "value": 0
    },
    "MGT.child_panic": {
        "value": 0
    },
    "MGT.child_start": {
        "value": 1
    },
    "MGT.child_stop": {
        "value": 0
    },
    "MGT.uptime": {
        "value": 13763
    },
    "SMA.Transient.c_bytes": {
        "value": 0
    },
    "SMA.Transient.c_fail": {
        "value": 0
    },
    "SMA.Transient.c_freed": {
        "value": 0
    },
    "SMA.Transient.c_req": {
        "value": 0
    },
    "SMA.Transient.g_alloc": {
        "value": 0
    },
    "SMA.Transient.g_bytes": {
        "value": 0
    },
    "SMA.Transient.g_space": {
        "value": 0
    },
    "SMA.Transient.g_ykey_keys": {
        "value": 0
    },
    "SMA.s0.c_bytes": {
        "value": 0
    },
    "SMA.s0.c_fail": {
        "value": 0
    },
    "SMA.s0.c_freed": {
        "value": 0
    },
    "SMA.s0.c_req": {
        "value": 0
    },
    "SMA.s0.g_alloc": {
        "value": 0
    },
    "SMA.s0.g_bytes": {
        "value": 0
    },
    "SMA.s0.g_space": {
        "value": 104857600
    },
    "SMA.s0.g_ykey_keys": {
        "value": 0
    },
    "VBE.boot.default.bereq_bodybytes": {
        "value": 0
    },
    "VBE.boot.default.bereq_hdrbytes": {
        "value": 0
    },
    "VBE.boot.default.beresp_bodybytes": {
        "value": 0
    },
    "VBE.boot.default.beresp_hdrbytes": {
        "value": 0
    },
    "VBE.boot.default.busy": {
        "value": 0
    },
    "VBE.boot.default.conn": {
        "value": 0
    },
    "VBE.boot.default.fail": {
        "value": 0
    },
    "VBE.boot.default.fail_eacces": {
        "value": 0
    },
    "VBE.boot.default.fail_eaddrnotavail": {
        "value": 0
    },
    "VBE.boot.default.fail_econnrefused": {
        "value": 0
    },
    "VBE.boot.default.fail_enetunreach": {
        "value": 0
    },
    "VBE.boot.default.fail_etimedout": {
        "value": 0
    },
    "VBE.boot.default.fail_other": {
        "value": 0
    },
    "VBE.boot.default.happy": {
        "value": 0
    },
    "VBE.boot.default.helddown": {
        "value": 0
    },
    "VBE.boot.default.pipe_hdrbytes": {
        "value": 0
    },
    "VBE.boot.default.pipe_in": {
        "value": 0
    },
    "VBE.boot.default.pipe_out": {
        "value": 0
    },
    "VBE.boot.default.req": {
        "value": 0
    },
    "VBE.boot.default.unhealthy": {
        "value": 0
    },
    "timestamp": {
        "value": 0
    }
}


GET

Child in state running


GET

active      auto/warm          0 boot


GET

vcl 4.0;

backend default {
	.host="localhost";
	.port="9001";
}

acl purge {
        "localhost";
}

sub vcl_recv {
        if (req.method == "PURGE") {
                if (!client.ip ~ purge) {
                        return(synth(405,"Not allowed."));
                }
                return (purge);
        }
}


GET

active      auto/warm          0 boot


GET

{
    "vcls": [
        {
            "mode": "",
            "name": "boot",
            "status": "active",
            "temp": "auto/warm"
        }
    ]
}


GET

{
    "revision": "0aace84169eca8d72edab4678e4751ace953ac5b",
    "version": "6.0.3r3"
}


POST

The following example will issue a ban request for all urls:

$ curl -is -X POST http://localhost:6085/ban -u agent:pass -d "req.url ~ '/'"


POST

The following example will issue a ban request for all urls, but it will send it as a full command directly to varnishadm:

$ curl -is -X POST http://localhost:6085/direct -u agent:pass -d "ban req.url ~ '/'"


POST

Child in state running


Stop varnish's child process.

POST

PUT

Child in state running


PUT


POST

The following example will store the specified vcl onto disk and then load it into varnish:

$ curl -is -X POST http://localhost:6085/vcl -u agent:pass -d "@/path/to/example.vcl"

The uploaded vcl will be saved by default at the path specified by the agent’s -p option.

Upon succesfull vcl load, the agent will generate and assign a name for the saved vcl so it will be able to load it into varnish. The agent’s response will contain a X-Vcl-Name header telling the vcl’s name.


PUT

The following example will store the vcl on disk with the specified name and then load it into varnish:

$ curl -is -X PUT http://localhost:6085/vcl/foo -u agent:pass -d "@/path/to/example.vcl"

The uploaded vcl will be saved by default at the path specified by the agent’s -p option.


DELETE

The following example will discard the specified vcl from varnish:

$ curl -is -X DELETE http://localhost:6085/vcl/foo -u agent:pass

Note that you cannot discard a vcl currently being used.


PUT

The following example will tell varnish to use the vcl named foo:

$ curl -is -X PUT http://localhost:6085/vcldeploy/foo -u agent:pass

Note that you can only use a vcl which has previosuly been loaded into varnish.