Skip site
navigation (1)Skip section navigation (2)
FreeBSD Manual Pages
- Q_SIGNED(3), Q_LTZ(3), Q_PRECEQ(3), Q_QLTQ(3), Q_QLEQ(3), Q_QGTQ(3), Q_QGEQ(3), Q_QEQ(3), Q_QNEQ(3), Q_OFLOW(3), Q_RELPREC(3)
- fixed-point math comparison and logic functions
- flowctl(8)
- ng_netflow(4) control utility
- ibv_create_flow(3), ibv_destroy_flow(3)
- create or destroy flow steering rules
- ng_netflow(4)
- Cisco's NetFlow implementation
- ng_split(4)
- netgraph node to separate incoming and outgoing flows
- pflow(4)
- kernel interface for pflow data export
- pflowctl(8)
- control pflow data export
- tcsendbreak(3), tcdrain(3), tcflush(3), tcflow(3)
- line control functions
- SSL_get_conn_close_info(3), SSL_CONN_CLOSE_FLAG_LOCAL(3), SSL_CONN_CLOSE_FLAG_TRANSPORT(3), OSSL_QUIC_ERR_NO_ERROR(3), OSSL_QUIC_ERR_INTERNAL_ERROR(3), OSSL_QUIC_ERR_CONNECTION_REFUSED(3), OSSL_QUIC_ERR_FLOW_CONTROL_ERROR(3), OSSL_QUIC_ERR_STREAM_LIMIT_ERROR(3), OSSL_QUIC_ERR_STREAM_STATE_ERROR(3), OSSL_QUIC_ERR_FINAL_SIZE_ERROR(3), OSSL_QUIC_ERR_FRAME_ENCODING_ERROR(3), OSSL_QUIC_ERR_TRANSPORT_PARAMETER_ERROR(3), OSSL_QUIC_ERR_CONNECTION_ID_LIMIT_ERROR(3), OSSL_QUIC_ERR_PROTOCOL_VIOLATION(3), OSSL_QUIC_ERR_INVALID_TOKEN(3), OSSL_QUIC_ERR_APPLICATION_ERROR(3), OSSL_QUIC_ERR_CRYPTO_BUFFER_EXCEEDED(3), OSSL_QUIC_ERR_KEY_UPDATE_ERROR(3), OSSL_QUIC_ERR_AEAD_LIMIT_REACHED(3), OSSL_QUIC_ERR_NO_VIABLE_PATH(3), OSSL_QUIC_ERR_CRYPTO_ERR_BEGIN(3), OSSL_QUIC_ERR_CRYPTO_ERR_END(3), OSSL_QUIC_ERR_CRYPTO_ERR(3), OSSL_QUIC_LOCAL_ERR_IDLE_TIMEOUT(3)
- get information about why a QUIC connection was closed
- SSL_get_conn_close_info(3ossl), SSL_CONN_CLOSE_FLAG_LOCAL(3ossl), SSL_CONN_CLOSE_FLAG_TRANSPORT(3ossl), OSSL_QUIC_ERR_NO_ERROR(3ossl), OSSL_QUIC_ERR_INTERNAL_ERROR(3ossl), OSSL_QUIC_ERR_CONNECTION_REFUSED(3ossl), OSSL_QUIC_ERR_FLOW_CONTROL_ERROR(3ossl), OSSL_QUIC_ERR_STREAM_LIMIT_ERROR(3ossl), OSSL_QUIC_ERR_STREAM_STATE_ERROR(3ossl), OSSL_QUIC_ERR_FINAL_SIZE_ERROR(3ossl), OSSL_QUIC_ERR_FRAME_ENCODING_ERROR(3ossl), OSSL_QUIC_ERR_TRANSPORT_PARAMETER_ERROR(3ossl), OSSL_QUIC_ERR_CONNECTION_ID_LIMIT_ERROR(3ossl), OSSL_QUIC_ERR_PROTOCOL_VIOLATION(3ossl), OSSL_QUIC_ERR_INVALID_TOKEN(3ossl), OSSL_QUIC_ERR_APPLICATION_ERROR(3ossl), OSSL_QUIC_ERR_CRYPTO_BUFFER_EXCEEDED(3ossl), OSSL_QUIC_ERR_KEY_UPDATE_ERROR(3ossl), OSSL_QUIC_ERR_AEAD_LIMIT_REACHED(3ossl), OSSL_QUIC_ERR_NO_VIABLE_PATH(3ossl), OSSL_QUIC_ERR_CRYPTO_ERR_BEGIN(3ossl), OSSL_QUIC_ERR_CRYPTO_ERR_END(3ossl), OSSL_QUIC_ERR_CRYPTO_ERR(3ossl), OSSL_QUIC_LOCAL_ERR_IDLE_TIMEOUT(3ossl)
- get information about why a QUIC connection was closed
- Sympa::Internals::Workflow(3Sympa)
- Overview on workflow of Sympa
- Sympa::Spindle(3Sympa)
- Base class of subclasses to define Sympa workflows
- Sympa::Spindle::AuthorizeMessage(3Sympa)
- Workflow to authorize messages bound for lists
- Sympa::Spindle::AuthorizeRequest(3Sympa)
- Workflow to authorize requests in command messages
- Sympa::Spindle::DispatchRequest(3Sympa)
- Workflow to dispatch requests
- Sympa::Spindle::DistributeMessage(3Sympa)
- Workflow to distribute messages to list members
- Sympa::Spindle::DoCommand(3Sympa)
- Workflow to handle command messages
- Sympa::Spindle::DoForward(3Sympa)
- Workflow to forward messages to administrators
- Sympa::Spindle::DoMessage(3Sympa)
- Workflow to handle messages bound for lists
- Sympa::Spindle::ProcessArchive(3Sympa)
- Workflow of archive storage
- Sympa::Spindle::ProcessAuth(3Sympa)
- Workflow of request confirmation
- Sympa::Spindle::ProcessAutomatic(3Sympa)
- Workflow of automatic list creation
- Sympa::Spindle::ProcessBounce(3Sympa)
- Workflow of bounce processing
- Sympa::Spindle::ProcessDigest(3Sympa)
- Workflow of digest sending
- Sympa::Spindle::ProcessHeld(3Sympa)
- Workflow of message confirmation
- Sympa::Spindle::ProcessIncoming(3Sympa)
- Workflow of processing incoming messages
- Sympa::Spindle::ProcessMessage(3Sympa)
- Workflow of command processing
- Sympa::Spindle::ProcessModeration(3Sympa)
- Workflow of message moderation
- Sympa::Spindle::ProcessOutgoing(3Sympa)
- Workflow of message distribution
- Sympa::Spindle::ProcessRequest(3Sympa)
- Workflow of request processing
- Sympa::Spindle::ProcessTask(3Sympa)
- Workflow of task processing
- Sympa::Spindle::ProcessTemplate(3Sympa)
- Workflow of template sending
- Sympa::Spindle::ResendArchive(3Sympa)
- Workflow of resending messages in archive
- barman-cloud-backup(1) - Barman-cloud Commands Synopsis 0.0 3.5 barman-cloud-backup [ { -V | --version } ] [ --help ] [ { { -v | --verbose } | { -q | --quiet } } ] [ { -t | --test } ] [ --cloud-provider { aws-s3 | azure-blob-storage | google-cloud-storage } ] [ { { -z | --gzip } | { -j | --bzip2 } | --snappy } ] [ { -h | --host } HOST ] [ { -p | --port } PORT ] [ { -U | --user } USER ] [ { -d | --dbname } DBNAME ] [ { -n | --name } BACKUP_NAME ] [ { -J | --jobs } JOBS ] [ { -S | --max-archive-size } MAX_ARCHIVE_SIZE ] [ --immediate-checkpoint ] [ --min-chunk-size MIN_CHUNK_SIZE ] [ --max-bandwidth MAX_BANDWIDTH ] [ --snapshot-instance SNAPSHOT_INSTANCE ] [ --snapshot-disk NAME [ --snapshot-disk NAME ... ] ] [ --snapshot-zone GCP_ZONE ] [ -snapshot-gcp-project GCP_PROJECT ] [ --tag KEY,VALUE [ --tag KEY,VALUE ... ] ] [ --endpoint-url ENDPOINT_URL ] [ { -P | --aws-profile } AWS_PROFILE ] [ --profile AWS_PROFILE ] [ --read-timeout READ_TIMEOUT ] [ { -e | --encryption } { AES256 | aws:kms } ] [ --sse-kms-key-id SSE_KMS_KEY_ID ] [ --aws-region AWS_REGION ] [ --aws-await-snapshots-timeout AWS_AWAIT_SNAPSHOTS_TIMEOUT ] [ --aws-snapshot-lock-mode { compliance | governance } ] [ --aws-snapshot-lock-duration DAYS ] [ --aws-snapshot-lock-cool-off-period HOURS ] [ --aws-snapshot-lock-expiration-date DATETIME ] [ { --azure-credential | --credential } { azure-cli | managed-identity | default } ] [ --encryption-scope ENCRYPTION_SCOPE ] [ --azure-subscription-id AZURE_SUBSCRIPTION_ID ] [ --azure-resource-group AZURE_RESOURCE_GROUP ] [ --gcp-project GCP_PROJECT ] [ --kms-key-name KMS_KEY_NAME ] [ --gcp-zone GCP_ZONE ] DESTINATION_URL SERVER_NAME Description The barman-cloud-backup script is used to create a local backup of a Postgres server and transfer it to a supported cloud provider, bypassing the Barman server. It can also be utilized as a hook script for copying Barman backups from the Barman server to one of the supported clouds (post_backup_retry_script). This script requires read access to PGDATA and tablespaces, typically run as the postgres user. When used on a Barman server, it requires read access to the directory where Barman backups are stored. If --snapshot- arguments are used and snapshots are supported by the selected cloud provider, the backup will be performed using snapshots of the specified disks (--snapshot-disk). The backup label and metadata will also be uploaded to the cloud. NOTE: 0.0 3.5 For GCP, only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported. IMPORTANT: 0.0 3.5 The cloud upload may fail if any file larger than the configured --max-archive-size is present in the data directory or tablespaces. However, Postgres files up to 1GB are always allowed, regardless of the --max-archive-size setting. Parameters 0.0 SERVER_NAME Name of the server to be backed up. DESTINATION_URL URL of the cloud destination, such as a bucket in AWS S3. For example: s3://bucket/path/to/folder. -V / --version Show version and exit. --help show this help message and exit. -v / --verbose Increase output verbosity (e.g., -vv is more than -v). -q / --quiet Decrease output verbosity (e.g., -qq is less than -q). -t / --test Test cloud connectivity and exit. --cloud-provider The cloud provider to use as a storage backend. Allowed options: 7.0 (bu 2 aws-s3. (bu 2 azure-blob-storage. (bu 2 google-cloud-storage. -z / --gzip gzip-compress the backup while uploading to the cloud (should not be used with python < 3.2). -j / --bzip2 bzip2-compress the backup while uploading to the cloud (should not be used with python < 3.3). --snappy snappy-compress the backup while uploading to the cloud (requires optional python-snappy library). -h / --host Host or Unix socket for Postgres connection (default: libpq settings). -p / --port Port for Postgres connection (default: libpq settings). -U / --user User name for Postgres connection (default: libpq settings). -d / --dbname Database name or conninfo string for Postgres connection (default: (dqpostgres(dq). -n / --name A name which can be used to reference this backup in commands such as barman-cloud-restore and barman-cloud-backup-delete. -J / --jobs Number of subprocesses to upload data to cloud storage (default: 2). -S / --max-archive-size Maximum size of an archive when uploading to cloud storage (default: 100GB). --immediate-checkpoint Forces the initial checkpoint to be done as quickly as possible. --min-chunk-size Minimum size of an individual chunk when uploading to cloud storage (default: 5MB for aws-s3, 64KB for azure-blob-storage, not applicable for google-cloud-storage). --max-bandwidth The maximum amount of data to be uploaded per second when backing up to object storages (default: 0 - no limit). --snapshot-instance Instance where the disks to be backed up as snapshots are attached. --snapshot-disk Name of a disk from which snapshots should be taken. --tag Tag to be added to all uploaded files in cloud storage, and/or to snapshots created, if snapshots are used. --tags Tags to be added to all uploaded files in cloud storage, and/or to snapshots created, if snapshots are used. NOTE: 0.0 3.5 If you are using --tags before positional arguments, you must insert -- after it to indicate the end of optional arguments. This tells the parser to treat everything after -- as positional arguments. Without the --, Barman may misinterpret positional arguments as values for the last option. Deprecated since version 3.15: --tags is deprecated. Use --tag instead. Extra options for the AWS cloud provider 0.0 --endpoint-url Override default S3 endpoint URL with the given one. -P / --aws-profile Profile name (e.g. INI section in AWS credentials file). --profile (deprecated) Profile name (e.g. INI section in AWS credentials file)
- replaced by --aws-profile. --read-timeout The time in seconds until a timeout is raised when waiting to read from a connection (defaults to 60 seconds). -e / --encryption The encryption algorithm used when storing the uploaded data in S3. Allowed options: 7.0 (bu 2 AES256. (bu 2 aws:kms. --sse-kms-key-id The AWS KMS key ID that should be used for encrypting the uploaded data in S3. Can be specified using the key ID on its own or using the full ARN for the key. Only allowed if -e / --encryption is set to aws:kms. --aws-region The name of the AWS region containing the EC2 VM and storage volumes defined by the --snapshot-instance and --snapshot-disk arguments. --aws-await-snapshots-timeout The length of time in seconds to wait for snapshots to be created in AWS before timing out (default: 3600 seconds). --aws-snapshot-lock-mode The lock mode for the snapshot. This is only valid if --snapshot-instance and --snapshot-disk are set. Allowed options: 7.0 (bu 2 compliance. (bu 2 governance. --aws-snapshot-lock-duration The lock duration is the period of time (in days) for which the snapshot is to remain locked, ranging from 1 to 36,500. Set either the lock duration or the expiration date (not both). --aws-snapshot-lock-cool-off-period The cooling-off period is an optional period of time (in hours) that you can specify when you lock a snapshot in compliance mode, ranging from 1 to 72. --aws-snapshot-lock-expiration-date The lock duration is determined by an expiration date in the future. It must be at least 1 day after the snapshot creation date and time, using the format YYYY-MM-DDTHH:MM:SS.sssZ. Set either the lock duration or the expiration date (not both). Extra options for the Azure cloud provider 0.0 --azure-credential / --credential Optionally specify the type of credential to use when authenticating with Azure. If omitted then Azure Blob Storage credentials will be obtained from the environment and the default Azure authentication flow will be used for authenticating with all other Azure services. If no credentials can be found in the environment then the default Azure authentication flow will also be used for Azure Blob Storage. Allowed options: 7.0 (bu 2 azure-cli. (bu 2 managed-identity. (bu 2 default. --encryption-scope The name of an encryption scope defined in the Azure Blob Storage service which is to be used to encrypt the data in Azure. --azure-subscription-id The ID of the Azure subscription which owns the instance and storage volumes defined by the --snapshot-instance and --snapshot-disk arguments. --azure-resource-group The name of the Azure resource group to which the compute instance and disks defined by the --snapshot-instance and --snapshot-disk arguments belong. Extra options for GCP cloud provider 0.0 --gcp-project GCP project under which disk snapshots should be stored. --snapshot-gcp-project (deprecated) GCP project under which disk snapshots should be stored - replaced by --gcp-project. --kms-key-name The name of the GCP KMS key which should be used for encrypting the uploaded data in GCS. --gcp-zone Zone of the disks from which snapshots should be taken. --snapshot-zone (deprecated) Zone of the disks from which snapshots should be taken - replaced by --gcp-zone
- barman-cloud-backup-delete(1) - Barman-cloud Commands Synopsis 0.0 3.5 barman-cloud-backup-delete [ { -V | --version } ] [ --help ] [ { { -v | --verbose } | { -q | --quiet } } ] [ { -t | --test } ] [ --cloud-provider { aws-s3 | azure-blob-storage | google-cloud-storage } ] [ --endpoint-url ENDPOINT_URL ] [ { -r | --retention-policy } RETENTION_POLICY ] [ { -m | --minimum-redundancy } MINIMUM_REDUNDANCY ] [ { -b | --backup-id } BACKUP_ID] [ --dry-run ] [ { -P | --aws-profile } AWS_PROFILE ] [ --profile AWS_PROFILE ] [ --read-timeout READ_TIMEOUT ] [ { --azure-credential | --credential } { azure-cli | managed-identity | default } ] [--batch-size DELETE_BATCH_SIZE] SOURCE_URL SERVER_NAME Description The barman-cloud-backup-delete script is used to delete one or more backups created with the barman-cloud-backup command from cloud storage and to remove the associated WAL files. Backups can be specified for deletion either by their backup ID (as obtained from barman-cloud-backup-list) or by a retention policy. Retention policies mirror those used by the Barman server, deleting all backups that are not required to meet the specified policy. When a backup is deleted, any unused WAL files associated with that backup are also removed. WALs are considered unused if: 0.0 (bu 2 The WALs predate the begin_wal value of the oldest remaining backup. (bu 2 The WALs are not required by any archival backups stored in the cloud. NOTE: 0.0 3.5 For GCP, only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported. IMPORTANT: 0.0 3.5 Each backup deletion involves three separate requests to the cloud provider: one for the backup files, one for the backup.info file, and one for the associated WALs. Deleting by retention policy may result in a high volume of delete requests if a large number of backups are accumulated in cloud storage. Parameters 0.0 SERVER_NAME Name of the server that holds the backup to be deleted. SOURCE_URL URL of the cloud source, such as a bucket in AWS S3. For example: s3://bucket/path/to/folder. -V / --version Show version and exit. --help show this help message and exit. -v / --verbose Increase output verbosity (e.g., -vv is more than -v). -q / --quiet Decrease output verbosity (e.g., -qq is less than -q). -t / --test Test cloud connectivity and exit. --cloud-provider The cloud provider to use as a storage backend. Allowed options are: 7.0 (bu 2 aws-s3. (bu 2 azure-blob-storage. (bu 2 google-cloud-storage. -b / --backup-id ID of the backup to be deleted. You can use a shortcut instead of the backup ID. -m / --minimum-redundancy The minimum number of backups that should always be available. -r / --retention-policy If specified, delete all backups eligible for deletion according to the supplied retention policy. Syntax: REDUNDANCY value | RECOVERY WINDOW OF value { DAYS | WEEKS | MONTHS } --batch-size The maximum number of objects to be deleted in a single request to the cloud provider. If unset then the maximum allowed batch size for the specified cloud provider will be used (1000 for aws-s3, 256 for azure-blob-storage and 100 for google-cloud-storage). --dry-run Find the objects which need to be deleted but do not delete them. Extra options for the AWS cloud provider 0.0 --endpoint-url Override default S3 endpoint URL with the given one. -P / --aws-profile Profile name (e.g. INI section in AWS credentials file). --profile (deprecated) Profile name (e.g. INI section in AWS credentials file)
- replaced by --aws-profile. --read-timeout The time in seconds until a timeout is raised when waiting to read from a connection (defaults to 60 seconds). Extra options for the Azure cloud provider 0.0 --azure-credential / --credential Optionally specify the type of credential to use when authenticating with Azure. If omitted then Azure Blob Storage credentials will be obtained from the environment and the default Azure authentication flow will be used for authenticating with all other Azure services. If no credentials can be found in the environment then the default Azure authentication flow will also be used for Azure Blob Storage. Allowed options are: 7.0 (bu 2 azure-cli. (bu 2 managed-identity. (bu 2 default
- barman-cloud-backup-keep(1) - Barman-cloud Commands Synopsis 0.0 3.5 barman-cloud-backup-keep [ { -V | --version } ] [ --help ] [ { { -v | --verbose } | { -q | --quiet } } ] [ { -t | --test } ] [ --cloud-provider { aws-s3 | azure-blob-storage | google-cloud-storage } ] [ --endpoint-url ENDPOINT_URL ] [ { -P | --aws-profile } AWS_PROFILE ] [ --profile AWS_PROFILE ] [ --read-timeout READ_TIMEOUT ] [ { --azure-credential | --credential } { azure-cli | managed-identity | default } ] [ { { -r | --release } | { -s | --status } | --target { full | standalone } } ] SOURCE_URL SERVER_NAME BACKUP_ID Description Use this script to designate backups in cloud storage as archival backups, ensuring their indefinite retention regardless of retention policies. This script allows you to mark backups previously created with barman-cloud-backup as archival backups. Once flagged as archival, these backups are preserved indefinitely and are not subject to standard retention policies. NOTE: 0.0 3.5 For GCP, only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported. Parameters 0.0 SERVER_NAME Name of the server that holds the backup to be kept. SOURCE_URL URL of the cloud source, such as a bucket in AWS S3. For example: s3://bucket/path/to/folder. BACKUP_ID The ID of the backup to be kept. You can use a shortcut instead of the backup ID. -V / --version Show version and exit. --help show this help message and exit. -v / --verbose Increase output verbosity (e.g., -vv is more than -v). -q / --quiet Decrease output verbosity (e.g., -qq is less than -q). -t / --test Test cloud connectivity and exit. --cloud-provider The cloud provider to use as a storage backend. Allowed options are: 7.0 (bu 2 aws-s3. (bu 2 azure-blob-storage. (bu 2 google-cloud-storage. -r / --release If specified, the command will remove the keep annotation and the backup will be eligible for deletion. -s / --status Print the keep status of the backup. --target Specify the recovery target for this backup. Allowed options are: 7.0 (bu 2 full (bu 2 standalone Extra options for the AWS cloud provider 0.0 --endpoint-url Override default S3 endpoint URL with the given one. -P / --aws-profile Profile name (e.g. INI section in AWS credentials file). --profile (deprecated) Profile name (e.g. INI section in AWS credentials file)
- replaced by --aws-profile. --read-timeout The time in seconds until a timeout is raised when waiting to read from a connection (defaults to 60 seconds). Extra options for the Azure cloud provider 0.0 --azure-credential / --credential Optionally specify the type of credential to use when authenticating with Azure. If omitted then Azure Blob Storage credentials will be obtained from the environment and the default Azure authentication flow will be used for authenticating with all other Azure services. If no credentials can be found in the environment then the default Azure authentication flow will also be used for Azure Blob Storage. Allowed options are: 7.0 (bu 2 azure-cli. (bu 2 managed-identity. (bu 2 default
- barman-cloud-backup-list(1) - Barman-cloud Commands Synopsis 0.0 3.5 barman-cloud-backup-list [ { -V | --version } ] [ --help ] [ { { -v | --verbose } | { -q | --quiet } } ] [ { -t | --test } ] [ --cloud-provider { aws-s3 | azure-blob-storage | google-cloud-storage } ] [ --endpoint-url ENDPOINT_URL ] [ { -P | --aws-profile } AWS_PROFILE ] [ --profile AWS_PROFILE ] [ --read-timeout READ_TIMEOUT ] [ { --azure-credential | --credential } { azure-cli | managed-identity | default } ] [ --format FORMAT ] SOURCE_URL SERVER_NAME Description This script lists backups stored in the cloud that were created using the barman-cloud-backup command. NOTE: 0.0 3.5 For GCP, only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported. Parameters 0.0 SERVER_NAME Name of the server that holds the backup to be listed. SOURCE_URL URL of the cloud source, such as a bucket in AWS S3. For example: s3://bucket/path/to/folder. -V / --version Show version and exit. --help show this help message and exit. -v / --verbose Increase output verbosity (e.g., -vv is more than -v). -q / --quiet Decrease output verbosity (e.g., -qq is less than -q). -t / --test Test cloud connectivity and exit. --cloud-provider The cloud provider to use as a storage backend. Allowed options are: 7.0 (bu 2 aws-s3. (bu 2 azure-blob-storage. (bu 2 google-cloud-storage. --format Output format (console or json). Default console. Extra options for the AWS cloud provider 0.0 --endpoint-url Override default S3 endpoint URL with the given one. -P / --aws-profile Profile name (e.g. INI section in AWS credentials file). --profile (deprecated) Profile name (e.g. INI section in AWS credentials file)
- replaced by --aws-profile. --read-timeout The time in seconds until a timeout is raised when waiting to read from a connection (defaults to 60 seconds). Extra options for the Azure cloud provider 0.0 --azure-credential / --credential Optionally specify the type of credential to use when authenticating with Azure. If omitted then Azure Blob Storage credentials will be obtained from the environment and the default Azure authentication flow will be used for authenticating with all other Azure services. If no credentials can be found in the environment then the default Azure authentication flow will also be used for Azure Blob Storage. Allowed options are: 7.0 (bu 2 azure-cli. (bu 2 managed-identity. (bu 2 default
- barman-cloud-backup-show(1) - Barman-cloud Commands Synopsis 0.0 3.5 barman-cloud-backup-show [ { -V | --version } ] [ --help ] [ { { -v | --verbose } | { -q | --quiet } } ] [ { -t | --test } ] [ --cloud-provider { aws-s3 | azure-blob-storage | google-cloud-storage } ] [ --endpoint-url ENDPOINT_URL ] [ { -P | --aws-profile } AWS_PROFILE ] [ --profile AWS_PROFILE ] [ --read-timeout READ_TIMEOUT ] [ { --azure-credential | --credential } { azure-cli | managed-identity | default } ] [ --format FORMAT ] SOURCE_URL SERVER_NAME BACKUP_ID Description This script displays detailed information about a specific backup created with the barman-cloud-backup command. The output is similar to the barman show-backup from the %barman show-backup command reference, but it has fewer information. NOTE: 0.0 3.5 For GCP, only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported. Parameters 0.0 BACKUP_ID The ID of the backup. You can use a shortcut instead of the backup ID. SERVER_NAME Name of the server that holds the backup to be displayed. SOURCE_URL URL of the cloud source, such as a bucket in AWS S3. For example: s3://bucket/path/to/folder. -V / --version Show version and exit. --help show this help message and exit. -v / --verbose Increase output verbosity (e.g., -vv is more than -v). -q / --quiet Decrease output verbosity (e.g., -qq is less than -q). -t / --test Test cloud connectivity and exit. --cloud-provider The cloud provider to use as a storage backend. Allowed options are: 7.0 (bu 2 aws-s3. (bu 2 azure-blob-storage. (bu 2 google-cloud-storage. --format Output format (console or json). Default console. Extra options for the AWS cloud provider 0.0 --endpoint-url Override default S3 endpoint URL with the given one. -P / --aws-profile Profile name (e.g. INI section in AWS credentials file). --profile (deprecated) Profile name (e.g. INI section in AWS credentials file)
- replaced by --aws-profile. --read-timeout The time in seconds until a timeout is raised when waiting to read from a connection (defaults to 60 seconds). Extra options for the Azure cloud provider 0.0 --azure-credential / --credential Optionally specify the type of credential to use when authenticating with Azure. If omitted then Azure Blob Storage credentials will be obtained from the environment and the default Azure authentication flow will be used for authenticating with all other Azure services. If no credentials can be found in the environment then the default Azure authentication flow will also be used for Azure Blob Storage. Allowed options are: 7.0 (bu 2 azure-cli. (bu 2 managed-identity. (bu 2 default
- barman-cloud-check-wal-archive(1) - Barman-cloud Commands Synopsis 0.0 3.5 barman-cloud-check-wal-archive [ { -V | --version } ] [ --help ] [ { { -v | --verbose } | { -q | --quiet } } ] [ { -t | --test } ] [ --cloud-provider { aws-s3 | azure-blob-storage | google-cloud-storage } ] [ --endpoint-url ENDPOINT_URL ] [ { -P | --aws-profile } AWS_PROFILE ] [ --profile AWS_PROFILE ] [ --read-timeout READ_TIMEOUT ] [ { --azure-credential | --credential } { azure-cli | managed-identity | default } ] [ --timeline TIMELINE ] DESTINATION_URL SERVER_NAME Description Verify that the WAL archive destination for a server is suitable for use with a new Postgres cluster. By default, the check will succeed if the WAL archive is empty or if the target bucket is not found. Any other conditions will result in a failure. NOTE: 0.0 3.5 For GCP, only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported. Parameters 0.0 SERVER_NAME Name of the server that needs to be checked. DESTINATION_URL URL of the cloud destination, such as a bucket in AWS S3. For example: s3://bucket/path/to/folder. -V / --version Show version and exit. --help show this help message and exit. -v / --verbose Increase output verbosity (e.g., -vv is more than -v). -q / --quiet Decrease output verbosity (e.g., -qq is less than -q). -t / --test Test cloud connectivity and exit. --cloud-provider The cloud provider to use as a storage backend. Allowed options are: 7.0 (bu 2 aws-s3. (bu 2 azure-blob-storage. (bu 2 google-cloud-storage. --timeline The earliest timeline whose WALs should cause the check to fail. Extra options for the AWS cloud provider 0.0 --endpoint-url Override default S3 endpoint URL with the given one. -P / --aws-profile Profile name (e.g. INI section in AWS credentials file). --profile (deprecated) Profile name (e.g. INI section in AWS credentials file)
- replaced by --aws-profile. --read-timeout The time in seconds until a timeout is raised when waiting to read from a connection (defaults to 60 seconds). Extra options for the Azure cloud provider 0.0 --azure-credential / --credential Optionally specify the type of credential to use when authenticating with Azure. If omitted then Azure Blob Storage credentials will be obtained from the environment and the default Azure authentication flow will be used for authenticating with all other Azure services. If no credentials can be found in the environment then the default Azure authentication flow will also be used for Azure Blob Storage. Allowed options are: 7.0 (bu 2 azure-cli. (bu 2 managed-identity. (bu 2 default
- barman-cloud-restore(1) - Barman-cloud Commands Synopsis 0.0 3.5 barman-cloud-restore [ { -V | --version } ] [ --help ] [ { { -v | --verbose } | { -q | --quiet } } ] [ { -t | --test } ] [ --cloud-provider { aws-s3 | azure-blob-storage | google-cloud-storage } ] [ --endpoint-url ENDPOINT_URL ] [ { -P | --aws-profile } AWS_PROFILE ] [ --profile AWS_PROFILE ] [ --read-timeout READ_TIMEOUT ] [ { --azure-credential | --credential } { azure-cli | managed-identity | default } ] [ --snapshot-recovery-instance SNAPSHOT_RECOVERY_INSTANCE ] [ --snapshot-recovery-zone GCP_ZONE ] [ --aws-region AWS_REGION ] [ --gcp-zone GCP_ZONE ] [ --azure-resource-group AZURE_RESOURCE_GROUP ] [ --tablespace NAME:LOCATION [ --tablespace NAME:LOCATION ... ] ] [ --target-lsn LSN ] [ --target-time TIMESTAMP ] [ --target-tli TLI ] SOURCE_URL SERVER_NAME BACKUP_ID RECOVERY_DESTINATION Description Use this script to restore a backup directly from cloud storage that was created with the barman-cloud-backup command. Additionally, this script can prepare for recovery from a snapshot backup by verifying that attached disks were cloned from the correct snapshots and by downloading the backup label from object storage. This command does not automatically prepare Postgres for recovery. You must manually manage any %PITR options, custom restore_command values, signal files, or required WAL files to ensure Postgres starts, either manually or using external tools. NOTE: 0.0 3.5 For GCP, only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported. Parameters 0.0 SERVER_NAME Name of the server that holds the backup to be restored. SOURCE_URL URL of the cloud source, such as a bucket in AWS S3. For example: s3://bucket/path/to/folder. BACKUP_ID The ID of the backup to be restored. You can use a shortcut instead of the backup ID. Besides that, you can use auto to have Barman automatically find the most suitable backup for the restore operation. RECOVERY_DESTINATION The path to a directory for recovery. -V / --version Show version and exit. --help show this help message and exit. -v / --verbose Increase output verbosity (e.g., -vv is more than -v). -q / --quiet Decrease output verbosity (e.g., -qq is less than -q). -t / --test Test cloud connectivity and exit. --cloud-provider The cloud provider to use as a storage backend. Allowed options are: 7.0 (bu 2 aws-s3. (bu 2 azure-blob-storage. (bu 2 google-cloud-storage. --snapshot-recovery-instance Instance where the disks recovered from the snapshots are attached. --tablespace Tablespace relocation rule. --target-lsn The recovery target lsn, e.g., 3/64000000. --target-time The recovery target timestamp with or without timezone, in the format %Y-%m-%d %H:%M:%S. --target-tli The recovery target timeline. Extra options for the AWS cloud provider 0.0 --endpoint-url Override default S3 endpoint URL with the given one. -P / --aws-profile Profile name (e.g. INI section in AWS credentials file). --profile (deprecated) Profile name (e.g. INI section in AWS credentials file)
- replaced by --aws-profile. --read-timeout The time in seconds until a timeout is raised when waiting to read from a connection (defaults to 60 seconds). --aws-region The name of the AWS region containing the EC2 VM and storage volumes defined by the --snapshot-instance and --snapshot-disk arguments. Extra options for the Azure cloud provider 0.0 --azure-credential / --credential Optionally specify the type of credential to use when authenticating with Azure. If omitted then Azure Blob Storage credentials will be obtained from the environment and the default Azure authentication flow will be used for authenticating with all other Azure services. If no credentials can be found in the environment then the default Azure authentication flow will also be used for Azure Blob Storage. Allowed options are: 7.0 (bu 2 azure-cli. (bu 2 managed-identity. (bu 2 default. --azure-resource-group The name of the Azure resource group to which the compute instance and disks defined by the --snapshot-instance and --snapshot-disk arguments belong. Extra options for GCP cloud provider 0.0 --gcp-zone Zone of the disks from which snapshots should be taken. --snapshot-recovery-zone (deprecated) Zone containing the instance and disks for the snapshot recovery - replaced by --gcp-zone
- barman-cloud-wal-archive(1) - Barman-cloud Commands Synopsis 0.0 3.5 barman-cloud-wal-archive [ { -V | --version } ] [ --help ] [ { { -v | --verbose } | { -q | --quiet } } ] [ { -t | --test } ] [ --cloud-provider { aws-s3 | azure-blob-storage | google-cloud-storage } ] [ { { -z | --gzip } | { -j | --bzip2 } | --xz | --snappy | --zstd | --lz4 } ] [ --compression-level COMPRESSION_LEVEL ] [ --tag KEY,VALUE [ --tag KEY,VALUE ... ] ] [ --history-tag KEY,VALUE [ --history-tag KEY,VALUE ... ] ] [ --endpoint-url ENDPOINT_URL ] [ { -P | --aws-profile } AWS_PROFILE ] [ --profile AWS_PROFILE ] [ --read-timeout READ_TIMEOUT ] [ { -e | --encryption } ENCRYPTION ] [ --sse-kms-key-id SSE_KMS_KEY_ID ] [ { --azure-credential | --credential } { azure-cli | managed-identity | default } ] [ --encryption-scope ENCRYPTION_SCOPE ] [ --max-block-size MAX_BLOCK_SIZE ] [ --max-concurrency MAX_CONCURRENCY ] [ --max-single-put-size MAX_SINGLE_PUT_SIZE ] [ --kms-key-name KMS_KEY_NAME ] DESTINATION_URL SERVER_NAME [ WAL_PATH ] Description The barman-cloud-wal-archive command is designed to be used in the archive_command of a Postgres server to directly ship WAL files to cloud storage. NOTE: 0.0 3.5 If you are using Python 2 or unsupported versions of Python 3, avoid using the compression options --gzip or --bzip2. The script cannot restore gzip-compressed WALs on Python < 3.2 or bzip2-compressed WALs on Python < 3.3. This script enables the direct transfer of WAL files to cloud storage, bypassing the Barman server. Additionally, it can be utilized as a hook script for WAL archiving (pre_archive_retry_script). NOTE: 0.0 3.5 For GCP, only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported. Parameters 0.0 SERVER_NAME Name of the server that will have the WALs archived. DESTINATION_URL URL of the cloud destination, such as a bucket in AWS S3. For example: s3://bucket/path/to/folder. WAL_PATH The value of the (aq%p(aq keyword (according to archive_command). -V / --version Show version and exit. --help show this help message and exit. -v / --verbose Increase output verbosity (e.g., -vv is more than -v). -q / --quiet Decrease output verbosity (e.g., -qq is less than -q). -t / --test Test cloud connectivity and exit. --cloud-provider The cloud provider to use as a storage backend. Allowed options are: 7.0 (bu 2 aws-s3. (bu 2 azure-blob-storage. (bu 2 google-cloud-storage. -z / --gzip gzip-compress the WAL while uploading to the cloud. -j / --bzip2 bzip2-compress the WAL while uploading to the cloud. --xz xz-compress the WAL while uploading to the cloud. --snappy snappy-compress the WAL while uploading to the cloud (requires the python-snappy Python library to be installed). --zstd zstd-compress the WAL while uploading to the cloud (requires the zstandard Python library to be installed). --lz4 lz4-compress the WAL while uploading to the cloud (requires the lz4 Python library to be installed). --compression-level A compression level to be used by the selected compression algorithm. Valid values are integers within the supported range of the chosen algorithm or one of the predefined labels: low, medium, and high. The range of each algorithm as well as what level each predefined label maps to can be found in %compression_level. --tag Tag to be added to archived WAL files in cloud storage. --tags Tag to be added to archived WAL files in cloud storage. NOTE: 0.0 3.5 If you are using --tags before positional arguments, you must insert -- after it to indicate the end of optional arguments. This tells the parser to treat everything after -- as positional arguments. Without the --, Barman may misinterpret positional arguments as values for the last option. Deprecated since version 3.15: --tags is deprecated. Use --tag instead. 0.0 --history-tag Tag to be added to archived history files in cloud storage. --history-tags Tags to be added to archived history files in cloud storage. NOTE: 0.0 3.5 If you are using --history-tags before positional arguments, you must insert -- after it to indicate the end of optional arguments. This tells the parser to treat everything after -- as positional arguments. Without the --, Barman may misinterpret positional arguments as values for the last option. Deprecated since version 3.15: --history-tags is deprecated. Use --history-tag instead. Extra options for the AWS cloud provider 0.0 --endpoint-url Override default S3 endpoint URL with the given one. -P / --aws-profile Profile name (e.g. INI section in AWS credentials file). --profile (deprecated) Profile name (e.g. INI section in AWS credentials file)
- replaced by --aws-profile. --read-timeout The time in seconds until a timeout is raised when waiting to read from a connection (defaults to 60 seconds). -e / --encryption The encryption algorithm used when storing the uploaded data in S3. Allowed options: 7.0 (bu 2 AES256. (bu 2 aws:kms. --sse-kms-key-id The AWS KMS key ID that should be used for encrypting the uploaded data in S3. Can be specified using the key ID on its own or using the full ARN for the key. Only allowed if -e / --encryption is set to aws:kms. Extra options for the Azure cloud provider 0.0 --azure-credential / --credential Optionally specify the type of credential to use when authenticating with Azure. If omitted then Azure Blob Storage credentials will be obtained from the environment and the default Azure authentication flow will be used for authenticating with all other Azure services. If no credentials can be found in the environment then the default Azure authentication flow will also be used for Azure Blob Storage. Allowed options are: 7.0 (bu 2 azure-cli. (bu 2 managed-identity. (bu 2 default. --encryption-scope The name of an encryption scope defined in the Azure Blob Storage service which is to be used to encrypt the data in Azure. --max-block-size The chunk size to be used when uploading an object via the concurrent chunk method (default: 4MB). --max-concurrency The maximum number of chunks to be uploaded concurrently (default: 1). --max-single-put-size Maximum size for which the Azure client will upload an object in a single request (default: 64MB). If this is set lower than the Postgres WAL segment size after any applied compression then the concurrent chunk upload method for WAL archiving will be used. Extra options for GCP cloud provider 0.0 --kms-key-name The name of the GCP KMS key which should be used for encrypting the uploaded data in GCS
- barman-cloud-wal-restore(1) - Barman-cloud Commands Synopsis 0.0 3.5 barman-cloud-wal-restore [ { -V | --version } ] [ --help ] [ { { -v | --verbose } | { -q | --quiet } } ] [ { -t | --test } ] [ --cloud-provider { aws-s3 | azure-blob-storage | google-cloud-storage } ] [ --endpoint-url ENDPOINT_URL ] [ { -P | --aws-profile } AWS_PROFILE ] [ --profile AWS_PROFILE ] [ --read-timeout READ_TIMEOUT ] [ { --azure-credential | --credential } { azure-cli | managed-identity | default } ] [ --no-partial ] SOURCE_URL SERVER_NAME WAL_NAME WAL_DEST Description The barman-cloud-wal-restore script functions as the restore_command for retrieving WAL files from cloud storage and placing them directly into a Postgres standby server, bypassing the Barman server. This script is used to download WAL files that were previously archived with the barman-cloud-wal-archive command. Disable automatic download of .partial files by calling --no-partial option. IMPORTANT: 0.0 3.5 On the target Postgres node, when pg_wal and the spool directory are on the same filesystem, files are moved via renaming, which is faster than copying and deleting. This speeds up serving WAL files significantly. If the directories are on different filesystems, the process still involves copying and deleting, so there(aqs no performance gain in that case. NOTE: 0.0 3.5 For GCP, only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported. Parameters 0.0 SERVER_NAME Name of the server that will have WALs restored. SOURCE_URL URL of the cloud source, such as a bucket in AWS S3. For example: s3://bucket/path/to/folder. WAL_NAME The value of the (aq%f(aq keyword (according to restore_command). WAL_DEST The value of the (aq%p(aq keyword (according to restore_command). -V / --version Show version and exit. --help show this help message and exit. -v / --verbose Increase output verbosity (e.g., -vv is more than -v). -q / --quiet Decrease output verbosity (e.g., -qq is less than -q). -t / --test Test cloud connectivity and exit. --cloud-provider The cloud provider to use as a storage backend. Allowed options are: 7.0 (bu 2 aws-s3. (bu 2 azure-blob-storage. (bu 2 google-cloud-storage. --no-partial Do not download partial WAL files Extra options for the AWS cloud provider 0.0 --endpoint-url Override default S3 endpoint URL with the given one. -P / --aws-profile Profile name (e.g. INI section in AWS credentials file). --profile (deprecated) Profile name (e.g. INI section in AWS credentials file)
- replaced by --aws-profile. --read-timeout The time in seconds until a timeout is raised when waiting to read from a connection (defaults to 60 seconds). Extra options for the Azure cloud provider 0.0 --azure-credential / --credential Optionally specify the type of credential to use when authenticating with Azure. If omitted then Azure Blob Storage credentials will be obtained from the environment and the default Azure authentication flow will be used for authenticating with all other Azure services. If no credentials can be found in the environment then the default Azure authentication flow will also be used for Azure Blob Storage. Allowed options are: 7.0 (bu 2 azure-cli. (bu 2 managed-identity. (bu 2 default
- cerfcx(3), erfcx(3)
- underflow-compensated complementary error function
- cflow(1)
- generate a C-language flowgraph
- cityflow(6)
- waves of boxes
- darktable(1)
- a digital photography workflow application
- deskflow(1), Deskflow:(1)
- Keyboard and mouse sharing utility (GUI)
- deskflow-core(1), deskflow-core:(1)
- Keyboard and mouse sharing utility
- ehnt(1)
- Extreme Happy Netflow Tool. Client part
- ehntserv(8)
- Extreme Happy Netflow Tool. Server part
- energystream(6)
- a flow of particles which form an energy stream
- euler2d(6)
- two dimensional incompressible inviscid fluid flow
- expire(8cn), doexpire(8cn), expireiflow(8cn)
- expire old news upact - update news active file
- explain_tcflow(3)
- explain tcflow(3) errors require_index { }
- explain_tcflow_or_die(3)
- terminal flow control and report errors require_index { "terminal flow control and report errors" }
- fixadd(3)
- Safe function to add fixed point numbers clamping overflow. Allegro game programming library
- fixsub(3)
- Safe function to subtract fixed point numbers clamping underflow. Allegro game programming library
- flow(1)
- show subroutine dependence of Fortran77 programs
- flow(6)
- strange attractors
- flow-capture(1)
- Manage storage of flow file archives by expiring old data
- flow-cat(1)
- Concatenate flow files
- flow-dscan(1)
- Detect scanning and other suspicious network activity
- flow-expire(1)
- Manage storage of flow file archives by expiring old data
- flow-export(1)
- Export flow-tools files into other NetFlow packages
- flow-fanout(1)
- Fanout (replicate) flow exports to many destinations
- flow-filter(1)
- Filter flows
- flow-gen(1)
- Generate test flows
- flow-header(1)
- Display meta information in flow file
- flow-import(1)
- Import flows into flow-tools from other NetFlow packages
- flow-log2rrd(1)
- Process the STAT: lines from flow-capture and flow-fanout to RRD files
- flow-mask(1)
- Apply tags to flow files
- flow-merge(1)
- Merge flow files
- flow-nfilter(1)
- Filter flows
- flow-print(1)
- Display flows in formatted ASCII
- flow-receive(1)
- Receive flow data with the NetFlow protocol
- flow-report(1)
- Generate reports from flow data
- flow-rpt2rrd(1)
- Convert flow-report CSV output to RRDtool format
- flow-rptfmt(1)
- Format the output of flow-report in ASCII or HTML
- flow-send(1)
- Transmit flow data with the NetFlow protocol
- flow-split(1)
- Split flow files into smaller files
- flow-stat(1)
- Generate reports with flow data
- flow-tag(1)
- Apply tags to flow files
- flow-tools(1)
- Tool set for working with NetFlow data
- flow-tools-examples(1)
- Example usage of flow-tools
- flow-xlate(1)
- Apply translations to selected fields of a flow
- flow6(1)
- A security assessment tool for the IPv6 Flow Label field
- flowcap(8)
- Capture network flow data and write it to temporary files
- flowd(8)
- NetFlow collector daemon
- flowd-reader(8)
- Read, filter and concatenate binary flowd logfiles
- flowd.conf(5)
- NetFlow Collector daemon configuration file
- flowgrind(1)
- advanced TCP traffic generator for Linux, FreeBSD, and Mac OS X
- flowgrind-stop(1)
- helper tool for the advanced TCP traffic generator flowgrind
- flowgrindd(1)
- advanced TCP traffic generator for Linux, FreeBSD, and Mac OS X
- flowkey(3)
- SiLK plug-in providing YAF flow key filter and field
- flowrate(3)
- SiLK plug-in providing payload and rate filters and fields
- fprobe(8)
- a NetFlow probe
- fuzzyflakes(6)
- falling snowflakes/flower shapes
- getFlowKeyHash(1)
- YAF Flow Key Calculator
- gh-run(1)
- View details about workflow runs
- gh-run-cancel(1)
- Cancel a workflow run
- gh-run-delete(1)
- Delete a workflow run
- gh-run-download(1)
- Download artifacts generated by a workflow run
- gh-run-list(1)
- List recent workflow runs
- gh-run-view(1)
- View a summary of a workflow run
- gh-workflow(1)
- View details about GitHub Actions workflows
- gh-workflow-disable(1)
- Disable a workflow
- gh-workflow-enable(1)
- Enable a workflow
- gh-workflow-list(1)
- List workflows
- gh-workflow-run(1)
- Run a workflow by creating a workflow_dispatch event
- gh-workflow-view(1)
- View the summary of a workflow
- gitworkflows(7)
- An overview of recommended workflows with Git
- gnuradio-companion(1)
- GNU Radio Companion (GRC) is a graphical tool for creating GNU Radio signal flowgraphs
- goto-analyzer(1)
- Data-flow analysis for C programs and goto binaries
- gvcolor(1)
- flow colors through a ranked digraph
- ipfixcol2-fds-output(7)
- Flow Data Storage (output plugin)
- labad(3)
- labad: over/underflow on obsolete pre-IEEE machines
- larmm(3)
- larmm: scale factor to avoid overflow, step in latrs
- lassq(3)
- lassq: sum-of-squares, avoiding over/underflow
- libowfat_add_of(3), add_of(3)
- add two integers, check for arithmetic overflow
- libowfat_imult16(3), imult16(3)
- integer multiplication with overflow checking
- libowfat_imult32(3), imult32(3)
- integer multiplication with overflow checking
- libowfat_imult64(3), imult64(3)
- integer multiplication with overflow checking