Hi Davide,

Thanks for the -A switch! Still need to learn some more Kubernetes. :)))

So no network policy in the "some" namespace but the -A switch was working.
:)

kubectl get networkpolicy -A
NAMESPACE          NAME              POD-SELECTOR     AGE
calico-apiserver   allow-apiserver   apiserver=true   232d


kubectl describe networkpolicy allow-apiserver -n calico-apiserver
Name:         allow-apiserver
Namespace:    calico-apiserver
Created on:   2022-05-25 16:39:58 +0200 CEST
Labels:       <none>
Annotations:  <none>
Spec:
 PodSelector:     apiserver=true
 Allowing ingress traffic:
   To Port: 5443/TCP
   From: <any> (traffic not restricted by source)
 Not affecting egress traffic
 Policy Types: Ingress

But as you mentioned earlier, this one doesn't seem to be involved in the
issue.

CNI:
I guess it's Calico. A colleague of mine installed our Kubernetes cluster
but I've found this config at /etc/cni/net.d/10-calico.conflict

{
  "name": "k8s-pod-network",
  "cniVersion": "0.3.1",
  "plugins": [
    {
      "type": "calico",
      "datastore_type": "kubernetes",
      "mtu": 0,
      "nodename_file_optional": false,
      "log_level": "Info",
      "log_file_path": "/var/log/calico/cni/cni.log",
      "ipam": { "type": "calico-ipam", "assign_ipv4" : "true",
"assign_ipv6" : "false"},
      "container_settings": {
          "allow_ip_forwarding": false
      },
      "policy": {
          "type": "k8s"
      },
      "kubernetes": {
          "k8s_api_root":"https://10.X.Y.1:443";, (I've changed)
          "kubeconfig": "/etc/cni/net.d/calico-kubeconfig"
      }
    },
    {
      "type": "bandwidth",
      "capabilities": {"bandwidth": true}
    },
    {"type": "portmap", "snat": true, "capabilities": {"portMappings":
true}}
  ]
}


Aaand I have some other news. Someone on Bacula-users list is helping me
investigate the issue. She thinks it's some database issue. Some INSERT run
into error. We'll see.

Best regards,
Zsolt

On Fri, Jan 13, 2023 at 5:16 AM Davide F. <bacula-...@dflc.ch> wrote:

> Hello Zsolt,
>
> You’re really welcome
>
> On Thu, 12 Jan 2023 at 16:29 Zsolt Kozak <koza...@gmail.com> wrote:
>
>> Hello Davide!
>>
>> I really appreciate your kind help!
>>
>> kubectl get networkpolicy gave the following:
>>
>>
>> "No resources found in default namespace."
>>
>>
> This is because network policies CR are relative to their namespace.
>
> You can use -A to list a specific resource kind for all namespaces.
>
>>
>>
>> Actually I've tried to run the Kubernetes plugin in a so-called "some" 
>> namespace, but there is no networkpolicy in "some" NS neither. (I've changed 
>> the name of the NS to some.)
>>
>>
>> kubectl get networkpolicy -n some
>> No resources found in some namespace.
>>
>>
> I’d suggest you to have a look at Kubernetes documentation about Network
> Policies
>
> https://kubernetes.io/docs/concepts/services-networking/network-policies/
>
> Another question, which CNI (container network interface ) are you using
> in your cluster ?
>
>>
>>
>> Best regards,
>>
>> Zsolt
>>
>>
> Again, I’ll have a try on my side and keep you updated.
>
> Best regards
>
> Davide
>
>
>>
>>
>> On Thu, Jan 12, 2023 at 8:00 AM Davide F. <bacula-...@dflc.ch> wrote:
>>
>>> Hello Zsolt,
>>>
>>> Indeed, the NetworkPolicy you've provided doesn't seem to be involved in
>>> the issue your facing.
>>>
>>> Let's keep trying to figure out what's going on with your setup
>>>
>>> Could you run these commands below
>>>
>>> kubectl get networkpolicy
>>>
>>>
>>> and if you get some result, run
>>>
>>> kubectl describe networkpolicy <networkpolicy-name>
>>>
>>>
>>> In the meantime, I'll setup a "test" environement and see if I'm facing
>>> the problem.
>>>
>>> I'll keep you updated.
>>>
>>> Best regards
>>>
>>> Davide
>>>
>>> On Wed, Jan 11, 2023 at 5:54 PM Zsolt Kozak <koza...@gmail.com> wrote:
>>>
>>>> Hi!
>>>>
>>>> Yes, but only one tiny:
>>>>
>>>> kind: NetworkPolicy
>>>> apiVersion: networking.k8s.io/v1
>>>> metadata:
>>>>   name: allow-apiserver
>>>>   namespace: calico-apiserver
>>>>   ownerReferences:
>>>>     - apiVersion: operator.tigera.io/v1
>>>>       kind: APIServer
>>>>       name: default
>>>>       controller: true
>>>>       blockOwnerDeletion: true
>>>>   managedFields:
>>>>     - manager: operator
>>>>       operation: Update
>>>>       apiVersion: networking.k8s.io/v1
>>>> spec:
>>>>   podSelector:
>>>>     matchLabels:
>>>>       apiserver: 'true'
>>>>   ingress:
>>>>     - ports:
>>>>         - protocol: TCP
>>>>           port: 5443
>>>>   policyTypes:
>>>>     - Ingress
>>>> status: {}
>>>>
>>>> But I guess it's an allow, not a block policy. (I'm somewhat new to
>>>> Kubernetes and not too familiar with network policies...)
>>>>
>>>> Best regards,
>>>> Zsolt
>>>>
>>>> On Wed, Jan 11, 2023 at 5:47 PM Davide F. <bacula-...@dflc.ch> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> Are you using some kind of network policy within your cluster ?
>>>>>
>>>>> Best,
>>>>>
>>>>> Davide
>>>>>
>>>>> On Wed, 11 Jan 2023 at 10:53 Zsolt Kozak <koza...@gmail.com> wrote:
>>>>>
>>>>>> Hello Davide!
>>>>>>
>>>>>> I am running the File Daemon on the master node, on the control
>>>>>> plane. It's Kubernetes vanilla, version 1.25.4.
>>>>>> No, the master node is running on the same subnet as the workers.
>>>>>>
>>>>>> It should be some network issue, I think.
>>>>>>
>>>>>> Best regards,
>>>>>> Zsolt
>>>>>>
>>>>>> On Wed, Jan 11, 2023 at 8:45 AM Davide F. <bacula-...@dflc.ch> wrote:
>>>>>>
>>>>>>> Hello Kozak,
>>>>>>>
>>>>>>> I haven’t tried k8s plugin but let me try to understand what could
>>>>>>> be the root cause of your problem.
>>>>>>>
>>>>>>> Could you explain further point 1 please ?
>>>>>>> On which node are you running the file daemon ?
>>>>>>>
>>>>>>> Which version / flavor of Kubernetes are you using ?
>>>>>>>
>>>>>>> Is it Kubernetes vanilla ? OpenShift ? Tansu ?
>>>>>>>
>>>>>>> Depending on your feedback from the first question, does master
>>>>>>> nodes runs in a different subnet than worker’s ?
>>>>>>>
>>>>>>> Thanks for your feedback
>>>>>>>
>>>>>>> Best,
>>>>>>>
>>>>>>> Davide
>>>>>>>
>>>>>>> On Tue, 10 Jan 2023 at 21:12 Zsolt Kozak <koza...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Hello,
>>>>>>>>
>>>>>>>> I have some problems with backuping Kubernetes PVCs with Bacula
>>>>>>>> Kubernetes Plugin. (I have asked it on bacula-users mailing list but 
>>>>>>>> got no
>>>>>>>> answer.)
>>>>>>>>
>>>>>>>> I am using the latest 13.0.1 Bacula from the community builds on
>>>>>>>> Debian Bullseye hosts.
>>>>>>>>
>>>>>>>> Backuping only the Kubernetes objects except Persistent Volume
>>>>>>>> Claims (PVC) works like a charm. I've installed the Kubernetes plugin 
>>>>>>>> and
>>>>>>>> the latest Bacula File Daemon on the master node (control plane) of our
>>>>>>>> Kubernetes cluster. Bacula can access the Kubernetes cluster and backup
>>>>>>>> every single object as YAML files.
>>>>>>>>
>>>>>>>> The interesting part comes with trying to backup a PVC...
>>>>>>>>
>>>>>>>> First of all I could build my own Bacula Backup Proxy Pod Image
>>>>>>>> from the source and it's deployed into our local Docker image 
>>>>>>>> repository
>>>>>>>> (repo). The Bacula File Daemon is configured properly I guess. Backup
>>>>>>>> process started and the following things happened.
>>>>>>>>
>>>>>>>> 1. Bacula File Daemon deployed Bacula Backup Proxy Pod Image into
>>>>>>>> the Kubernetes cluster, so Bacula-backup container pod started.
>>>>>>>> 2. I got into the pod and I could see the Baculatar application
>>>>>>>> started and running.
>>>>>>>> 3. The k8s_backend application started on the Bacula File Daemon
>>>>>>>> host (kubernetes.server) in 2 instances.
>>>>>>>> 4. From the Bacula-backup pod I could check that Baculatar could
>>>>>>>> connect to the k8s_backend at the default 9104 port
>>>>>>>> (kubernetes.server:9104).
>>>>>>>> 5. I checked the console messages of the job with Bat that Bacula
>>>>>>>> File Daemon started to process the configured PVC, started to write a
>>>>>>>> pvc.tar but nothing happened.
>>>>>>>> 6. After default 600 sec, after timeout the job was cancelled.
>>>>>>>> 7. It may be important that Bacula File Daemon could not delete the
>>>>>>>> Bacula-backup pod. (It could create it but could not delete it.)
>>>>>>>>
>>>>>>>>
>>>>>>>> Could you please tell me what's wrong?
>>>>>>>>
>>>>>>>>
>>>>>>>> Here are some log parts. (I've changed some sensitive data.)
>>>>>>>>
>>>>>>>>
>>>>>>>> Bacula File Daemon configuration:
>>>>>>>>
>>>>>>>> FileSet {
>>>>>>>>     Name = "Kubernetes Set"
>>>>>>>>     Include {
>>>>>>>>         Options {
>>>>>>>>             signature = SHA512
>>>>>>>>             compression = GZIP
>>>>>>>>             Verify = pins3
>>>>>>>>         }
>>>>>>>>         Plugin = "kubernetes: \
>>>>>>>>             debug=1 \
>>>>>>>>             baculaimage=repo/bacula-backup:04jan23 \
>>>>>>>>             namespace=namespace \
>>>>>>>>             pvcdata \
>>>>>>>>             pluginhost=kubernetes.server \
>>>>>>>>             timeout=120 \
>>>>>>>>             verify_ssl=0 \
>>>>>>>>             fdcertfile=/etc/bacula/certs/bacula-backup.cert \
>>>>>>>>             fdkeyfile=/etc/bacula/certs/bacula-backup.key"
>>>>>>>>     }
>>>>>>>> }
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Bacula File Daemon debug log (parts):
>>>>>>>>
>>>>>>>>
>>>>>>>> DEBUG:[baculak8s/jobs/estimation_job.py:134 in processing_loop]
>>>>>>>> processing get_annotated_namespaced_pods_data:namespace:nrfound:0
>>>>>>>> DEBUG:[baculak8s/plugins/kubernetes_plugin.py:319 in
>>>>>>>> list_pvcdata_for_namespace] list pvcdata for namespace:namespace
>>>>>>>> pvcfilter=True estimate=False
>>>>>>>> DEBUG:[baculak8s/plugins/k8sbackend/pvcdata.py:108 in
>>>>>>>> pvcdata_list_namespaced] pvcfilter: True
>>>>>>>> DEBUG:[baculak8s/plugins/k8sbackend/pvcdata.py:112 in
>>>>>>>> pvcdata_list_namespaced] found:some-claim
>>>>>>>> DEBUG:[baculak8s/plugins/k8sbackend/pvcdata.py:127 in
>>>>>>>> pvcdata_list_namespaced] add pvc: {'name': 'some-claim', 'node_name': 
>>>>>>>> None,
>>>>>>>> 'storage_class_name': 'nfs-client', 'capacity': '2Gi', 'fi':
>>>>>>>> <baculak8s.entities.file_info.FileInfo object at 0x7ffaa55bfcc0>}
>>>>>>>> DEBUG:[baculak8s/jobs/estimation_job.py:165 in processing_loop]
>>>>>>>> processing list_pvcdata_for_namespace:namespace:nrfound:1
>>>>>>>> DEBUG:[baculak8s/jobs/estimation_job.py:172 in processing_loop]
>>>>>>>> PVCDATA:some-claim:{'name': 'some-claim', 'node_name': 'node1',
>>>>>>>> 'storage_class_name': 'nfs-client', 'capacity': '2Gi', 'fi':
>>>>>>>> <baculak8s.entities.file_info.FileInfo object at 0x7ffaa55bfcc0>}
>>>>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet
>>>>>>>> I000041
>>>>>>>> Start backup volume claim: some-claim
>>>>>>>>
>>>>>>>> DEBUG:[baculak8s/jobs/job_pod_bacula.py:298 in prepare_bacula_pod]
>>>>>>>> prepare_bacula_pod:token=xx88M5oggQJ....4YDbSwBRxTOhT 
>>>>>>>> namespace=namespace
>>>>>>>> DEBUG:[baculak8s/jobs/job_pod_bacula.py:136 in prepare_pod_yaml]
>>>>>>>> pvcdata: {'name': 'some-claim', 'node_name': 'node1', 
>>>>>>>> 'storage_class_name':
>>>>>>>> 'nfs-client', 'capacity': '2Gi', 'fi':
>>>>>>>> <baculak8s.entities.file_info.FileInfo object at 0x7ffaa55bfcc0>}
>>>>>>>> DEBUG:[baculak8s/plugins/k8sbackend/baculabackup.py:102 in
>>>>>>>> prepare_backup_pod_yaml] host:kubernetes.server port:9104
>>>>>>>> namespace:namespace image:repo/bacula-backup:04jan23
>>>>>>>> job:KubernetesBackup.2023-01-04_21.05.03_10:410706
>>>>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet
>>>>>>>> I000149
>>>>>>>> Prepare Bacula Pod on: node1 with: repo/bacula-backup:04jan23
>>>>>>>> <IfNotPresent> kubernetes.server:9104
>>>>>>>>
>>>>>>>> DEBUG:[baculak8s/jobs/job_pod_bacula.py:198 in
>>>>>>>> prepare_connection_server] prepare_connection_server:New 
>>>>>>>> ConnectionServer:
>>>>>>>> 0.0.0.0:9104
>>>>>>>> DEBUG:[baculak8s/util/sslserver.py:180 in listen]
>>>>>>>> ConnectionServer:Listening...
>>>>>>>> DEBUG:[baculak8s/jobs/job_pod_bacula.py:307 in prepare_bacula_pod]
>>>>>>>> prepare_bacula_pod:start pod
>>>>>>>> INFO:[baculak8s/plugins/kubernetes_plugin.py:771 in
>>>>>>>> backup_pod_isready] backup_pod_status:isReady: False / 0
>>>>>>>> INFO:[baculak8s/plugins/kubernetes_plugin.py:771 in
>>>>>>>> backup_pod_isready] backup_pod_status:isReady: True / 1
>>>>>>>> DEBUG:[baculak8s/jobs/estimation_job.py:183 in _estimate_file]
>>>>>>>> {'name': 'some-claim', 'node_name': 'node1', 'storage_class_name':
>>>>>>>> 'nfs-client', 'capacity': '2Gi', 'fi':
>>>>>>>> <baculak8s.entities.file_info.FileInfo object at 0x7ffaa55bfcc0>}
>>>>>>>> DEBUG:[baculak8s/jobs/estimation_job.py:190 in _estimate_file]
>>>>>>>> file_info: {FileInfo
>>>>>>>> name:/@kubernetes/namespaces/namespace/persistentvolumeclaims/some-claim.tar
>>>>>>>> namespace:None type:F objtype:pvcdata cached:False}
>>>>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet
>>>>>>>> C000079
>>>>>>>>
>>>>>>>> FNAME:/@kubernetes/namespaces/namespace/persistentvolumeclaims/some-claim.tar
>>>>>>>>
>>>>>>>>
>>>>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet
>>>>>>>> C000040
>>>>>>>> TSTAMP:1672861077 1672861077 1672861077
>>>>>>>>
>>>>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet
>>>>>>>> C000031
>>>>>>>> STAT:F 2147483648 0 0 100640 1
>>>>>>>>
>>>>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet
>>>>>>>> F000000
>>>>>>>> (EOD PACKET)
>>>>>>>>
>>>>>>>> DEBUG:[baculak8s/jobs/backup_job.py:77 in __backup_pvcdata]
>>>>>>>> backup_pvcdata:data recv
>>>>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet
>>>>>>>> C000005
>>>>>>>> DATA
>>>>>>>>
>>>>>>>> DEBUG:[baculak8s/util/sslserver.py:193 in handle_connection]
>>>>>>>> ConnectionServer:Connection from: ('192.168.XX.YY', 10541)
>>>>>>>> DEBUG:[baculak8s/util/sslserver.py:145 in gethello] ['Hello',
>>>>>>>> 'KubernetesBackup.2023-01-04_21.05.03_10', '410706']
>>>>>>>> DEBUG:[baculak8s/util/token.py:57 in check_auth_data]
>>>>>>>> AUTH_DATA:Token: xx88M5oggQJuGsPbtD........ohQjeU7PkA4YDbSwBRxTOhT
>>>>>>>> DEBUG:[baculak8s/util/token.py:59 in check_auth_data]
>>>>>>>> RECV_TOKEN_DATA:Token: xx88M5oggQJuGsPbtD....ohQjeU7PkA4YDbSwBRxTOhT
>>>>>>>> DEBUG:[baculak8s/util/sslserver.py:105 in authenticate]
>>>>>>>> ConnectionServer:Authenticated
>>>>>>>>
>>>>>>>> .... after timeout
>>>>>>>>
>>>>>>>> DEBUG:[baculak8s/jobs/job_pod_bacula.py:121 in
>>>>>>>> handle_pod_data_recv] handle_pod_data_recv:EOT
>>>>>>>> DEBUG:[baculak8s/util/sslserver.py:201 in handle_connection]
>>>>>>>> ConnectionServer:Finish - disconnect.
>>>>>>>> DEBUG:[baculak8s/jobs/backup_job.py:85 in __backup_pvcdata]
>>>>>>>> backup_pvcdata:logs recv
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Job messages:
>>>>>>>>
>>>>>>>> bacula-dir No prior or suitable Full backup found in catalog for
>>>>>>>> the current FileSet. Doing FULL backup.
>>>>>>>>  The FileSet "Kubernetes Set" was modified on 2023-01-04 20:20:41,
>>>>>>>> this is after the last successful backup on 2023-01-04 19:19:49.
>>>>>>>> bacula-sd Ready to append to end of Volume "Full-XXX"
>>>>>>>> size=3,838,161,002
>>>>>>>> bacula-fd Connected to Storage at bacula.server:9103 with TLS
>>>>>>>> bacula-sd Volume "Full-XXXX" previously written, moving to end of
>>>>>>>> data.
>>>>>>>> bacula-dir Connected to Client "bacula-fd" at
>>>>>>>> kubernetes.server:9102 with TLS
>>>>>>>>  Using Device "FileStorageEeyoreFull" to write.
>>>>>>>>  Connected to Storage "InternalStorageFull" at bacula.server:9103
>>>>>>>> with TLS
>>>>>>>>  Start Backup JobId 410830,
>>>>>>>> Job=KubernetesBackup.2023-01-04_21.05.03_10
>>>>>>>> bacula-fd kubernetes: Prepare Bacula Pod on: node with:
>>>>>>>> repo/bacula-backup:04jan23  kubernetes.server:9104
>>>>>>>>  kubernetes: Processing namespace: namespace
>>>>>>>>  kubernetes: Start backup volume claim: some-claim
>>>>>>>>  kubernetes: Connected to Kubernetes 1.25 - v1.25.4.
>>>>>>>> bacula-dir
>>>>>>>> Error: Bacula Enterprise bacula-dir 13.0.1 (05Aug22):
>>>>>>>>   Build OS:               x86_64-pc-linux-gnu-bacula-enterprise
>>>>>>>> debian 11.2
>>>>>>>>   JobId:                  410830
>>>>>>>>   Job:                    KubernetesBackup.2023-01-04_21.05.03_10
>>>>>>>>   Backup Level:           Full (upgraded from Differential)
>>>>>>>>   Client:                 "bacula-fd" 13.0.1 (05Aug22)
>>>>>>>> x86_64-pc-linux-gnu-bacula-enterprise,debian,10.11
>>>>>>>>   FileSet:                "Kubernetes Set" 2023-01-04 20:20:41
>>>>>>>>   Pool:                   "Full-Pool" (From Job FullPool override)
>>>>>>>>   Catalog:                "MyCatalog" (From Client resource)
>>>>>>>>   Storage:                "InternalStorageFull" (From Pool resource)
>>>>>>>>   Scheduled time:         04-Jan-2023 21:05:03
>>>>>>>>   Start time:             04-Jan-2023 21:27:04
>>>>>>>>   End time:               04-Jan-2023 21:29:06
>>>>>>>>   Elapsed time:           2 mins 2 secs
>>>>>>>>   Priority:               10
>>>>>>>>   FD Files Written:       23
>>>>>>>>   SD Files Written:       0
>>>>>>>>   FD Bytes Written:       52,784 (52.78 KB)
>>>>>>>>   SD Bytes Written:       0 (0 B)
>>>>>>>>   Rate:                   0.4 KB/s
>>>>>>>>   Software Compression:   100.0% 1.0:1
>>>>>>>>   Comm Line Compression:  5.6% 1.1:1
>>>>>>>>   Snapshot/VSS:           no
>>>>>>>>   Encryption:             yes
>>>>>>>>   Accurate:               yes
>>>>>>>>   Volume name(s):         Full-XXXX
>>>>>>>>   Volume Session Id:      43
>>>>>>>>   Volume Session Time:    1672853724
>>>>>>>>   Last Volume Bytes:      3,838,244,105 (3.838 GB)
>>>>>>>>   Non-fatal FD errors:    3
>>>>>>>>   SD Errors:              0
>>>>>>>>   FD termination status:  OK
>>>>>>>>   SD termination status:  SD despooling Attributes
>>>>>>>>   Termination:            *** Backup Error ***
>>>>>>>>  Fatal error: catreq.c:680 Restore object create error.
>>>>>>>> bacula-fd
>>>>>>>> Error: kubernetes: PTCOMM cannot get packet header from backend.
>>>>>>>> bacula-dir Fatal error: sql_create.c:1273 Create db Object record
>>>>>>>> INSERT INTO RestoreObject
>>>>>>>> (ObjectName,PluginName,RestoreObject,ObjectLength,ObjectFullLength,ObjectIndex,ObjectType,ObjectCompression,FileIndex,JobId)
>>>>>>>> VALUES ('RestoreOptions','kubernetes: \n            debug=1 \n
>>>>>>>>        baculaimage=repo/bacula-backup:04jan23 \n
>>>>>>>>  namespace=namespace \n            pvcdata \n
>>>>>>>> pluginhost=kubernetes.server \n            timeout=120 \n
>>>>>>>>  verify_ssl=0 \n            
>>>>>>>> fdcertfile=/etc/bacula/certs/bacula-backup.cert
>>>>>>>> \n
>>>>>>>> fdkeyfile=/etc/bacula/certs/bacula-backup.key','# Plugin
>>>>>>>> configuration file\n# Version 1\nOptPrompt=\"K8S config
>>>>>>>> file\"\nOptDefault=\"*None*\"\nconfig=@STR@\n\n
>>>>>>>> OptPrompt=\"K8S API server
>>>>>>>> URL/Host\"\nOptDefault=\"*None*\"\nhost=@STR@\n\nOptPrompt=\"K8S
>>>>>>>> Bearertoken\"\nOptDefault=\"*None*\"\ntoken=@STR@\n\nOptPrompt=\"K8S
>>>>>>>> API server cert verification\"\n
>>>>>>>> OptDefault=\"True\"\nverify_ssl=@BOOL@\n\nOptPrompt=\"Custom CA
>>>>>>>> Certs file to 
>>>>>>>> use\"\nOptDefault=\"*None*\"\nssl_ca_cert=@STR@\n\nOptPrompt=\"Output
>>>>>>>> format when saving to file (JSON, YAML)\"\n
>>>>>>>> OptDefault=\"RAW\"\noutputformat=@STR@\n\nOptPrompt=\"The address
>>>>>>>> for listen to incoming backup pod
>>>>>>>> data\"\nOptDefault=\"*FDAddress*\"\nfdaddress=@STR@\n\n
>>>>>>>> OptPrompt=\"The port for opening socket for
>>>>>>>> listen\"\nOptDefault=\"9104\"\nfdport=@INT32@\n\nOptPrompt=\"The
>>>>>>>> endpoint address for backup pod to connect\"\n
>>>>>>>> OptDefault=\"*FDAddress*\"\npluginhost=@STR@\n\nOptPrompt=\"The
>>>>>>>> endpoint port to connect\"\nOptDefault=\"9104\"\n
>>>>>>>> pluginport=@INT32@\n\n',859,859,0,27,0,1,410830) failed. ERR=Data
>>>>>>>> too long for column 'PluginName' at row 1
>>>>>>>>
>>>>>>>> bacula-sd Sending spooled attrs to the Director. Despooling 8,214
>>>>>>>> bytes ...
>>>>>>>> bacula-fd
>>>>>>>> Error: kubernetes: Error closing backend. Err=Child exited with
>>>>>>>> code 1
>>>>>>>>  Fatal error: kubernetes: Wrong backend response to JobEnd command.
>>>>>>>> bacula-sd Elapsed time=00:02:02, Transfer rate=659  Bytes/second
>>>>>>>> bacula-fd
>>>>>>>> Error: kubernetes: PTCOMM cannot get packet header from backend.
>>>>>>>>
>>>>>>>> Error: kubernetes: Cannot successfully start bacula-backup pod in
>>>>>>>> expected time!
>>>>>>>>
>>>>>>>> Error: kubernetes: Job already running in 'namespace' namespace.
>>>>>>>> Check logs or delete bacula-backup Pod manually.
>>>>>>>>
>>>>>>>>
>>>>>>>> Best regards,
>>>>>>>> Zsolt
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Bacula-devel mailing list
>>>>>>>> Bacula-devel@lists.sourceforge.net
>>>>>>>> https://lists.sourceforge.net/lists/listinfo/bacula-devel
>>>>>>>>
>>>>>>>
_______________________________________________
Bacula-devel mailing list
Bacula-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-devel

Reply via email to