I’ve set “forks = 1” and rerun my test which does an ansible-playbook and an
uptime.
Could it be the register rather than a memory leak?
Here free(1) keeps going down:
> [meggleston@pc1uepsiadm01 ~]$ while true
> > do
> > date
> > free -m
> > sleep 60
> > done
> Fri Dec 9 16:30:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1139 6190 18 490
> 6408
> Swap: 4095 2897 1198
> Fri Dec 9 16:31:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1149 6127 19 544
> 6398
> Swap: 4095 2896 1199
> Fri Dec 9 16:32:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1163 6112 19 544
> 6383
> Swap: 4095 2896 1199
> Fri Dec 9 16:33:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1173 6100 19 546
> 6373
> Swap: 4095 2896 1199
> Fri Dec 9 16:34:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1184 6081 19 554
> 6362
> Swap: 4095 2895 1200
> Fri Dec 9 16:35:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1215 6042 20 562
> 6329
> Swap: 4095 2885 1210
> Fri Dec 9 16:36:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1225 6032 20 563
> 6319
> Swap: 4095 2884 1211
> Fri Dec 9 16:37:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1229 6028 20 563
> 6315
> Swap: 4095 2880 1215
> Fri Dec 9 16:38:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1257 5999 20 563
> 6287
> Swap: 4095 2861 1234
> Fri Dec 9 16:39:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1286 5971 21 563
> 6259
> Swap: 4095 2860 1235
> Fri Dec 9 16:40:31 EST 2022
> total used free shared buff/cache
> available
> Mem: 7821 1295 5961 21 563
> 6249
> Swap: 4095 2860 1235
>
The playbook is really stupid:
[meggleston@pc1uepsiadm01 playbooks]$ cat y.yml
# $Id$
# $Log$
# set cyberark to a known point
# :!ansible-playbook --syntax-check %
# :!ansible-playbook --check --limit pc1uepsiadm01.res.prod.global %
# :!ansible-playbook --limit pc1uepsiadm01.res.prod.global %
# :!ansible-playbook %
---
- hosts: all
become: no
gather_facts: true
vars:
reportfile:
"/tmp/inventory-report-{{ansible_date_time.year}}{{ansible_date_time.month}}{{ansible_date_time.day}}.csv"
uptime_host: "UNKNOWN"
tasks:
- name: "get the uptime"
command: uptime
register: uptime
- set_fact:
uptime_host="{{uptime.stdout}}"
when: uptime is match("")
- name: "create the file and write the header"
lineinfile:
path="{{reportfile}}"
state="present"
insertafter="EOF"
create="true"
line="\"hostname\",\"uptime\""
delegate_to: localhost
- name: "write the line for this host"
lineinfile:
path="{{reportfile}}"
state="present"
insertafter="EOF"
create="true"
line="\"{{ansible_host}}\",\"{{uptime_host}}\""
delegate_to: localhost
Mike
> On Dec 9, 2022, at 15:29, Mike Eggleston <[email protected]> wrote:
>
> Does Ansible have a memory leak (that only shows up with a high number of
> hosts)?
>
> Mike
>
> [meggleston@pc1uepsiadm01 ~]$ rpm -qa | grep -i ansible
> ansible-2.9.27-1.el7.noarch
>
>> On Dec 9, 2022, at 09:16, Mike Eggleston <[email protected]
>> <mailto:[email protected]>> wrote:
>>
>> I changed forks = back to 5 (commented out my change) and still I get the
>> out of memory error. I removed all hosts that are in AWS so I’m not using
>> the proxy in ssh(1). My inventory is down to 4400 hosts. I wonder what’s
>> eating the memory…? Any ideas….?
>>
>> Current stack trace:
>> 82"]}, "sectors": "7812935680", "start": "2048", "holders": [], "size":
>> "3.64 TB"}}, "sas_device_handle": null, "sas_address": null, "virtual": 1,
>> "host": "RAID bus controller: Broadcom / LSI MegaRAID SAS-3 3108 [Invader]
>> (rev 02)", "sectorsize": "512", "removable": "0", "support_discard": "0",
>> "model": "PERC H730P Mini", "wwn": "0x61866da06192eb0024e6a07712d7ee30",
>> "holders": [], "size": "3.64 TB"}, "dm-4": {"scheduler_mode": "",
>> "rotational": "1", "vendor": null, "sectors": "104857600", "links":
>> {"masters": [], "labels": [], "ids": ["dm-name-rootvg-optvol",
>> "dm-uuid-LVM-h8Zoe5OZiBjf9awu8HyY4OuQIIZK52yneJbOfXRJ6QddDY581MzfUj6Ai4MOtle8"],
>> "uuids": ["51166620-cd67-4954-8b5f-cf91926b036d"]}, "sas_device_handle":
>> null, "sas_address": null, "virtual": 1, "host": "", "sectorsize": "512",
>> "removable": "0", "support_discard": "0", "model": null, "partitions": {},
>> "holders": [], "size": "50.00 GB"}, "dm-5": {"scheduler_mode": "",
>> "rotational": "1", "vendor": null, "sectors": "62914560", "links":
>> {"masters": [], "labels": [], "ids": ["dm-name-rootvg-tmpvol",
>> "dm-uuid-LVM-h8Zoe5OZiBjf9awu8HyY4OuQIIZK52ynDjjkCPQW51kaWpzKqwJkcPy2qbRW0Fxm"],
>> "uuids": ["38f0cd51-d7e1-4bca-a062-1b39ede2fed2"]}, "sas_device_handle":
>> null, "sas_address": null, "virtual": 1, "host": "", "sectorsize": "512",
>> "removable": "0", "support_discard": "0", "model": null, "partitions": {},
>> "holders": [], "size": "30.00 GB"}, "dm-2": {"scheduler_mode": "",
>> "rotational": "1", "vendor": null, "sectors": "125829120", "links":
>> {"masters": [], "labels": [], "ids": ["dm-name-rootvg-varvol",
>> "dm-uuid-LVM-h8Zoe5OZiBjf9awu8HyY4OuQIIZK52ynRn8k3kCCl3ICeXjPbpYKBa1d9B7s2bhs"],
>> "uuids": ["5cb9ffc3-fd77-4906-98d7-edb27aa63f40"]}, "sas_device_handle":
>> null, "sas_address": null, "virtual": 1, "host": "", "sectorsize": "512",
>> "removable": "0", "support_discard": "0", "model": null, "partitions": {},
>> "holders": [], "size": "60.00 GB"}, "dm-3": {"scheduler_mode": "",
>> "rotational": "1", "vendor": null, "sectors": "8388608", "links":
>> {"masters": [], "labels": [], "ids": ["dm-name-rootvg-homevol",
>> "dm-uuid-LVM-h8Zoe5OZiBjf9awu8HyY4OuQIIZK52ynVWpzyMe28x7F4igthHHVgvTM2K8ZI08R"],
>> "uuids": ["8627acb7-4c2b-4394-95d0-6a084066a23a"]}, "sas_device_handle":
>> null, "sas_address": null, "virtual": 1, "host": "", "sectorsize": "512",
>> "removable": "0", "support_discard": "0", "model": null, "partitions": {},
>> "holders": [], "size": "4.00 GB"}, "dm-0": {"scheduler_mode": "",
>> "rotational": "1", "vendor": null, "sectors": "8388608", "links":
>> {"masters": [], "labels": [], "ids": ["dm-name-rootvg-lv_swap",
>> "dm-uuid-LVM-h8Zoe5OZiBjf9awu8HyY4OuQIIZK52ynqqkyDshARxIJhfyP1hRtk5SrMN3BK79c"],
>> "uuids": ["799e361d-7fee-4f6b-ae45-75d75f518985"]}, "sas_device_handle":
>> null, "sas_address": null, "virtual": 1, "host": "", "sectorsize": "512",
>> "removable": "0", "support_discard": "0", "model": null, "partitions": {},
>> "holders": [], "size": "4.00 GB"}, "dm-1": {"scheduler_mode": "",
>> "rotational": "1", "vendor": null, "sectors": "41943040", "links":
>> {"masters": [], "labels": [], "ids": ["dm-name-rootvg-rootvol",
>> "dm-uuid-LVM-h8Zoe5OZiBjf9awu8HyY4OuQIIZK52ynPxpNeSPUUNGqorDA6GDRwK4jFcd9IzuW"],
>> "uuids": ["798d9b72-8fc1-475f-92f0-0fad71bd3e5a"]}, "sas_device_handle":
>> null, "sas_address": null, "virtual": 1, "host": "", "sectorsize": "512",
>> "removable": "0", "support_discard": "0", "model": null, "partitions": {},
>> "holders": [], "size": "20.00 GB"}}, "ansible_user_uid": 2101067335,
>> "ansible_ssh_host_key_dsa_public":
>> "AAAAB3NzaC1kc3MAAACBAPPCrb44cIbwiG15T60D7doNgsOgwLOW4N76U3gvkeiJUafrLqGexH0XMMEwRhFnGGxckQGhgQE3O2ZKmlgTAYFG+qaCDBjHPGBIxKE9PcMO+enFTUYKHd4KY+xid9f3J4QnpauJZXoB4Et2GGwE0Q8fBJB7bLevybjAgAbMfM51AAAAFQCFf6SYNVwXyG0c1RYjCzeaLMB22wAAAIBm8je+yytTJ7DigfHYoleH4LrWKD0g0PeSBFVKG0snNlorhBtCGa5QIKwgR9OE+BNXQddwcqHf1jwmn54wcROWicNbdJFdIrDHHSnbzBm2tOkiNqovTLx92676L45uOZlBzNHi/bqOSzbSem9Piukn6pDu2XsfLmXfd4wz1Z3XagAAAIEA4B7lnz4xWwgjZCnX2oXiOPOOkVH2Xo7MG3YibLr8DnuK1L8n3m/pkX3WhAqrfw87OHECkCE3Kg4EPnXwW9FfNLR4YQnJBXWCU5IJ5M+HSOE5IDSTyNlj3HEs3SaGC0EU8APei7SvRc4k+TlonHu3m1XeKsB6yCNYZdtGhm5q4Ps=",
>> "ansible_bios_date": "11/26/2019", "ansible_system_capabilities":
>> [""]}}\r\n', 'Connection to pc1udtlhhad561.prodc1.harmony.global
>> closed.\r\n')
>> <pc1udtlhhad561.prodc1.harmony.global> ESTABLISH SSH CONNECTION FOR USER:
>> None
>> <pc1udtlhhad561.prodc1.harmony.global> SSH: EXEC ssh -C -o ControlMaster=no
>> -o ControlPersist=30s -o ConnectTimeout=15s -o
>> KbdInteractiveAuthentication=no -o
>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o
>> PasswordAuthentication=no -o ConnectTimeout=10 -o
>> ControlPath=/home/meggleston/.ansible/cp/e963c4e7ae
>> pc1udtlhhad561.prodc1.harmony.global '/bin/sh -c '"'"'rm -f -r
>> /home/meggleston/.ansible/tmp/ansible-tmp-1670598305.51-22544-115269321697104/
>> > /dev/null 2>&1 && sleep 0'"'"''
>> <pc1udtlhhad561.prodc1.harmony.global> (0, '', '')
>> ERROR! Unexpected Exception, this is probably a bug: [Errno 12] Cannot
>> allocate memory
>> the full traceback was:
>>
>> Traceback (most recent call last):
>> File "/usr/bin/ansible-playbook", line 123, in <module>
>> exit_code = cli.run()
>> File "/usr/lib/python2.7/site-packages/ansible/cli/playbook.py", line 128,
>> in run
>> results = pbex.run()
>> File
>> "/usr/lib/python2.7/site-packages/ansible/executor/playbook_executor.py",
>> line 169, in run
>> result = self._tqm.run(play=play)
>> File
>> "/usr/lib/python2.7/site-packages/ansible/executor/task_queue_manager.py",
>> line 282, in run
>> play_return = strategy.run(iterator, play_context)
>> File
>> "/usr/lib/python2.7/site-packages/ansible/plugins/strategy/linear.py", line
>> 311, in run
>> self._queue_task(host, task, task_vars, play_context)
>> File
>> "/usr/lib/python2.7/site-packages/ansible/plugins/strategy/__init__.py",
>> line 390, in _queue_task
>> worker_prc.start()
>> File
>> "/usr/lib/python2.7/site-packages/ansible/executor/process/worker.py", line
>> 100, in start
>> return super(WorkerProcess, self).start()
>> File "/usr/lib64/python2.7/multiprocessing/process.py", line 130, in start
>> self._popen = Popen(self)
>> File "/usr/lib64/python2.7/multiprocessing/forking.py", line 121, in
>> __init__
>> self.pid = os.fork()
>> OSError: [Errno 12] Cannot allocate memory
>>
>>
>>> On Dec 6, 2022, at 11:55, Todd Lewis <[email protected]
>>> <mailto:[email protected]>> wrote:
>>>
>>> I'm not particularly fluent in instrumenting resource consumption, but I'm
>>> going out on a limb and guessing that 50 or so ssh connections is a lot
>>> more light-weight than 50 forks of ansible-playbook. So ignoring ssh as a
>>> possible resource limit for the moment, try changing forks back to 5 and
>>> running your playbook. At the same time, in another window, monitor (in a
>>> way to be determined by you) resource consumption. I'd expect it to work 5
>>> forks, just not as fast as with more forks.
>>>
>>> If it *does* work, then try it again with, say, 10 forks and compare
>>> resources during that run to the 5 fork run. I expect this to also work,
>>> barely, and that you'll be almost out of … something. But you'll also have
>>> a much better picture of where the walls are in this black box.
>>>
>>> On Tuesday, December 6, 2022 at 9:32:30 AM UTC-5 [email protected]
>>> <http://gmail.com/> wrote:
>>> I’ve not changed “strategy”, but I did change “forks” from 5 to 50.
>>> I have copied /etc/ansible.cfg to ~/.ansible.cfg and changed forks = 50,
>>> inventory = $HOME/src/ansible/inventory and log_path =
>>> /tmp/${USER}_ansible.log.
>>>
>>>
>>>> On Dec 5, 2022, at 17:21, Todd Lewis <[email protected]
>>>> <applewebdata://7878B376-AE82-4780-A2EC-D7F82E325719>> wrote:
>>>>
>>>
>>>> Have you changed any defaults for "strategy" or "forks"?
>>>>
>>>> Also I see your ssh is config'd for "-o ControlMaster=auto -o
>>>> ControlPersist=60s". I'm not sure how many hosts you're caching
>>>> connections for during any give 60 second window, or how much memory that
>>>> would eat, but it may be a significant factor.
>>>>
>>>> On Monday, December 5, 2022 at 6:03:42 PM UTC-5 [email protected]
>>>> <http://gmail.com/> wrote:
>>>> 5709
>>>>
>>>>
>>>>> On Dec 5, 2022, at 15:56, Todd Lewis <[email protected] <>> wrote:
>>>>>
>>>>
>>>>> How many hosts are in your inventory?
>>>>>
>>>>> On Monday, December 5, 2022 at 4:52:39 PM UTC-5 [email protected]
>>>>> <http://gmail.com/> wrote:
>>>>> I’m getting:
>>>>>
>>>>> e/meggleston/.ansible/tmp/ansible-tmp-1670263852.31-8085-175291763336523/
>>>>> > /dev/null 2>&1 && sleep 0'"'"''
>>>>> <pa2udtlhsql602.prod.harmony.aws2> ESTABLISH SSH CONNECTION FOR USER:
>>>>> None
>>>>> <pa2udtlhsql602.prod.harmony.aws2> SSH: EXEC ssh -C -o ControlMaster=auto
>>>>> -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o
>>>>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>>>>> -o PasswordAuthentication=no -o ConnectTimeout=10 -o
>>>>> ControlPath=/home/meggleston/.ansible/cp/f012ac57b9
>>>>> pa2udtlhsql602.prod.harmony.aws2 '/bin/sh -c '"'"'rm -f -r
>>>>> /home/meggleston/.ansible/tmp/ansible-tmp-1670263852.91-8148-50804275661258/
>>>>> > /dev/null 2>&1 && sleep 0'"'"''
>>>>> <pa2udtlhsql1023.prod.harmony.aws2> ESTABLISH SSH CONNECTION FOR USER:
>>>>> None
>>>>> <pa2udtlhsql604.prod.harmony.aws2> ESTABLISH SSH CONNECTION FOR USER:
>>>>> None
>>>>> <pa2udtlhsql1023.prod.harmony.aws2> SSH: EXEC ssh -C -o
>>>>> ControlMaster=auto -o ControlPersist=60s -o
>>>>> KbdInteractiveAuthentication=no -o
>>>>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>>>>> -o PasswordAuthentication=no -o ConnectTimeout=10 -o
>>>>> ControlPath=/home/meggleston/.ansible/cp/754f3010c5
>>>>> pa2udtlhsql1023.prod.harmony.aws2 '/bin/sh -c '"'"'rm -f -r
>>>>> /home/meggleston/.ansible/tmp/ansible-tmp-1670263852.31-8085-175291763336523/
>>>>> > /dev/null 2>&1 && sleep 0'"'"''
>>>>> <pa2udtlhsql604.prod.harmony.aws2> SSH: EXEC ssh -C -o ControlMaster=auto
>>>>> -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o
>>>>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>>>>> -o PasswordAuthentication=no -o ConnectTimeout=10 -o
>>>>> ControlPath=/home/meggleston/.ansible/cp/09c53a2792
>>>>> pa2udtlhsql604.prod.harmony.aws2 '/bin/sh -c '"'"'rm -f -r
>>>>> /home/meggleston/.ansible/tmp/ansible-tmp-1670263853.52-8164-79599240649234/
>>>>> > /dev/null 2>&1 && sleep 0'"'"''
>>>>> <pa2udtlhsql1020.prod.harmony.aws2> ESTABLISH SSH CONNECTION FOR USER:
>>>>> None
>>>>> <pa2udtlhsql1020.prod.harmony.aws2> SSH: EXEC ssh -C -o
>>>>> ControlMaster=auto -o ControlPersist=60s -o
>>>>> KbdInteractiveAuthentication=no -o
>>>>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>>>>> -o PasswordAuthentication=no -o ConnectTimeout=10 -o
>>>>> ControlPath=/home/meggleston/.ansible/cp/3301cea578
>>>>> pa2udtlhsql1020.prod.harmony.aws2 '/bin/sh -c '"'"'rm -f -r
>>>>> /home/meggleston/.ansible/tmp/ansible-tmp-1670263852.15-8057-21113899783559/
>>>>> > /dev/null 2>&1 && sleep 0'"'"''
>>>>> <pa2udtlhsql602.prod.harmony.aws2> ESTABLISH SSH CONNECTION FOR USER:
>>>>> None
>>>>> <pa2udtlhsql602.prod.harmony.aws2> SSH: EXEC ssh -C -o ControlMaster=auto
>>>>> -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o
>>>>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>>>>> -o PasswordAuthentication=no -o ConnectTimeout=10 -o
>>>>> ControlPath=/home/meggleston/.ansible/cp/f012ac57b9
>>>>> pa2udtlhsql602.prod.harmony.aws2 '/bin/sh -c '"'"'rm -f -r
>>>>> /home/meggleston/.ansible/tmp/ansible-tmp-1670263852.91-8148-50804275661258/
>>>>> > /dev/null 2>&1 && sleep 0'"'"''
>>>>> <pa2udtlhsql1022.prod.harmony.aws2> ESTABLISH SSH CONNECTION FOR USER:
>>>>> None
>>>>> <pa2udtlhsql1022.prod.harmony.aws2> SSH: EXEC ssh -C -o
>>>>> ControlMaster=auto -o ControlPersist=60s -o
>>>>> KbdInteractiveAuthentication=no -o
>>>>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>>>>> -o PasswordAuthentication=no -o ConnectTimeout=10 -o
>>>>> ControlPath=/home/meggleston/.ansible/cp/a501b68168
>>>>> pa2udtlhsql1022.prod.harmony.aws2 '/bin/sh -c '"'"'rm -f -r
>>>>> /home/meggleston/.ansible/tmp/ansible-tmp-1670263852.07-8072-136961495388876/
>>>>> > /dev/null 2>&1 && sleep 0'"'"''
>>>>> ERROR! Unexpected Exception, this is probably a bug: [Errno 12] Cannot
>>>>> allocate memory
>>>>> the full traceback was:
>>>>>
>>>>> Traceback (most recent call last):
>>>>> File "/usr/bin/ansible-playbook", line 123, in <module>
>>>>> exit_code = cli.run()
>>>>> File "/usr/lib/python2.7/site-packages/ansible/cli/playbook.py", line
>>>>> 128, in run
>>>>> results = pbex.run()
>>>>> File
>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/playbook_executor.py",
>>>>> line 169, in run
>>>>> result = self._tqm.run(play=play)
>>>>> File
>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/task_queue_manager.py",
>>>>> line 282, in run
>>>>> play_return = strategy.run(iterator, play_context)
>>>>> File
>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/strategy/linear.py",
>>>>> line 311, in run
>>>>> self._queue_task(host, task, task_vars, play_context)
>>>>> File
>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/strategy/__init__.py",
>>>>> line 390, in _queue_task
>>>>> worker_prc.start()
>>>>> File
>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/process/worker.py",
>>>>> line 100, in start
>>>>> return super(WorkerProcess, self).start()
>>>>> File "/usr/lib64/python2.7/multiprocessing/process.py", line 130, in
>>>>> start
>>>>> self._popen = Popen(self)
>>>>> File "/usr/lib64/python2.7/multiprocessing/forking.py", line 121, in
>>>>> __init__
>>>>> self.pid = os.fork()
>>>>>
>>>>> When I run a stupid playbook with the command: anssible-playbook -vvv
>>>>> 1.yml
>>>>>
>>>>> for the playbook:
>>>>> —
>>>>> - hosts: all
>>>>> become: yes
>>>>> gather_facts: true
>>>>>
>>>>> tasks:
>>>>> - name: “remove the file”
>>>>> file:
>>>>> path=/tmp/getspace.sh
>>>>> state=absent
>>>>>
>>>>> I added the “-vvv” when the previous run told me to. Any ideas what’s
>>>>> going on (besides the obvious “out of memory”) and how to fix this?
>>>>>
>>>>> Mike
>>>>>
>>>>
>>>>> --
>>>>> You received this message because you are subscribed to the Google Groups
>>>>> "Ansible Project" group.
>>>>> To unsubscribe from this group and stop receiving emails from it, send an
>>>>> email to [email protected] <>.
>>>>> To view this discussion on the web visit
>>>>> https://groups.google.com/d/msgid/ansible-project/83646132-5874-4829-9152-127bee997178n%40googlegroups.com
>>>>>
>>>>> <https://groups.google.com/d/msgid/ansible-project/83646132-5874-4829-9152-127bee997178n%40googlegroups.com?utm_medium=email&utm_source=footer>.
>>>>
>>>>
>>>> --
>>>> You received this message because you are subscribed to the Google Groups
>>>> "Ansible Project" group.
>>>> To unsubscribe from this group and stop receiving emails from it, send an
>>>> email to [email protected]
>>>> <applewebdata://7878B376-AE82-4780-A2EC-D7F82E325719>.
>>>
>>>> To view this discussion on the web visit
>>>> https://groups.google.com/d/msgid/ansible-project/bda55e7c-e0ac-4da2-ab19-5ce504218d9dn%40googlegroups.com
>>>>
>>>> <https://groups.google.com/d/msgid/ansible-project/bda55e7c-e0ac-4da2-ab19-5ce504218d9dn%40googlegroups.com?utm_medium=email&utm_source=footer>.
>>>
>>>
>>> --
>>> You received this message because you are subscribed to the Google Groups
>>> "Ansible Project" group.
>>> To unsubscribe from this group and stop receiving emails from it, send an
>>> email to [email protected]
>>> <mailto:[email protected]>.
>>> To view this discussion on the web visit
>>> https://groups.google.com/d/msgid/ansible-project/79eec9ab-8973-4be5-8a66-b122c96c6fa7n%40googlegroups.com
>>>
>>> <https://groups.google.com/d/msgid/ansible-project/79eec9ab-8973-4be5-8a66-b122c96c6fa7n%40googlegroups.com?utm_medium=email&utm_source=footer>.
>>
>
--
You received this message because you are subscribed to the Google Groups
"Ansible Project" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/ansible-project/738861CE-F899-46B2-9C9F-2532BA2D0DA6%40gmail.com.