Hi -

I am experiencing the following situation while evaluating airflow;
wondering if this rings any bells with anyone:


   1. *airflow 1.7.1.2* is running in a *docker* container on amazon *ECS*
   2. *dagrun_timeout* = timedelta(hours=24) for this DAG
   3. running redshift queries with the PostGresOperator works fine

*except when that query takes over two hours*, and then the query is *cut
off* *exactly at the two hour mark* with this message

[2016-09-02 00:32:14,307] {models.py:1286} ERROR - server closed the
connection unexpectedly
This probably means the server terminated abnormally
before or while processing the request.
Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/airflow/models.py", line
1245, in run
    result = task_copy.execute(context=context)
  File
"/usr/local/lib/python2.7/dist-packages/airflow/operators/postgres_operator.py",
line 39, in execute
    self.hook.run(self.sql, self.autocommit, parameters=self.parameters)
  File
"/usr/local/lib/python2.7/dist-packages/airflow/hooks/dbapi_hook.py", line
132, in run
    cur.execute(s)
OperationalError: server closed the connection unexpectedly
This probably means the server terminated abnormally
before or while processing the request.



We are 100% sure that this account is able to run long queries in redshift
because we run these same queries from outside airflow from a regular AWS
EC2 instance (not a docker container) without issue.

The *exactly *two hours says to me that there is some setting somewhere -
either in airflow itself or in the postgresoperator or in the docker TCP
settings (or somewhere else) - that is going awry, but I cannot sort out
what it could be.  It doesn't seem to be an issue with celery, because
other jobs are running that polling for 6+ hours.

If this rings any bells for anyone I would love to hear about it.

Thanks for reading.

Reply via email to