[
https://issues.apache.org/jira/browse/AIRFLOW-1987?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Ash Berlin-Taylor closed AIRFLOW-1987.
--------------------------------------
Resolution: Cannot Reproduce
Fix Version/s: (was: 1.10.4)
This appears to have been fixed between 1.9 and 1.10
> Illegal query generation with MS SQL Server as backend
> ------------------------------------------------------
>
> Key: AIRFLOW-1987
> URL: https://issues.apache.org/jira/browse/AIRFLOW-1987
> Project: Apache Airflow
> Issue Type: Bug
> Components: scheduler, webserver
> Affects Versions: 1.9.0
> Environment: Linux, MS Sql Server 2017
> Reporter: Konstantin Privezentsev
> Priority: Minor
>
> Sheduler and webserver is failed on startup with error like this:
> {code:none}
> [2018-01-09 14:56:11,965] {{models.py:189}} INFO - Filling up the DagBag
> from /mnt/data01/
> [2018-01-09 14:56:11,971] {{jobs.py:1551}} INFO - Exited execute loop
> [2018-01-09 14:56:11,980] {{jobs.py:1565}} INFO - Terminating child PID: 53
> [2018-01-09 14:56:11,981] {{jobs.py:1569}} INFO - Waiting up to 5 seconds
> for processes to exit...
> [2018-01-09 14:56:11,981] {{jobs.py:384}} ERROR - Got an exception!
> Propagating...
> Traceback (most recent call last):
> File "/usr/local/lib/python2.7/dist-packages/airflow/jobs.py", line 376,
> in helper
> pickle_dags)
> File "/usr/local/lib/python2.7/dist-packages/airflow/utils/db.py", line
> 50, in wrapper
> result = func(*args, **kwargs)
> File "/usr/local/lib/python2.7/dist-packages/airflow/jobs.py", line 1752,
> in process_file
> dagbag = models.DagBag(file_path)
> File "/usr/local/lib/python2.7/dist-packages/airflow/models.py", line 202,
> in __init__
> self.collect_dags(dag_folder)
> File "/usr/local/lib/python2.7/dist-packages/airflow/models.py", line 409,
> in collect_dags
> self.process_file(dag_folder, only_if_updated=only_if_updated)
> File "/usr/local/lib/python2.7/dist-packages/airflow/models.py", line 285,
> in process_file
> m = imp.load_source(mod_name, filepath)
> File "/mnt/data01/airflow-dags/his/udns_aggr.py", line 57, in <module>
> image='his-sdc-dev-01.avp.ru/his/udns-aggregator:prod')
> File "/opt/airflow/plugins/his.py", line 99, in __init__
> super(SparkOperator, self).__init__(*args, **kwargs)
> File "/opt/airflow/plugins/his.py", line 66, in __init__
> super(DockerOperator, self).__init__(*args, **kwargs)
> File "/usr/local/lib/python2.7/dist-packages/airflow/utils/decorators.py",
> line 86, in wrapper
> result = func(*args, **kwargs)
> File
> "/usr/local/lib/python2.7/dist-packages/airflow/operators/docker_operator.py",
> line 121,
> super(DockerOperator, self).__init__(*args, **kwargs)
> File "/usr/local/lib/python2.7/dist-packages/airflow/utils/decorators.py",
> line 86, in wrapper
> result = func(*args, **kwargs)
> File "/usr/local/lib/python2.7/dist-packages/airflow/models.py", line
> 2219, in __init__
> self.resources = Resources(**(resources or {}))
> File
> "/usr/local/lib/python2.7/dist-packages/airflow/utils/operator_resources.py",
> line 104, in
> cpus = configuration.getint('operators', 'default_cpus')
> File "/usr/local/lib/python2.7/dist-packages/airflow/configuration.py",
> line 420, in getint
> return conf.getint(section, key)
> File "/usr/local/lib/python2.7/dist-packages/airflow/configuration.py",
> line 227, in getint
> return int(self.get(section, key))
> File "/usr/local/lib/python2.7/dist-packages/airflow/configuration.py",
> line 186, in get
> section = str(section).lower()
> File "/usr/local/lib/python2.7/dist-packages/future/types/newstr.py", line
> 102, in __new__
> return super(newstr, cls).__new__(cls, value)
> File "/usr/local/lib/python2.7/dist-packages/airflow/bin/cli.py", line 73,
> in sigint_handler
> sys.exit(0)
> SystemExit: 0
> Traceback (most recent call last):
> File "/usr/local/bin/airflow", line 27, in <module>
> args.func(args)
> File "/usr/local/lib/python2.7/dist-packages/airflow/bin/cli.py", line
> 826, in scheduler
> job.run()
> File "/usr/local/lib/python2.7/dist-packages/airflow/jobs.py", line 203,
> in run
> self._execute()
> File "/usr/local/lib/python2.7/dist-packages/airflow/jobs.py", line 1549,
> in _execute
> self._execute_helper(processor_manager)
> File "/usr/local/lib/python2.7/dist-packages/airflow/jobs.py", line 1664,
> in _execute_helper
> (State.SCHEDULED,))
> File "/usr/local/lib/python2.7/dist-packages/airflow/utils/db.py", line
> 50, in wrapper
> result = func(*args, **kwargs)
> File "/usr/local/lib/python2.7/dist-packages/airflow/jobs.py", line 1337,
> in
> session=session)
> File "/usr/local/lib/python2.7/dist-packages/airflow/utils/db.py", line
> 50, in wrapper
> result = func(*args, **kwargs)
> File "/usr/local/lib/python2.7/dist-packages/airflow/jobs.py", line 1079,
> in
> task_instances_to_examine = ti_query.all()
> File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/orm/query.py",
> line 2726, in all
> return list(self)
> File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/orm/query.py",
> line 2878, in __iter__
> return self._execute_and_instances(context)
> File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/orm/query.py",
> line 2901, in
> result = conn.execute(querycontext.statement, self._params)
> File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py",
> line 948, in execute
> return meth(self, multiparams, params)
> File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/sql/elements.py",
> line 269, in
> return connection._execute_clauseelement(self, multiparams, params)
> File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py",
> line 1060, in
> compiled_sql, distilled_params
> File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py",
> line 1200, in
> context)
> File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py",
> line 1413, in
> exc_info
> File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/util/compat.py",
> line 203, in
> reraise(type(exception), exception, tb=exc_tb, cause=cause)
> File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py",
> line 1193, in
> context)
> File
> "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/default.py", line
> 507, in
> cursor.execute(statement, parameters)
> sqlalchemy.exc.ProgrammingError: (pyodbc.ProgrammingError) ('42000',
> "[42000] [Microsoft][ODBC ecified in a context where a condition is expected,
> near ')'. (4145) (SQLExecDirectW)") [SQL: u'SELECT task_instance.try_number
> AS task_instance_try_number, task_instance.task_id AS task_instance_task_id,
> task_instance.dag_id AS task_instance_dag_id, task_instance.execution_date AS
> task_instance_execution_date, task_instance.start_date AS
> task_instance_start_date, task_instance.end_date AS task_instance_end_date,
> task_instance.duration AS task_instance_duration, task_instance.state AS
> task_instance_state, task_instance.max_tries AS task_instance_max_tries,
> task_instance.hostname AS task_instance_hostname, task_instance.unixname AS
> task_instance_unixname, task_instance.job_id AS task_instance_job_id,
> task_instance.pool AS task_instance_pool, task_instance.queue AS
> task_instance_queue, task_instance.priority_weight AS
> task_instance_priority_weight, task_instance.operator AS
> task_instance_operator, task_instance.queued_dttm AS
> task_instance_queued_dttm, task_instance.pid AS task_instance_pid \nFROM
> task_instance LEFT OUTER JOIN dag_run ON dag_run.dag_id =
> task_instance.dag_id AND dag_run.execution_date =
> task_instance.execution_date LEFT OUTER JOIN dag ON dag.dag_id =
> task_instance.dag_id \nWHERE task_instance.dag_id IN (?) AND (dag_run.run_id
> IS NULL OR dag_run.run_id NOT LIKE ?) AND (dag.dag_id IS NULL OR NOT
> dag.is_paused) AND task_instance.state IN (?)'] [parameters:
> ('udns_kudu_partition_manager', u'backfill_%', u'scheduled')] (Background on
> this error at: http://sqlalche.me/e/f405)
> {code}
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)