Olivier Girardot created AIRFLOW-699:
----------------------------------------

             Summary: Dag can't be triggered at the same second due to 
constraint on dag_id + execution_date
                 Key: AIRFLOW-699
                 URL: https://issues.apache.org/jira/browse/AIRFLOW-699
             Project: Apache Airflow
          Issue Type: Bug
          Components: db
    Affects Versions: Airflow 1.7.1.3
            Reporter: Olivier Girardot


We have a system that triggers Dags when several files arrive in HDFS, we have 
crafted a correct run_id to trace the trigger but since the schema of dag_run 
table is : 

{code:sql}
CREATE TABLE `dag_run` (
  `id` int(11) NOT NULL AUTO_INCREMENT,
  `dag_id` varchar(250) DEFAULT NULL,
  `execution_date` datetime DEFAULT NULL,
  `state` varchar(50) DEFAULT NULL,
  `run_id` varchar(250) DEFAULT NULL,
  `external_trigger` tinyint(1) DEFAULT NULL,
  `conf` blob,
  `end_date` datetime DEFAULT NULL,
  `start_date` datetime DEFAULT NULL,
  PRIMARY KEY (`id`),
  UNIQUE KEY `dag_id` (`dag_id`,`execution_date`),
  UNIQUE KEY `dag_id_2` (`dag_id`,`run_id`)
) ENGINE=InnoDB AUTO_INCREMENT=2998 DEFAULT CHARSET=latin1 {code}

We end up with DuplicateEntry exception
{noformat}
sqlalchemy.exc.IntegrityError: (_mysql_exceptions.IntegrityError) (1062, 
"Duplicate entry 'my-job-2016-12-13 19:52:33' for key 'dag_id'") [SQL: u'INSERT 
INTO dag_run (dag_id, execution_date, start_date, end_date, state, run_id, 
external_trigger, conf) VALUES (%s, %s, now(), %s, %s, %s, %s, %s)'] 
[parameters: ('my-job', datetime.datetime(2016, 12, 13, 19, 52, 33, 210790), 
None, u'running', 
'my-job-custom-run-id_2016-12-13T19:52:32.291_785a2860-c622-47f6-a29c-4c6394f931fa',
 1, "...")
{noformat}

Is there any need for this constraint ? The "datetime" precision is problematic 
for us because it's usual that some dags get triggered at the same second.




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to