I thought that we were under the impression that rather than losing data
it's likely better to update your pipeline to handle these?

On Wed, Apr 13, 2016 at 10:59 AM, Luke Cwik (JIRA) <j...@apache.org> wrote:

>
>     [
> https://issues.apache.org/jira/browse/BEAM-190?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15239701#comment-15239701
> ]
>
> Luke Cwik commented on BEAM-190:
> --------------------------------
>
> I believe this can easily extend beyond BigQuery to having a dead letter
> feature for failing DoFns of any kind.
>
> > Dead-letter drop for bad BigQuery records
> > -----------------------------------------
> >
> >                 Key: BEAM-190
> >                 URL: https://issues.apache.org/jira/browse/BEAM-190
> >             Project: Beam
> >          Issue Type: Bug
> >          Components: runner-core
> >            Reporter: Mark Shields
> >            Assignee: Frances Perry
> >
> > If a BigQuery insert fails for data-specific rather than structural
> reasons (eg cannot parse a date) then the bundle will be retried
> indefinitely, first by BigQueryTableInserter.insertAll then by the overall
> production retry logic of the underlying runner.
> > Better would be to allow customer to specify a dead-letter store for
> records such as those so that overall processing can continue while bad
> records are quarantined.
>
>
>
> --
> This message was sent by Atlassian JIRA
> (v6.3.4#6332)
>

Reply via email to