How do I chain multiple Google Cloud DataPrep flows?











up vote
1
down vote

favorite












I've created two Flows in Cloud DataPrep - the first outputs to a BigQuery table and also creates a reference dataset. The second flow takes the reference dataset and processes it further before outputting to a second BigQuery table.



Is it possible to schedule these two Flows to run sequentially? At the moment I have to estimate the time taken for the first flow, and schedule the second one to run XX minutes after the first.



Either a recipe in the first flow to trigger the second, or a way to schedule them in sequence would be ideal.



This question shows how to create a reference dataset, but doesn't explain if it's possible to run them automatically/sequentially.










share|improve this question


























    up vote
    1
    down vote

    favorite












    I've created two Flows in Cloud DataPrep - the first outputs to a BigQuery table and also creates a reference dataset. The second flow takes the reference dataset and processes it further before outputting to a second BigQuery table.



    Is it possible to schedule these two Flows to run sequentially? At the moment I have to estimate the time taken for the first flow, and schedule the second one to run XX minutes after the first.



    Either a recipe in the first flow to trigger the second, or a way to schedule them in sequence would be ideal.



    This question shows how to create a reference dataset, but doesn't explain if it's possible to run them automatically/sequentially.










    share|improve this question
























      up vote
      1
      down vote

      favorite









      up vote
      1
      down vote

      favorite











      I've created two Flows in Cloud DataPrep - the first outputs to a BigQuery table and also creates a reference dataset. The second flow takes the reference dataset and processes it further before outputting to a second BigQuery table.



      Is it possible to schedule these two Flows to run sequentially? At the moment I have to estimate the time taken for the first flow, and schedule the second one to run XX minutes after the first.



      Either a recipe in the first flow to trigger the second, or a way to schedule them in sequence would be ideal.



      This question shows how to create a reference dataset, but doesn't explain if it's possible to run them automatically/sequentially.










      share|improve this question













      I've created two Flows in Cloud DataPrep - the first outputs to a BigQuery table and also creates a reference dataset. The second flow takes the reference dataset and processes it further before outputting to a second BigQuery table.



      Is it possible to schedule these two Flows to run sequentially? At the moment I have to estimate the time taken for the first flow, and schedule the second one to run XX minutes after the first.



      Either a recipe in the first flow to trigger the second, or a way to schedule them in sequence would be ideal.



      This question shows how to create a reference dataset, but doesn't explain if it's possible to run them automatically/sequentially.







      google-cloud-dataprep






      share|improve this question













      share|improve this question











      share|improve this question




      share|improve this question










      asked Nov 9 at 10:53









      angusham

      234




      234
























          1 Answer
          1






          active

          oldest

          votes

















          up vote
          1
          down vote



          accepted










          The documentation for reference datasets sort of alludes to the behaviour, but could be clearer.




          • The second flow that uses the first flow as reference dataset WILL RUN the first flow job and therefore use updated data for flow 2.


          • BUT the export to BigQuery would not perform the export from flow 1.


          The easiest solution, but not necessarily what you want is that can choose to perform the BigQuery export from Flow 1 in Flow 2, i.e. have a simple recipe that does nothing but host the export job.



          The solution I use, where recipes/jobs are stable, is to run the resultant Dataflow jobs rather than use the Dataprep tool itself for execution: Run Job on Cloud Dataflow



          You have a few options for scheduling the two Dataflow jobs. (Cloud Scheduler could be a good one - its a new project and I'm looking at replacing custom solution I have right now hosted as Cloud Functions)



          In terms of running them in sequence, however, BigQuery doesn't have update event that could trigger Job2 so you can either schedule them long enough apart and hope Job 1 finishes soon enough, OR you can poll the BigQuery table metadata to see if the modified date changes.






          share|improve this answer





















            Your Answer






            StackExchange.ifUsing("editor", function () {
            StackExchange.using("externalEditor", function () {
            StackExchange.using("snippets", function () {
            StackExchange.snippets.init();
            });
            });
            }, "code-snippets");

            StackExchange.ready(function() {
            var channelOptions = {
            tags: "".split(" "),
            id: "1"
            };
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function() {
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled) {
            StackExchange.using("snippets", function() {
            createEditor();
            });
            }
            else {
            createEditor();
            }
            });

            function createEditor() {
            StackExchange.prepareEditor({
            heartbeatType: 'answer',
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader: {
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            },
            onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            });


            }
            });














             

            draft saved


            draft discarded


















            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53224329%2fhow-do-i-chain-multiple-google-cloud-dataprep-flows%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown

























            1 Answer
            1






            active

            oldest

            votes








            1 Answer
            1






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes








            up vote
            1
            down vote



            accepted










            The documentation for reference datasets sort of alludes to the behaviour, but could be clearer.




            • The second flow that uses the first flow as reference dataset WILL RUN the first flow job and therefore use updated data for flow 2.


            • BUT the export to BigQuery would not perform the export from flow 1.


            The easiest solution, but not necessarily what you want is that can choose to perform the BigQuery export from Flow 1 in Flow 2, i.e. have a simple recipe that does nothing but host the export job.



            The solution I use, where recipes/jobs are stable, is to run the resultant Dataflow jobs rather than use the Dataprep tool itself for execution: Run Job on Cloud Dataflow



            You have a few options for scheduling the two Dataflow jobs. (Cloud Scheduler could be a good one - its a new project and I'm looking at replacing custom solution I have right now hosted as Cloud Functions)



            In terms of running them in sequence, however, BigQuery doesn't have update event that could trigger Job2 so you can either schedule them long enough apart and hope Job 1 finishes soon enough, OR you can poll the BigQuery table metadata to see if the modified date changes.






            share|improve this answer

























              up vote
              1
              down vote



              accepted










              The documentation for reference datasets sort of alludes to the behaviour, but could be clearer.




              • The second flow that uses the first flow as reference dataset WILL RUN the first flow job and therefore use updated data for flow 2.


              • BUT the export to BigQuery would not perform the export from flow 1.


              The easiest solution, but not necessarily what you want is that can choose to perform the BigQuery export from Flow 1 in Flow 2, i.e. have a simple recipe that does nothing but host the export job.



              The solution I use, where recipes/jobs are stable, is to run the resultant Dataflow jobs rather than use the Dataprep tool itself for execution: Run Job on Cloud Dataflow



              You have a few options for scheduling the two Dataflow jobs. (Cloud Scheduler could be a good one - its a new project and I'm looking at replacing custom solution I have right now hosted as Cloud Functions)



              In terms of running them in sequence, however, BigQuery doesn't have update event that could trigger Job2 so you can either schedule them long enough apart and hope Job 1 finishes soon enough, OR you can poll the BigQuery table metadata to see if the modified date changes.






              share|improve this answer























                up vote
                1
                down vote



                accepted







                up vote
                1
                down vote



                accepted






                The documentation for reference datasets sort of alludes to the behaviour, but could be clearer.




                • The second flow that uses the first flow as reference dataset WILL RUN the first flow job and therefore use updated data for flow 2.


                • BUT the export to BigQuery would not perform the export from flow 1.


                The easiest solution, but not necessarily what you want is that can choose to perform the BigQuery export from Flow 1 in Flow 2, i.e. have a simple recipe that does nothing but host the export job.



                The solution I use, where recipes/jobs are stable, is to run the resultant Dataflow jobs rather than use the Dataprep tool itself for execution: Run Job on Cloud Dataflow



                You have a few options for scheduling the two Dataflow jobs. (Cloud Scheduler could be a good one - its a new project and I'm looking at replacing custom solution I have right now hosted as Cloud Functions)



                In terms of running them in sequence, however, BigQuery doesn't have update event that could trigger Job2 so you can either schedule them long enough apart and hope Job 1 finishes soon enough, OR you can poll the BigQuery table metadata to see if the modified date changes.






                share|improve this answer












                The documentation for reference datasets sort of alludes to the behaviour, but could be clearer.




                • The second flow that uses the first flow as reference dataset WILL RUN the first flow job and therefore use updated data for flow 2.


                • BUT the export to BigQuery would not perform the export from flow 1.


                The easiest solution, but not necessarily what you want is that can choose to perform the BigQuery export from Flow 1 in Flow 2, i.e. have a simple recipe that does nothing but host the export job.



                The solution I use, where recipes/jobs are stable, is to run the resultant Dataflow jobs rather than use the Dataprep tool itself for execution: Run Job on Cloud Dataflow



                You have a few options for scheduling the two Dataflow jobs. (Cloud Scheduler could be a good one - its a new project and I'm looking at replacing custom solution I have right now hosted as Cloud Functions)



                In terms of running them in sequence, however, BigQuery doesn't have update event that could trigger Job2 so you can either schedule them long enough apart and hope Job 1 finishes soon enough, OR you can poll the BigQuery table metadata to see if the modified date changes.







                share|improve this answer












                share|improve this answer



                share|improve this answer










                answered Nov 16 at 14:19









                Jonathon

                2,97031940




                2,97031940






























                     

                    draft saved


                    draft discarded



















































                     


                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function () {
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53224329%2fhow-do-i-chain-multiple-google-cloud-dataprep-flows%23new-answer', 'question_page');
                    }
                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    Schultheiß

                    Verwaltungsgliederung Dänemarks

                    Liste der Kulturdenkmale in Wilsdruff