Spark job failed to write to Alluxio due to DeadlineExceededException










1















I am running a Spark job writing to an Alluxio cluster with 20 workers (Alluxio 1.6.1). Spark job failed to write its output due to alluxio.exception.status.DeadlineExceededException. The worker is still alive from Alluxio WebUI. How can I avoid this failure?



alluxio.exception.status.DeadlineExceededException: Timeout writing to WorkerNetAddresshost=spark-74-44.xxxx, rpcPort=51998, dataPort=51999, webPort=51997, domainSocketPath= for request type: ALLUXIO_BLOCK
id: 3209355843338240
tier: 0
worker_group
host: "spark6-64-156.xxxx"
rpc_port: 51998
data_port: 51999
web_port: 51997
socket_path: ""










share|improve this question


























    1















    I am running a Spark job writing to an Alluxio cluster with 20 workers (Alluxio 1.6.1). Spark job failed to write its output due to alluxio.exception.status.DeadlineExceededException. The worker is still alive from Alluxio WebUI. How can I avoid this failure?



    alluxio.exception.status.DeadlineExceededException: Timeout writing to WorkerNetAddresshost=spark-74-44.xxxx, rpcPort=51998, dataPort=51999, webPort=51997, domainSocketPath= for request type: ALLUXIO_BLOCK
    id: 3209355843338240
    tier: 0
    worker_group
    host: "spark6-64-156.xxxx"
    rpc_port: 51998
    data_port: 51999
    web_port: 51997
    socket_path: ""










    share|improve this question
























      1












      1








      1








      I am running a Spark job writing to an Alluxio cluster with 20 workers (Alluxio 1.6.1). Spark job failed to write its output due to alluxio.exception.status.DeadlineExceededException. The worker is still alive from Alluxio WebUI. How can I avoid this failure?



      alluxio.exception.status.DeadlineExceededException: Timeout writing to WorkerNetAddresshost=spark-74-44.xxxx, rpcPort=51998, dataPort=51999, webPort=51997, domainSocketPath= for request type: ALLUXIO_BLOCK
      id: 3209355843338240
      tier: 0
      worker_group
      host: "spark6-64-156.xxxx"
      rpc_port: 51998
      data_port: 51999
      web_port: 51997
      socket_path: ""










      share|improve this question














      I am running a Spark job writing to an Alluxio cluster with 20 workers (Alluxio 1.6.1). Spark job failed to write its output due to alluxio.exception.status.DeadlineExceededException. The worker is still alive from Alluxio WebUI. How can I avoid this failure?



      alluxio.exception.status.DeadlineExceededException: Timeout writing to WorkerNetAddresshost=spark-74-44.xxxx, rpcPort=51998, dataPort=51999, webPort=51997, domainSocketPath= for request type: ALLUXIO_BLOCK
      id: 3209355843338240
      tier: 0
      worker_group
      host: "spark6-64-156.xxxx"
      rpc_port: 51998
      data_port: 51999
      web_port: 51997
      socket_path: ""







      apache-spark alluxio






      share|improve this question













      share|improve this question











      share|improve this question




      share|improve this question










      asked Nov 15 '18 at 18:20









      apc999apc999

      1115




      1115






















          1 Answer
          1






          active

          oldest

          votes


















          1














          This error indicates that your Spark job timed out while trying to write data to an Alluxio worker. The worker could be under high load, or have a slow connection to your UFS.



          The default timeout is 30 seconds. To increase the timeout, configure alluxio.user.network.netty.timeout on the Spark side.



          For example, to increase the timeout to 5 minutes, use the --conf option to spark-submit



          $ spark-submit --conf 'spark.executor.extraJavaOptions=-Dalluxio.user.network.netty.timeout=5min' 
          --conf 'spark.driver.extraJavaOptions=-Dalluxio.user.network.netty.timeout=5min'
          ...


          You can also set these properties in your spark-defaults.conf file to have them automatically applied to all jobs.



          Source: https://www.alluxio.org/docs/1.6/en/Configuration-Settings.html#spark-jobs






          share|improve this answer






















            Your Answer






            StackExchange.ifUsing("editor", function ()
            StackExchange.using("externalEditor", function ()
            StackExchange.using("snippets", function ()
            StackExchange.snippets.init();
            );
            );
            , "code-snippets");

            StackExchange.ready(function()
            var channelOptions =
            tags: "".split(" "),
            id: "1"
            ;
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function()
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled)
            StackExchange.using("snippets", function()
            createEditor();
            );

            else
            createEditor();

            );

            function createEditor()
            StackExchange.prepareEditor(
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader:
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            ,
            onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            );



            );













            draft saved

            draft discarded


















            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53325658%2fspark-job-failed-to-write-to-alluxio-due-to-deadlineexceededexception%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown

























            1 Answer
            1






            active

            oldest

            votes








            1 Answer
            1






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            1














            This error indicates that your Spark job timed out while trying to write data to an Alluxio worker. The worker could be under high load, or have a slow connection to your UFS.



            The default timeout is 30 seconds. To increase the timeout, configure alluxio.user.network.netty.timeout on the Spark side.



            For example, to increase the timeout to 5 minutes, use the --conf option to spark-submit



            $ spark-submit --conf 'spark.executor.extraJavaOptions=-Dalluxio.user.network.netty.timeout=5min' 
            --conf 'spark.driver.extraJavaOptions=-Dalluxio.user.network.netty.timeout=5min'
            ...


            You can also set these properties in your spark-defaults.conf file to have them automatically applied to all jobs.



            Source: https://www.alluxio.org/docs/1.6/en/Configuration-Settings.html#spark-jobs






            share|improve this answer



























              1














              This error indicates that your Spark job timed out while trying to write data to an Alluxio worker. The worker could be under high load, or have a slow connection to your UFS.



              The default timeout is 30 seconds. To increase the timeout, configure alluxio.user.network.netty.timeout on the Spark side.



              For example, to increase the timeout to 5 minutes, use the --conf option to spark-submit



              $ spark-submit --conf 'spark.executor.extraJavaOptions=-Dalluxio.user.network.netty.timeout=5min' 
              --conf 'spark.driver.extraJavaOptions=-Dalluxio.user.network.netty.timeout=5min'
              ...


              You can also set these properties in your spark-defaults.conf file to have them automatically applied to all jobs.



              Source: https://www.alluxio.org/docs/1.6/en/Configuration-Settings.html#spark-jobs






              share|improve this answer

























                1












                1








                1







                This error indicates that your Spark job timed out while trying to write data to an Alluxio worker. The worker could be under high load, or have a slow connection to your UFS.



                The default timeout is 30 seconds. To increase the timeout, configure alluxio.user.network.netty.timeout on the Spark side.



                For example, to increase the timeout to 5 minutes, use the --conf option to spark-submit



                $ spark-submit --conf 'spark.executor.extraJavaOptions=-Dalluxio.user.network.netty.timeout=5min' 
                --conf 'spark.driver.extraJavaOptions=-Dalluxio.user.network.netty.timeout=5min'
                ...


                You can also set these properties in your spark-defaults.conf file to have them automatically applied to all jobs.



                Source: https://www.alluxio.org/docs/1.6/en/Configuration-Settings.html#spark-jobs






                share|improve this answer













                This error indicates that your Spark job timed out while trying to write data to an Alluxio worker. The worker could be under high load, or have a slow connection to your UFS.



                The default timeout is 30 seconds. To increase the timeout, configure alluxio.user.network.netty.timeout on the Spark side.



                For example, to increase the timeout to 5 minutes, use the --conf option to spark-submit



                $ spark-submit --conf 'spark.executor.extraJavaOptions=-Dalluxio.user.network.netty.timeout=5min' 
                --conf 'spark.driver.extraJavaOptions=-Dalluxio.user.network.netty.timeout=5min'
                ...


                You can also set these properties in your spark-defaults.conf file to have them automatically applied to all jobs.



                Source: https://www.alluxio.org/docs/1.6/en/Configuration-Settings.html#spark-jobs







                share|improve this answer












                share|improve this answer



                share|improve this answer










                answered Nov 15 '18 at 19:06









                AAudibertAAudibert

                161213




                161213





























                    draft saved

                    draft discarded
















































                    Thanks for contributing an answer to Stack Overflow!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid


                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.

                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function ()
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53325658%2fspark-job-failed-to-write-to-alluxio-due-to-deadlineexceededexception%23new-answer', 'question_page');

                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    Top Tejano songwriter Luis Silva dead of heart attack at 64

                    ReactJS Fetched API data displays live - need Data displayed static

                    政党