Hyperparameter Optimization Using Gaussian Processes












5














I have a dataset that is divided into training and validation dataset. I am using Gaussian Processes to perform hyperparameter optimization. So I am using the accuracy recorded on the validation dataset to tune the hyperparameters of the DNN model. Is that considered cheating? Will the last reported results be considered credible?



Any help is much appreciated!!










share|cite|improve this question



























    5














    I have a dataset that is divided into training and validation dataset. I am using Gaussian Processes to perform hyperparameter optimization. So I am using the accuracy recorded on the validation dataset to tune the hyperparameters of the DNN model. Is that considered cheating? Will the last reported results be considered credible?



    Any help is much appreciated!!










    share|cite|improve this question

























      5












      5








      5


      0





      I have a dataset that is divided into training and validation dataset. I am using Gaussian Processes to perform hyperparameter optimization. So I am using the accuracy recorded on the validation dataset to tune the hyperparameters of the DNN model. Is that considered cheating? Will the last reported results be considered credible?



      Any help is much appreciated!!










      share|cite|improve this question













      I have a dataset that is divided into training and validation dataset. I am using Gaussian Processes to perform hyperparameter optimization. So I am using the accuracy recorded on the validation dataset to tune the hyperparameters of the DNN model. Is that considered cheating? Will the last reported results be considered credible?



      Any help is much appreciated!!







      neural-networks gaussian-process hyperparameter






      share|cite|improve this question













      share|cite|improve this question











      share|cite|improve this question




      share|cite|improve this question










      asked Dec 18 '18 at 17:05









      I. A

      1385




      1385






















          1 Answer
          1






          active

          oldest

          votes


















          13














          As a result of doing that you will also overfit the validation set (the more so the more you tuned the hyperparameters - if you tried two or three configurations, the effect is less than if you did some systematic search e.g. using the Gaussian process approach). The standard solution to this would be to not just have a training and validation set, but a third set (a test set). You would only ever look at the test set once with you very final model after hyperparameter tuning.






          share|cite|improve this answer

















          • 3




            Depending on computational limitations, it may also be possible to evaluate a hyperparameter configuration via cross validation on the training set.
            – John Madden
            Dec 18 '18 at 21:58






          • 1




            I've always referred to a non-final test set as a "test set" and validation to be only the final set to test to validate the entire system. Is there a standard on this that I've been ignoring or is it person to person?
            – Poik
            Dec 19 '18 at 16:20










          • @JohnMadden Also depends on data limitations. We have very little to hold out for proper validation for my task, sadly. Cross validation helps in this scenario.
            – Poik
            Dec 19 '18 at 16:23













          Your Answer





          StackExchange.ifUsing("editor", function () {
          return StackExchange.using("mathjaxEditing", function () {
          StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
          StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
          });
          });
          }, "mathjax-editing");

          StackExchange.ready(function() {
          var channelOptions = {
          tags: "".split(" "),
          id: "65"
          };
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function() {
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled) {
          StackExchange.using("snippets", function() {
          createEditor();
          });
          }
          else {
          createEditor();
          }
          });

          function createEditor() {
          StackExchange.prepareEditor({
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: false,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: null,
          bindNavPrevention: true,
          postfix: "",
          imageUploader: {
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          },
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          });


          }
          });














          draft saved

          draft discarded


















          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f383621%2fhyperparameter-optimization-using-gaussian-processes%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown

























          1 Answer
          1






          active

          oldest

          votes








          1 Answer
          1






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes









          13














          As a result of doing that you will also overfit the validation set (the more so the more you tuned the hyperparameters - if you tried two or three configurations, the effect is less than if you did some systematic search e.g. using the Gaussian process approach). The standard solution to this would be to not just have a training and validation set, but a third set (a test set). You would only ever look at the test set once with you very final model after hyperparameter tuning.






          share|cite|improve this answer

















          • 3




            Depending on computational limitations, it may also be possible to evaluate a hyperparameter configuration via cross validation on the training set.
            – John Madden
            Dec 18 '18 at 21:58






          • 1




            I've always referred to a non-final test set as a "test set" and validation to be only the final set to test to validate the entire system. Is there a standard on this that I've been ignoring or is it person to person?
            – Poik
            Dec 19 '18 at 16:20










          • @JohnMadden Also depends on data limitations. We have very little to hold out for proper validation for my task, sadly. Cross validation helps in this scenario.
            – Poik
            Dec 19 '18 at 16:23


















          13














          As a result of doing that you will also overfit the validation set (the more so the more you tuned the hyperparameters - if you tried two or three configurations, the effect is less than if you did some systematic search e.g. using the Gaussian process approach). The standard solution to this would be to not just have a training and validation set, but a third set (a test set). You would only ever look at the test set once with you very final model after hyperparameter tuning.






          share|cite|improve this answer

















          • 3




            Depending on computational limitations, it may also be possible to evaluate a hyperparameter configuration via cross validation on the training set.
            – John Madden
            Dec 18 '18 at 21:58






          • 1




            I've always referred to a non-final test set as a "test set" and validation to be only the final set to test to validate the entire system. Is there a standard on this that I've been ignoring or is it person to person?
            – Poik
            Dec 19 '18 at 16:20










          • @JohnMadden Also depends on data limitations. We have very little to hold out for proper validation for my task, sadly. Cross validation helps in this scenario.
            – Poik
            Dec 19 '18 at 16:23
















          13












          13








          13






          As a result of doing that you will also overfit the validation set (the more so the more you tuned the hyperparameters - if you tried two or three configurations, the effect is less than if you did some systematic search e.g. using the Gaussian process approach). The standard solution to this would be to not just have a training and validation set, but a third set (a test set). You would only ever look at the test set once with you very final model after hyperparameter tuning.






          share|cite|improve this answer












          As a result of doing that you will also overfit the validation set (the more so the more you tuned the hyperparameters - if you tried two or three configurations, the effect is less than if you did some systematic search e.g. using the Gaussian process approach). The standard solution to this would be to not just have a training and validation set, but a third set (a test set). You would only ever look at the test set once with you very final model after hyperparameter tuning.







          share|cite|improve this answer












          share|cite|improve this answer



          share|cite|improve this answer










          answered Dec 18 '18 at 17:22









          Björn

          9,7951937




          9,7951937








          • 3




            Depending on computational limitations, it may also be possible to evaluate a hyperparameter configuration via cross validation on the training set.
            – John Madden
            Dec 18 '18 at 21:58






          • 1




            I've always referred to a non-final test set as a "test set" and validation to be only the final set to test to validate the entire system. Is there a standard on this that I've been ignoring or is it person to person?
            – Poik
            Dec 19 '18 at 16:20










          • @JohnMadden Also depends on data limitations. We have very little to hold out for proper validation for my task, sadly. Cross validation helps in this scenario.
            – Poik
            Dec 19 '18 at 16:23
















          • 3




            Depending on computational limitations, it may also be possible to evaluate a hyperparameter configuration via cross validation on the training set.
            – John Madden
            Dec 18 '18 at 21:58






          • 1




            I've always referred to a non-final test set as a "test set" and validation to be only the final set to test to validate the entire system. Is there a standard on this that I've been ignoring or is it person to person?
            – Poik
            Dec 19 '18 at 16:20










          • @JohnMadden Also depends on data limitations. We have very little to hold out for proper validation for my task, sadly. Cross validation helps in this scenario.
            – Poik
            Dec 19 '18 at 16:23










          3




          3




          Depending on computational limitations, it may also be possible to evaluate a hyperparameter configuration via cross validation on the training set.
          – John Madden
          Dec 18 '18 at 21:58




          Depending on computational limitations, it may also be possible to evaluate a hyperparameter configuration via cross validation on the training set.
          – John Madden
          Dec 18 '18 at 21:58




          1




          1




          I've always referred to a non-final test set as a "test set" and validation to be only the final set to test to validate the entire system. Is there a standard on this that I've been ignoring or is it person to person?
          – Poik
          Dec 19 '18 at 16:20




          I've always referred to a non-final test set as a "test set" and validation to be only the final set to test to validate the entire system. Is there a standard on this that I've been ignoring or is it person to person?
          – Poik
          Dec 19 '18 at 16:20












          @JohnMadden Also depends on data limitations. We have very little to hold out for proper validation for my task, sadly. Cross validation helps in this scenario.
          – Poik
          Dec 19 '18 at 16:23






          @JohnMadden Also depends on data limitations. We have very little to hold out for proper validation for my task, sadly. Cross validation helps in this scenario.
          – Poik
          Dec 19 '18 at 16:23




















          draft saved

          draft discarded




















































          Thanks for contributing an answer to Cross Validated!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid



          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.


          Use MathJax to format equations. MathJax reference.


          To learn more, see our tips on writing great answers.





          Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


          Please pay close attention to the following guidance:


          • Please be sure to answer the question. Provide details and share your research!

          But avoid



          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f383621%2fhyperparameter-optimization-using-gaussian-processes%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          Morgemoulin

          Scott Moir

          Souastre