[madlib] branch master updated: correct disk space comment for gp5 and 6 in keras multi fit user docs

2020-03-26 Thread fmcquillan
This is an automated email from the ASF dual-hosted git repository.

fmcquillan pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/madlib.git


The following commit(s) were added to refs/heads/master by this push:
 new 2f0bb2e  correct disk space comment for gp5 and 6 in keras multi fit 
user docs
2f0bb2e is described below

commit 2f0bb2e0b01e060150b443c43f00c5e1d664a5c6
Author: Frank McQuillan 
AuthorDate: Thu Mar 26 15:25:49 2020 -0700

correct disk space comment for gp5 and 6 in keras multi fit user docs
---
 .../modules/deep_learning/madlib_keras_fit_multiple_model.sql_in | 5 +++--
 1 file changed, 3 insertions(+), 2 deletions(-)

diff --git 
a/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
 
b/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
index 4d1eb09..9238652 100644
--- 
a/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
+++ 
b/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
@@ -88,10 +88,11 @@ You can set up the models and hyperparameters to try with 
the
 Model Selection utility to define the unique combinations
 of model architectures, compile and fit parameters.
 
-@note If 'madlib_keras_fit_multiple_model()' is running on GPDB 5, the 
database will
+@note If 'madlib_keras_fit_multiple_model()' is running on GPDB 5 and some 
versions
+of GPDB 6, the database will
 keep adding to the disk space (in proportion to model size) and will only
 release the disk space once the fit multiple query has completed execution.
-This is not the case for GPDB 6+ where disk space is released during the
+This is not the case for GPDB 6.5.0+ where disk space is released during the
 fit multiple query.
 
 @note CUDA GPU memory cannot be released until the process holding it is 
terminated.



[madlib] branch master updated: DL: Clean up user docs

2020-03-26 Thread okislal
This is an automated email from the ASF dual-hosted git repository.

okislal pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/madlib.git


The following commit(s) were added to refs/heads/master by this push:
 new 1a9b557  DL: Clean up user docs
1a9b557 is described below

commit 1a9b557ab3c221a2c3956a201cfe2ef060977f39
Author: Orhan Kislal 
AuthorDate: Thu Mar 26 16:53:33 2020 -0400

DL: Clean up user docs
---
 doc/mainpage.dox.in  |  1 -
 src/ports/postgres/modules/deep_learning/madlib_keras.sql_in | 12 ++--
 .../deep_learning/madlib_keras_fit_multiple_model.sql_in | 12 ++--
 3 files changed, 12 insertions(+), 13 deletions(-)

diff --git a/doc/mainpage.dox.in b/doc/mainpage.dox.in
index 82be4a5..189b948 100644
--- a/doc/mainpage.dox.in
+++ b/doc/mainpage.dox.in
@@ -296,7 +296,6 @@ Interface and implementation are subject to change.
 @brief Train multiple deep learning models at the same time for model 
architecture search and hyperparameter selection.
 @details Train multiple deep learning models at the same time for 
model architecture search and hyperparameter selection.
 @{
-@defgroup grp_automl AutoML
 @defgroup grp_keras_run_model_selection Run Model Selection
 @defgroup grp_keras_setup_model_selection Setup Model Selection
 @}
diff --git a/src/ports/postgres/modules/deep_learning/madlib_keras.sql_in 
b/src/ports/postgres/modules/deep_learning/madlib_keras.sql_in
index 75fa56a..b0c77a3 100644
--- a/src/ports/postgres/modules/deep_learning/madlib_keras.sql_in
+++ b/src/ports/postgres/modules/deep_learning/madlib_keras.sql_in
@@ -87,15 +87,15 @@ Note that the following MADlib functions are targeting a 
specific Keras
 version (2.2.4) with a specific TensorFlow kernel version (1.14).
 Using a newer or older version may or may not work as intended.
 
-@note CUDA GPU memory cannot be released until the process holding it is 
terminated. 
-When a MADlib deep learning function is called with GPUs, Greenplum internally 
-creates a process (called a slice) which calls TensorFlow to do the 
computation. 
+@note CUDA GPU memory cannot be released until the process holding it is 
terminated.
+When a MADlib deep learning function is called with GPUs, Greenplum internally
+creates a process (called a slice) which calls TensorFlow to do the 
computation.
 This process holds the GPU memory until one of the following two things happen:
-query finishes and user logs out of the Postgres client/session; or, 
-query finishes and user waits for the timeout set by 
`gp_vmem_idle_resource_timeout`.  
+query finishes and user logs out of the Postgres client/session; or,
+query finishes and user waits for the timeout set by 
gp_vmem_idle_resource_timeout.
 The default value for this timeout is 18 sec [8].  So the recommendation is:
 log out/reconnect to the session after every GPU query; or
-wait for `gp_vmem_idle_resource_timeout` before you run another GPU query (you 
can 
+wait for gp_vmem_idle_resource_timeout before you run another GPU query (you 
can
 also set it to a lower value).
 
 @anchor keras_fit
diff --git 
a/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
 
b/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
index b929724..4d1eb09 100644
--- 
a/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
+++ 
b/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
@@ -94,15 +94,15 @@ release the disk space once the fit multiple query has 
completed execution.
 This is not the case for GPDB 6+ where disk space is released during the
 fit multiple query.
 
-@note CUDA GPU memory cannot be released until the process holding it is 
terminated. 
-When a MADlib deep learning function is called with GPUs, Greenplum internally 
-creates a process (called a slice) which calls TensorFlow to do the 
computation. 
+@note CUDA GPU memory cannot be released until the process holding it is 
terminated.
+When a MADlib deep learning function is called with GPUs, Greenplum internally
+creates a process (called a slice) which calls TensorFlow to do the 
computation.
 This process holds the GPU memory until one of the following two things happen:
-query finishes and user logs out of the Postgres client/session; or, 
-query finishes and user waits for the timeout set by 
`gp_vmem_idle_resource_timeout`.  
+query finishes and user logs out of the Postgres client/session; or,
+query finishes and user waits for the timeout set by 
gp_vmem_idle_resource_timeout.
 The default value for this timeout is 18 sec [8].  So the recommendation is:
 log out/reconnect to the session after every GPU query; or
-wait for `gp_vmem_idle_resource_timeout` before you run another GPU query (you 
can 
+wait for gp_vmem_idle_resource_timeout before you run another GPU query (you 
can