summaryrefslogtreecommitdiff
path: root/ml_exp/kernels.py
diff options
context:
space:
mode:
Diffstat (limited to 'ml_exp/kernels.py')
-rw-r--r--ml_exp/kernels.py85
1 files changed, 25 insertions, 60 deletions
diff --git a/ml_exp/kernels.py b/ml_exp/kernels.py
index 488232392..11679460b 100644
--- a/ml_exp/kernels.py
+++ b/ml_exp/kernels.py
@@ -30,15 +30,17 @@ except ImportError:
TF_AV = False
-def gaussian_kernel(X1,
+def laplauss_kernel(X1,
X2,
sigma,
+ laplauss='gauss',
use_tf=True):
"""
- Calculates the Gaussian Kernel.
+ Calculates the Lpalacian or Gaussian Kernel.
X1: first representations.
X2: second representations.
sigma: kernel width.
+ laplauss: which kernel to calculate.
use_tf: if tensorflow should be used.
"""
# If tf is to be used but couldn't be imported, don't try to use it.
@@ -47,62 +49,12 @@ def gaussian_kernel(X1,
X1_size = X1.shape[0]
X2_size = X2.shape[0]
- i_sigma = -0.5 / (sigma*sigma)
-
- if use_tf:
- if tf.config.experimental.list_physical_devices('GPU'):
- with tf.device('GPU:0'):
- X1 = tf.convert_to_tensor(X1)
- X2 = tf.convert_to_tensor(X2)
- X2r = tf.rank(X2)
-
- def cond(i, _):
- return tf.less(i, X1_size)
-
- def body(i, K):
- if X2r == 3:
- norm = tf.norm(X2 - X1[i], axis=(1, 2))
- else:
- norm = tf.norm(X2 - X1[i], axis=-1)
-
- return (i + 1,
- K.write(i, tf.exp(i_sigma * tf.square(norm))))
-
- K = tf.TensorArray(dtype=tf.float64,
- size=X1_size)
- i_state = (0, K)
- n, K = tf.while_loop(cond, body, i_state)
- K = K.stack()
+ if laplauss == 'gauss':
+ i_sigma = -0.5 / (sigma*sigma)
+ elif laplauss == 'laplace':
+ i_sigma = -0.5 / sigma
else:
- K = np.zeros((X1_size, X2_size), dtype=np.float64)
- # Faster way of calculating the kernel (no numba support).
- for i in range(X1_size):
- if X2.ndim == 3:
- norm = np.linalg.norm(X2 - X1[i], axis=(1, 2))
- else:
- norm = np.linalg.norm(X2 - X1[i], axis=-1)
- K[i, :] = np.exp(i_sigma * np.square(norm))
- return K
-
-
-def laplacian_kernel(X1,
- X2,
- sigma,
- use_tf=True):
- """
- Calculates the Laplacian Kernel.
- X1: first representations.
- X2: second representations.
- sigma: kernel width.
- use_tf: if tensorflow should be used.
- """
- # If tf is to be used but couldn't be imported, don't try to use it.
- if use_tf and not TF_AV:
- use_tf = False
-
- X1_size = X1.shape[0]
- X2_size = X2.shape[0]
- i_sigma = -0.5 / sigma
+ i_sigma = -0.5 / (sigma**2)
if use_tf:
if tf.config.experimental.list_physical_devices('GPU'):
@@ -120,8 +72,15 @@ def laplacian_kernel(X1,
else:
norm = tf.norm(X2 - X1[i], axis=-1)
- return (i + 1,
- K.write(i, tf.exp(i_sigma * norm)))
+ if laplauss == 'gauss':
+ return (i + 1,
+ K.write(i, tf.exp(i_sigma * tf.square(norm))))
+ elif laplauss == 'laplace':
+ return (i + 1,
+ K.write(i, tf.exp(i_sigma * norm)))
+ else:
+ return (i + 1,
+ K.write(i, tf.exp(i_sigma * tf.square(norm))))
K = tf.TensorArray(dtype=tf.float64,
size=X1_size)
@@ -136,5 +95,11 @@ def laplacian_kernel(X1,
norm = np.linalg.norm(X2 - X1[i], axis=(1, 2))
else:
norm = np.linalg.norm(X2 - X1[i], axis=-1)
- K[i, :] = np.exp(i_sigma * norm)
+ if laplauuss == 'gauss':
+ K[i, :] = np.exp(i_sigma * np.square(norm))
+ elif laplauce == 'laplace':
+ K[i, :] = np.exp(i_sigma * norm)
+ else:
+ K[i, :] = np.exp(i_sigma * np.square(norm))
+
return K