Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
81 commits
Select commit Hold shift + click to select a range
52e1489
redocument
t-kalinowski Sep 25, 2025
28ba55c
fix freeze_weights example, workaournd efficientnetb0 bug
t-kalinowski Sep 25, 2025
b0f80c3
`activation_elu()`: document alpha slope
t-kalinowski Sep 25, 2025
45c335e
`callback_lambda()` docs: `logs` is keyword
t-kalinowski Sep 25, 2025
157df0e
internal tether change: `EarlyStopping` uses `MonitorCallback`
t-kalinowski Sep 25, 2025
41baa8b
`callback_model_checkpoint()`: clarify auto direction
t-kalinowski Sep 25, 2025
5eb3228
internal tether change: `ReduceLROnPlateau` uses `MonitorCallback`
t-kalinowski Sep 25, 2025
ca15277
`Layer`: `add_weight()` gains `overwrite_with_gradient`; new `symbol…
t-kalinowski Sep 25, 2025
f30b452
`layer_rescaling()`: doc per-channel scale/offset
t-kalinowski Sep 25, 2025
30926e1
`layer_dense()`: add `lora_alpha`
t-kalinowski Sep 25, 2025
05d512f
`layer_embedding()`: add `lora_alpha` doc
t-kalinowski Sep 25, 2025
53e0d0f
`layer_layer_normalization()`: drop `rms_scaling`
t-kalinowski Sep 25, 2025
a7beb92
`layer_average_pooling_2d()`: fix `padding="same"` shape
t-kalinowski Sep 25, 2025
9dfb905
internal tether change: drop `Discretization.build()`
t-kalinowski Sep 25, 2025
56ecd08
`layer_rnn()` example: drop `self$built`
t-kalinowski Sep 25, 2025
a73c0f3
`loss_dice()`: use class-based example
t-kalinowski Sep 25, 2025
eea32a5
`loss_sparse_categorical_crossentropy()`: update examples, add `axis`
t-kalinowski Sep 25, 2025
30c97ee
`layer_conv*_transpose()`: add `output_padding` arg
t-kalinowski Sep 25, 2025
30bf422
`op_conv_transpose()`: default `strides=1`
t-kalinowski Sep 25, 2025
53e93d6
internal tether changes to `op_vectorized_map()`
t-kalinowski Sep 25, 2025
0d943c4
`op_dot_product_attention()`: add `attn_logits_soft_cap` arg
t-kalinowski Sep 25, 2025
4398732
`op_einsum()`: mention extra kwargs
t-kalinowski Sep 25, 2025
e59d45a
`op_rms_normalization()`: default `scale` to `NULL`
t-kalinowski Sep 25, 2025
8662462
`layer_einsum_dense()`: add `lora_alpha` arg
t-kalinowski Sep 25, 2025
e64bdf7
`layer_torch_module_wrapper()`: add `output_shape` arg
t-kalinowski Sep 25, 2025
9247e2e
`optimizer_lion()` docs: grammar tweak
t-kalinowski Sep 25, 2025
ba8400d
internal tether change: `to_categorical` example reshape
t-kalinowski Sep 25, 2025
5028107
`load_model_weights()`: document sharded manifests
t-kalinowski Sep 25, 2025
060028e
`metric_binary_focal_crossentropy()`: align docs
t-kalinowski Sep 25, 2025
97886e7
internal tether change: update SensitivityAtSpecificity example
t-kalinowski Sep 25, 2025
3af08ae
internal tether change: update SpecificityAtSensitivity example
t-kalinowski Sep 25, 2025
39eb0d4
`save_model_weights()` gains arg `max_shard_size`.
t-kalinowski Sep 25, 2025
75c7a50
internal tether change: update applications module path
t-kalinowski Sep 25, 2025
f9fb925
internal tether changes
t-kalinowski Sep 25, 2025
953e5cc
internal tether changes
t-kalinowski Sep 25, 2025
60a050c
`keras_variable()` gains arg synchronization
t-kalinowski Sep 25, 2025
07bf72f
internal tether change: allow Sequential call kwargs
t-kalinowski Sep 25, 2025
0ff476b
redocument
t-kalinowski Sep 25, 2025
f332596
add `op_sparse_sigmoid()` and `activation_sparse_sigmoid()`
t-kalinowski Sep 26, 2025
147e48e
add `tools/make-r-wrapper.R` script
t-kalinowski Sep 26, 2025
8d0e5ab
add `op_angle()`
t-kalinowski Sep 26, 2025
a49f307
add `Arg()` method for tensors that calls `op_angle()`
t-kalinowski Sep 26, 2025
73a17e2
add `op_bartlett()`
t-kalinowski Sep 26, 2025
69c831e
add `op_blackman()`
t-kalinowski Sep 26, 2025
2b6f3df
add `op_hamming()`
t-kalinowski Sep 26, 2025
0383f94
add `op_hanning()`
t-kalinowski Sep 26, 2025
0d6cf91
add `op_heaviside()`
t-kalinowski Sep 26, 2025
90e6078
expand docs for window functions
t-kalinowski Sep 26, 2025
5793c85
add `op_kaiser()`
t-kalinowski Sep 26, 2025
5bd3987
add `op_cbrt` `op_corrcoef` `op_deg2rad`
t-kalinowski Sep 26, 2025
5c93463
add `op_view_as_real()` and `op_view_as_complex()`
t-kalinowski Sep 26, 2025
d233cae
add `op_layer_normalization()`
t-kalinowski Sep 26, 2025
811b426
internal tether changes: modules moved '.api.' -> '.'
t-kalinowski Sep 26, 2025
fc81ba4
add `loss_categorical_generalized_cross_entropy()`
t-kalinowski Sep 26, 2025
f8cb769
redocument
t-kalinowski Sep 26, 2025
52a3823
add `op_image_elastic_transform()`
t-kalinowski Sep 26, 2025
1a10cd4
update `op_searchsorted()` docs
t-kalinowski Sep 26, 2025
dff5ed0
remove dup tether
t-kalinowski Sep 26, 2025
0816b43
add `optimizer_muon()`
t-kalinowski Sep 26, 2025
4175115
add `layer_random_elastic_transform()`
t-kalinowski Sep 26, 2025
aba497e
internal tether change
t-kalinowski Sep 26, 2025
9840004
internal tether changes
t-kalinowski Sep 26, 2025
4bf0c91
internal tether changes
t-kalinowski Sep 26, 2025
9125798
internal tether diff
t-kalinowski Sep 26, 2025
7460fac
internal tether change
t-kalinowski Sep 26, 2025
05497ca
redocument
t-kalinowski Sep 26, 2025
d55d807
add `config_max_epochs()`, `config_set_max_epochs()`, `config_max_ste…
t-kalinowski Sep 26, 2025
9708bb5
redocument
t-kalinowski Sep 27, 2025
cf36a21
internal tether changes: quantizers
t-kalinowski Sep 27, 2025
c687e5b
internal tether changes: distribution
t-kalinowski Sep 27, 2025
033c605
update NAMESPACE
t-kalinowski Sep 27, 2025
eb76f94
check in new tethers
t-kalinowski Sep 27, 2025
01eb4b0
retether guides
t-kalinowski Sep 27, 2025
8f185ae
add missing family tag for optimizer_muon
t-kalinowski Sep 27, 2025
159480f
add NEWS
t-kalinowski Sep 27, 2025
ad15437
knit examples and guides
t-kalinowski Sep 27, 2025
39d3f56
rebuild site
t-kalinowski Sep 27, 2025
8aeb96a
R CMD check fixes
t-kalinowski Sep 27, 2025
37fa8fd
redocument
t-kalinowski Sep 27, 2025
8dab59b
roxygen warnings fixes
t-kalinowski Sep 28, 2025
8790244
more roxygen fixes
t-kalinowski Sep 28, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
12 changes: 11 additions & 1 deletion .tether/man/Layer.txt
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
Help on class Layer in module keras.src.layers.layer:

class Layer(keras.src.backend.tensorflow.layer.TFLayer, keras.src.ops.operation.Operation, keras.src.saving.keras_saveable.KerasSaveable)
class Layer(keras.src.backend.tensorflow.layer.TFLayer, keras.src.ops.operation.Operation)
| Layer(*args, **kwargs)
|
| This is the class from which all layers inherit.
Expand Down Expand Up @@ -243,6 +243,7 @@ class Layer(keras.src.backend.tensorflow.layer.TFLayer, keras.src.ops.operation.
| regularizer=None,
| constraint=None,
| aggregation='none',
| overwrite_with_gradient=False,
| name=None
| )
| Add a weight variable to the layer.
Expand Down Expand Up @@ -274,6 +275,9 @@ class Layer(keras.src.backend.tensorflow.layer.TFLayer, keras.src.ops.operation.
| the type of multi-replica aggregation to be used for this
| variable when writing custom data parallel training loops.
| Defaults to `"none"`.
| overwrite_with_gradient: Boolean, whether to overwrite the variable
| with the computed gradient. This is useful for float8 training.
| Defaults to `False`.
| name: String name of the variable. Useful for debugging purposes.
|
| build(self, input_shape)
Expand Down Expand Up @@ -448,6 +452,12 @@ class Layer(keras.src.backend.tensorflow.layer.TFLayer, keras.src.ops.operation.
| ref_var.assign(value)
| ```
|
| symbolic_call(
| self,
| *args,
| **kwargs
| )
|
| ----------------------------------------------------------------------
| Static methods defined here:
|
Expand Down
1 change: 1 addition & 0 deletions .tether/man/activation_elu.txt
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,7 @@ and the information that is propagated to the next layer.

Args:
x: Input tensor.
alpha: A scalar, slope of positive section. Defaults to `1.0`.

Reference:

Expand Down
18 changes: 18 additions & 0 deletions .tether/man/activation_sparse_sigmoid.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
__signature__
keras.activations.sparse_sigmoid(x)
__doc__
Sparse sigmoid activation function.

It is defined as

`f(x) = 0` for `x <= -1`,
`f(x) = 0.5 * (x + 1)` for `-1 < x < 1`,
`f(x) = 1` for `x >= 1`.

Args:
x: Input tensor.

Reference:

- [M. Blondel, A. F. T. Martins, V. Niculae, 2019](https://arxiv.org/pdf/1901.02324)

3 changes: 2 additions & 1 deletion .tether/man/callback_early_stopping.txt
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
Help on class EarlyStopping in module keras.src.callbacks.early_stopping:

class EarlyStopping(keras.src.callbacks.callback.Callback)
class EarlyStopping(keras.src.callbacks.monitor_callback.MonitorCallback)
| EarlyStopping(monitor='val_loss', min_delta=0, patience=0, verbose=0, mode='auto', baseline=None, restore_best_weights=False, start_from_epoch=0)
|
| Stop training when a monitored metric has stopped improving.
Expand Down Expand Up @@ -60,6 +60,7 @@ class EarlyStopping(keras.src.callbacks.callback.Callback)
|
| Method resolution order:
| EarlyStopping
| keras.src.callbacks.monitor_callback.MonitorCallback
| keras.src.callbacks.callback.Callback
| builtins.object
|
Expand Down
5 changes: 3 additions & 2 deletions .tether/man/callback_lambda.txt
Original file line number Diff line number Diff line change
Expand Up @@ -13,8 +13,8 @@ class LambdaCallback(keras.src.callbacks.callback.Callback)
| `epoch`, `logs`
| - `on_train_begin` and `on_train_end` expect one positional argument:
| `logs`
| - `on_train_batch_begin` and `on_train_batch_end` expect two positional
| arguments: `batch`, `logs`
| - `on_train_batch_begin` and `on_train_batch_end` expect a positional
| argument `batch` and a keyword argument `logs`
| - See `Callback` class definition for the full list of functions and their
| expected arguments.
|
Expand Down Expand Up @@ -79,3 +79,4 @@ class LambdaCallback(keras.src.callbacks.callback.Callback)
| )
| Initialize self. See help(type(self)) for accurate signature.
|

8 changes: 4 additions & 4 deletions .tether/man/callback_model_checkpoint.txt
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
Help on class ModelCheckpoint in module keras.src.callbacks.model_checkpoint:

class ModelCheckpoint(keras.src.callbacks.callback.Callback)
class ModelCheckpoint(keras.src.callbacks.monitor_callback.MonitorCallback)
| ModelCheckpoint(filepath, monitor='val_loss', verbose=0, save_best_only=False, save_weights_only=False, mode='auto', save_freq='epoch', initial_value_threshold=None)
|
| Callback to save the Keras model or model weights at some frequency.
Expand Down Expand Up @@ -95,9 +95,8 @@ class ModelCheckpoint(keras.src.callbacks.callback.Callback)
| decision to overwrite the current save file is made based on either
| the maximization or the minimization of the monitored quantity.
| For `val_acc`, this should be `"max"`, for `val_loss` this should be
| `"min"`, etc. In `"auto"` mode, the mode is set to `"max"` if the
| quantities monitored are `"acc"` or start with `"fmeasure"` and are
| set to `"min"` for the rest of the quantities.
| `"min"`, etc. In `"auto"` mode, the direction is automatically
| inferred from the name of the monitored quantity.
| save_weights_only: if `True`, then only the model's weights will be
| saved (`model.save_weights(filepath)`), else the full model is
| saved (`model.save(filepath)`).
Expand All @@ -116,6 +115,7 @@ class ModelCheckpoint(keras.src.callbacks.callback.Callback)
|
| Method resolution order:
| ModelCheckpoint
| keras.src.callbacks.monitor_callback.MonitorCallback
| keras.src.callbacks.callback.Callback
| builtins.object
|
Expand Down
3 changes: 2 additions & 1 deletion .tether/man/callback_reduce_lr_on_plateau.txt
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
Help on class ReduceLROnPlateau in module keras.src.callbacks.reduce_lr_on_plateau:

class ReduceLROnPlateau(keras.src.callbacks.callback.Callback)
class ReduceLROnPlateau(keras.src.callbacks.monitor_callback.MonitorCallback)
| ReduceLROnPlateau(monitor='val_loss', factor=0.1, patience=10, verbose=0, mode='auto', min_delta=0.0001, cooldown=0, min_lr=0.0, **kwargs)
|
| Reduce learning rate when a metric has stopped improving.
Expand Down Expand Up @@ -39,6 +39,7 @@ class ReduceLROnPlateau(keras.src.callbacks.callback.Callback)
|
| Method resolution order:
| ReduceLROnPlateau
| keras.src.callbacks.monitor_callback.MonitorCallback
| keras.src.callbacks.callback.Callback
| builtins.object
|
Expand Down
9 changes: 9 additions & 0 deletions .tether/man/config_is_nnx_enabled.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,9 @@
__signature__
keras.config.is_nnx_enabled()
__doc__
Checks whether NNX specific features are enabled for the JAX backend.

Returns:
bool: `True` if NNX backend features are enabled, `False` otherwise.
Defaults to `False`.

13 changes: 13 additions & 0 deletions .tether/man/config_max_epochs.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
__signature__
keras.config.max_epochs()
__doc__
Get the maximum number of epochs for any call to fit.

Retrieves the limit on the number of epochs set by
`keras.config.set_max_epochs` or the `KERAS_MAX_EPOCHS` environment
variable.

Returns:
The integer limit on the number of epochs or `None`, if no limit has
been set.

13 changes: 13 additions & 0 deletions .tether/man/config_max_steps_per_epoch.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
__signature__
keras.config.max_steps_per_epoch()
__doc__
Get the maximum number of steps for any call to fit/evaluate/predict.

Retrieves the limit on the number of epochs set by
`keras.config.set_max_steps_per_epoch` or the `KERAS_MAX_STEPS_PER_EPOCH`
environment variable.

Args:
max_epochs: The integer limit on the number of epochs or `None`. If
`None`, no limit is applied.

13 changes: 13 additions & 0 deletions .tether/man/config_set_max_epochs.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
__signature__
keras.config.set_max_epochs(max_epochs)
__doc__
Limit the maximum number of epochs for any call to fit.

This will cap the number of epochs for any training run using `model.fit()`.
This is purely for debugging, and can also be set via the `KERAS_MAX_EPOCHS`
environment variable to quickly run a script without modifying its source.

Args:
max_epochs: The integer limit on the number of epochs or `None`. If
`None`, no limit is applied.

14 changes: 14 additions & 0 deletions .tether/man/config_set_max_steps_per_epoch.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,14 @@
__signature__
keras.config.set_max_steps_per_epoch(max_steps_per_epoch)
__doc__
Limit the maximum number of steps for any call to fit/evaluate/predict.

This will cap the number of steps for single epoch of a call to `fit()`,
`evaluate()`, or `predict()`. This is purely for debugging, and can also be
set via the `KERAS_MAX_STEPS_PER_EPOCH` environment variable to quickly run
a scrip without modifying its source.

Args:
max_epochs: The integer limit on the number of epochs or `None`. If
`None`, no limit is applied.

1 change: 1 addition & 0 deletions .tether/man/keras.activations.txt
Original file line number Diff line number Diff line change
Expand Up @@ -31,6 +31,7 @@ softmax(x, axis=-1)
softplus(x)
softsign(x)
sparse_plus(x)
sparse_sigmoid(x)
sparsemax(x, axis=-1)
squareplus(x, b=4)
swish(x)
Expand Down
34 changes: 17 additions & 17 deletions .tether/man/keras.applications.txt
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
convnext: Module(keras.api.applications.convnext)
convnext: Module(keras.applications.convnext)
ConvNeXtBase(
include_top=True,
include_preprocessing=True,
Expand Down Expand Up @@ -54,7 +54,7 @@ ConvNeXtXLarge(
classifier_activation='softmax',
name='convnext_xlarge'
)
densenet: Module(keras.api.applications.densenet)
densenet: Module(keras.applications.densenet)
DenseNet121(
include_top=True,
weights='imagenet',
Expand Down Expand Up @@ -85,8 +85,8 @@ DenseNet201(
classifier_activation='softmax',
name='densenet201'
)
efficientnet: Module(keras.api.applications.efficientnet)
efficientnet_v2: Module(keras.api.applications.efficientnet_v2)
efficientnet: Module(keras.applications.efficientnet)
efficientnet_v2: Module(keras.applications.efficientnet_v2)
EfficientNetB0(
include_top=True,
weights='imagenet',
Expand Down Expand Up @@ -244,9 +244,9 @@ EfficientNetV2S(
include_preprocessing=True,
name='efficientnetv2-s'
)
imagenet_utils: Module(keras.api.applications.imagenet_utils)
inception_resnet_v2: Module(keras.api.applications.inception_resnet_v2)
inception_v3: Module(keras.api.applications.inception_v3)
imagenet_utils: Module(keras.applications.imagenet_utils)
inception_resnet_v2: Module(keras.applications.inception_resnet_v2)
inception_v3: Module(keras.applications.inception_v3)
InceptionResNetV2(
include_top=True,
weights='imagenet',
Expand All @@ -267,7 +267,7 @@ InceptionV3(
classifier_activation='softmax',
name='inception_v3'
)
mobilenet: Module(keras.api.applications.mobilenet)
mobilenet: Module(keras.applications.mobilenet)
MobileNet(
input_shape=None,
alpha=1.0,
Expand All @@ -281,8 +281,8 @@ MobileNet(
classifier_activation='softmax',
name=None
)
mobilenet_v2: Module(keras.api.applications.mobilenet_v2)
mobilenet_v3: Module(keras.api.applications.mobilenet_v3)
mobilenet_v2: Module(keras.applications.mobilenet_v2)
mobilenet_v3: Module(keras.applications.mobilenet_v3)
MobileNetV2(
input_shape=None,
alpha=1.0,
Expand Down Expand Up @@ -322,7 +322,7 @@ MobileNetV3Small(
include_preprocessing=True,
name='MobileNetV3Small'
)
nasnet: Module(keras.api.applications.nasnet)
nasnet: Module(keras.applications.nasnet)
NASNetLarge(
input_shape=None,
include_top=True,
Expand All @@ -343,8 +343,8 @@ NASNetMobile(
classifier_activation='softmax',
name='nasnet_mobile'
)
resnet: Module(keras.api.applications.resnet)
resnet_v2: Module(keras.api.applications.resnet_v2)
resnet: Module(keras.applications.resnet)
resnet_v2: Module(keras.applications.resnet_v2)
ResNet101(
include_top=True,
weights='imagenet',
Expand Down Expand Up @@ -385,7 +385,7 @@ ResNet152V2(
classifier_activation='softmax',
name='resnet152v2'
)
resnet50: Module(keras.api.applications.resnet50)
resnet50: Module(keras.applications.resnet50)
ResNet50(
include_top=True,
weights='imagenet',
Expand All @@ -406,7 +406,7 @@ ResNet50V2(
classifier_activation='softmax',
name='resnet50v2'
)
vgg16: Module(keras.api.applications.vgg16)
vgg16: Module(keras.applications.vgg16)
VGG16(
include_top=True,
weights='imagenet',
Expand All @@ -417,7 +417,7 @@ VGG16(
classifier_activation='softmax',
name='vgg16'
)
vgg19: Module(keras.api.applications.vgg19)
vgg19: Module(keras.applications.vgg19)
VGG19(
include_top=True,
weights='imagenet',
Expand All @@ -428,7 +428,7 @@ VGG19(
classifier_activation='softmax',
name='vgg19'
)
xception: Module(keras.api.applications.xception)
xception: Module(keras.applications.xception)
Xception(
include_top=True,
weights='imagenet',
Expand Down
5 changes: 5 additions & 0 deletions .tether/man/keras.config.txt
Original file line number Diff line number Diff line change
Expand Up @@ -12,10 +12,15 @@ floatx()
image_data_format()
is_flash_attention_enabled()
is_interactive_logging_enabled()
is_nnx_enabled()
is_traceback_filtering_enabled()
max_epochs()
max_steps_per_epoch()
set_backend(backend)
set_dtype_policy(policy)
set_epsilon(value)
set_floatx(value)
set_image_data_format(data_format)
set_max_epochs(max_epochs)
set_max_steps_per_epoch(max_steps_per_epoch)

16 changes: 8 additions & 8 deletions .tether/man/keras.datasets.txt
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
boston_housing: Module(keras.api.datasets.boston_housing)
california_housing: Module(keras.api.datasets.california_housing)
cifar10: Module(keras.api.datasets.cifar10)
cifar100: Module(keras.api.datasets.cifar100)
fashion_mnist: Module(keras.api.datasets.fashion_mnist)
imdb: Module(keras.api.datasets.imdb)
mnist: Module(keras.api.datasets.mnist)
reuters: Module(keras.api.datasets.reuters)
boston_housing: Module(keras.datasets.boston_housing)
california_housing: Module(keras.datasets.california_housing)
cifar10: Module(keras.datasets.cifar10)
cifar100: Module(keras.datasets.cifar100)
fashion_mnist: Module(keras.datasets.fashion_mnist)
imdb: Module(keras.datasets.imdb)
mnist: Module(keras.datasets.mnist)
reuters: Module(keras.datasets.reuters)

1 change: 1 addition & 0 deletions .tether/man/keras.distribution.txt
Original file line number Diff line number Diff line change
Expand Up @@ -21,6 +21,7 @@ ModelParallel(
*,
layout_map=None,
batch_dim_name=None,
auto_shard_dataset=True,
**kwargs
)
set_distribution(value)
Expand Down
Loading
Loading