Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Avoid common TensorFlow mistakes — tf.function retracing, GPU memory, data pipeline bottlenecks, and gradient traps.
Avoid common TensorFlow mistakes — tf.function retracing, GPU memory, data pipeline bottlenecks, and gradient traps.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
New input shape/dtype causes retrace — expensive, prints warning Use input_signature for fixed shapes — @tf.function(input_signature=[tf.TensorSpec(...)]) Python values retrace — pass as tensors, not Python ints/floats Avoid Python side effects in tf.function — only runs once during tracing
TensorFlow grabs all GPU memory by default — set memory_growth=True before any ops tf.config.experimental.set_memory_growth(gpu, True) — must be called before GPU init OOM with large models — reduce batch size or use gradient checkpointing CUDA_VISIBLE_DEVICES="" to force CPU — for testing without GPU
tf.data.Dataset without .prefetch() — CPU/GPU idle time between batches .cache() after expensive ops — but before random augmentation .batch() before .map() for vectorized ops — faster than per-element num_parallel_calls=tf.data.AUTOTUNE — parallel preprocessing Dataset iteration in eager mode is slow — use in tf.function or model.fit
First dimension is batch — None for variable batch size in Input layer model.build(input_shape) if not using Input layer — or first call errors Reshape errors unclear — tf.debugging.assert_shapes() for debugging Broadcasting silently succeeds — may hide shape bugs
Variables watched by default — tensors need tape.watch(tensor) persistent=True for multiple gradients — otherwise tape consumed after first use tape.gradient returns None if no path — check for disconnected graph @tf.custom_gradient for custom backward — not all ops have gradients
model.trainable = False after compile does nothing — set before compile BatchNorm behaves differently in training vs inference — training=True/False matters model.fit shuffles by default — shuffle=False for time series validation_split takes from end — shuffle data first if order matters
model.save() saves everything — architecture, weights, optimizer state model.save_weights() only weights — need model code to restore SavedModel format for serving — tf.saved_model.save(model, path) H5 format limited — doesn't save custom objects well, use SavedModel
Mixing Keras and raw tf ops incorrectly — use layers.Lambda to wrap tf ops in Sequential tf.print vs Python print — Python print only runs at trace time in tf.function NumPy ops in graph — use tf ops, numpy executes eagerly only Loss returns scalar per sample — Keras averages, custom loops may need tf.reduce_mean
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.