diff --git a/site/en/hub/caching.md b/site/en/hub/caching.md index f9e9f5d89a3..678b2c22af0 100644 --- a/site/en/hub/caching.md +++ b/site/en/hub/caching.md @@ -29,7 +29,7 @@ location `/tmp/tfhub_modules` (or whatever `os.path.join(tempfile.gettempdir(), Users who prefer persistent caching across system reboots can instead set `TFHUB_CACHE_DIR` to a location in their home directory. For example, a user of the bash shell on a Linux system can add a line like the following to -`~/.bashrc` +`~/.bashrc`: ```bash export TFHUB_CACHE_DIR=$HOME/.cache/tfhub_modules @@ -41,7 +41,7 @@ persistent location, be aware that there is no automatic cleanup. ### Reading from remote storage Users can instruct the `tensorflow_hub` library to directly read models from -remote storage (GCS) instead of downloading the models locally with +remote storage (GCS) instead of downloading the models locally with: ```shell os.environ["TFHUB_MODEL_LOAD_FORMAT"] = "UNCOMPRESSED" @@ -64,7 +64,7 @@ location by default. There are two workarounds for this situation: The easiest solution is to instruct the `tensorflow_hub` library to read the models from TF Hub's GCS bucket as explained above. Users with their own GCS bucket can instead specify a directory in their bucket as the cache location -with code like +with code like: ```python import os @@ -83,4 +83,4 @@ load_options = tf.saved_model.LoadOptions(experimental_io_device='/job:localhost') reloaded_model = hub.load("https://tfhub.dev/...", options=load_options) ``` -**Note:** See more information regarding valid handles [here](tf2_saved_model.md#model_handles). \ No newline at end of file +**Note:** See more information regarding valid handles [here](tf2_saved_model.md#model_handles).