Skip to content

Commit b0588cc

Browse files
authored
minor cosmetic updates to lpg and checkpoint conversion syntax (#846)
minor updates to lpg and checkpoint conversion
1 parent 559e4c8 commit b0588cc

File tree

3 files changed

+12
-6
lines changed

3 files changed

+12
-6
lines changed

benchmarks/benchmark/tools/profile-generator/README.md

+9-3
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ It currently supports the following frameworks:
2525
- text generation inference (tgi)
2626
- vllm
2727
- sax
28-
-jetstream
28+
- jetstream
2929

3030
## Instructions
3131

@@ -50,6 +50,12 @@ not logged into gcloud, run the following:
5050
gcloud auth application-default login
5151
```
5252

53+
If you do not already have an output bucket, create one by running:
54+
55+
```
56+
gcloud storage buckets create gs://OUTPUT_BUCKET
57+
```
58+
5359
To give viewer permissions on the gcs bucket to the gcloud service account,
5460
run the following:
5561

@@ -88,13 +94,13 @@ gcloud artifacts repositories create ai-benchmark --location=us-central1 --repos
8894

8995
### Step 4: create and configure terraform.tfvars
9096

91-
Create a `terraform.tfvars` file. `./sample-tfvars` is provided as an example
97+
Create a `terraform.tfvars` file. `./sample.tfvars` is provided as an example
9298
file. You can copy the file as a starting point.
9399
Note that at a minimum you will have to change the existing
94100
`credentials_config`, `project_id`, and `artifact_registry`.
95101

96102
```bash
97-
cp ./sample-tfvars terraform.tfvars
103+
cp ./sample.tfvars terraform.tfvars
98104
```
99105

100106
Fill out your `terraform.tfvars` with the desired model and server configuration, referring to the list of required and optional variables [here](#variables). The following variables are required:

benchmarks/benchmark/tools/profile-generator/container/benchmark_serving.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -555,8 +555,8 @@ def main(args: argparse.Namespace):
555555

556556
# NOTE: The latency below includes requests awaiting time on server side.
557557
# It's not comparable with the model inference latency for batch size 1.
558-
**(get_stats_for_set("latency", "Average milliseconds/request (includes waiting time on server)" ,[1000 * latency for _, _, latency in REQUEST_LATENCY])),
559-
**(get_stats_for_set("per_output_token_latency", "Average milliseconds/output_token (includes waiting time on server)", [1000 * latency / output_len for _, output_len, latency in REQUEST_LATENCY])),
558+
**(get_stats_for_set("latency", "milliseconds/request (includes waiting time on server)" ,[1000 * latency for _, _, latency in REQUEST_LATENCY])),
559+
**(get_stats_for_set("per_output_token_latency", "milliseconds/output_token (includes waiting time on server)", [1000 * latency / output_len for _, output_len, latency in REQUEST_LATENCY])),
560560
**(get_stats_for_set("input_len", "input length", [float(prompt_len) for prompt_len, _, _ in REQUEST_LATENCY])),
561561
**(get_stats_for_set("output_len", "output length", [float(output_len) for _, output_len, _ in REQUEST_LATENCY]))
562562
}

tutorials-and-examples/inference-servers/checkpoints/checkpoint_converter.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -114,7 +114,7 @@ convert_maxtext_checkpoint() {
114114
git checkout ${VERSION}
115115
python3 -m pip install -r requirements.txt
116116

117-
if [ $VERSION == "jetstream-v0.2.2" || $VERSION == "jetstream-v0.2.1" || $VERSION == "jetstream-v0.2.0" ]; then
117+
if [[ $VERSION == "jetstream-v0.2.2" || $VERSION == "jetstream-v0.2.1" || $VERSION == "jetstream-v0.2.0" ]]; then
118118
pip3 install orbax-checkpoint==0.5.20
119119
else
120120
pip3 install orbax-checkpoint==0.6.0

0 commit comments

Comments
 (0)