yeandy commented on code in PR #22069:
URL: https://github.com/apache/beam/pull/22069#discussion_r907842950


##########
sdks/python/apache_beam/examples/inference/README.md:
##########
@@ -108,27 +110,27 @@ This writes the output to the `predictions.csv` with 
contents like:
 ---
 ## Image segmentation
 
-[`pytorch_image_segmentation.py`](./pytorch_image_segmentation.py) contains an 
implementation for a RunInference pipeline that performs image segementation 
using the maskrcnn_resnet50_fpn architecture.
+[`pytorch_image_segmentation.py`](./pytorch_image_segmentation.py) contains an 
implementation for a RunInference pipeline that performs image segementation 
using the `maskrcnn_resnet50_fpn` architecture.
 
-The pipeline reads images, performs basic preprocessing, passes them to the 
PyTorch implementation of RunInference, and then writes the predictions to a 
text file.
+The pipeline reads images, performs basic preprocessing, passes the images to 
the PyTorch implementation of RunInference, and then writes predictions to a 
text file.
 
 ### Dataset and model for image segmentation
-You will need to create or download images, and place them into your 
`IMAGES_DIR` directory. Another popular dataset is from 
[Coco](https://cocodataset.org/#home). Please follow their instructions to 
download the images.
-- **Required**: A path to a file called `IMAGE_FILE_NAMES` that contains the 
absolute paths of each of the images in `IMAGES_DIR` on which you want to run 
image segmentation. Paths can be different types of URIs such as your local 
file system, a AWS S3 bucket or GCP Cloud Storage bucket. For example:
+Create a directory named `IMAGES_DIR`. Create or download images and put them 
in this directory. The directory is not required if image names in the input 
file `IMAGE_FILE_NAMES` have absolute paths.

Review Comment:
   ```suggestion
   Create a directory named `IMAGES_DIR`. Create or download images and put 
them in this directory. The directory is not required if image names in the 
input file `IMAGE_FILE_NAMES` (see below) have absolute paths.
   ```



##########
sdks/python/apache_beam/examples/inference/README.md:
##########
@@ -60,27 +61,28 @@ for details."
 ---
 ## Image classification
 
-[`pytorch_image_classification.py`](./pytorch_image_classification.py) 
contains an implementation for a RunInference pipeline that performs image 
classification using the mobilenet_v2 architecture.
+[`pytorch_image_classification.py`](./pytorch_image_classification.py) 
contains an implementation for a RunInference pipeline that performs image 
classification using the `mobilenet_v2` architecture.
 
-The pipeline reads the images, performs basic preprocessing, passes them to 
the PyTorch implementation of RunInference, and then writes the predictions to 
a text file.
+The pipeline reads the images, performs basic preprocessing, passes the images 
to the PyTorch implementation of RunInference, and then writes the predictions 
to a text file.
 
 ### Dataset and model for image classification
 
-You will need to create or download images, and place them into your 
`IMAGES_DIR` directory. One popular dataset is from 
[ImageNet](https://www.image-net.org/). Please follow their instructions to 
download the images.
-- **Required**: A path to a file called `IMAGE_FILE_NAMES` that contains the 
absolute paths of each of the images in `IMAGES_DIR` on which you want to run 
image segmentation. Paths can be different types of URIs such as your local 
file system, a AWS S3 bucket or GCP Cloud Storage bucket. For example:
+Create a directory named `IMAGES_DIR`. Create or download images and put them 
in this directory. The directory is not required if image names in the input 
file `IMAGE_FILE_NAMES` have absolute paths.

Review Comment:
   ```suggestion
   Create a directory named `IMAGES_DIR`. Create or download images and put 
them in this directory. The directory is not required if image names in the 
input file `IMAGE_FILE_NAMES` (see below) have absolute paths.
   ```



##########
sdks/python/apache_beam/examples/inference/README.md:
##########
@@ -108,27 +110,27 @@ This writes the output to the `predictions.csv` with 
contents like:
 ---
 ## Image segmentation
 
-[`pytorch_image_segmentation.py`](./pytorch_image_segmentation.py) contains an 
implementation for a RunInference pipeline that performs image segementation 
using the maskrcnn_resnet50_fpn architecture.
+[`pytorch_image_segmentation.py`](./pytorch_image_segmentation.py) contains an 
implementation for a RunInference pipeline that performs image segementation 
using the `maskrcnn_resnet50_fpn` architecture.
 
-The pipeline reads images, performs basic preprocessing, passes them to the 
PyTorch implementation of RunInference, and then writes the predictions to a 
text file.
+The pipeline reads images, performs basic preprocessing, passes the images to 
the PyTorch implementation of RunInference, and then writes predictions to a 
text file.
 
 ### Dataset and model for image segmentation
-You will need to create or download images, and place them into your 
`IMAGES_DIR` directory. Another popular dataset is from 
[Coco](https://cocodataset.org/#home). Please follow their instructions to 
download the images.
-- **Required**: A path to a file called `IMAGE_FILE_NAMES` that contains the 
absolute paths of each of the images in `IMAGES_DIR` on which you want to run 
image segmentation. Paths can be different types of URIs such as your local 
file system, a AWS S3 bucket or GCP Cloud Storage bucket. For example:
+Create a directory named `IMAGES_DIR`. Create or download images and put them 
in this directory. The directory is not required if image names in the input 
file `IMAGE_FILE_NAMES` have absolute paths.
+A popular dataset is from [Coco](https://cocodataset.org/#home). Follow their 
instructions to download the images.
+- **Required**: A path to a file named `IMAGE_FILE_NAMES` that contains the 
absolute paths of each of the images in `IMAGES_DIR` that you want to use to 
run image segmentation. Paths can be different types of URIs such as your local 
file system, an AWS S3 bucket, or a GCP Cloud Storage bucket. For example:
 ```
 /absolute/path/to/image1.jpg
 /absolute/path/to/image2.jpg
 ```
-- **Required**: A path to a file called `MODEL_STATE_DICT` that contains the 
saved parameters of the maskrcnn_resnet50_fpn model. You will need to download 
the [maskrcnn_resnet50_fpn](https://pytorch.org/vision/0.12/models.html#id70)
-model from Pytorch's repository of pretrained models. Note that this requires 
`torchvision` library.
+- **Required**: Download the 
[maskrcnn_resnet50_fpn](https://pytorch.org/vision/0.12/models.html#id70) model 
from Pytorch's repository of pretrained models. This model requires the 
torchvision library. To download this model, run the following commands:

Review Comment:
   ```suggestion
   - **Required**: Download the 
[maskrcnn_resnet50_fpn](https://pytorch.org/vision/0.12/models.html#id70) model 
from Pytorch's repository of pretrained models. This model requires the 
`torchvision` library. To download this model, run the following commands:
   ```



##########
sdks/python/apache_beam/examples/inference/README.md:
##########
@@ -157,21 +159,22 @@ Each line has data separated by a semicolon ";". The 
first item is the file name
 ---
 ## Language modeling
 
-[`pytorch_language_modeling.py`](./pytorch_language_modeling.py) contains an 
implementation for a RunInference pipeline that performs masked language 
modeling (i.e. decoding a masked token in a sentence) using the BertForMaskedLM 
architecture from Hugging Face.
+[`pytorch_language_modeling.py`](./pytorch_language_modeling.py) contains an 
implementation for a RunInference pipeline that performs masked language 
modeling (that is, decoding a masked token in a sentence) using the 
`BertForMaskedLM` architecture from Hugging Face.
 
 The pipeline reads sentences, performs basic preprocessing to convert the last 
word into a `[MASK]` token, passes the masked sentence to the PyTorch 
implementation of RunInference, and then writes the predictions to a text file.
 
 ### Dataset and model for language modeling
 
-- **Required**: A path to a file called `MODEL_STATE_DICT` that contains the 
saved parameters of the BertForMaskedLM model. You will need to download the 
[BertForMaskedLM](https://huggingface.co/docs/transformers/model_doc/bert#transformers.BertForMaskedLM)
 model from Hugging Face's repository of pretrained models. Make sure you have 
installed `transformers` too.
+- **Required**: Download the 
[BertForMaskedLM](https://huggingface.co/docs/transformers/model_doc/bert#transformers.BertForMaskedLM)
 model from Hugging Face's repository of pretrained models. You must already 
have `transformers` installed.

Review Comment:
   ```suggestion
   - **Required**: Download the 
[BertForMaskedLM](https://huggingface.co/docs/transformers/model_doc/bert#transformers.BertForMaskedLM)
 model from Hugging Face's repository of pretrained models. You must already 
have `transformers` installed. To download this model, run the following 
commands:
   ```



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to