Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion 3d_segmentation/swin_unetr_brats21_segmentation_3d.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -455,7 +455,7 @@
"source": [
"## Create Swin UNETR model\n",
"\n",
"In this scetion, we create Swin UNETR model for the 3-class brain tumor semantic segmentation. We use a feature size of 48. We also use gradient checkpointing (use_checkpoint) for more memory-efficient training. However, use_checkpoint for faster training if enough GPU memory is available. "
"In this section, we create Swin UNETR model for the 3-class brain tumor semantic segmentation. We use a feature size of 48. We also use gradient checkpointing (use_checkpoint) for more memory-efficient training. However, use_checkpoint for faster training if enough GPU memory is available. "
]
},
{
Expand Down
4 changes: 2 additions & 2 deletions 3d_segmentation/swin_unetr_btcv_segmentation_3d.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -96,7 +96,7 @@
"source": [
"# Pre-trained Swin UNETR Encoder\n",
"\n",
"We use weights from self-supervised pre-training of Swin UNETR encoder (3D Swin Tranformer) on a cohort of 5050 CT scans from publicly available datasets. The encoder is pre-trained using reconstructin, rotation prediction and contrastive learning pre-text tasks as shown below. For more details, please refer to [1] (CVPR paper) and see this [repository](https://github.com/Project-MONAI/research-contributions/tree/main/SwinUNETR/Pretrain). \n",
"We use weights from self-supervised pre-training of Swin UNETR encoder (3D Swin Transformer) on a cohort of 5050 CT scans from publicly available datasets. The encoder is pre-trained using reconstructin, rotation prediction and contrastive learning pre-text tasks as shown below. For more details, please refer to [1] (CVPR paper) and see this [repository](https://github.com/Project-MONAI/research-contributions/tree/main/SwinUNETR/Pretrain). \n",
"\n",
"Please download the pre-trained weights from this [link](https://github.com/Project-MONAI/MONAI-extra-test-data/releases/download/0.8.1/model_swinvit.pt) and place it in the root directory of this tutorial. \n",
"\n",
Expand Down Expand Up @@ -452,7 +452,7 @@
"source": [
"### Initialize Swin UNETR encoder from self-supervised pre-trained weights\n",
"\n",
"In this section, we intialize the Swin UNETR encoder from pre-trained weights. The weights can be downloaded using the wget command below, or by following [this link](https://github.com/Project-MONAI/MONAI-extra-test-data/releases/download/0.8.1/model_swinvit.pt) to GitHub. If training from scratch is desired, please skip this section."
"In this section, we initialize the Swin UNETR encoder from pre-trained weights. The weights can be downloaded using the wget command below, or by following [this link](https://github.com/Project-MONAI/MONAI-extra-test-data/releases/download/0.8.1/model_swinvit.pt) to GitHub. If training from scratch is desired, please skip this section."
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion auto3dseg/docs/gpu_opt.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ Sometimes the low GPU utilization is because that GPU capacities is not fully ut
Our proposed solution is capable to automatically estimate hyper-parameters in model training configurations maximizing utilities of the available GPU capacities.
The solution is leveraging hyper-parameter optimization algorithms to search for optimital hyper-parameters with any given GPU devices.

The following hyper-paramters in model training configurations are optimized in the process.
The following hyper-parameters in model training configurations are optimized in the process.

1. **num_images_per_batch:** Batch size determines how many images are in each mini-batch and how many training iterations per epoch. Large batch size can reduce training time per epoch and increase GPU memory usage with decent CPU capacities for I/O;
2. **num_sw_batch_size:** Batch size in sliding-window inference directly relates to how many patches are in one pass of model feedforward operation. Large batch size in sliding-window inference can reduce overall inference time and increase GPU memory usage;
Expand Down
2 changes: 1 addition & 1 deletion deployment/Triton/models/mednist_class/1/model.py
Original file line number Diff line number Diff line change
Expand Up @@ -89,7 +89,7 @@ def initialize(self, args):
"""
`initialize` is called only once when the model is being loaded.
Implementing `initialize` function is optional. This function allows
the model to intialize any state associated with this model.
the model to initialize any state associated with this model.
"""

# Pull model from google drive
Expand Down
2 changes: 1 addition & 1 deletion deployment/Triton/models/monai_covid/1/model.py
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@ def initialize(self, args):
"""
`initialize` is called only once when the model is being loaded.
Implementing `initialize` function is optional. This function allows
the model to intialize any state associated with this model.
the model to initialize any state associated with this model.
"""

# Pull model from google drive
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -1313,7 +1313,7 @@
"# loop over tokens\n",
"for i in range(1, latent.shape[1]):\n",
" if mask_flattened[i - 1]:\n",
" # if token is low probability, replace with tranformer's most likely token\n",
" # if token is low probability, replace with Transformer's most likely token\n",
" logits = transformer_model(latent_healed[:, :i])\n",
" probs = F.softmax(logits, dim=-1)\n",
" # don't sample beginning of sequence token\n",
Expand Down
2 changes: 1 addition & 1 deletion generation/maisi/scripts/sample.py
Original file line number Diff line number Diff line change
Expand Up @@ -597,7 +597,7 @@ def __init__(
label_dict = json.load(f)
self.all_anatomy_size_condtions_json = all_anatomy_size_condtions_json

# intialize variables
# initialize variables
self.body_region = body_region
self.anatomy_list = [label_dict[organ] for organ in anatomy_list]
self.all_mask_files_json = all_mask_files_json
Expand Down
2 changes: 1 addition & 1 deletion monailabel/monailabel_HelloWorld_radiology_3dslicer.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@
"\n",
"***The Active Learning Process with MONAI Label***\n",
"\n",
"In this notebook, we provide a hello world example of MONAI Label use case. Using Spleen segmentation in Radiology app as the demonstration, 3D Slicer as the client viewer, we show how MONAI Label workflow serves as interacitve AI-Assisted tool for labeling CT scans. \n",
"In this notebook, we provide a hello world example of MONAI Label use case. Using Spleen segmentation in Radiology app as the demonstration, 3D Slicer as the client viewer, we show how MONAI Label workflow serves as interactive AI-Assisted tool for labeling CT scans. \n",
"\n",
"![workflow](./figures/monailabel_radiology_3dslicer/teaser_img.png)\n",
"\n",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@
"An example of images and corresponding reports in Open-I dataset is presented as follows [2]:\n",
"![image](../../figures/openi_sample.png)\n",
"\n",
"In this tutorial, we use the TransCheX model with 2 layers for each of vision, language mixed modality encoders respectively. As an input to the TransCheX, we use the patient **report** and corresponding **chest X-ray image**. The image itself will be divided into non-overlapping patches with a specified patch resolution and projected into an embedding space. Similarly the reports are tokenized and projected into their respective embedding space. The language and vision encoders seperately encode their respective features from the projected embeddings in each modality. Furthmore, the output of vision and language encoders are fed into a mixed modality encoder which extraxts mutual information. The output of the mixed modality encoder is then utilized for the classification application. \n",
"In this tutorial, we use the TransCheX model with 2 layers for each of vision, language mixed modality encoders respectively. As an input to the TransCheX, we use the patient **report** and corresponding **chest X-ray image**. The image itself will be divided into non-overlapping patches with a specified patch resolution and projected into an embedding space. Similarly the reports are tokenized and projected into their respective embedding space. The language and vision encoders seperately encode their respective features from the projected embeddings in each modality. Furthmore, the output of vision and language encoders are fed into a mixed modality encoder which extracts mutual information. The output of the mixed modality encoder is then utilized for the classification application. \n",
"\n",
"[1] : \"Hatamizadeh et al.,TransCheX: Self-Supervised Pretraining of Vision-Language Transformers for Chest X-ray Analysis\"\n",
"\n",
Expand Down
2 changes: 1 addition & 1 deletion self_supervised_pretraining/vit_unetr_ssl/ssl_train.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -260,7 +260,7 @@
"\n",
"model = model.to(device)\n",
"\n",
"# Define Hyper-paramters for training loop\n",
"# Define Hyper-parameters for training loop\n",
"max_epochs = 500\n",
"val_interval = 2\n",
"batch_size = 4\n",
Expand Down
Loading