Skip to content

Commit

Permalink
Merge branch 'main' into main
Browse files Browse the repository at this point in the history
  • Loading branch information
johko authored Aug 14, 2024
2 parents a412664 + 59e8b16 commit b240fba
Show file tree
Hide file tree
Showing 3 changed files with 8 additions and 5 deletions.
8 changes: 6 additions & 2 deletions Makefile
Original file line number Diff line number Diff line change
@@ -1,5 +1,9 @@
.PHONY: quality
.PHONY: quality style

# Check code formatting
quality:
python utils/code_formatter.py --check_only
python utils/code_formatter.py --check_only

# Format code samples automatically and check if there are any problems left that need manual fixing
style:
python utils/code_formatter.py
2 changes: 1 addition & 1 deletion chapters/en/unit2/cnns/vgg.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -99,7 +99,7 @@ class VGG19(nn.Module):

def forward(self, x):
x = self.feature_extractor(x) # Pass input through the feature extractor layers
x = self.avgpool(x) # Pass Data through a pooling layer
x = self.avgpool(x) # Pass Data through a pooling layer
x = x.view(x.size(0), -1) # Flatten the output for the fully connected layers
x = self.classifier(x) # Pass flattened output through the classifier layers
return x
Expand Down
3 changes: 1 addition & 2 deletions chapters/en/unit4/multimodal-models/a_multimodal_world.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -48,8 +48,7 @@ A dataset consisting of multiple modalities is a multimodal dataset. Out of the
- Vision + Audio: [VGG-Sound Dataset](https://www.robots.ox.ac.uk/~vgg/data/vggsound/), [RAVDESS Dataset](https://zenodo.org/records/1188976), [Audio-Visual Identity Database (AVID)](https://www.avid.wiki/Main_Page).
- Vision + Audio + Text: [RECOLA Database](https://diuf.unifr.ch/main/diva/recola/), [IEMOCAP Dataset](https://sail.usc.edu/iemocap/).

Now let us see what kind of tasks can be performed using a multimodal dataset? There are many examples, but we will focus generally on tasks that contains the visual and textual
A multimodal dataset will require a model which is able to process data from multiple modalities, such a model is a multimodal model.
Now, let us see what kind of tasks can be performed using a multimodal dataset. There are many examples, but we will generally focus on tasks that contain both visual and textual elements. A multimodal dataset requires a model that is able to process data from multiple modalities. Such a model is called a multimodal model.

## Multimodal Tasks and Models

Expand Down

0 comments on commit b240fba

Please sign in to comment.