description |
---|
β βBasics of Sequence Models |
- Sequences are data structures where each example could be seen as a series of data points, for example π§:
Task | Input X | Output Y | Type |
---|---|---|---|
π¬ Speech Recognition | Wave sequence | Text sequence | Sequence-to-Sequence |
πΆ Music Generation | Nothing / Integer | Wave Sequence | One-to_Sequence |
π Sentiment Classification | Text Sequence | Integer Rating (1β‘5) | Sequence-to-One |
π Machine Translation | Text Sequence | Text Sequence | Sequence-to-Sequence |
πΉ Video Activity Recognition | Video Frames | Label | Sequence-to-One |
- Since we have labeled data X and Y so all of these tasks are addressed as Supervised Learning π©βπ«
- Even in Sequence-to-Sequence tasks lengths of input and output can be different β
- Machine learning algorithms typically require the text input to be represented as a fixed-length vector π
- Thus, to model sequences, we need a specific learning framework able to:
- β Deal with variable-length sequences
- β Maintain sequence order
- β Keep track of long-term dependencies rather than cutting input data too short
- β Share parameters across the sequence (so not re-learn things across the sequence)