Multimodal learning
Jump to navigation
Jump to search
Multi-modal learning is a type of machine learning that involves the use of multiple types of data, such as images, text and audio. By using different modalities of data robowaifus can learn the context of a situation and interpret the meanings of various inputs better. This allows them to more adequately respond to commands and requests, as well as to better anticipate the needs of users. Additionally, they can also use the various data received to more accurately predict the user’s next move.