Release date: 2020-02-25
Vector length: 2048 (default)
#pip install vectorhub[encoders-audio-tfhub] from vectorhub.encoders.audio.tfhub import TrillDistilled2Vec model = TrillDistilled2Vec() sample = model.read('https://vecsearch-bucket.s3.us-east-2.amazonaws.com/voices/common_voice_en_2.wav') model.encode(sample)
Index and search your vectors easily on the cloud using 1 line of code!
username = '<your username>' email = '<your email>' # You can request an api_key using - type in your username and email. api_key = model.request_api_key(username, email) # Index in 1 line of code items = ['https://vecsearch-bucket.s3.us-east-2.amazonaws.com/voices/common_voice_en_69.wav', 'https://vecsearch-bucket.s3.us-east-2.amazonaws.com/voices/common_voice_en_99.wav', 'https://vecsearch-bucket.s3.us-east-2.amazonaws.com/voices/common_voice_en_10.wav', 'https://vecsearch-bucket.s3.us-east-2.amazonaws.com/voices/common_voice_en_5.wav'] model.add_documents(user, api_key, items) # Search in 1 line of code and get the most similar results. model.search('https://vecsearch-bucket.s3.us-east-2.amazonaws.com/voices/common_voice_en_69.wav') # Add metadata to your search metadata = None model.add_documents(user, api_key, items, metadata=metadata)
The ultimate goal of transfer learning is to reduce labeled data requirements by exploiting a pre-existing embedding model trained for different datasets or tasks. The visual and language communities have established benchmarks to compare embeddings, but the speech community has yet to do so. This paper proposes a benchmark for comparing speech representations on non-semantic tasks, and proposes a representation based on an unsupervised triplet-loss objective. The proposed representation outperforms other representations on the benchmark, and even exceeds state-of-the-art performance on a number of transfer learning tasks. The embedding is trained on a publicly available dataset, and it is tested on a variety of low-resource downstream tasks, including personalization tasks and medical domain. The benchmark, models, and evaluation code are publicly released.
If you are using this in colab and want to save this so you don't have to reload, use:
import os os.environ['TFHUB_CACHE_DIR'] = "drive/MyDrive/" os.environ["TFHUB_MODEL_LOAD_FORMAT"] = "COMPRESSED"