Hacker Newsnew | past | comments | ask | show | jobs | submit | _josh_meyer_'s commentslogin

excellent post -- I turned it into a video :) https://supabase.manatee.work/storage/v1/object/public/video...


Okay, this is mind blowing!


complete overview of the Model Context Protocol


Code and a recipe for XTTS_v1.1 GPT encoder training is released under the Mozilla Public License 2.0


XTTS model release (Text-to-Speech and voice cloning)

# From the release notes:

This model is trained on top of XTTS v1, using output masking. We mask the part of the output that is used as the audio prompt while training and don't compute loss for that segment. This helps us to resolve the hallucination issue that V1 experienced.

- Add Japanese - Resolve the hallucination issue (repeating the audio prompt) - Increased expressivity - Added ne_hifigan that was trained without denoising that brought some EQ and compression profile that might be unwanted for some use-cases


XTTS + Whisper + Mistral 7B


I didn't hear about xtts till now, cheers.


Coqui releases model weights for XTTS generative Voice model. Demo live on Huggingface


an example pipeline for midjourney / runwayML / coqui to create a video



Oh this is cool:)


hi there! what kind of better API support do you mean?


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: