[go: nahoru, domu]

Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is it possible to use transformers.js to implement audio source separation tasks? #788

Open
asasas234 opened this issue Jun 2, 2024 · 2 comments
Labels
question Further information is requested

Comments

@asasas234
Copy link

Question

Hello, I have a beginner's question.

I want to implement the task of removing the human voice from the audio in the video and retaining the background sound in the browser. The idea is to load the model for audio source separation related to transformers.js to achieve the separation of the background sound and human voice, and then only return the background sound.

But I couldn't find relevant examples in the documentation, so I was wondering if this can be implemented? If so, what are the learning or research paths?

Looking forward to your reply

@asasas234 asasas234 added the question Further information is requested label Jun 2, 2024
@xenova
Copy link
Owner
xenova commented Jun 3, 2024

Hi there 👋 This library serves as a JavaScript port of the Python transformers library, so if you know of a model where you can do this, we can certainly look into it! Is something like https://huggingface.co/speechbrain/sepformer-wham what you're looking for?

@asasas234
Copy link
Author

Yes, speechbrain looks good, but I think demucs would be the best. However, I'm a beginner in machine learning and Python, so I'm looking for the simplest solution that can achieve my goal

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants