Recent comments in /f/MachineLearning
rising_pho3nix t1_jd2jjsz wrote
Reply to [Project] Machine Learning for Audio: A library for audio analysis, feature extraction, etc by Leo_D517
This is nice.. I'm doing MIR as part of my Thesis work. Will definitely use this.
Straight-Comb-6956 t1_jd2iwp6 wrote
Reply to comment by currentscurrents in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
> Llamma.cpp uses the neural engine,
Does it?
starstruckmon t1_jd2hzl6 wrote
Reply to comment by Carrasco_Santo in [P] OpenAssistant is now live on reddit (Open Source ChatGPT alternative) by pixiegirl417
It's unlikely that the main problem is the RHLF data and not the base model.
1azytux OP t1_jd2ho88 wrote
Reply to comment by aozorahime in Recent advances in multimodal models: What are your thoughts on chain of thoughts models? [D] by 1azytux
i'm looking for ideas based on the papers given :
- Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering
- Multimodal Chain-of-Thought Reasoning in Language Models
and such .. with general chain of thought idea for language can be looked at this paper.
I'm not sure if the link you provided will work, but as it's huge I might have missed (I've glanced on it) can you point out the parts which you think should be paid attention?
fanjink t1_jd2ho6o wrote
Reply to comment by Leo_D517 in [Project] Machine Learning for Audio: A library for audio analysis, feature extraction, etc by Leo_D517
Thank you, I’ll try it later
Leo_D517 OP t1_jd2hhov wrote
Reply to comment by fanjink in [Project] Machine Learning for Audio: A library for audio analysis, feature extraction, etc by Leo_D517
First of all, we have noticed this issue and it will be resolved in the upcoming next version. For now, you can install by compiling the source code.
Please follow the steps in the Document to compile the source code.
The steps are as follows:
- Installing dependencies on macOS
Install Command Line Tools for Xcode. Even if you install Xcode from the app store you must configure command-line compilation by running:
xcode-select --install - Python setup:
$ python setup.py build
$ python setup.py install
fanjink t1_jd2ghpk wrote
Reply to [Project] Machine Learning for Audio: A library for audio analysis, feature extraction, etc by Leo_D517
This library looks great, but I get this:
OSError: dlopen(/Users/***/opt/anaconda3/envs/audio/lib/python3.9/site-packages/audioflux/lib/libaudioflux.dylib, 0x0006): tried: '/Users/***/opt/anaconda3/envs/audio/lib/python3.9/site-packages/audioflux/lib/libaudioflux.dylib' (mach-o file, but is an incompatible architecture (have (x86_64), need (arm64e)))
AlexMan777 t1_jd2g76o wrote
Is their actual dataset available which the current model was trained on? It would be great to try in on other models to compare results.
Leo_D517 OP t1_jd2g6pg wrote
Reply to comment by CheekProfessional146 in [Project] Machine Learning for Audio: A library for audio analysis, feature extraction, etc by Leo_D517
First, librosa is a very good audio feature library.
The difference between audioflux and librosa is that:
- Systematic and multi-dimensional feature extraction and combination can be flexibly used for various task research and analysis.
- High performance, core part C implementation, FFT hardware acceleration based on different platforms, convenient for large-scale data feature extraction.
- It supports the mobile end and meets the real-time calculation of audio stream at the mobile end.
Our team wants to do audio MIR related business at mobile end, all operations of feature extraction must be fast and cross-platform support for the mobile end.
For training, we used the librosa method to extract CQT-related features at that time. It took about 3 hours for 10000 sample data, which was really slow.
Here is a simple performance comparison
Server hardware:
- CPU: AMD Ryzen Threadripper 3970X 32-Core Processor
- Memory: 128GB
Each sample data is 128ms(sampling rate: 32000, data length: 4096).
The total time it takes to extract features from 1000 sample data.
| Package | audioFlux | librosa | pyAudioAnalysis | python_speech_features |
|---|---|---|---|---|
| Mel | 0.777s | 2.967s | -- | -- |
| MFCC | 0.797s | 2.963s | 0.805s | 2.150s |
| CQT | 5.743s | 21.477s | -- | -- |
| Chroma | 0.155s | 2.174s | 1.287s | -- |
Finally, audioflux has been developed for about half a year, and open source has only been more than two months. There must be some deficiencies and improvements. The team will continue to work hard to listen to community opinions and feedback.
Thank you for your participation and support. We hope that the follow-up of the project will be better and better.
CheekProfessional146 t1_jd2faom wrote
Reply to [Project] Machine Learning for Audio: A library for audio analysis, feature extraction, etc by Leo_D517
Very good, but what is the difference between it and librosa
[deleted] t1_jd2f97u wrote
Reply to [D] Simple Questions Thread by AutoModerator
[deleted]
xbcslzy t1_jd2eyo7 wrote
Reply to [Project] Machine Learning for Audio: A library for audio analysis, feature extraction, etc by Leo_D517
Nice, hope it helps me in my work
Nikelui t1_jd2eapi wrote
Reply to comment by TimelySuccess7537 in [P] TherapistGPT by SmackMyPitchHup
>People are going to consult with tools like ChatGPT about their mental health anyway regardless of what he does, people are already doing it with Google so why not ChatGPT that can actually talk to you, remember what you said etc.
Because that's outside the scope of both Google and chatGPT. If you are marketing your tool as a therapist aid and you don't have a license, you are probably breaking more laws than you can afford to.
baffo32 t1_jd2carr wrote
Reply to comment by Ayacyte in [P] OpenAssistant is now live on reddit (Open Source ChatGPT alternative) by pixiegirl417
sign in to the website to contribute data and corrections. join the discord to contribute in other ways.
gliptic t1_jd2bsc7 wrote
Reply to comment by lurkinginboston in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
In fact, GPT3 is 175B. But GPT3 is old now and doesn't make effective use of those parameters.
Nezarah t1_jd297zo wrote
Reply to Smarty-GPT: wrapper of prompts/contexts [P] by usc-ur
Is this essentially In-Context Learning?
You condense additional knowledge as a prefix to the prompt as “context” so that the question/input can use that information to create a more accurate/useful output?
TimelySuccess7537 t1_jd27pk3 wrote
Reply to [P] TherapistGPT by SmackMyPitchHup
Looks like such tools will eventually exist and be widely used, it's inevitable. Whether you are the one to succeed doing that, that's a matter of ambition, market fit, luck etc. It's not clear the people are ready for this now but they will be eventually.
Good luck!
TimelySuccess7537 t1_jd27klp wrote
Reply to comment by save_the_panda_bears in [P] TherapistGPT by SmackMyPitchHup
How so ?
He can make the users sign some waiver. People are going to consult with tools like ChatGPT about their mental health anyway regardless of what he does, people are already doing it with Google so why not ChatGPT that can actually talk to you, remember what you said etc.
Sure this thing needs to be tested thoroughly but I really don't see why everyone is so outraged about this - psychotherapy is expensive and is not a right fit for everyone, maybe these tools can help people.
If some psychologist tested this app you would be cool with it? I'm sure some psychologist will eventually vouch for such a tool.
btw actual psychotherapy is not only expensive but ineffective way too often https://www.psychreg.org/why-most-psychotherapies-equally-ineffective/
ReasonablyBadass t1_jd26flf wrote
What was the hardware this was trained on? Boinc like distribution?
And what are the hardware requirements for running it locally?
aozorahime t1_jd26b23 wrote
Reply to Recent advances in multimodal models: What are your thoughts on chain of thoughts models? [D] by 1azytux
is it similar to multi modal deep learning? because this is what I am currently studying. you can check this paper for a brief explanation https://arxiv.org/pdf/2301.04856.pdf
for chain of thought models, could you elaborate about this?
timedacorn369 t1_jd25c76 wrote
Reply to comment by yahma in [P] OpenAssistant is now live on reddit (Open Source ChatGPT alternative) by pixiegirl417
Yeah I hope so. Compared to other chat assistants this seems to have a much better rlhf part but somewhat bad text completion based on my limited analysis of using them . So if they use LlaMa I think the output would substantially improve .
yahma t1_jd24u9z wrote
Would switching the base pythia-12b model for llama-13b improve things?
shafall t1_jd2380o wrote
Reply to comment by Enturbulated in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
To give some more specifics, most of the time its not the CPU that copies the data on modern systems, it is the PCI DMA chip (that may be on the same die though). CPU just sends address ranges to DMA Info
BarockMoebelSecond t1_jd22r8w wrote
Reply to comment by Ayacyte in [P] OpenAssistant is now live on reddit (Open Source ChatGPT alternative) by pixiegirl417
I do hope that we can disable these limitations in the future. That's what I'm really looking forward to with OSS LLMs.
Leo_D517 OP t1_jd2k6u1 wrote
Reply to comment by rising_pho3nix in [Project] Machine Learning for Audio: A library for audio analysis, feature extraction, etc by Leo_D517
Thank you for your support. If you are interested, you can join our project. Suggestions and feedback are welcome.