Recent comments in /f/MachineLearning
LeN3rd t1_jcgrxfp wrote
Reply to comment by denxiaopin in [D] Simple Questions Thread by AutoModerator
Strongly depends on your constraints. There are ways to get 3d geometry from a photo/video. If you have the geometry of your glasses you should be able to see if they fit, though you might have some problems with actually adjusting the glasses to fit on the face geometry. But you could also just do what you optician does and take a frontal photo of your face in a controlled environment.
LeN3rd t1_jcgrhlm wrote
Reply to comment by towsif110 in [D] Simple Questions Thread by AutoModerator
Be a little more coherent in your question please. No one has any idea about your specific setup unless you tell us what you want to achieve. I.e. RF is usually short for reinforcement learning in the AI community, not radiofrequency. If you want to classify data streams coming from drones, take a look at pattern matching and nearest neighbour methods, before you start to train up a large neural network.
LeN3rd t1_jcgqzvo wrote
Reply to comment by DreamMidnight in [D] Simple Questions Thread by AutoModerator
If you have more variables than datapoints, you will run into problems, if your model starts learning by heart. Your models overfits to the training data: https://en.wikipedia.org/wiki/Overfitting
You can either reduce the number of parameters in your model, or apply a prior (a constraint on your model parameters) to improve test dataset performance.
Since neural networks (the standard emperical machine learning tools nowadays) have a structure for their parameters, this means they can have much more parameters than simple linear regression models, but seem to run into problems, when the number of parameters in the network matches the number of datapoints. This is just empirically shown, i do not know any mathematical proves for it.
cathie_burry t1_jcgqwkn wrote
Reply to [P] nanoT5 - Inspired by Jonas Geiping's Cramming and Andrej Karpathy's nanoGPT, we fill the gap of a repository for pre-training T5-style "LLMs" under a limited budget in PyTorch by korec1234
How does it compare to current large language models in terms of efficacy etc.
cathie_burry t1_jcgqrcx wrote
LeN3rd t1_jcgq97y wrote
Reply to comment by No_Complaint_1304 in [D] Simple Questions Thread by AutoModerator
You will need more than a week. If you just want to predict the next word in a sentence, take a look at large language models. ChatGPT being one of them. BERT is a research alternative afaik. If you aim to learn the probabilities yourself, you will need at least a few months.
In general what you want is a generative model that can sample from the conditional probability distribution. In sequences usually transformers like BERT and chatgpt are state of the art. You can also take a look at normalizing flows and diffusion models to learn probability distributions. But this needs some maths, and i unfortunatly do not know what smaller models can be used for computational linguistic applications like this.
LeN3rd t1_jcgp44s wrote
Reply to comment by Sonicxc in [D] Simple Questions Thread by AutoModerator
How big is your dataset? Before you start anything wild, i would look at kernel clustering methods. Or even clustering without kernels. Just cluster your broken and non broken images and calculate some distance (can be done with kernels if it needs to be nonlinear).
Also Nearest neighbor could work pretty well in your case. Just compare your new image to the closest (according to some metric) in your two datasets and bobs your uncle.
If you need a number, look at simple CNNs. you need more training data though for this to work well.
impossiblefork t1_jcgp1dt wrote
Reply to [P] nanoT5 - Inspired by Jonas Geiping's Cramming and Andrej Karpathy's nanoGPT, we fill the gap of a repository for pre-training T5-style "LLMs" under a limited budget in PyTorch by korec1234
So this is actually cheap? About 20 USD?
LeN3rd t1_jcgo5ro wrote
Reply to comment by PhysZhongli in [D] Simple Questions Thread by AutoModerator
You should take a look at uncertainty in general. What you are trying to do is calculate epistemic uncertainty. (google epistemic vs aleatoric uncertainty).
One thing that works well is to have a dropout layer, that is active during prediction!! (in tensorflow you have to feed training=True into the call to activate it during prediction). Sample like 100 times and calculate the standard deviation. This gives you a general "i do not know" function from the network. You can also do so by training 20 models and letting them output 20 different results. With this you can assign the 101 label, when the uncertainty is too high.
In my experience you should stay away from bayesian neural networks, since the are extremly hard to train, and cannot model multimodal uncertainty. (dropout can neither, but is WAAAAYYY easier to train).
throwawaychives t1_jcgnhuf wrote
Reply to comment by LightbulbChanger25 in [N] PyTorch 2.0: Our next generation release that is faster, more Pythonic and Dynamic as ever by [deleted]
PyTorch docs are more than enough to learn torch, especially if you have good experience in other ML frameworks. Nothing will beat implementing an actual model in torch and there are plenty of GitHub repos out there you can use as a reference
LeN3rd t1_jcgn73n wrote
Reply to [D] Simple Questions Thread by AutoModerator
Can anyone recommend a good, maintained and well organized MCMC python package? Everything i found was either not maintained, had only a single research group behind it, or had to many bugs for me to continue with that project. I want Tensorflow/Pytorch, but for MCMC sampling please.
sayoonarachu t1_jcgjosz wrote
Reply to comment by CyberDainz in [N] PyTorch 2.0: Our next generation release that is faster, more Pythonic and Dynamic as ever by [deleted]
Doesn't work on Windows in 2.1 dev either, fyi.
LightbulbChanger25 t1_jcgir88 wrote
Reply to [N] PyTorch 2.0: Our next generation release that is faster, more Pythonic and Dynamic as ever by [deleted]
I think 2.0 is a good moment to add pytorch to my list of skills. Are there any good resources to learn pytorch 2.0 yet? I would consider myself between intermediate and advanced in tensorflow.
bohreffect t1_jcgcr96 wrote
Seeing who constitutes high profile names in "AI Ethics", having watched individual debacles unfold in the industry over the last few years, it doesn't instill confidence that they're particularly valuable. It's very, very hard not to have cynical takes about the people bubbling to the top.
In general I don't think unaccountable focus groups of people are the best moral arbiters either. In this instance I feel we have to go with our least worst option deferring to the wisdom of crowds.
[deleted] t1_jcgbolc wrote
[deleted]
fteem t1_jcg3zlh wrote
Reply to [D] Simple Questions Thread by AutoModerator
What happened with the WAYR (What Are You Reading) threads?
SomewhereAtWork t1_jcg3kak wrote
Reply to comment by UnusualClimberBear in [D] Is it possible to train LLaMa? by New_Yak1645
Thank you!
janpaul123 t1_jcg2kqk wrote
Reply to comment by mostancient in [N] A $250k contest to read ancient Roman papyrus scrolls with ML by nat_friedman
Given that the villa was likely owned by a Roman consul and senator, that could make for some exciting accounting!
janpaul123 t1_jcg1w91 wrote
Reply to comment by Disastrous_Elk_6375 in [N] A $250k contest to read ancient Roman papyrus scrolls with ML by nat_friedman
Yes, see for example the "carbon phantom scroll" used in this paper: https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0215775
Though I don't think attempts at the same resolution (4-8µm) have been made.
1F9 t1_jcfxc5b wrote
Reply to comment by -Rizhiy- in [N] PyTorch 2.0: Our next generation release that is faster, more Pythonic and Dynamic as ever by [deleted]
That reason is that they replaced Lua with Python as the high-level language that wrapped Torch's core, and needed to differentiate that from the original Torch. But it seems as though you believe the "py" prefix means the correct design decision for the project is to replace ever more parts of torch with Python. Perhaps you could elaborate more on your thinking there?
Empty-Revolution7570 OP t1_jcfx1wv wrote
Reply to comment by ml_head in [P] Multimedia GPT: Can ChatGPT/GPT-4 be used for vision / audio tasks just by prompt engineering? by Empty-Revolution7570
That makes sense!
Based on how it works it think original stories would also work.
thedabking123 t1_jcfupuc wrote
Reply to comment by Hydreigon92 in In your experience, are AI Ethics teams valuable/effective? [D] by namey-name-name
thank god that only applies to giant platforms... Our firm would crumble in the face of that.
WikiSummarizerBot t1_jcfub6d wrote
Reply to comment by currentscurrents in Modern language models refute Chomsky’s approach to language [R] by No_Draft4778
>In neuropsychology, linguistics, and philosophy of language, a natural language or ordinary language is any language that has evolved naturally in humans through use and repetition without conscious planning or premeditation. Natural languages can take different forms, such as speech or signing. They are distinguished from constructed and formal languages such as those used to program computers or to study logic.
>In logic, mathematics, computer science, and linguistics, a formal language consists of words whose letters are taken from an alphabet and are well-formed according to a specific set of rules. The alphabet of a formal language consists of symbols, letters, or tokens that concatenate into strings of the language. Each string concatenated from symbols of this alphabet is called a word, and the words that belong to a particular formal language are sometimes called well-formed words or well-formed formulas. A formal language is often defined by means of a formal grammar such as a regular grammar or context-free grammar, which consists of its formation rules.
^([ )^(F.A.Q)^( | )^(Opt Out)^( | )^(Opt Out Of Subreddit)^( | )^(GitHub)^( ] Downvote to remove | v1.5)
currentscurrents t1_jcfu9l8 wrote
Reply to comment by sam__izdat in Modern language models refute Chomsky’s approach to language [R] by No_Draft4778
That's why it's a natural language instead of a formal language.
LeN3rd t1_jcgs51v wrote
Reply to comment by wikipedia_answer_bot in [D] Simple Questions Thread by AutoModerator
don't hurt me