Recent comments in /f/MachineLearning
darkshenron t1_jdc4wvn wrote
Reply to comment by currentscurrents in [D] Do you have a free and unlimited chat that specializes only in teaching programming or computing in general? by Carrasco_Santo
Yes from an efficient angle for sure
Armanoth t1_jdc3lyf wrote
Reply to comment by MisterManuscript in [R] Introducing SIFT: A New Family of Sparse Iso-FLOP Transformations to Improve the Accuracy of Computer Vision and Language Models by CS-fan-101
Yeah, whenever there is papers that try to redefine/takeover existing well known acronyms, i just get the sense that the goal is publicity through controversy.
I dont believe its just a coincidence, especially not when its an acronym so prominent. I mean who tries to coin a term without doing a basic Google search, let alone pick an acronym that is so well-known in the same field.
RemarkableGuidance44 t1_jdc3hut wrote
Reply to comment by immune_star in [P] CodeAlpaca Code and Data release by immune_star
Yeah I would to know what the difference is from LoRA to just Full finetune?
Armanoth t1_jdc2vt0 wrote
Reply to comment by Tejalapeno in [R] Introducing SIFT: A New Family of Sparse Iso-FLOP Transformations to Improve the Accuracy of Computer Vision and Language Models by CS-fan-101
While the paper is good and definetly presents some novel approach. Re-using existing acronyms, especially such prominent ones. The main purpose of these acronyms to allow for readers to easily identify and reference existing methods.
If your choice of acronym forces all subsequent research to have to elaborate on which SIFT is mentioned, it is not only a poor choice but also a point of confusion. And existing papers that mention SIFT are retroactively affected.
As many in this thread has pointed out, there are other equally catchy, non-overlapping acronyms that could have been chosen.
RemarkableGuidance44 t1_jdc2opy wrote
Reply to comment by immune_star in [P] CodeAlpaca Code and Data release by immune_star
Yeah I was wondering why you did not release them as its allowed as you are not selling it. :)
of_patrol_bot t1_jdc2kt9 wrote
Reply to comment by RemarkableGuidance44 in GPT-4 For SQL Schema Generation + Unstructured Feature Extraction [D] by Mental-Egg-2078
Hello, it looks like you've made a mistake.
It's supposed to be could've, should've, would've (short for could have, would have, should have), never could of, would of, should of.
Or you misspelled something, I ain't checking everything.
Beep boop - yes, I am a bot, don't botcriminate me.
RemarkableGuidance44 t1_jdc2k75 wrote
Reply to comment by breadbrix in GPT-4 For SQL Schema Generation + Unstructured Feature Extraction [D] by Mental-Egg-2078
haha exactly, the guy has never worked with data. Just imagine getting an Audit and not knowing if your data is right or not. It could of messed up big time and cost 100's of thousands.
Individual-Road-5784 t1_jdc0z0j wrote
Reply to comment by _Arsenie_Boca_ in [P] Open-source GPT4 & LangChain Chatbot for large PDF docs by radi-cho
Instead of FAISS, you can also use a truly vector search database like Qdrant. It's open-source and also offers a generous free tier offering in the cloud https://qdrant.tech
_Arsenie_Boca_ t1_jdc0ko2 wrote
Reply to comment by Icko_ in [P] Open-source GPT4 & LangChain Chatbot for large PDF docs by radi-cho
True, but im not sure how much cheaper that would really be.
Icko_ t1_jdc09e5 wrote
Reply to comment by _Arsenie_Boca_ in [P] Open-source GPT4 & LangChain Chatbot for large PDF docs by radi-cho
You could use faiss instead of pinecone and alpaca instead of gpt-4
weaponized_lazyness t1_jdbvolq wrote
Reply to [D] Simple Questions Thread by AutoModerator
Is there a subreddit for more academic discussions on ML? This space has now been swarmed by LLM enthusiasts, which is fine but it's not the content I was looking for.
MisterManuscript t1_jdbuih4 wrote
Reply to comment by tdgros in [R] Introducing SIFT: A New Family of Sparse Iso-FLOP Transformations to Improve the Accuracy of Computer Vision and Language Models by CS-fan-101
I stand corrected regarding the patent. The naming conflict, on the other hand, is here to stay.
tdgros t1_jdbu6w7 wrote
Reply to comment by MisterManuscript in [R] Introducing SIFT: A New Family of Sparse Iso-FLOP Transformations to Improve the Accuracy of Computer Vision and Language Models by CS-fan-101
the SIFT patent expired in March, 2020. It's included in openCV now (it used to be in a "non free" extension of openCV)
fnordstar t1_jdbtg4x wrote
Reply to comment by GaryS2000 in [D] Simple Questions Thread by AutoModerator
That is an image dataset. What are you even training on if you're not using the images?
_Arsenie_Boca_ t1_jdbsl4b wrote
What are the costs for all the services? I assume GPT-4 is billed per request and Pinecone per hour?
KerfuffleV2 t1_jdbrkc1 wrote
Reply to comment by Unlucky_Excitement_2 in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
Uh, did you reply to the wrong person or something? Your post doesn't have anything to do with either mine or the parent.
Mxbonn t1_jdbr3gz wrote
Reply to comment by midasp in [D] ICML 2023 Reviewer-Author Discussion by zy415
Is there anything I can do if they don't reply by the saturday? Should I message the AC?
BreadSugar t1_jdbqq4d wrote
I love these chatbots especially for their cute portraits. lol
[deleted] t1_jdbpst7 wrote
localhost80 t1_jdbmrtf wrote
Reply to [P] One of the best ChatGPT-like models (possibly better than OpenAssistant, Stanford Alpaca, ChatGLM and others) by [deleted]
Nice clickbait title!
I wish you the best of luck on your journey as a student but no need to hype up your project with insane claims.
Disastrous_Elk_6375 t1_jdblohr wrote
Reply to comment by crt09 in [P] One of the best ChatGPT-like models (possibly better than OpenAssistant, Stanford Alpaca, ChatGLM and others) by [deleted]
Hahaha, took me a couple of prompts to see what you did there. Interesting approach, I guess you're looking to see if it can "learn" labels from the prompt and overcome the bias of training (e.g. associating good&awesome, etc with "positive").
[deleted] OP t1_jdbiemc wrote
msgs t1_jdbi9r2 wrote
Reply to [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
magnet:?xt=urn:btih:6K5O4J7DCKAMMMAJHWXQU72OYFXPZQJG&dn=ggml-alpaca-30b-q4.bin&xl=20333638921&tr=udp%3A%2F%2Ftracker.openbittorrent.com%3A80%2Fannounce
I hope this magnet link works properly. I've never created one before. This the alpaca.cpp 30B 4-bit weight file. Same file downloaded from huggingface. Apologies if it doesn't work. Ping me if it doesn't.
[deleted] t1_jdbg14k wrote
Reply to comment by jakderrida in [R] Introducing SIFT: A New Family of Sparse Iso-FLOP Transformations to Improve the Accuracy of Computer Vision and Language Models by CS-fan-101
[removed]
jcansdale2 t1_jdc5lte wrote
Reply to comment by sam__izdat in Modern language models refute Chomsky’s approach to language [R] by No_Draft4778
> Literally none of these cut-and-paste color-texture-animal-verb-adverb responses satisfies the query, and to get even a little bit closer with GPT, you have to explicitly beg for examples of contradictions and category errors.
What do you think of this exchange?
Q: Why is the sentence “colorless green ideas sleep furiously” interesting?
The sentence "Colorless green ideas sleep furiously" is interesting because it was constructed by Noam Chomsky, a renowned linguist and philosopher, as an example of a sentence that is grammatically correct but semantically nonsensical. It appears in his 1957 book "Syntactic Structures" to illustrate the distinction between syntax and semantics.
Syntax refers to the rules governing how words are combined to form sentences, whereas semantics deals with the meaning of words and sentences. Chomsky's sentence demonstrates that a sentence can follow the rules of grammar and be well-formed syntactically, but still not convey any coherent meaning or make logical sense.
This example has become famous in the fields of linguistics, cognitive science, and philosophy of language, as it highlights the importance of studying both syntax and semantics in order to understand human language fully.
Q: Generate ten grammatically correct but semantically nonsensical sentences.
Q: Sort these in order of nonsensicalness.
Sorting these sentences in order of nonsensicalness can be subjective, as different people may perceive various aspects of these sentences differently. However, here's one possible ordering, starting with the least nonsensical to the most nonsensical:
Remember that this ordering is subjective and open to interpretation. Others might find a different arrangement more fitting.
Q: Explain your reasoning.
My reasoning for the ordering of these sentences is based on the degree of disconnection between the words and concepts in each sentence. I've considered the relationship between subjects, verbs, and objects, as well as the overall coherence of the images or ideas evoked.
Keep in mind that this reasoning is subjective, and other interpretations might lead to a different ordering of nonsensicalness.
Edit: fixed formatting