Recent comments in /f/MachineLearning
[deleted] t1_jdgvy9w wrote
Reply to comment by michaelthwan_ai in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
[removed]
elprophet t1_jdgvfby wrote
Reply to comment by Emergency_Apricot_77 in [N] ChatGPT plugins by Singularian2501
Highly recommend the Wolfram blog post
paulgavrikov t1_jdguvmk wrote
Reply to [D] ICML 2023 Reviewer-Author Discussion by zy415
Same here, 3 out of 4 reviewers didn’t respond. And with all due respect, there is no excuse for that. If you agree to serve as reviewer, you also agree to do it faithfully. Unfortunately, the last part is often missed.
kross00 t1_jdgutr0 wrote
Reply to [D] Simple Questions Thread by AutoModerator
Is it feasible to train Llama 65B (or smaller models) to engage in chit-chatting in a manner that would not readily reveal whether one is conversing with an AI or a human? The AI does not need to answer highly complex questions and could decline them similarly to how a human would.
utopiah t1_jdgu9aa wrote
Reply to comment by deepneuralnetwork in [N] ChatGPT plugins by Singularian2501
Does ChatGPT actually do that currently, namely keep track of your past prompts and makes a model of your tastes or values, so that "me" here is meaningful?
PS: not sure why the downvote. Is it an offensive or idiotic question?
westingtyler t1_jdgtfb2 wrote
Reply to comment by currentscurrents in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
tell me about it. i never cared about video cards before this ai stuff, but now I want an awesome video card.
devzaya t1_jdgt8i8 wrote
Reply to comment by killver in [N] ChatGPT plugins by Singularian2501
Here is the description how https://qdrant.tech/articles/chatgpt-plugin/ To make it as you suggest, ChatGPT would need first crawl the whole documentation.
bias_guy412 t1_jdgt504 wrote
Reply to comment by hassan789_ in [D] What is the best open source chatbot AI to do transfer learning on? by to4life4
No, flan-t5 isn’t instruct fine tuned.
killver t1_jdgt1sn wrote
Reply to comment by devzaya in [N] ChatGPT plugins by Singularian2501
How exactly are you using the vector database there? It seems rather like querying the web for this info and the first example is about the docs.
sam__izdat t1_jdgsyr3 wrote
Reply to comment by SuperTimmyH in [N] ChatGPT plugins by Singularian2501
the site was kind of a buzz topic when it came out
ReasonablyBadass t1_jdgsfv5 wrote
Reply to comment by Maleficent_Refuse_11 in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
Auto-regressive and external knowledge hub aren't contraditions though, are they?
Theory of Mind: there is a recent Edan Mayer video about this exact topic
ReasonablyBadass t1_jdgs88s wrote
Reply to comment by visarga in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
Someone trying to get the word out? Or PR stunt?
Jean-Porte t1_jdgs7u3 wrote
Reply to [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
Isn't Davinci-3 GPT3 ? Is GPT-4 GPT3 trained much longer ?
[deleted] t1_jdgrv2v wrote
ghostfaceschiller t1_jdgrba9 wrote
Reply to comment by willer in [N] ChatGPT plugins by Singularian2501
Fr??? Wow what an insane oversight
Or I guess maybe they don’t wanna rack up all the extra embeddings calls, bc I assume like 100% if users would turn that feature on
dotnethero t1_jdgqv13 wrote
Reply to [D] Simple Questions Thread by AutoModerator
Hey everyone, I'm trying to figure out which parts of my code are using CPU and which are using GPU. During training, I've noticed that only about 5% of my usage is on the GPU, while the CPU usage is high. Any tips on how I can better understand what's going on with my code? Thanks in advance!
Miserable_Movie_4358 t1_jdgqcy4 wrote
Reply to comment by Nickvec in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
if you follow the argumentation line this person is referring to the model described on the published paper. In addition to that I invite you to investigate what knowledge means (Ps is not having access only to data)
bert0ld0 t1_jdgpv9k wrote
Reply to comment by fishybird in [N] ChatGPT plugins by Singularian2501
What is Langchain?
willer t1_jdgps4b wrote
Reply to comment by ghostfaceschiller in [N] ChatGPT plugins by Singularian2501
I read through the docs, and in this release, ChatGPT only calls the /query API. So you can't implement long term memory of your chats yourself, as it won't send your messages and the responses to this service. Your retrieval API acts in effect as a readonly memory store of external memories, like a document library.
bert0ld0 t1_jdgpr23 wrote
Reply to comment by SuperTimmyH in [N] ChatGPT plugins by Singularian2501
I mean Wolfram has always amazed me, it's power is insane! But I never used it much and always forgot about its existence. ChatGPT+Wolfram is next level thing! Never been more excited
blackvrocky t1_jdgpg24 wrote
Reply to comment by Danoman22 in [N] ChatGPT plugins by Singularian2501
there's a writing assistant tool called Lex that has gpt4 integrated into it.
nraw t1_jdgp1gp wrote
Reply to comment by lIllIIIllIllIIlIlllI in [N] ChatGPT plugins by Singularian2501
It's only uphill from here until the singularity my dude!
Username912773 t1_jdgp0u0 wrote
Reply to comment by SomeGuyInDeutschland in [D] Which AI model for RTX 3080 10GB? by SomeGuyInDeutschland
Just use a static website generator
Nickvec t1_jdgnrr5 wrote
Reply to comment by Maleficent_Refuse_11 in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
With the recent addition of plug-ins, GPT-4 effectively has access to the entire Internet. Doesn’t this contradict your assertion that it has no external knowledge hub?
Latter-Personality-6 t1_jdgvzzj wrote
Reply to [D] What is the best open source chatbot AI to do transfer learning on? by to4life4
You can try Haystack - https://haystack.deepset.ai/tutorials/01_basic_qa_pipeline