Recent comments in /f/MachineLearning

linverlan t1_jd9vckl wrote

I just wrote this computer science domain chatbot, it’s probably SOTA. You can just copy the code below and run it locally on your own machine. Let me know if you have any dependency issues, I can share a yaml file.

from googlesearch import search
import sys

query = ' '.join(sys.argv[1:]) + ' stackoverflow'
out = list(search(query, num=1, stop=1))

print(f"Answer is probably at: {out[0]}")
3

GaryS2000 t1_jd9tanf wrote

For my final year uni project I need to train a TensorFlow CNN on the FER-2013 dataset. When training the model on data from the .csv file instead of locally stored images the model trains significantly faster, with around 10 seconds per epoch as opposed to 10 minutes or so for the images. My question is it okay for me to use .csv data instead of locally stored images for this image classification task? I know I won't be able to apply data augmentation as easily but I can't think of any other downsides which would disqualify me from using the .csv data instead of the images

1

Nameless1995 t1_jd9scqe wrote

There would be a period of internal discussion after author-reviewer discussion period. So my guess would be (if there is no author-reviewer engangement beyond just the first rebuttal), AC (if they are willing to do their due dilligence) will simply push the reviewers privately and ask what their take is on the rebuttal. If in that private discussion nothing really happens (for example, all reviewers just go MIA), then really it might be upto meta-reviewer's personal judgment how they are going to take into account the strength of rebuttals.

2

darkshenron t1_jd9lmz6 wrote

I was looking for something similar and realised you can just apply an appropriate system prompt to GPT4 to narrow its focus. Some variant of “you are a helpful programming assistant. You help users answer questions related to programming in python language. If the question is not related to programming you decline to answer. “

3

Astaligorn t1_jd9evse wrote

That is pretty much my rebuttal, along with previous work in CV suggesting that it might not work as well because one the main hypothesis does not seem to hold in image.

​

But they decreased their grade because we did not provide experiment in CV :(

2

Desticheq t1_jd9c1fe wrote

I'm looking to apply the PEFT technique for some llm to use in my Regis AI extension that works on top of leetcode. While GPT's fine for the hints and general conversation, there are other applications like code improvement or complexity estimation where I might benefit from a customized model

2

Astaligorn t1_jd93sbx wrote

I had a difficult discussion on an NLP paper I submitted with a reviewer whose sole argument for rejecting the paper is, "you did not try your method in computer vision". (Even though it beats all baselines and SOTA in NLP, on a broader NLP benchmark than all previous works).

I hope the AC will step in and will not take this review into account :(

2

UnusualClimberBear t1_jd9109w wrote

First, they know publication is now a big circus and that most papers are clever solutions to problems that don't exist or beautiful explanations that cannot be leveraged. Acceptance is random if your work is not in the top 2% but still in the top 60%.

Publication as proof of work is toxic

1

Nondzu t1_jd8uh3n wrote

I'm looking for the same thing as you. A model designed specifically for programming or at least similar to the capabilities that ChatGPT with DAN has. Training such a model from scratch would be an incredible challenge. However, it seems to me that a similar model may already exist, I just haven't found it yet. It would be great to simply be able to upload it to LLaMA and use it

5