Recent comments in /f/MachineLearning

aadityaubhat OP t1_jdxhe79 wrote

Sure, currently it supports aggregation and visualization, I am working to add more functionality to it.
The core process of Auto-Analyst consists of several steps:

  1. ⁠Parsing the data, description, and question: The tool takes your data and a plain English question as input, then parses and understands the context.
  2. ⁠Basic data cleaning: Before diving into the analysis, Auto-Analyst cleans the data to ensure it's ready for processing.
  3. ⁠Determining the answer type: Based on the input question, Auto-Analyst figures out if the answer can be provided through aggregation or visualization.
  4. ⁠Aggregation: If the question requires an aggregated answer, Auto-Analyst leverages the OpenAI API to generate an SQL query. It then tries running the query on the data. If it fails, the OpenAI API is used to correct the query. This process continues until a working query is obtained or the user-defined maximum number of tries is reached. The aggregation results are then returned to the user.
  5. ⁠Visualization: If the question calls for a plot, Auto-Analyst first identifies the aggregated data needed for the visualization. It uses the aggregation steps described above to obtain this data. Next, it employs the OpenAI API to generate Python code for the plot and returns the visualization to the user.
7

suflaj t1_jdxh3kc wrote

That would be breaching copyright. Depending on the company and the product, you'd get anywhere from a pretty nasty debt to straight up ruining your whole life (and potentially the lives of your family and people associated with you).

The same way you wouldn't steal from the mob, you would not steal from a company that makes money on a product FOSS can't compete with. Aside from that, decompilers exist for a very long time yet we have not witnessed such vigilantism.

−5

suflaj t1_jdxgd3z wrote

In most cases yes, but inherently no. Understand that compilers, as part of their optimization step, might compile high level code into something that you can't really connect with the actual code. Part of the information is lost in the optimization step and so in a general case you will not be able to revert the compilation step. At least not fully, of course you will be able to get something resembling the solution, but it is not guaranteed to be the exact code that compiled into your starting input.

This is, of course, after taking into consideration you will not be able to recover dead source code if it's never compiled into something. Because if you take this into account, even if a language does not optimize the source code otherwise, if it only discards dead code: you are also losing information.

And also, this is disregarding name mangling. Obviously name mangling can be done in a way you have information loss, but this is probably irrelevant since concrete entity names are not that relevant.

2

a_marklar t1_jdx7ryn wrote

In a limited sense, we're already there. For example, the microsoft avatar generation.

I'd guess that its very unlikely that generative models will use triangles. Point clouds, SDFs, parametric surfaces all seem to be better data formats for these types of things. Those can all be converted to triangle meshes if that's required.

3

12max345 t1_jdx4sdm wrote

I feel like LLMs have encoded sort of law of a languages in their latent space through texts and responding accordingly, anything that follows a law isnt called concious for e.g inaminate objects follow law of physics,but that doesnt mean that it indicates an intelligent behvaiour.

After all texts are medium to represent our thoughts, its the thoughts that matter not the medium.

The concept of causality , fundamental reality , and dcesion making is much more than following laws of languages which are just a means.

These LLMs cant question you until you ask them explicitly,they cant interject you , knowledge was never consciousness ,its these abilities that compose consciousness

I dont know how much sense i make to others or maybe i am at loss of good words,in a nutshell any model that fundamentally predicts tokens based of weightage of previous tokens can never achieve consciousness. We

1

was_der_Fall_ist t1_jdwz4qw wrote

I think you make a good point. We probably need better methods of post-training LLMs. But it does seem like the current regime is still sometimes more useful than the pre-trained model, which Christiano also says. It's only in some contexts that this behavior is worse. I'm not sure if it's really better than top-p sampling, though. I'm not sure that it is. But RLHF models do seem pretty useful.

2