101010.pl is one of the many independent Mastodon servers you can use to participate in the fediverse.
101010.pl czyli najstarszy polski serwer Mastodon. Posiadamy wpisy do 2048 znaków.

Server stats:

516
active users

#neuralnets

0 posts0 participants0 posts today

#PromptEngineering is a task best left to #AI models | The Register @theregister / Thomas Claburn @thomasclaburn
theregister.com/2024/02/22/pro

"(...) Machine-learning boffins find #OpenSource #NeuralNets can optimize their own queries (...)

enlisting an #LLM to refine prompts for improved performance on benchmark tests. (...)

"The point of the research was to discover if smaller, open source models can also be used as optimizers," explained Battle, "And the answer turned out to be yes." (...)"

The Register · Prompt engineering is a task best left to AI modelsBy Thomas Claburn

I'm giving a short advertisement for my work on #NeuralNets at #JMM2024 in about 15 minutes! I applied a bit late and ended up in one of the contributed paper sessions. These are already a bit more subdued, but this one is particularly poorly attended. Perhaps it's for the best, since the videos on my YouTube channel (link in bio) are much more detailed than what I'm about to say.

My talk from yesterday on a categorical semantics for neural nets for the New York Category Theory Seminar has already been posted on YouTube! You can find it at youtube.com/watch?v=FKkpVKuspm and you can see more about this seminar at sci.brooklyn.cuny.edu/~noson/S. The preprint I mention is arxiv.org/abs/2308.00677 and the talk by Joyal which was mentioned at the end can be found at youtube.com/watch?v=MxClaWFiGK.

I made my first post to the #AI section of the arXiv this week! You can find the preprint "Discrete neural nets and polymorphic learning" at arxiv.org/abs/2308.00677.

In this paper a learning algorithm based on polymorphisms of finite structures is described. This provides a systematic way to choose activation functions for neural nets whose neurons can only act on a fixed finite set of values. These polymorphisms preserve any specified constraints imposed by the learning task in question.

This paper is the result of a 2021 REU at the University of #Rochester. I am working with a great group of students on a follow-up project right now, so videos of talks and a sequel preprint should be out soon!

arXiv.orgDiscrete neural nets and polymorphic learningTheorems from universal algebra such as that of Murskiĭ from the 1970s have a striking similarity to universal approximation results for neural nets along the lines of Cybenko's from the 1980s. We consider here a discrete analogue of the classical notion of a neural net which places these results in a unified setting. We introduce a learning algorithm based on polymorphisms of relational structures and show how to use it for a classical learning task.

After the initial excitement has set by now, more and more people have raised the issue that the GPT-3 based chatbots are able to make statements that sound convincing, but may have no bases in reality.

This makes a lot of sense to me, as AI, in its core, is based on one thing: pattern recognition.
So really, all it does is cargo cult, not learn or necessarily understand.

#gpt3#chatbot#ai

Conjecture that is likely true, and damning for large language models presuming that it is: An LLM trained strictly on truth will still confabulate, because it will break the bindings in what it saw, in the interpolation process, and continue to fabricate.

Further conjecture; if the above is (as I suspect true), will we never see fully honest LLMs (though they might be used as components in larger systems)

#gpt3#llm#language