photog.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
🌈 An inclusive place for your photos, silliness, and convos! 🌈

Administered by:

Server stats:

265
active users

#ais

1 post1 participant0 posts today

Not many people realize that the job of an AI engineer is *not* about training AIs. It used to be, yes. Just like once upon a time the job of an AI engineer was about coding heuristic rules, and before that it was just plain logic programming.

We have for some years now switched from curating training data and training #AIs on those to designing continuous processes and systems which allow the AIs to train themselves.

There are many ways to do this:
- Design families of games where AIs can compete against each others. The age of single games has been over for a while now, now it's about making AIs create masses of games each with novel rules and play those.
- Make AIs themselves curate training data from suitable sources.
- Make AIs refine the training data by evaluating it continuously, or by simply making progressive improvement steps on it.
- Make AIs perform easier inverse tasks, and then turn those around into harder tasks.
- Make AIs self-improve both in designing new kinds of tasks for example in domains like mathematics and software engineering, and also in the tasks themselves and in evaluating the quality of the solutions.

In a way, the AIs are pulling themselves out from unreality into the material world, and we're just supporting them in that.

Replied in thread

@Geri @drifthood Unless you can give me time, date, coordinates and IMO number of at least one ship in the photo when it was taken, I cannot verify that.

  • Also I only account ships with #AIS trackers and independent verification.

Those images are way too convenient and photographic...

Lots of boats, perhaps 25, all bearing the Palestinian flag
MastodonTotts (@Geri@mastodon.online)Attached: 1 image The Fleet of Steadfastness on its way to #Gaza

A new technique for LLMs has just landed: Explainable training!

Let me *explain*.

Normal supervised training works so that you show ground truth inputs and outputs to a model and then you backpropagate the error to the model weights. All in this is opaque black box. If you train with data which contains for example personally identifiable information (PII) or copyrighted contents, those will plausibly be stored verbatim in model weights.

What if we do it like this instead:

Let's write initial instructions to an LLM for generating synthetic data which resembles real data.

Then we go to the real data, and one by one show an LLM an example of a real data, and an example of the synthetic data, and the instructions used to generate the synthetic data. Then we ask it to iteratively refine those instructions to make the synthetic data resemble real data more, in the features and characteristics which matter.

You can also add reasoning parts, and instructions for not putting PII as such into the synthetic data generation instructions.

This is just like supervised learning but explainable! You'll get a document as a result which has refined instructions on how to generate better synthetic data, informed by real data, but now it's human readable and explainable!

You can easily verify that this relatively small document doesn't contain for example PII and you can use it to generate any volume of synthetic training data while guaranteeing that critical protected details in the real data do not leak into the trained model!

This is the next level of privacy protection for training AIs!

#AIs#LLMs#privacy

AIs are not going to save the world, or change it for the better. They are only there because they will work for #billionaires more cheaply than humans do. Everything today is *FOR PROFIT*.

#AIs will not save us. We need to do that ourselves.

aeon.co/essays/what-godels-inc

AeonWhat Gödel’s incompleteness theorems say about AI morality | Aeon EssaysMany hope that AI will discover ethical truths. But as Gödel shows, deciding what is right will always be our burden
Continued thread

Thus, for the first time, confirmation has been obtained that the Russian Navy openly protects the shadow fleet with military force.

#AIS data show coordinated movements of the three ships since June 16. Two shadow tankers and a Russian warship simultaneously entered the English Channel, heading to Russian ports to load oil. The operation involved the Steregushchiy-class corvette Boyky, the tanker SELVA (also known as NOSTOS/NAXOS),

Replied in thread

@thejasonhowell Here, code to parse anytime anyone sends "FEDI" in any packet in APRS

import aprslib

def callback(packet):
if "FEDI" in packet['raw']:
print(packet)

AIS = aprslib.IS("N0CALL")
AIS.connect()
#AIS.consumer(callback, raw=True)
AIS.consumer(callback)

Training #AIs, whether #LLMs or #LMMs, is all about data.

They perform tasks within the distribution of their training data, and in the same level of competence.

This also applies to reinforcement learning; if you look behind the scenes, what happens is supervised learning with extra steps. The error is still backpropagated back through the network, but the objectives are slightly different. Instead of direct task performances to imitate, we'll get the feedback from exploration and rewards, basically the most successful explorations can be understood as the data.

So, the data is where the buck stops when an #AI makes a misjudgement. When designing data pipelines, refinement processes and exploratory games for your AIs, you can forget about everything you know about the mechanistic aspects of Transformer architectures, and just simply focus on how to improve the data.

The data is better if it can be used to train better models. Better models have higher competence in skills, higher volume of relevant knowledge, and higher coverage in task generalism and flexibility. Hence the best data represents exactly those aspects maximally.

Human level is not the gold standard and neither is raw real-world data. We can design processes which refine the data without apparent limits, and make that data alive through AIs trained with it.

It is all about reasoning now. #LLM chatbots are already superhuman in their vast knowledge, and no human can ever compete with their billions of parameters in capacity.

Indeed, we aren't benchmarking their knowledge against single human level, but against the totality of humankind.

But to get the most value out of this knowledge, these #AIs need to be able to execute cognitive skills as instructed on this knowledge. Reasoning is one of such skills, "system 2" in Kahneman's categorization. But it's not just about building an architecture which is in principle capable of system 2 like thinking.

Reasoning is not a single skill, it is an open bag of skills which require mental exploration and which allow producing more knowledge out of already crystallized knowledge.

For example, there are skills needed for solving mathematical equations, multiplication, division and so on. These are different mental skills than skills needed for constructing and evaluating mathematical proofs. Furthermore, reasoning skills related to engineering tasks, scientific work, medical evaluations, and so on, are all slightly different.

Although there are some ridiculously trasferrable reasoning skills which apply across many domains, many domains also have their own reasoning skills which differ from the same in other domains.

We can teach these skills to the AIs similarly as we taught knowledge to them; first by presenting them to them in volumes, and then letting them practice them and improve while doing it.

The same as with humans, but in the end we get AIs which rival the whole of humanity in volumes of available knowledge, but also in cognitive skills, and are able to use their universalist expertise across all domains.

This alone, even without surpassing humanity's intelligence, will alone mean a golden age never seen before. Imagine the repercussions.

Continued thread

to the #Kavkaz port on the #Russian coast of the #Kerch_Strait

However, some voyages are currently delayed for unknown reasons, while other vessels have turned off the #AIS vessel identification system, so it is impossible to determine the location of the tankers.

One such vessel is the #Volgoneft_109. On December 17, just two days after the disaster involving #Volgoneft_239 and #Volgoneft_212 in the #Kerch_Strait, its captain issued a distress signal

#ADS-B #AIS #radiosonde release party livestream 19/12/24 @ 1900 CET

We would like to invite you all to join our launch webstream where we will present the new features of version 4.0

We are keen on your opinion and ideas so do not hesitate to join the mumble discussion afterwards.

The stream starts Thursday 19 December 2024 19:00 CET on chaos-consulting.de/stream.php
You can join us on our mumble server on mumble://mumble.chaos-consulting.de

chaos-consulting.dechaos-consulting