youtube

There’s a great video on Youtube(above) called ‘The War on Science’ by ASAPScience which outlines an oft misunderstood conflict. That conflict is when:

“Science and society are often at odds”.

Putting the conflict in these terms clearly shows the wrong-headed thinking of those who are undercutting the intellectual values of science with the social values of society. Current social norms may be more convenient to defend and continue for society but it not intelligent to continue thinking the same thing when evidence shows otherwise.

In fact, rather than simply wrong-headed, such defence of social values in the face of intellectual values to the contrary, is immoral and not supported by the MOQ.

The historical risk though, is that without the Metaphysics of Quality the intellectual level can start to undercut the quality of society and defend biological values at the risk of social cohesion. This could well explain why many a political conflict throughout the world simply are between those who defend social values vs those who support intellectual ones.

The MOQ however, shows there is a more nuanced way to view social vs intellectual conflicts such as this. Within the structure of the MOQ is the ability to morally defend intellectual values while not risking social decay in the process. This is clearly shown with the MOQ’s ‘Codes of Morality’ and in the difference between ‘The Law’ and ‘Intellectual Morality’ the latter of which is not acknowledged with our current Metaphysics.

DON'T FORGET TO SHARE THIS POST!

TOPICS:
youtube

I’ve seen lots of talk recently about the moral threat of AI. So, what does the MOQ have to say about it?

To start with, how about a fact which appears to be lost in much of the discussion.

No computer has ever made a moral judgement which it hasn’t been told to make and so there is no reason to think this will ever change. Believing this will change spontaneously as a result of improved intelligence of machines is just that, a leap of faith, and not supported by evidence. As it stands, it is the human programmer making all moral judgements of consequence. Computers, being 0’s and 1’s, are simply the inorganic tools of the culturally moral programmer.

Unfortunately though, this isn’t likely to be appreciated any time soon because of a philosophical blind spot our culture has. That blind spot is our metaphysics which neglects the fundamental nature of morality and in doing so gets confused about both where morality comes from and whether machines can make moral judgements independently of being instructed to do so.

For example, in the case of a recent foreign affairs article – Nayed Al-Rodhan appears to believe that AI will start making moral judgements as a result of more ‘sophistication’ and learning and experience.

“Eventually, a more sophisticated robot capable of writing its own source code could start off by being amoral and develop its own moral compass through learning and experience.”

The MOQ however makes no such claim which, as already mentioned, is contrary to our experience. According to our experience it is only human beings and higher primates who can make social moral judgments in response to Dynamic Quality. Machines are simply inorganic tools and their components only make ‘moral decisions’ at the inorganic level.

That’s not to say though, that there aren’t any dangers of AI and that all risks are overblown. AI – being loosely defined as advanced computational/mechanical decision not requiring frequent human input – threatens society if it is either poorly programmed and a catastrophic sequence of decisions occurs or if it is well programmed by a morally corrupt programmer. However each of these scenarios aren’t fundamentally technological but philosophical, psychological & legal in nature.

The unique threat of AI is this aforementioned increase in freedom of machines to make decisions without human intervention making them both more powerful and more dangerous. The sooner our culture realises this, the sooner our culture can start to discuss these moral challenges and stop worrying about the machines ‘taking over’ in some kind of singularity apocalypse. Because unfortunately, if we don’t understand the problem, a solution will be wanting, and therein lies the real threat of AI.

DON'T FORGET TO SHARE THIS POST!

TOPICS:
_video

The New Age movement and the MOQ on first glance are well suited for one another. As Wikipedia says:

‘The movement aims to create “a spirituality without borders or confining dogmas” that is inclusive and pluralistic. It holds to “a holistic worldview,” emphasizing that the Mind, Body, and Spirit are interrelated.’

However the problem begins with the addition of the term ‘spirit’. It’s a slippery word which can lead to folks getting burned at the stake because of their bad ‘spirits’. Once ‘spirit’ is brought into the conversation you can say goodbye to the empirically verifiable strength of science and hello to pseudoscience and anyone bringing in whatever pseudo-scientific thing they see fit.

The term value on the other hand cannot be co-opted in such a way. Anything can be valuable but using the MOQ we can break value up into a hierarchy of values, each level of which are part of experience and can be empirically verified.

There is an advantage to the New Age movement however and that’s in its ability to create a space where science, and specifically quantum mechanics, can be discussed in an easily accessible way way that won’t shy away from any apparent logical inconsistencies.

In a popular New Age movie called ‘What the bleep do we know?’ there is a great explanation (above) of the famous double slit experiment. The explanation ends with the statement “The observer collapsed the wave function simply by observing!” and from a value free objective scientist perspective that would be the correct way of putting it. However a better way to put it is that the “values of the observer dictated what was observed” because , unlike a scientist may assume, values are more fundamental than physical matter.

DON'T FORGET TO SHARE THIS POST!

TOPICS: