This needs far more context for an educated discussion.
In the picture you can see organizations moving in the public sphere around AI. On the left you have right-wing and libertarian think tanks, corporations and frontline actors that fuel a sense of panic around AI, either to sabotage their business competitors or to leverage this panic to project an idea of being sellers of a very powerful tool while at the same time deflecting responsibility. If the AI is dangerous and sentient, you won’t care much about the engineers behind.
On the right you have several public orgs or NGOs operating in the field of algorithmic accountability, digital rights and so on. They push the opposite of the AI panic, pointing the finger at the corporations and powers that create and govern AI
What is this based on? Some kind of paper? Were there objective criteras, which were choosen beforehand, and on which the companies were rated, leading to there grouping in these groups?
Or did you just made up 4 cool sounding categories, which you fitted various companies in, based on your personal opinion?
It’s not from me but from AlgorithmWatch, one of the most famous and respected NGOs in the field of Algorithmic accountability. They published plenty of stuff on these topics and human rights threats from these companies.
Also this is an ecosystem analysis of political positioning. These companies and think tanks are going on newspapers with their names to say we should panic about AI. It’s not a secret, just open Google News and you fill find a landslide of news on these topics sponsored by these companies with a simple search.
If by panic you mean AI hype, then maybe.
For example, this post is just as sensationalist.
“AI panic” is such a broad term that it is really meaningless.
Can I get an explanation as to how these companies are “marketers of ai panic”?
They are directly selling AI-based products and services. They release or boost sensational stories about those capabilities through their various channels of media influence so they can make their products seem more powerful and useful than they really are. The sensationalisation widens the window on what seems possible even if it’s nowhere near the reality. Even people who don’t buy into those notions about society-destroying automation or humanity-threatening emergence are more likely to buy into stuff that seems tamer but still lacks any substantial proof of viability like AI driving or AI written movie scripts.