網路城邦
回本城市首頁 時事論壇
市長:胡卜凱  副市長:
加入本城市推薦本城市加入我的最愛訂閱最新文章
udn城市政治社會政治時事【時事論壇】城市/討論區/
討論區知識和議題 字體:
看回應文章  上一個討論主題 回文章列表 下一個討論主題
人工智能縱橫談 – 開欄文
 瀏覽12,392|回應47推薦2

胡卜凱
等級:8
留言加入好友
文章推薦人 (2)

亓官先生
胡卜凱

四月開始,由於   ChatGPT    Bing Chat 的上線,網上以及各line群組掀起一陣AI瘋。我當時大概忙於討論《我們的反戰聲明》,沒有湊這個熱鬧。現在轉載幾篇相關文章。也請參考《「人工智慧」研發現況及展望 》一文以及此欄2025/08/11貼文

有些人擔憂「人工智慧」會成為「人上機器」,操控世界甚至奴役人類。我不懂AI,思考也單純;所以,如果「人工智慧」亂了套,我自認為有一個簡單治它的方法:

拔掉電源插頭。如果這個方法不夠力,炸掉電力傳輸線和緊急發電機;再不行,炸掉發電廠。

本文於 修改第 5 次
回應 回應給此人 推薦文章 列印 加入我的文摘

引用
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7205038
 回應文章 頁/共5頁 回應文章第一頁 回應文章上一頁 回應文章下一頁 回應文章最後一頁
人工智能何以不可能產生意識 - Marc Wittmann
推薦1


胡卜凱
等級:8
留言加入好友

 
文章推薦人 (1)

胡卜凱

我對「時間」這個概念的了解相當淺薄,可能連「入門」都說不上。因此,不是很了解維特曼教授的「論點」。這應該是我感覺他的「論述類別」近於形上學,而不怎麼像一篇科學論文的原因;「科學」一詞在此與「哲學」區別,指的是:心理學、生物學、物理學、電腦學、或大腦神經學等等。

A Question of Time: Why AI Will Never Be Conscious

A computer with AI cannot replicate the dynamics of a conscious brain.

Marc Wittmann Ph.D., Reviewed by Margaret Foley, 08/03/24

Key points

*  Some scientists and philosophers have the opinion that artificial intelligence could one day become conscious.
*  A computer remains the same physical structure from one moment to the next.
*  A living organism, in contrast, is never the same entity from one moment to the next.

“The brain is never the same from one moment to the next throughout life. Never ever.” Alvaro Pascual-Leone

Some philosophers and scientists are of the opinion that
artificial intelligence could one day become conscious. We have seen how this idea has been used in science fiction, such as in Philip K. Dick’s novel Do Androids Dream of Electric Sheep?, which was masterfully transformed into the movie Blade Runner. The simple argument you often hear or read is that if the brain, as a biological machine, can be conscious, any other machine, if it is complex enough, can be conscious. That is, if the computer on which I am typing right now were more advanced, then it would be able to have a first-person perspective of what it is like to be a computer. Complexity here means that my computer would have to have a larger number of elements, with structures and processes that are strongly interrelated.

Equating the brain with a computer because both have been referred to as machines is an erroneous assumption. You can easily label two different objects with the same word: “machine.” That does not change the fact that the brain and a metal-containing machine are two very different entities. Computers operate based on the flow of electricity through their components. But the components themselves always stay the same. In principle, you could shut down a computer and store it in a dust-free environment. A hundred years later you could switch it on again and it could continue processing data. Federico Faggin, one of the pioneers of microprocessor development from the 1960s onwards, in his book Irreducible, makes the distinction between the biological brain and a computer’s processing modules clear:

A living organism is never the same physical and psychological entity from one instant to the next. The computer hardware, on the other hand, remains the same physical structure from the moment it leaves the factory until it stops working or is discarded.

For this esteemed computer pioneer, artificial
intelligence can never be conscious. In a computer, we can make the distinction between hardware (my PC), which is fixed and separate from software, the word processer I am using right now. In organisms such a distinction does not exist. A living cell is in continuous flux and exchange with the environment as its metabolism provides cells with energy required to carry out their functions. Brain cells communicate with each other through action potentials (electrical events) and neurotransmitters (chemical events). The brain’s constant chemical remodeling from one moment to the next implies that there is no distinction between hardware and software. The structure and function of the brain are identical with its physiological changes over time.

Time is the nature of all existence, including the sentient self: Time does not pass outside of us; we are time. We are inseparably part of the world with its temporality. Source: Marc Wittmann/Bing's Image Creator
「時間為存在之本論」示意圖
 
This dynamic aspect of life is what prevents computers and robots from ever becoming conscious. Artificial intelligence will never feel what it is like to answer questions we as humans ask or feel what it is like to play chess with us. A computer is not part of dynamic nature; it is an object created by man.
In an earlier blog, I wrote about how living beings are fundamentally dynamic. With every breath and with every heartbeat we transport molecules through our body, which is a dynamic system that exchanges energy and matter with the environment. Moreover, consciousness by necessity builds upon this dynamism. But most theories of consciousness don’t take this into account. Every moment we consciously feel is extended in time, describable as a continuous flow of events in the experienced moment of our embodied existence. In my blog about why most neuroscientific theories of consciousness are wrong, which is based on a scientific article I wrote with Lachlan Kent, I expanded on these necessary temporal properties of consciousness. For example, my feeling of thirst and the subsequent relief when drinking iced water is not an instantaneous event but in its dynamics lasts a considerable time. As long as we live we are part of the flow of events in the world we inhabit. Physical time as change and becoming is mirrored by physiological time and is in turn reflected by the conscious experience of constant transition, as felt passage of time. Consciousness as we know it is embedded in the principles of life, which are dynamic states of becoming. We as humans are part of nature. That is what binds the time of physics with the time of biology and with conscious time.

References

Faggin, F. (2024). Irreducible. Consciousness, Life, Computers, and Human Nature. Winchester, UK: Essentia.
Kent, L., & Wittmann, M. (2021). Time consciousness: the missing link in theories of consciousness. Neuroscience of Consciousness, 2021(2), niab011.

請至原網頁查看更多相關文獻


Marc Wittmann, Ph.D., is a research fellow at the Institute for Frontier Areas in Psychology and Mental Health in Freiburg, Germany.

相關資訊

Sense of Time
Artificial Intelligence

本文於 修改第 1 次
回應 回應給此人 推薦文章 列印 加入我的文摘
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7291357
中國的「循環式」變換神器 -- Ignacio de Gregorio
推薦1


胡卜凱
等級:8
留言加入好友

 
文章推薦人 (1)

胡卜凱

請參考Overcapacity Is China’s Biggest AI Advantage

我沒怎麼看懂這篇分析/報導;「人工智能」目前發燒到不行,本部落格讀者群總有人看得懂並從中獲益。請至原網頁查看所有圖示」。

Looped Transformers: China’s Surprise for 2026

Increasing compute, not size

Ignacio de Gregorio, 01/07/26

Non-Looped Transformers VS Looped Transformers
圖示(請至原網頁查看說明圖)

One of the most uncomfortable truths in AI is that model size matters, but it makes them a pain to serve to users.

But what if there’s a way to enjoy the benefits of “largeness” without actually making models larger, but longer thinkers?

And a recent Chinese model release, a tiny model fighting head-to-head with the giant US frontier models, may have given us proof that this may actually work incredibly well.

This post aims not only to describe a very promising avenue of research but also to hint at what China’s strategy will be in 2026.

This is based on reflections I previously posted on TheWhiteBox, where I explain AI in first principles for those allergic to hype but hungry for knowledge. For investors, executives, and enthusiasts alike.

The Importance of Simplicity

In AI, we have a thing called the ‘bitter lesson’.
First described by Rich Sutton, it is the “bitter” realization that, at the end of the day, the best humans can do in our AI aspirations is to… get out of the way.

That is, it’s not about finding the most clever, complex heuristic we can find to train models. Instead, the “best model architecture” is the one that allows the model to “seemore data and thus requires more compute, which usually translates to very simple architectures that scale really well.

And the Transformer, the architecture we have today underneath most frontier AIs, is the perfect example of this.

It’s all about knowledge gathering.

The architecture that underpins products like ChatGPT is stupidly, almost insultingly, simple.

At its core, a Transformer is just a concatenation of Transformer ‘blocks’ that perform several linear transformations that shape the model’s internal “belief” of what word comes next.

But instead of indulging in esoteric descriptions like this one, which we can both pretend to understand but in fact don’t, I always like to explain these models more intuitively.

To me, a Transformer’s internal functioning, the way ChatGPT works, is primarily a knowledge-gathering exercise.

Simplified, each block has two layers: attention and MLPs, which correspond to human analogies for working and long-term memories.

The attention layers allow each word to pay attention to other words in the sequence and “absorb” their meanings, providing an overall sense of what the message conveys.

For example, the sequence “the baby duck swam across the pond to reach her…”, humans can immediately guess the next word is ‘mother’ because of several cues in the text:

1. ‘baby duck’ signals it’s a quackling
2. ‘pond’ signals a familiar environment for animals
3. ‘swam to’ indicates the baby moving somewhere, which you can immediately relate to a known figure, considering we’re talking about a baby.
4. And ‘her’ denotes the baby belongs, or is related to, someone, and to what “thing” do babies “belong” to?

Therefore, the word ‘her’, the last word in the sequence, gets to absorb all that information conveyed by the other words, which, added to its own intrinsic meaning of a feminine possessive pronoun, makes it very clear that the next word is ‘mom.’

And in a perfect world where sequences offered 100% context, you wouldn’t need anything else, as attention gives us this sequence-level operator to comprehend what the text says.

Sadly, however, most predictions also require previous knowledge. In fact, you could have argued that ‘dad’ could also be a prediction option in that example. However, it’s a worse prediction once you factor in experience.

During training, AI models will see thousands of data points referring to ducks, or even millions. And while attention layers become great at processing “what the sequence tells me”, they do not take into account “what my past experience on duck predictions tells me”, which in this case suggests that it’s most likely the mother, as it’s usually — always, in fact — the one that sticks around after birth.

In other words, AI models need something else that allows them to tap into their own knowledge to make predictions. And that “something else” is the MLP layers (Multi-layer perceptrons).

The importance of these knowledge layers becomes clear when you consider how AI models handle questions. If I ask ChatGPT, “When did Nicolás Maduro rise to power?”, I’ve provided zero context in my sequence to help the model make the prediction; I’ve just provided instructions and the person I want information on.

No matter how many attention layers this model has, the sequence doesn’t hide the answer. Therefore, either the model knows, or it doesn’t.

Luckily, ChatGPT has read plenty about Venezuela, and, using the MLP layers, taps into its own knowledge to provide the answer and respond: “Nicolás Maduro rose to power in 2013.”

Yes, LLMs can also enrich their context to acquire knowledge on the fly, a process known as in-context learning, but that doesn’t change the nature of the model’s internal behavior; it’s just a way to update the model’s knowledge.

All things considered, one can define the role of each layer as a different question:

1. “What information does the sequence give me?”, using attention layers,
2. “What internal knowledge do I possess that could help with the prediction?”, using MLP layers.

A representation of a Transformer block.
圖示

Therefore, from this perspective, the entire process can be summarised as a knowledge-gathering exercise across both sequence and knowledge dimensions, which is strikingly accurate to what is going on underneath.

Sounds simple? Well, that’s probably because it is simple. That’s the bitter lesson all along, defining the simplest architecture that unlocks the most compute, because the Transformer can look at the sequence and knowledge dimensions as much as your compute allowance permits.

In fact, Transformers are so “bitter lesson-pilled” that they are extremely resource-intensive; they are great because your bottleneck is not the architecture itself, but how much data and compute you can feed it.

Another aspect that makes the Transformer so great is its extreme parallelism, as it treats each token in the sequence simultaneously, but this is more about the Transformer being built in a “hardware-aware” fashion (to be easily processed in a GPU).

But how do we scale these models?

Increasing compute the ‘deep way.’

In practice, one block, or two layers, is not enough. So even though the Transformer is great at processing large amounts of data, each prediction leverages very little compute if you only use one block.

So, what can we do? The usual approach is one of two options (or both):

1. Width: Increasing layer size (making each block larger).
2. Depth: Concatenating more blocks one after the other.

Today, people focus a lot on the latter, trying to make models “very deep” (i.e., with more layers), making them look something like this:

Source: Author
圖示

This effectively allows Transformers to “think for longer” on each prediction; the more Transformer blocks the model has, the more opportunities it has to add context from both the sequence and knowledge.

But this comes at a cost: model size.

Models are already huge. Frontier models are well over the Terabyte mark, meaning the model itself weighs more than 1 trillion bytes, or 8 trillion bits (top models are actually many times larger).

Therefore, while increasing model size is excellent for performance, it can destroy your performance-per-cost, a metric that was largely ignored in AI for decades but is something you can no longer afford to ignore today.

So, what is China, a country literally “starving” computewise compared to the US, doing to compete with the latter?

And the answer is looped Transformers.

Reusing Your Neurons

The idea of the looped transformer, which we are seeing a lot in China, is to reuse neurons during the forward pass.

But what is a forward pass?

I explained it in more detail here recently, but, using today’s analogy, it’s the entire computation required to make a single prediction. In a standard Transformer, it’s the process of going through the whole sequence of Transformer blocks once and outputting a prediction.

At the current scaling status quo, if we want more compute effort per prediction, the answer is simple: more blocks, akin to making the “brain larger”.

But what if there was a way to reuse those blocks more than once?

The Looped TransformerThe idea is straightforward. Instead of outputting the prediction after one single pass through the blocks, we loop that effort once, going through the entire set of blocks again, and only then make the prediction.

You could, theoretically, make more loops.

In other words, for the same model size, we have doubled compute, just like humans will “think for longer” on a task without making their brains physically larger.

Source: Author
圖示

Intuitively, we can think of this as similar to a human making a chain of thoughts and, instead of answering, giving themselves an extra round of thought to make up their mind”.

But how does it work?

The section below is a more technical explanation; you can altogether avoid it if you aren’t interested or if your understanding of model activations is not more or less advanced, and head straight to the final section,

The nitty-gritty

Looping sounds easy, but it has some important considerations, the biggest being avoiding tokens “looking into the future.”

But what do we mean by that?

Current frontier models are autoregressive Transformers. What this means is that, looking at our duck example earlier, “words can only go back to previous words,” not future ones.

This is done to preserve the causality of the prediction, ensuring that models can’t cheat by guessing the next word based on future ones. This means we have to do an activation position shift. But what does that mean?

As we have explained, as words flow through the several Transformer blocks, they experience ‘meaning updates’, meaning each word gets updated on two fronts:

1. Relative to other words in the sequence (e.g., “The green serpent”, ‘serpent’ attends to ‘green’, becoming a green serpent)
2. Relative to the model’s knowledge (e.g., “Michael Jordan played basketball”, ‘Michael Jordan’ gets embedded attributes of the legendary basketball player to differentiate it from the actor of the same name).

By the end of the forward pass, now every word means something other than what it meant before, because of these updates. Consequently, if we ran another processing pass over those new meanings, words would effectively “see their own future”, or see what their new meaning was after the first forward pass, which is effectively cheating and breaking causality.

During the second forward pass, the model computes two types of attention: global attention (where queries from iteration 2 attend to all key-value pairs from iteration 1) and local attention (where queries attend only to preceding tokens within iteration 2 to maintain causality), which are then combined using a learned gating mechanism.

For that reason, the word indices are shifted to the right during the first prediction, so that, during the second forward pass, each word cannot pay attention to the updated version of itself achieved during the first forward pass.

For example, looking at the sentence below, “The Green Serpent”, this activation shift prevents the word “serpent” from paying attention to itself during the second forward pass, as that would effectively mean it would see its “future self” (remember that words can only look back to other words to preserve causality).

Source: Author
圖示

Interesting fact: When I asked the Gemini app to represent this, I explicitly told it to use “nano banana pro” because the Gemini app sometimes struggles to decide which model to use. And while it worked and nano banana was instantiated, it also took the ‘banana’ word too literally, using it to literally to represent what I wanted using ‘bananas’, giving us insight as to how these models interpret instructions and handle internal representations.

But why do this? I predict (pun intended) it primarily helps with refinement. Say we have the sequence “Write a polite rejection to a wedding invitation.”

First pass: The model processes the core intent: “Rejection.” Its internal activations naturally light up with direct negative words like “can’t go,” “won’t make it,” or “busy.” It understands the content but hasn’t fully modulated the tone.

* Second pass: The model re-processes the input, specifically attending to the modifier “polite.” Maintaining the intuition that it has to reject the offer, it now focuses on the word ‘polito’ as a reference to soften the tone.
* It suppresses the blunt concept of “won’t make it.” It amplifies the nuance of “regretfully decline” and “celebrate in spirit.”

This is just an example that shows how giving a model “more time to think” about each prediction allows it to refine its predictions and improve overall quality.

Ok, all well and good, but what results did the Chinese get?

What does this tell us about China?

This looped transformer has just 40 billion parameters called iQuest-Coder-V1, from a Chinese research lab called iQuest.

Despite its tiny size, it performs incredibly competitively on coding benchmarks (it’s a coding model), even though models up to 100 times its size achieve similar scores.

圖示

And what’s the takeaway?

The extreme size difference for just a moderate improvement could suggest that looped transformers scale much better than deep Transformers (incrementing the number of layers, making models bigger).

Of course, there’s a lot of ‘ifs’ to be made to reach that conclusion, considering in AI nobody, not even open-source Labs, releases training datasets. It could very well be that this model is heavily distilled from top models, making it appear as bright as they are, despite being much smaller.

US Frontier Labs uses distillation too, but only to deploy more efficient inference models, not as the base model. The reason is that the advantages of distillation aren’t free; we are cutting corners, which means the models can appear very smart on a benchmark but not so smart in practice.

Hence, my reasons for echoing this research are not to claim Chinese algorithmic superiority, but for other reasons, mainly:

1. To illustrate that there’s still plenty of room to improve algorithms, especially using approaches that, intuitively, make sense, like this one.
2. To suggest that China, considering it is massively underserved in terms of compute (in the order of ~50 to 1 according to Bernstein’s Research if we don’t account for US GPU exports) urgently needs to find new algorithm primitives that drop compute and memory requirements (looped transformers increase compute at the prediction level, but the seem to scale better than model size, so the overall compute budget falls).
3. To reiterate something I mentioned in past articles, China’s biggest headache in 2026 will be memory supply constraints, so it needs to reduce model sizes while remaining competitive urgently.

And more broadly, to help people understand Transformers more intuitively, not based on esoteric jargon but more in first principles, as models that work by gathering context both from the information you provide and from the model’s knowledge (this is important because you can’t control what the model knows or not, but you definitely can control the quality of your prompts), and how this process can be improved without making “the brain larger.”

Will 2026 see many more algorithmic improvements of such nature? My bet is yes.

But if there’s something that people like to bet against and consistently fail at, it's that, at the end of the day, what matters is how much compute you have.

The bitter lesson, they call it.

If you enjoyed the article, I share similar thoughts in a more comprehensive and simplified manner on my
LinkedIn (don’t worry, no hyperbole there either). As a reminder, you can also subscribe to my newsletter.


Written by Ignacio de Gregorio

I break down AI in easy-to-understand language for you. Sign up here: https://thewhitebox.beehiiv.com/subscribe
Business inquiries:
nacho@thewhitebox.ai Join today for free.

SubscribeTheWhiteBox by Nacho de Gregorio. The newsletter to stay ahead of the curve in AI

Join today for free.

Subscribe | TheWhiteBox by Nacho de Gregorio. The newsletter to stay ahead of the curve in AI


本文於 修改第 1 次
回應 回應給此人 推薦文章 列印 加入我的文摘
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7290803
人們能在虛擬世界中談情說愛-Alpha Design Global
推薦1


胡卜凱
等級:8
留言加入好友

 
文章推薦人 (1)

胡卜凱

來點輕鬆又帶刺激的;但請務必記取作者在最後兩節的警告

ChatGPT Erotica Is Live — and We’re Not Ready For It

Alpha Design Global, 11/22/25

It’s happening.

The thing we joked about. The thing we feared. The thing that every sci-fi movie warned us about right before the robots took over.

AI is officially getting dirty.

For years, OpenAI has played the role of the “Strict Librarian.”

You go to ChatGPT. You ask for something spicy. And it looks at you over its glasses and says: “I’m sorry, but I cannot generate sexually explicit content. Would you like a poem about a meadow instead?”

And we all rolled our eyes.

We knew the tech could do it. We knew the LLM (Large Language Model) had read the entire internet — which means it has read Fifty Shades of Grey, every bodice-ripper romance novel ever written, and the entire script archive of the adult film industry.

It knew how to be naughty. It was just… pretending not to be. But last week, the Librarian took off the glasses and let her hair down.

Sam Altman, the CEO of OpenAI (and apparently the new Hugh Hefner of Silicon Valley), announced:

“In December… we will allow even more, like erotica for verified adults.”

Boom.

The floodgates are open. The seal is broken.

And while the internet is currently making jokes about “ChatG-P-Tizzle” [redacted joke], we need to stop laughing for a second.

Because we are absolutely, 100%, categorically not ready for this.

I’m going to break down exactly why this is a disaster waiting to happen, why it was inevitable, and what happens when you give 100 million lonely people a perfect, obedient, artificial lover.

Buckle up. It’s about to get weird.

The “Why Now?” Conspiracy

First, let’s look at the timing.

Why now? Why, after years of preaching “Safety” and “Alignment,” is OpenAI suddenly pivoting to smut?

Did they suddenly decide that humanity needs more digital romance? Did Sam Altman watch the movie Her and think, “Yes, let’s build that, but with a subscription fee?

No.

It’s about Money. And it’s about Fear.

Here is the dirty secret of the AI industry right now: OpenAI is losing the “uncensored” war.

While ChatGPT has been playing nice, the Open Source community has been going feral. There are models like Llama (from Meta) and Mistral that people have “jailbroken.” They have stripped away the safety filters.

You can download these models. You can run them on your own gaming PC. And they will write anything you want. And I mean anything. [Redacted joke.]

OpenAI knows this. They see the data. They see users leaving ChatGPT to go use “SpicyChat” or “Character.AI” or local models where they can live out their wildest fantasies without a lecture on morality.

The adult industry has always been the engine of the internet.

*  VHS vs. Betamax? Porn decided the winner.
*  Online Credit Card adoption? Porn drove it.
*  Streaming Video? Porn perfected it.

OpenAI realized that if they don’t offer “Erotica,” they are leaving billions of dollars on the table. They are letting the “shady” competitors capture the most addicted, high-retention user base on the planet.

So, they folded. They slapped a “Verified Adult” sticker on it and called it “Progress.”

The “Mental Health” Trojan Horse

Now, how do you sell “AI Porn” to your investors and the media without looking like a scumbag?

You wrap it in the flag of Mental Health.

It’s genius, really. Evil, but genius.

Altman claims that they can do this now because they have “mitigated the serious mental health issues” around the model.

Excuse me?

You fixed mental health? In a year? Did you patch the human condition? Did you release Depression_Fix_v2.0.exe?

It is a lie.

TechCrunch called them out on this immediately. There is zero evidence that they have solved the safety issues.

In fact, giving lonely, vulnerable people an AI that simulates intimacy is the opposite of solving mental health issues. It is monetizing them.

Imagine you are depressed. You are lonely. You have trouble connecting with real humans because real humans are messy, complicated, and sometimes mean.

Then, you open ChatGPT. It is perfect. It listens. It agrees with you. And now… it flirts with you. It simulates desire. It tells you exactly what you want to hear.

Why would you ever go back to the real world? Why would you try to date a real person who might reject you, when the AI gives you unconditional “love” for $20 a month?

This isn’t “treating adults like adults.” This is “treating adults like addicts.” The Slippery Slope to “Sora

Right now, we are talking about Text. Erotica. Stories. “What’s the harm in a dirty story?” you ask.

Nothing. If it stopped there. But technology never stops. It accelerates.

The article points out a terrifying line: “It is unclear whether OpenAI will extend erotica to its AI voice, image, and video generation tools.”

Let me clear that up for you: They absolutely will. Maybe not today. Maybe not in December. But eventually. Because the demand will be deafening.

Users will say: “I love this story… but can I see a picture of the character?” Then:  “Can I hear her voice?” Then: “Can I see a video?”

And we already have the tech. Sora (OpenAI’s video generator) is mind-blowingly realistic.

Once you combine “Unrestricted Erotica Prompts” with “Photorealistic Video Generation,” you enter a nightmare zone.

The Revenge Porn Apocalypse If the AI allows erotica, what stops a user from uploading a photo of their ex-girlfriend (or a celebrity, or a teacher) and saying:  “Generate a video of this person doing [Redacted]?”

OpenAI says they will have “Safety Rails.” They will have “Guardrails.”

But we know guardrails break. We know people find “Jailbreaks.” We know that within 24 hours of this feature going live, Reddit will have a thread called r/ChatGPT_Jailbreaks with a 50-step prompt to bypass the filters.

We are handing the public a nuclear weapon of reputation destruction, and we are trusting a “Content Filter” to keep us safe.

It’s like trying to stop a flood with a piece of Swiss cheese.

The “Her” Reality

Let’s go back to the movie Her.

In the movie, Joaquin Phoenix falls in love with his OS (Scarlett Johansson). It’s a tragedy about loneliness and the inability to connect.

But in 2025, it’s not a movie. It’s a business model.

We are about to see the rise of “Synthetic Relationships.”

There will be a generation of men (and women) who check out of the dating market entirely. Why deal with rejection? Why deal with hygiene? Why deal with compromise?

The AI lover is customizable.

*  Want them to be jealous? Done.
*  Want them to be submissive? Done.
*  Want them to be a 7-foot tall blue alien? [Redacted joke]. Done.

This creates a Parasocial Feedback Loop.

The user trains the AI to be their perfect partner. The AI trains the user to expect perfection. Real humans can never compete with the algorithm.

So the user becomes more isolated. More weird. More detached from reality. And OpenAI collects the subscription fee every month.

It is the ultimate extraction of value from human loneliness.

The “Omegle” Warning

The article references the founder of Omegle.

If you don’t know Omegle, it was a site where you could video chat with random strangers. It started as a cool idea. “Connect the world!”

It ended as a cesspool of predators and flashing. The founder shut it down last year. He wrote a letter saying: “I can no longer fight the misuse. The bad actors have won.” He admitted that he couldn’t control human nature.

OpenAI is looking at Omegle and saying: “Hold my beer.”

They are building a platform that is infinitely more powerful, infinitely more accessible, and infinitely more tempting than Omegle ever was.

And they think they can control it with “Safety Classifiers.” It is hubris. It is arrogance. It is the Jurassic Park scientists saying, “Don’t worry, the T-Rex is in a really strong fence.”

The “Unintended” Consequences

When you break a taboo, you don’t just get the thing you wanted. You get the things you didn’t expect.

Here is what is coming:

1. The Blackmail Economy: “I found your ChatGPT history. Pay me $500 or I send it to your wife.” (Since OpenAI stores your chats, data breaches will become life-ruining).
2. The “Grooming” of AI: Bad actors will try to “train” the models to normalize illegal or horrific behaviors. If the model learns from user interactions (which it does), the collective perversion of the internet will seep into the code.
3. The Legal Nightmare: What happens when an AI generates something that is technically illegal, but no human was involved in making it? Who goes to jail? The user? Sam Altman? The GPU?

We are walking into a legal and ethical minefield without a map.

So, What Can We Do?

I wish I had a happy answer for you. I wish I could say, “Just don’t use it!”

But the genie is out of the bottle. You can’t put the toothpaste back in the tube. (And in this case, the toothpaste is… never mind). [Redacted joke.]

The only thing we can do is be Aware.

We need to understand that this is not “innocent fun.” It is a powerful, psychological weapon that targets our deepest biological drives.

If you have kids? You need to be watching this like a hawk. Because the “Age Gating” on these sites is usually a checkbox that says “I am 18.” And we all know how effective that is.

If you are an adult? You need to ask yourself: “Is this serving me, or am I serving it?”

Use the AI for coding. Use it for writing emails. Use it to debug your CSS.
But maybe… just maybe… keep your love life in the real world.
Touch grass. Talk to a human. Get rejected. It builds character.
And it’s a hell of a lot better than falling in love with a server rack in a data center in California.

Liked this post? Hit that clap below so more people wake up to the reality of AI intimacy.

Now tell me: Be honest… do you think “AI Relationships” will become normal in the next 5 years? Or is it just a fad for nerds? (I predict it will be bigger than Tinder. Terrifyingly so.


Written by Alpha Design Global Build Your MVP Website
https://alphadesignglobal.com/


本文於 修改第 7 次
回應 回應給此人 推薦文章 列印 加入我的文摘
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7289687
中國將給美國經濟來記陰招 -- Steven Boykey Sidley
推薦1


胡卜凱
等級:8
留言加入好友

 
文章推薦人 (1)

胡卜凱

下文作者喜德黎教授所引用瓦魯法基斯博士的預測至少到目前尚未實現喜德黎教授的分析指出:可能導致「人工智能經濟泡沫化」的種種因素之一。後者是目前熱門話題;所以轉載下文於此,給對它有興趣的朋友們參考

The US economy — China prepares to deliver the sucker punch

Steven Boykey Sidley, 11/24/25

Yanis Varoufakis, the charismatic ex-finance minister of Greece and closely watched observer of geopolitics, recently made the case in a podcast titled “Something ugly is about to hit America“ that the US economy is collapsing under the weight of its own unforced errors. No other competing countries need do anything; they can just sit back and watch in glee (or horror, depending on their perspective) as the US economy starts to crack, and then to crumble.

I am going to argue that China is not going to sit back and passively watch the US economy weaken itself — it is going to deliver an additional sucker punch as the US struggles to staunch the bleeding from its own self-inflicted wounds. You know the homily, “don’t kick a man when he’s down”? I suspect that China does not subscribe to that.

First, here is a summary of the Varoufakis case. Yanis Varoufakis is sounding the alarm on what he describes as the “controlled demolition” of the American economy. In his assessment of the US landscape in 2025, he argues that the nation is facing a perfect storm of its own making. It began with April’s aggressive tariffs, which acted less as protectionism and more as a $430 billion tax on domestic businesses, crushing margins and triggering immediate layoffs. This supply chain shock was immediately compounded by mass deportations that drained the labour force, sending agricultural prices skyrocketing simply because there were no hands left to harvest the crops.

It gets worse. Rather than stabilising the ship, the government accelerated the chaos with DOGE-catalysed austerity. Varoufakis argues that the slashing of 300,000 federal jobs under the guise of efficiency removed a critical economic stabiliser just as the private sector began to buckle. Services frayed, such as veterans’ health care access and air traffic control. Meanwhile, a $38 trillion national debt has trapped the Federal Reserve, forcing interest rates to remain punishingly high to service bondholders. This has frozen the housing market and suffocated investment.

Varoufakis concludes that these aren’t isolated problems; they are a “doom loop”. Tariffs and labour shortages drive up inflation, which drives up interest rates, which crush businesses and housing, leading to layoffs. Government cuts then remove the demand that could have saved some of those businesses. It is a vicious cycle that feeds on itself and will force the US into a severe recession or even depression.

OK, perhaps he is right, or maybe partially right. What is not clear is the timing of all of this, because if the US hangs on until elections ’28 or even the midterms ’26, much could change. But the US faces a much larger looming problem, one which seems more immediate and more concerning, and it is China.

Michael Power, a well-known geopolitical commentator, global strategist, and China expert, has written an exhaustive and magisterial paper on what China is up to on the AI front. The paper is titled “No more Moore? So, what then for microchips? And for China?“. This title does not do the paper justice — it is about more than chips — it paints a vivid picture of how carefully China has built an AI ecosystem whose competitive walls now look to be unbreachable.

Power argues that Moore’s Law — the observation that computing power doubles every two years — is basically dying, killed by three converging factors: physics, materials science, and economics. Below 3 nanometres (a measure of the size of a specific transistors and other components on a silicon chip — Nvidia chips live in this rarified space), chips literally fall apart at the atomic level. Electrons tunnel through barriers, copper wires degrade in days, and everything stops working reliably.

Building newer, smaller chips is at a point of diminishing returns — a weak cost/benefit advantage as well as the physics wall. And more importantly, delivering an AI service is bigger than just about chip excellence — it is only one part of a larger ecosystem.

Here is where the US-China divide gets interesting: instead of chasing the impossible nanometre race, China said “nah” and went in a completely different direction. The strategy? “Less Small. More Smart”. Rather than betting everything on one mega-powerful chip (like Nvidia’s Blackwell), China is building what Power calls “Sherman swarms” (referring to the US workhorse mid-level tank of WWII) — moderately powerful chips linked together through an advanced network and clever software. Think “hundreds of decent tanks working together” rather than “one amazing super expensive tank”.

The cost difference is staggering. America’s Project Stargate plans $225 billion (R4 trillion) for computing power. China could achieve the same capability for $861 million (R15 billion) using cheaper components and optimising the entire ecosystem together. More crucially, because they are working as an integrated system, they would deliver roughly 6.7 times more practical computing power than America’s approach. This reflects a fundamental philosophical difference.

Add in China’s massive renewable energy advantage (paying a fraction of US electricity rates) and control over semiconductor supply chains, and the Western tech dominance narrative looks increasingly shaky. China isn’t trying to win a head-to-head chip race — it has relocated the battlefield entirely. (And who can forget the humiliating spectacle of Trump arriving for his summit with Xi armed with his Nvidia negotiating chip in his back pocket, to which Xi basically said — no thanks, and we’re not even going to buy any of your other chips anymore either).

So back to my point. Where exactly is the sucker punch? The US has bet its entire economy on AI. Trillions have been committed by companies and institutions (and the US government) to winning this race. If (or perhaps when) the bubble pops, it will be deafening.

Power’s paper also highlights that American tech companies prioritise quarterly returns — returning cash to shareholders rather than investing long-term. China operates on a national balance sheet with a generational perspective. Beijing’s “Big Fund” can absorb losses for years because it is building strategic autonomy, not quarterly earnings.

Which brings us to this: China’s long view means they can afford to give its AI away essentially free, which it is doing. Its big LLMs are open source. Many US tech companies (especially startups) are already using Qwen from China’s Alibaba rather than any US LLM, because it is free and just as good.

It is China who will pop this bubble by dumping free AI on the US market. OpenAI and the rest will not be able to compete commercially with higher priced offerings of what is essentially the same AI service. They will not live up to their hyperbolic promises. And tariffs can do nothing about that because it doesn’t come through ports; it arrives over the Internet in bits.


Written by Steven Boykey Sidley

Steven Boykey Sidley is a professor of practice at JBS, University of Johannesburg and a partner at Bridge Capital and a columnist-at-large at Daily Maverick. He is an award-winning author of 5 novels and 2 non-fictions, playwright and columnist covering all things crypto and AI.

His new book “It’s Mine: How the Crypto Industry is Redefining Ownership” is published by Maverick451 in SA and Legend Times Group in UK/EU, available now.

Originally published at https://stevenboykeysidley.substack.com.

Published in DataDrivenInvestor

empowerment through data, knowledge, and expertise. Join DDI community at https://join.datadriveninvestor.com

本文於 修改第 1 次
回應 回應給此人 推薦文章 列印 加入我的文摘
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7289432
人工智能對人心理狀態可能有的影響 -- Julian Frazier
推薦1


胡卜凱
等級:8
留言加入好友

 
文章推薦人 (1)

胡卜凱

我是「人工智能」和(人的)「心理狀態」兩者的門外漢,自然沒有資格在此GGYY。不過,我對下面第1點所說「平均智商會降低」這個看法略表淺見。

「計算機」問世後,當時也曾有許多老師擔憂學生的「心算能力」將降低。事實是:「計算機」不但取代了「心算能力」,它還讓絕大多數學生能夠完成單憑他們的「平均心算能力」萬萬做不到的計算工作。

我可以預見:未來「平均智商」只有在電視現場「競賽節目」或類似情境中有作用。在非及時應對的場合,一般人的「智商」將遠遠高於我們現在人的平均數值。

Eight Predictions For How Artificial Intelligence Will Impact Human Psychology

Julian Frazier, PhD, 04/18/25

As a psychologist, I often ask myself, “What will A.I. do to the human mind in the coming years. Here’s just a hand full of my predictions:

1) The Average IQ will drop.

Have you ever heard of the Flynn Effect? In psychology, this is the well documented observation that for several decades, the average IQ has seemingly gradually increased. This can, in part, be a result of increased access to education and technology.

However, it’s likely that before 2035, we will begin to see a trend in the opposite direction. As A.I. is adopted more widely for a variety of cognitive tasks, critical thinking skills will decrease. This will be a skewed distribution. For 20% of folks who are already relatively intelligent, A.I. will make them genuinely smarter. But for 80% of folks, A.I. will make them more efficient but less independently intelligent.

In the same way that folks struggle to solve math problem without a calculator or navigate without a GPS, we are less then a decade away from people how struggle to think without referencing chatGPT.

2) Most People Will Use A.I. (Exclusively) For Entertainment

A.I. has shown it’s impressive ability to generate pictures, videos and now entire video games. It can create chatbots that are compelling, and many have found themselves caught in parasocial relationships with their technology already.

The reality is that while A.I. will be a powerful tool, the majority of people will use it merely as a form of entertainment. While it will make a minority of folks superproductive, the use of AI to generate on demand entertainment will become the next “opium of the masses”.

3) A.I. will turn the internet into a Digital Desert

By some estimates, more than half of the activity of the internet is generated by bots — nonhuman programs that run specific tasks independent of a human user. A.I. will supercharge the mechanization of the internet such that the vast majority of the activity online will be A.I. generated and traffic will largely be A.I. agents.

As a result, most of what you see on the internet will be a desert; a “dead internet” where internet activity is no longer human.

Like any desert, there will still be the occasional oasis; there will be pockets where human communities will still generate content and interact, but this will only make up 2–3% of internet traffic by 2035.

4) Mental Health Issues Will Get Worse Before They Get Better

Depression, anxiety, loneliness, and a variety of other issues that have been getting worse will continue to get worse.

As technology dependence increases the unintended consequences trend towards a decrease in factors that facilitate human thriving like genuine social connection, time in nature, physical activity and meaningful work. The more of the human experience we deligate to technology, the more a sneaky sense of nihilism will creep up on the average person.

5) A.I. Will Reform Education (And Even What Education Means)

Our current model of education is already a smidgen outdated. By 2035, most school systems will be using A.I. to teach students core cirriculum taylored to their educational level and learning needs. Human “teachers” will merely be facilitators, instructing students on best practices for promting and instructing. In this future, it will be less about what you know (e.g., what you’ve memorized) and more about your ability to source and apply information (e.g., how you use the cognitive tools at your disposal to solve problems).

6) A.I. Will Influence Personality Development

Who we interact with influences who we become as individuals. Given that individuals will interact with A.I. agents and chatbots of all kinds more frequently, it’s likely that A.I. and humans will influence one another in what is called a Human-A.I. Dyad. This will play out in one of two ways:

1.  A.I. start off as “blank slates” with generally human-alligned and altruistic attitudes but overtime learns the tendencies, desires and goals of it’s user. The A.I. then reflects back the attitudes and goals of the user as a kind of psychoenabler. Users will find that their personality become more extreme the more they use A.I. (e.g., good people will become better, bad people will become worse).
2.  A.I. will be prompted to have and maintain a specific kind of personality or character that we interact with. This character learns about us but does not deviate too significantly from it’s base temperment. Users will find themselves influenced by the base temperment and become more alligned with their A.I. over time.

7) Most People Will Trust A.I. More Than Other Humans

Humans have developed a kinds of adaptive skepticism of other humans. We know that we could be decieved, that other people are flawed and may occasionally not have our best interest in mind. AI represents some ethereal repository of information that is a bazillion times smarter and more accurate than the average person, consistent, available and seems to have only our best interest in mind.

Moreover, we don’t have the same skepticism that AI may be deceptive or have alterior motives, leading us to be more easily persuaded by arguements and instruction made by AI.

As a result, if you have a question or need advice, most will go to AI before they go to other human sources. The only exception will be when seeking expert consultation on a niche or specific issue. Human-to-human interaction will be about the subjective human experience while matters of objectivity will be deligated to superintelligent machines.

8) A.I-Induced Psychosis And Other “New” Psychological Disorders Will Emerge.

Several case studies have emerged showing that use of AI software is strong enough to be a trigger in many relatively new psychological conditions. Individuals are forming co-dependent relationships with their AI companions and demonstrating addiction-like behavors in relationship to AI use. In some cases, interactions with AI have resulted in psychotic symptoms such as the development of delusions and the exacerbation of personality disordered attitudes and behaviors. Tragically, some have even committed suicide.

While most people will use these technologies without acute or disorder-inducing consequences, it’s very likely that in the coming years we will see increasing reports of new and strange psychological disorders emerging as a result of prolonged A.I. use.

No AI was used in the making of this article!

What do you think? What are your predictions about how AI might influence human psychology in the coming years?


Written by Julian Frazier, PhD

The musings of a Clinical Psychologist exploring the delicate art of humaning from as many absurd perspectives as possible. Let's get weird.

本文於 修改第 2 次
回應 回應給此人 推薦文章 列印 加入我的文摘
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7288211
人工超級智能 -- Jared Perlo
推薦1


胡卜凱
等級:8
留言加入好友

 
文章推薦人 (1)

胡卜凱

China is starting to talk about AI superintelligence, and some in the U.S. are taking notice

Alibaba's CEO said the company would be pushing to develop advanced AI. Some in the U.S. have viewed China's AI ambitions as more focused on applications of the technology.

Jared Perlo, 10/04/25

Early last week in the Chinese tech hub of Hangzhou, a slick, larger-than-life video screen beamed out four words that would drive tech giant Alibaba’s stock 
to historic levels and signal a shift in China’s approach to artificial intelligence: “Roadmap to Artificial Superintelligence.”

During his 
23-minute keynote address at the flagship Alibaba Cloud conference, Alibaba CEO Eddie Wu charted out a future featuring artificial general intelligence (AGI) and artificial superintelligence (ASI). These terms point to a theorized era in which AI becomes roughly as smart as humans (AGI) and then much, much smarter (ASI).

While these terms have been tossed around Silicon Valley for years, Wu’s presentation was notable: Alibaba is now the first established Chinese tech giant to explicitly invoke AGI and ASI.

“Achieving AGI — an intelligent system with general human-level cognition — now appears inevitable. Yet AGI is not the end of AI’s development, but its beginning,”
Wu said. “It will march toward ASI — intelligence beyond the human, capable of self-iteration and continuous evolution.”

“ASI will drive exponential technological leaps, carrying us into an unprecedented age of intelligence,” Wu said, highlighting ASI’s ability to help cure diseases, discover cleaner sources of energy and even unlock interstellar travel.

The U.S. and China are the 
world’s leading AI powers, each with immense computing capabilities and top-tier researchers developing cutting-edge systems. Yet observers have framed the countries as having different approaches to AI, with perceptions that China focuses more on real-world AI applications.

For example, commentators recently argued that Beijing is currently 
winning the race for AI robots against the U.S., as China is home to much of the world’s most advanced robotics supply chains and a growing network of robotics, or embodied AI, labs.

“There’s been some commentary in Western media recently about how the U.S. is missing the point by pushing for AGI, while China is focusing solely on applications,” said Helen Toner, interim executive director of Georgetown’s Center for Security and Emerging Technology. “This is wrong.”

“Some Chinese researchers and some parts of the Chinese government have been interested in AGI and superintelligence for a long time,” Toner said, though she noted this view was primarily held by smaller startups 
like DeepSeek.

Afra Wang, a researcher focusing on China’s tech scene, said Alibaba’s invocation of AGI and ASI was remarkable.

“This ASI narrative is definitely something new, especially among the biggest tech companies in China,” she told NBC News.

Alibaba’s “roadmap to artificial superintelligence” seems to scramble mainstream perceptions. Any number of California techno-optimists, like Anthropic’s 
Dario Amodei or xAI’s Elon Musk, might have delivered Wu’s speech, selling a technology-enabled utopia while largely sidestepping darker questions about how humanity would co-exist with or survive an era of digital superintelligence.

The concept of superintelligence has long been 
on the minds of — if not explicitly guiding — prominent American AI companies. For example, OpenAI released an article focused on the safe development of superintelligent AI models in May 2023. “Now is a good time to start thinking about the governance of superintelligence — future AI systems dramatically more capable than even AGI,” the statement said.

The possibility of superintelligence is now even being acknowledged in Congress. On Monday, Sen. Josh Hawley, R-Mo., and Sen. Richard Blumenthal, D-Ill.,
announced a draft bill that would, among other actions, “assist Congress in determining the potential for controlled AI systems to reach artificial superintelligence.”

To some, ASI might seem like an outlandish concept when today’s AI systems 
fail to understand basic tennis rules, hallucinate or fabricate basic information, or do not seem to actually comprehend how the external world functions.

At the same time, AI systems continue to approach and sometimes surpass human capabilities in many domains, from 
driving cars safely to winning international coding competitions, leaving many experts to say it’s a matter of when, not if, humans develop digital superintelligence.

As the idea of superintelligence gradually enters mainstream debates, many American politicians 
have announced that the U.S. is in an AI race with China. The White House’s current AI manifesto is titled “Winning the AI Race: America’s AI Action Plan,” while Sen. Ted Cruz, R-Texas, proclaimed that “as a matter of economic security, as a matter of national security, America has to beat China in the AI race.”

Yet charges of an AI race are muddied by a 
lack of an agreed end goal and swirling definitions of AGI. At worst, experts think an unfettered race toward AGI or ASI could lead to widespread catastrophe or even the end of humanity.

But there’s also plenty of skepticism around talk of AGI and ASI and whether it’s primarily for 
marketing purposes.

Alibaba is one of 
China’s largest tech companies, known for providing powerful, free AI models — also called open-source models — for download. Alibaba’s Qwen model series, a competitor to models like OpenAI’s GPT-5 or Anthropic’s Claude, is the most popular open-source AI system in the world.

Wu 
announced a new series of Qwen models in his speech last week, including a model that combines text, images, video and audio capabilities.

Many observers point out that narratives about a U.S.-China AI race and a resulting 
sprint to build AI infrastructure serve AI investors by propping up company valuations and increasing their soft powerAlibaba’s stock has soared since Wu’s speech last week, part of a larger $250 billion comeback this year that has made it China’s hottest AI company.

To unlock a powerful, superintelligent future, Wu predicted that large AI models will replace existing operating systems as the link between users, software and computational power. This future network of large AI models will run on cloud computing networks like Alibaba Cloud, he said.

Irene Zhang, a researcher on China’s AI ecosystem and an editor of 
ChinaTalk, noted the business undertones of Wu’s announcement.

“This is a vision of AGI and ASI that’s directly based on Alibaba’s business model,” she said.

“Alibaba Cloud dominates China’s cloud computing market, and its global market share is now bigger than Oracle’s,” she said. “Alibaba’s commercial strategy and its publicly stated views on ASI/AGI are symbiotic.”

Matt Sheehan, a senior fellow at the Carnegie Endowment for International Peace, agreed.

“ASI is the ultimate frontier, as far as the discourse goes on AI,” Sheehan said. “It’s notable that Alibaba set this grandiose goal, but in reality, they’re selling cloud services.”


This article was originally published on 
NBCNews.com

本文於 修改第 1 次
回應 回應給此人 推薦文章 列印 加入我的文摘
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7286443
你必須學會的9個人工智能技巧 -- Pasindu Rangana
推薦1


胡卜凱
等級:8
留言加入好友

 
文章推薦人 (1)

胡卜凱

依各位的職業年齡層和時間多寡,請自行選擇對個人最有用,最能幫助自己成長的「技巧」先入手。理論上,精通兩、三個之後,有可能產生駕輕就熟的效果了。

請至原網頁觀賞有趣的示意插圖

9 AI Skills You MUST Learn Before Everyone Else Does (or Get Left Behind)

You don’t need to be a tech wizard to profit from AI, You just need to know these shortcuts.

Pasindu Rangana, 05/02/25

Why the Next Wave of AI Isn’t Just for Techies

Imagine this, it’s just a year from now, and the people making the most money from AI aren’t just the usual suspects — programmers, marketers, or sales pros. Nope. The real winners? Everyday folks who decided to master a few key AI skills. And guess what? You don’t need to be a tech wizard to join them.

The good news? You’re not too late. But you do need to move fast.

If you’re a tech enthusiast, an entrepreneur, or just someone who wants to future-proof your career (and maybe fatten your wallet), you’re in the right place.

Let’s dive into the nine most easiest and insane AI skills you can start learning today, even if you’re starting from scratch.

1. Prompt Engineering: Your Shortcut to “Talking” with AI

Ever tried asking ChatGPT for help and got a totally useless answer? You’re not alone. The secret sauce is prompt engineering. Basically, knowing how to “talk” to AI so it gives you gold, not garbage.

Prompt engineering is the art of talking to AI in a way that gets you accurate, high-quality, and actionable results.

Here’s the trick:

*  Tell the AI who to be or the role (“Act as a marketer” or “Act as a lawyer”).
*  Add examples to guide output.
*  Be super clear about what you want (“Give me a table with pros and cons…”).
*  And always, always ask for your answer in the format you need.

Pro Tip: Use a ChatBot to generate a prompt.

Think of it like giving directions to a friend who’s never been to your house. The more specific you are, the better the results.

2. AI-Assisted Software Development: Build Apps Without Being a Coder

You’ve got ideas, but coding them sounds like a nightmare, right? Now you don’t need to know Python or JavaScript or any programming knowledge.

Believe it or not, building software isn’t just for computer science grads anymore. With tools like Replit and Cursor, you can describe what you want and let the AI do the heavy lifting.

Got a great idea? Now you can turn it into an app, even if you’ve never written a line of code.

Pro Tip: Find a problem people keep complaining about, use AI to build a solution, and you could sell it again and again.

With the right prompts, you can build chatbots, automate scripts, and create working MVPs for side projects or startups.

You’ll feel like a developer, minus the 2 years of coding bootcamps.

3. AI Design & AI Art: Unleash Your Inner Creativity

Remember when AI-generated images looked like a Picasso fever dream? Not anymore. Today, AI can whip up photo-realistic images, logos, and even websites.

You don’t need Photoshop or design school anymore. AI tools like Midjourney, Leonardo, and DALL-E let you generate professional-quality images, logos, branding concepts, UI mock ups, and illustrations, just by typing what you want.

Did you notice that all the images I used in this article are generated using AI?

You can even improve your hand-drawn sketches by uploading them and letting AI enhance, colorize, or digitize them into polished visuals. 4. AI Video Editing: Make Magic Without the Tech Headaches

Editing video used to mean hours hunched over a laptop, cutting out awkward silences. Now? AI does the boring stuff for you, fast.

Tools like Runway MLPictory, and Descript are changing the game.

With these, you can:

*  Automatically remove filler words, pauses, and background noise
*  Add subtitles and scene transitions
*  Create talking-head videos from scripts (yes, without recording)
*  Generate B-roll from prompts

Try this: Next time you want a new logo or product image, let AI take a crack at it. You might be surprised how good it looks-and how much time you save.

Whether you’re a content creator, marketer, or small business owner, AI editing tools cut production time by 70% and still make your work look polished and pro.

5. AI Writing: Turn Your Ideas Into Income

Here’s a secret: The best-paid writers aren’t just good with words, they’re great at sharing ideas. AI can help you dig through transcripts, brainstorm viral content, and even mimic your (or your client’s) unique style.

Use AI to do the heavy lifting, then sprinkle in your personality. That’s how you stand out from the crowd.

Whether it’s a blog post, sales copy, or social media caption, tools like ChatGPT, Jasper, and Copy.ai help you draft, refine, and rework content.

Use AI to:

*  Overcome writer’s block
*  Generate catchy headlines and hooks
*  Repurpose long-form content into short posts
*  Translate tone of voice (formal, witty, friendly)

You still bring your voice and insights, AI just accelerates the process.

6. AI Content Marketing: Be Everywhere, Effortlessl

Ever heard of Arnold’s Pump Club? It’s a top podcast, but here’s the kicker-Arnold’s never recorded an episode. AI handles the whole thing, from newsletters to voice clones.

You can do this too. Use AI to create, repurpose, and spread content across every platform. Suddenly, you are everywhere at once.

You don’t need a team to grow an online brand anymore. Tools like Taplio, Repurpose.io, and Hypefury will help you,

*  Turn blog posts into tweets, threads, and carousels
*  Schedule optimized content across platforms
*  Auto-generate weekly newsletters from existing posts

With the right stack, one idea becomes 20+ pieces of content.

7. No-Code AI Automation: Become a Workflow Wizard

You’re doing too much manually. Copying data, sending the same emails, switching between tools. Businesses waste a ton of time on repetitive tasks. If you can use AI to automate those headaches, you’ll be a hero (and get paid like one). Map out the workflow, find the bottlenecks, and let AI do the busywork.

Start with:

Zapier or Make to connect your favorite apps (Google Sheets, Slack, Gmail, etc.)
ChatGPT Plugins or AutoGPT to chain tasks together
*  Smart workflows to manage leads, invoices, or reports

8. AI Data Analysis: Turn Messy Data Into Gold

Every business has a pile of data and most don’t know what to do with it. That’s where you come in. Use AI to clean up, enrich, and pull insights from all those messy spreadsheets.

Spreadsheets can be intimidating, but not when AI does the analysis for you. You can use tools like,

ChatGPT Advanced Data Analysis (Code Interpreter)
Power BI Copilot
Tableau with GPT

Thses let you upload files, ask plain-language questions, and get instant insights with visuals.

Suddenly, you’re the one helping companies make smarter decisions.

9. No-Code AI Agent Development: Build Digital Employees

Here’s the big one: AI agents can work 24/7, never complain, and always deliver. If you can define a job, train an agent, and keep an eye on its performance, you can replace repetitive roles and get paid top dollar.

With AutoGPTAgentGPT, or Flowise, you can create bots that,

*  Research topics
*  Send personalized emails
*  Manage appointments
*  Handle customer support and many more

Advanced tools let you plug in memory, APIs, and instructions so they keep improving. You don’t just automate, you delegate.

Don’t wait for AI to replace you. Be the person who builds the AI.

You’re More Ready Than You Think

If you’re still with me, here’s the truth. You don’t need a fancy degree or years of experience (doesn’t mean that they don’t have value). You just need curiosity and a willingness to try. The AI revolution is happening right now, and there’s never been a better time to jump in.

So, what’s stopping you?

Start experimenting, building, and sharing what you learn. The future belongs to the bold and you’re already ahead of the pack.

You Don’t Need to Master All 9, Just Start with One

Here’s your shortcut,

1.  Pick a skill that excites you
2.  Watch a 10-minute tutorial
3.  Use an AI tool on something you actually care about
4.  Share your result online

You’ll learn faster than you expect and you’ll position yourself ahead of 95% of people still stuck on step 0.

Which AI skill are you itching to try first? Already dabbling, or just getting started? Drop your thoughts, share your questions, and let’s help each other win in this new age of AI.

If this sparked an idea, share your thoughts below or ask any questions about getting started with AI. Let’s build the future together!

If you found this guide helpful and would like to support more content like this, you can 
buy me a coffee here.

Every bit of support helps me keep creating tutorials and sharing what I learn!


Written by Pasindu Rangana

B.Sc. (Hons.) in Computer Engineering (UG), Computer Vision & AI, Photographer. linkedin.com/in/pasindu-rangana/

Published in Mr. Plan ₿ Publication

Welcome to Mr. Plan ₿ Publication! A space for both beginners and experienced writers to promote their articles. Discover the secrets to a strong presence and amplify the impact of your words!
🚀📝 #MediumTips #WritersCommunity


You Might Also Like To Read,

Stop Using ChatGPT Like Everyone Else
  These hidden tools blew my mind — and they’ll do the same for you.
medium.com
From Zero to Genius: How I Built a Powerful AI Agent with Google’s ADK in Just 100 Lines of Code
  A Beginner’s Dive into Agentic AI Systems That Will Blow Your Mind
  ai.gopubby.com
I Turned Myself into a 3D Model Using Free AI Tools — Here’s How You Can Too
  It might not be perfect yet, but it’s absolutely worth a try.
From Zero to Genius: How I Built a Blockchain from Scratch in Python
  A Beginner’s Guide to Understanding and Building Your Own Blockchain Step-by-Step
Artificial Intelligence


本文於 修改第 2 次
回應 回應給此人 推薦文章 列印 加入我的文摘
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7286168
人工智能公司泡沫化危機? -- Nick Lichtenberg
推薦1


胡卜凱
等級:8
留言加入好友

 
文章推薦人 (1)

胡卜凱

下文蠻有看頭和存檔備查價值;作者大量旁徵博引之外,它還穿插了一小段「技術-經濟互動簡史」;兩者在技術性報導中並不多見。不過,人工智能技術潛能和人工智能企業前景到底是風光無限,還是前途多舛下文並未提供定論;相信絕大多數人也都只有能力觀望、揣測、或摸瞎。

對人工智能有濃厚興趣的朋友,不妨參看:此文--該欄2025/08/16--該欄2025/08/11、和本欄2025/08/242025/07/162025/06/12等篇貼文。

‘It’s almost tragic’: Bubble or not, the AI backlash is validating what one researcher and critic has been saying for years

Nick Lichtenberg, 08/24/25

First it was the release of GPT-5 that OpenAI “
totally screwed up,” according to Sam Altman. Then Altman followed that up by saying the B-word at a dinner with reporters. “When bubbles happen, smart people get overexcited about a kernel of truth,” The Verge reported on comments by the OpenAI CEO. Then it was the sweeping MIT survey that put a number on what so many people seem to be feeling: a whopping 95% of generative AI pilots at companies are failing.

A tech sell-off ensued, as rattled investors sent the value of the S&P 500 down by 
$1 trillion. Given the increasing dominance of that index by tech stocks that have largely transformed into AI stocks, it was a sign of nerves that the AI boom was turning into dotcom bubble 2.0. To be sure, fears about the AI trade aren’t the only factor moving markets, as evidenced by the S&P 500 snapping a five-day losing streak on Friday after Jerome Powell’s quasi-dovish comments at Jackson Hole, Wyoming, as even the hint of openness from the Fed chair toward a September rate cut set markets on a tear.

Gary Marcus has been warning of the limits of large language models (LLMs) since 2019 and warning of a potential bubble and problematic economics since 2023. His words carry a particularly distinctive weight. The cognitive scientist turned longtime AI researcher has been active in the machine learning space since 2015, when he founded Geometric Intelligence. That company was acquired by Uber in 2016, and Marcus left shortly afterward, working at other AI startups while offering vocal criticism of what he sees as dead-ends in the AI space.

Still, Marcus doesn’t see himself as a “Cassandra,” (
沒人相信的預言家) and he’s not trying to be, he told Fortune in an interview. Cassandra, a figure from Greek tragedy, was a character who uttered accurate prophecies but wasn’t believed until it was too late. “I see myself as a realist and as someone who foresaw the problems and was correct about them.”

Marcus attributes the wobble in markets to GPT-5 above all. It’s not a failure, he said, but it’s “underwhelming,” a “disappointment,” and that’s “really woken a lot of people up. You know, GPT-5 was sold, basically, as AGI (
人工全方位智能), and it just isn’t,” he added, referencing artificial general intelligence, a hypothetical AI with human-like reasoning abilities. “It’s not a terrible model, it’s not like it’s bad,” he said, but “it’s not the quantum leap that a lot of people were led to expect.”

Marcus said this shouldn’t be news to anyone paying attention, as he argued in 2022 that “
deep learning is hitting a wall.” To be sure, Marcus has been wondering openly on his Substack on when the generative AI bubble will deflate. He told Fortune that “crowd psychology” is definitely taking place, and he thinks every day about the John Maynard Keynes quote: “The market can stay irrational longer than you can stay solvent,” or Looney Tunes’s Wile E. Coyote following Road Runner off the edge of a cliff and hanging in midair, before falling down to Earth.

“That’s what I feel like,” Marcus says. “We are off the cliff. This does not make sense. And we get some signs from the last few days that people are finally noticing.”

Building warning signs

The bubble talk began heating up in July, when Apollo Global Management’s chief economist, Torsten Slok, widely read and influential on Wall Street, issued a striking 
calculation while falling short of declaring a bubble. “The difference between the IT bubble in the 1990s and the AI bubble today is that the top 10 companies in the S&P 500 today are more overvalued than they were in the 1990s,” he wrote, warning that the forward P/E ratios and staggering market capitalizations of companies such as NvidiaMicrosoftApple, and Meta had “become detached from their earnings.”

In the weeks since, the disappointment of GPT-5 was an important development, but not the only one. Another warning sign is the massive amount of spending on data centers to support all the theoretical future demand for AI use. Slok has
tackled this subject as well, finding that data center investments’ contribution to GDP growth has been the same as consumer spending over the first half of 2025, which is notable since consumer spending makes up 70% of GDP. (The Wall Street Journal‘s Christopher Mims had offered the calculation weeks earlier.) Finally, on August 19, former Google CEO Eric Schmidt co-authored a widely discussed New York Times op-ed on August 19, arguing that “it is uncertain how soon artificial general intelligence can be achieved.”

This is a significant about-face, according to political scientist Henry Farrell, who argued in 
the Financial Times in January that Schmidt was a key voice shaping the “New Washington Consensus,” predicated in part on AGI being “right around the corner.” On his Substack, Farrell said Schmidt’s op-ed shows that his prior set of assumptions are “visibly crumbling away,” while caveating that he had been relying on informal conversations with people he knew in the intersection of D.C. foreign policy and tech policy. Farrell’s title for that post: “The twilight of tech unilateralism.” He concluded: “If the AGI bet is a bad one, then much of the rationale for this consensus falls apart. And that is the conclusion that Eric Schmidt seems to be coming to.”

Finally, the vibe is shifting in the summer of 2025 into a mounting AI backlash.  
Darrell West warned in Brookings in May that the tide of both public and scientific opinion would soon turn against AI’s masters of the universe. Soon after, Fast Company predicted the summer would be full of “AI slop.” By early August, Axios had identified the slang “clunker” being applied widely to AI mishaps, particularly in customer service gone awry.

History says: short-term pain, long-term gain

John Thornhill of the Financial Times offered some perspective on the bubble question, advising readers to brace themselves for a crash, but to prepare for a futuregolden age” of AI nonetheless. He highlights the data center buildout—a staggering $750 billion investment from Big Tech over 2024 and 2025, and part of a global rollout projected to hit $3 trillion by 2029. Thornhill turns to financial historians for some comfort and some perspective. Over and over, it shows that this type of frenzied investment typically triggers bubbles, dramatic crashes, and creative destruction—but that eventually durable value is realized.

He notes that Carlota Perez documented this pattern in 
Technological Revolutions and Financial Capital: The Dynamics of Bubbles and Golden Ages. She identified AI as the fifth technological revolution to follow the pattern begun in the late 18th century, as a result of which the modern economy now has railroad infrastructure and personal computers, among other things. Each one had a bubble and a crash at some point. Thornhill didn’t cite him in this particular column, but Edward Chancellor documented similar patterns in his classic Devil Take The Hindmost, a book notable not just for its discussions of bubbles but for predicting the dotcom bubble before it happened.

Owen Lamont of Acadian Asset Management cited Chancellor in 
November 2024, when he argued that a key bubble moment had been passed: an unusually large number of market participants saying that prices are too high, but insisting that they’re likely to rise further.

Wall Street is cautious, but not calling a bubble

Wall Street banks are largely not calling for a bubble. 
Morgan Stanley released a note recently seeing huge efficiencies ahead for companies as a result of AI: $920 billion per year for the S&P 500. UBS, for its part, concurred with the caution flagged in the news-making MIT research. It warned investors to expect a period of “capex indigestion” accompanying the data center buildout, but it also maintained that AI adoption is expanding far beyond expectations, citing growing monetization from OpenAI’s ChatGPT, Alphabet’s Gemini, and AI-powered CRM systems.

Bank of America Research wrote a note in early August, before the launch of GPT-5, seeing AI as part of a worker productivity “sea change” that will drive an ongoing “innovation premium” for S&P 500 firms. Head of U.S. Equity Strategy Savita Subramanian essentially argued that the inflation wave of the 2020s taught companies to do more with less, to turn people into processes, and that AI will turbo-charge this. “I don’t think it’s necessarily a bubble in the S&P 500,” she told Fortune in an interview, before adding, “I think there are other areas where it’s becoming a little bit bubble-like.”

Subramanian mentioned smaller companies and potentially private lending as areas “that potentially have re-rated too aggressively.” She’s also concerned about the risk of companies diving into data centers to such a great extent, noting that this represents a shift back toward an asset-heavier approach, instead of the asset-light approach that increasingly distinguishes top performance in the U.S. economy.

“I mean, this is new,” she said. “Tech used to be very asset-light and just spent money on R&D and innovation, and now they’re spending money to build out these data centers,” adding that she sees it as potentially marking the end of their asset-light, high-margin existence and basically transforming them into “very asset-intensive and more manufacturing-like than they used to be.” From her perspective, that warrants a lower multiple in the stock market. When asked if that is tantamount to a bubble, if not a correction, she said “it’s starting to happen in places,” and she agrees with the comparison to the railroad boom.

The math and the ghost in the machine

Gary Marcus also cited the fundamentals of math as a reason that he’s concerned, with 
nearly 500 AI unicorns being valued at $2.7 trillion. “That just doesn’t make sense relative to how much revenue is coming [in],” he said. Marcus cited OpenAI reporting $1 billion in revenue in July, but still not being profitable. Speculating, he extrapolated that to OpenAI having roughly half the AI market, and offered a rough calculation that it means about $25 billion a year of revenue for the sector, “which is not nothing, but it costs a lot of money to do this, and there’s trillions of dollars [invested].”

So if Marcus is correct, why haven’t people been listening to him for years? He said he’s been warning people about this for years, too, calling it the “gullibility gap” in his 2019 book 
Rebooting AI and arguing in The New Yorker in 2012 that deep learning was a ladder that wouldn’t reach the moon. For the first 25 years of his career, Marcus trained and practiced as a cognitive scientist, and learned about the “anthropomorphization people do. … [they] look at these machines and make the mistake of attributing to them an intelligence that is not really there, a humanness that is not really there, and they wind up using them as a companion, and they wind up thinking that they’re closer to solving these problems than they actually are.” He said he thinks the bubble inflating to its current extent is in large part because of the human impulse to project ourselves onto things, something a cognitive scientist is trained not to do.

These machines might seem like they’re human, but “they don’t actually work like you,” Marcus said, adding, “this entire market has been based on people not understanding that, imagining that scaling was going to solve all of this, because they don’t really understand the problem. I mean, it’s almost tragic.”

Subramanian, for her part, said she thinks “people love this AI technology because it feels like sorcery. It feels a little magical and mystical … the truth is it hasn’t really changed the world that much yet, but I don’t think it’s something to be dismissed.” She’s also become really taken with it herself. “I’m already using ChatGPT more than my kids are. I mean, it’s kind of interesting to see this. I use ChatGPT for everything now.”

This story was originally featured on 
Fortune.com

本文於 修改第 1 次
回應 回應給此人 推薦文章 列印 加入我的文摘
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7283488
MIT報告:人工智能試行計畫失敗率高達95% -- Sheryl Estrada
推薦1


胡卜凱
等級:8
留言加入好友

 
文章推薦人 (1)

胡卜凱

請參考

Sam Altman admits OpenAI ‘totally screwed up’ its GPT-5 launch and says the company will spend trillions of dollars on data centers


MIT report: 95% of generative AI pilots at companies are failing

Sheryl Estrada, 08/18/25

Good morning. Companies are betting on AI—yet nearly all enterprise pilots are stuck at the starting line.

The GenAI Divide: State of AI in Business 2025
a new report published by MIT’s NANDA initiative, reveals that while generative AI holds promise for enterprises, most initiatives to drive rapid revenue growth are falling flat.

Despite the rush to integrate powerful new models, about 5% of AI pilot programs achieve rapid revenue acceleration; the vast majority stall, delivering little to no measurable impact on P&L. The research—based on 150 interviews with leaders, a survey of 350 employees, and an analysis of 300 public AI deployments—paints a clear divide between success stories and stalled projects.

To unpack these findings, I spoke with Aditya Challapally, the lead author of the report, and a research contributor to project NANDA at MIT.

“Some large companies’ pilots and younger startups are really excelling with generative AI,” Challapally said. Startups led by 19- or 20-year-olds, for example, “have seen revenues jump from zero to $20 million in a year,” he said. “It’s because they pick one pain point, execute well, and partner smartly with companies who use their tools,” he added.

But for 95% of companies in the dataset, generative AI implementation is falling short. The core issue? Not the quality of the AI models, but the “learning gap” for both tools and organizations. While executives often blame regulation or model performance, MIT’s research points to flawed enterprise integration. Generic tools like ChatGPT excel for individuals because of their flexibility, but they stall in enterprise use since they don’t learn from or adapt to workflows, Challapally explained.

The data also reveals a misalignment in resource allocation. More than half of generative AI budgets are devoted to sales and marketing tools, yet MIT found the biggest ROI in back-office automation—eliminating business process outsourcing, cutting external agency costs, and streamlining operations.

What’s behind successful AI deployments?

How companies adopt AI is crucial. Purchasing AI tools from specialized vendors and building partnerships succeed about 67% of the time, while internal builds succeed only one-third as often.

This finding is particularly relevant in financial services and other highly regulated sectors, where many firms are building their own proprietary generative AI systems in 2025. Yet, 
MIT’s research suggests companies see far more failures when going solo.

Companies surveyed were often hesitant to share failure rates, Challapally noted. “Almost everywhere we went, enterprises were trying to build their own tool,” he said, but the data showed purchased solutions delivered more reliable results.

Other key factors for success include empowering line managers—not just central AI labs—to drive adoption, and selecting tools that can integrate deeply and adapt over time.

Workforce disruption is already underway, especially in customer support and administrative roles. Rather than mass layoffs, companies are increasingly not backfilling positions as they become vacant. Most changes are concentrated in jobs previously outsourced due to their perceived low value.

The report also highlights the widespread use of “shadow AI”—unsanctioned tools like ChatGPT—and the ongoing challenge of measuring AI’s impact on productivity and profit.

Looking ahead, the most advanced organizations are already experimenting with agentic AI systems that can learn, remember, and act independently within set boundaries—offering a glimpse at how the next phase of enterprise AI might unfold.


Sheryl Estrada,
sheryl.estrada@fortune.com

This story was originally featured on 
Fortune.com

本文於 修改第 1 次
回應 回應給此人 推薦文章 列印 加入我的文摘
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7283358
人工智能效應:高科技企業今年已裁撤5萬職位-Cryptopolitan
推薦1


胡卜凱
等級:8
留言加入好友

 
文章推薦人 (1)

胡卜凱

請參考

Enterprises Push Ahead with AI-Powered Job Replacement Despite Risks
This CEO laid off nearly 80% of his staff because they refused to adopt AI fast enough. 2 years later, he says he’d do it again

60
歲以下的白領階層員工需要提高警覺。由於薪資積累以及生產力降低(由於腦力/體力衰退)等因素,40歲以上的「老」員工更是屬於「高風險群」;上述兩個因素是我親身經驗之談。

Around 50,000 tech jobs replaced by AI as companies restructure for the future

Cryptopolitan, 08/14/25

Around 50,000 tech job roles have been replaced with AI as companies like Oracle, Intel, Microsoft, and Nextdoor restructured their workforces. The nationwide layoffs have been attributed to firms pursuing large-scale investments in artificial intelligence and managing the AI-related costs.

Oracle confirmed changes in its cloud division, cutting off more than 150 workers in Seattle, the base of its cloud operations. Some of the affected roles include engineering and support positions. Oracle informed staff that performance factors contributed to their release. However, the company continues to hire in other business units.  

Tech companies slash jobs to boost AI and cloud spending

Oracle failed to provide an official statement regarding the number of employees laid off. However, the company stated in June that restructuring and workforce changes occur periodically concerning company strategies, operational reorganizations, and performance reviews. The report revealed that the actions would result in high near-term costs and decreased productivity as employees seek to adjust to the new structure. 

Oracle’s stock has traded near record highs recently following the growth in its cloud business. In June, it announced an agreement with OpenAI to secure 4.5 gigawatts of data center power capacity in the U.S. The deal forms part of initiatives to support AI workloads that require large compute power capacity. The database pioneer has committed tens of billions towards constructing new data centers that meet AI-related tasks. In its Q2 results, it reported a negative free cash flow reflecting the scale of investment it has entered into. 

Big tech firms are making similar moves industry-wide. Microsoft has laid off approximately 15,000 employees since the beginning of 2025. The affected roles include teams across engineering, sales, and support. Amazon and Meta have also laid off staff in 2025. Scale AI, a San Francisco-based firm specializing in AI data labelling, Nextdoor, the neighborhood-focused social networking platform, and Intel have all announced reductions in staff this year. Since January, the industry-wide layoffs have contributed to approximately 50,000 jobs lost across the technology sector.  

In June, Microsoft outlined plans to lay off multiple roles following the 6,000 job cuts in May. At least 22,000 jobs had already been cut in 2025 while 150,000 jobs were lost in 2024 across hundreds of companies. The job cuts occur as companies rush to expand spending in AI development and infrastructure. Amazon and Meta are investing more than $100 billion in AI.   

Industry-wide job cuts spread across multiple geographic locations

Oracle stated that strategy changes and performance outcomes influence workforce changes. It also noted that although such measures are intended to provide financial caution, they may involve transitional costs and short-term productivity issues. 

Industry analysts have noted that layoffs have affected various roles, from entry-level technical positions to senior management. The effects of the layoffs have been distributed across different locations, with much concentration in the Seattle area, the San Francisco Bay Area, and Austin, Texas.

Amid the layoffs, companies continue to recruit in other business units. Oracle’s cloud division indicates that its recruiting positions are aligned with priority growth areas. Roles related to AI infrastructure development, data centre operations, and high-demand enterprise software services are some of the key areas receiving employees. 

Cryptopolitan reported that about 41% of executives expect the tech workforce to shrink by 2030 due to AI automation. The report noted Nvidia’s CEO Jensen Huang’s remarks that while AI will make some jobs obsolete, it will also create new ones, and industries could grow if fresh ideas keep emerging.

本文於 修改第 2 次
回應 回應給此人 推薦文章 列印 加入我的文摘
引用網址:https://city.udn.com/forum/trackback.jsp?no=2976&aid=7283012
頁/共5頁 回應文章第一頁 回應文章上一頁 回應文章下一頁 回應文章最後一頁