Introduction
2023 was the third and so far, hottest summer of artificial intelligence. But what should developers and businesses do? Is it worth cutting back on people and fully trusting the new technology? Or is it all alluvial, and we are waiting for a new AI winter? And how will AI be combined with classic management tools and practices, managing projects, products, and business processes? Does AI break all the rules of the game?
We will delve into these questions, and in the process of thinking and analyzing various factors, I will try to find answers to them.
I will say right away, there will be a minimum of technical information here. And even more so, there may be technical inaccuracies. But this book is not about what type of neural networks to choose for solving a particular problem. I offer you an analyst’s, manager’s, and entrepreneur’s perspective on what’s happening here and now, what to expect globally in the near future, and what to prepare for.
Who can use this book and what can it be useful for?
— Owner and top manager of am companies.
They will understand what AI is, how it works, where trends are going, and what to expect. In other words, they will be able to avoid the main mistake in digitalization — distorted expectations. This means, that they will be able to lead these areas, minimize costs, risks and deadlines.
— IT entrepreneur and founder of startups.
They will be able to understand where the industry is moving, what IT products and integrations should be developed, and what you will have to face in practice.
— IT specialist.
They will be able to look at the issue of development not only from a technical point of view, but also from an economic, and managerial point of view. They will understand why up to 90% of AI products remain unclaimed. Perhaps this will help them in their career development.
— Ordinary people.
They will understand what the future holds for them and whether they should be afraid that AI will replace them. Spoiler alert: creative professionals are under threat.
Our journey will go through 3 major parts.
— First, we will dive in and understand, what AI is, how it all started, what problems and opportunities it has, what trends are developing, and what the future holds for us.
— Next, we’ll look at the synergy of AI with systems-based tools. How they affect each other, in what scenarios AI will be applied in IT solutions.
— At the end, we will go through almost all examples and give recommendations.
In my favorite tradition, the book will contain QR-codes (for the printed version) and active hyperlinks (for the electronic version) to useful articles and interesting materials.
As for the AI itself, it was used in the writing of the book to demonstrate its work, and these fragments are highlighted, as well as to search for ideas. The level of technology is still not enough to take materials and insert them into a book. It is also important to note that my main language is Russian, so I will make all requests to the AI in Russian.
I would like to conclude my preface by thanking the people who helped me:
— to my wife Alisa and son Valery;
— to my parents;
— to my coach Evgeny Bazhov;
— my team, especially Alexander Peremyshlin;
— to my partners and clients who have given me time to think, especially Kirill Neyelov;
— to my project colleagues.
P.S. This book has been translated from my native language using artificial intelligence. Therefore, there may be grammatical and other errors
Part 1. Introduction to Artificial Intelligence
Chapter 1. Introduction to AI
What is artificial intelligence?
To begin with, there is no such definition. Let’s ask the AI itself, or rather several large language models (LLM — large language model). Below are some answers with the original spelling.
GPT4: “Artificial intelligence (AI) is the ability of artificial intelligent systems to perform creative functions that are traditionally considered the prerogative of humans. It is related to the science and technology of creating intelligent machines, especially intelligent computer programs. It’s important to note that AI isn’t necessarily limited to biologically plausible methods.
Existing intelligent systems have rather narrow application areas. For example, programs that can beat a human at chess usually can’t answer questions.
In general, artificial intelligence allows computer systems to mimic human cognitive abilities, such as learning and problem solving.
This is an important area in modern computer science and data science, and its application spans many areas, from finance and medicine to entertainment and transportation”.”
YandexGPT 2: “Artificial Intelligence — AI) is a set of computer science, mathematics, biology, and psychology techniques that develop systems that can perform tasks that usually require human intelligence. Their main task is to model the human mind.”
YandexGPT 3 Pro: “Artificial intelligence (AI) is the ability of intelligent systems to perform creative functions that are traditionally considered the prerogative of humans. It is the science and technology of creating intelligent machines, especially computer programs that can perform tasks that require human intelligence, such as solving complex problems, learning from experience, understanding natural language, and being able to adapt to new situations. AI is related to the use of computers to understand human intelligence, but is not limited to biologically plausible methods.”
GigaChat: “Artificial intelligence (AI) is the ability of intelligent systems to perform creative functions that are traditionally considered the prerogative of humans. It is the science and technology of creating intelligent machines, especially intelligent computer programs. Artificial intelligence is related to the similar task of using computers to understand human intelligence, but is not necessarily limited to biologically plausible methods. Today’s smart systems have rather narrow areas of application. Artificial intelligence was founded as an academic discipline in 1956 and has generated several waves of optimism in the decades since.”
But as a human, the simplest and most understandable definition is closer: AI is any mathematical method that allows you to imitate human or other intelligence from natural nature.
In other words, AI is a huge number of solutions, including primitive mathematical algorithms and expert systems based on rules.
And although this direction was born somewhere in the 50s of the 20th centuries, we are primarily interested in what we understand by this today, in the early 2020s. And tut has three main areas.
1. Neural networks — mathematical models created in the likeness of neural connections of the brain of living beings. Actually, the human brain is a super-complex neural network, the key feature of which is that our neurons are not limited to “on / off” states, but have many more parameters that cannot yet be digitized and fully applied.
2. Machine learning — ML) — statistical methods that allow computers to improve the quality of the task performed with the accumulation of experience and in the process of retraining. This direction has been known since the 1980s.
3. Deep learning (DL) is not only machine learning with the help of a person who says what is true and what is not (as we often raise children, this is called reinforcement learning), but also self-learning of systems (learning without reinforcement, without human participation). This is the simultaneous use of various training and data analysis techniques. This direction has been developing since the 2010s and is considered the most promising for solving creative tasks, and those tasks where the person himself does not understand clear relationships. But here we can’t predict at all what conclusions and results the neural network will come to. Here you can manipulate what data we “feed” to the AI model at the input.
How are AI models trained?
Currently, most AI-models are trained with reinforcement: a person sets input information, the neural network returns an answer, and then the person tells it whether it answered correctly or not. And so on, time after time.
Similarly, the so-called “CAPTCHA” (CAPTCHA, Completely Automated Public Turing test to tell Computers and Humans Apart) work, that is, graphical security tests on websites that calculate who the user is: a person or a computer. This is when, for example, you are shown a picture divided into parts and asked to specify the areas where bicycles are depicted. Or they ask you to enter numbers or letters displayed in an intricate way on the generated image. In addition to the main task (the Turing test), this data is then used to train AI.
At the same time, there is also unsupervised learning, in which the system learns without human feedback. These are the most complex projects, but they also allow you to solve the most complex and creative tasks.
General features of current AI-based solutions
Fundamentally, all AI-based solutions at the current level of development have common problems.
— Amount of training data.
Neural networks need huge amounts of high-quality and marked-up data for training. If a human can learn to distinguish dogs from cats in a couple of examples, then AI needs thousands.
— Dependence on data quality.
Any inaccuracies in the source data strongly affect the final result.
— The ethical component.
There is no ethics for AI. Only math and problem completion. As a result, complex ethical issues arise. For example, why should I knock down autopilot in a desperate situation: an adult, a child, or a pensioner? There are countless similar disputes. For artificial intelligence, there is neither good, nor evil, just like the concept of “common sense”.
— Neural networks cannot evaluate data for reality and logic, and they are also prone to generating poor-quality content and AI hallucinations.
Neural networks simply collect data and do not analyze facts or their connectedness. They make a large number of mistakes, which leads to two problems.
The first is the degradation of search engines. AI created so much low-quality content that search engines (Google and others) began to degrade. Just because there is more low-quality content, it dominates. This is especially helpful for SEO-optimizers of sites that simply outline popular queries for promotion.
The second is the degradation of AI models. Generative models also use the Internet for “retraining”. As a result, people, using AI and not checking for it, fill the Internet with low-quality content themselves. And the AI starts using it. The result is a vicious circle that leads to more and more problems.
An article on this topic is also available by using the QR code and hyperlink.
Realizing the problem of generating the largest amount of disinformation content by AI, Google conducted a study on this topic. Scientists analyzed about two hundred media articles (from January 2023 to March 2024) about cases when artificial intelligence was used for other purposes. According to the results, most often AI is used to generate fake images of people and false evidence of something.
— The quality of “teachers”.
Almost all neural networks are taught by people: they form requests and give feedback. And there are many limitations here: who teaches you what, based on what data, and for what purpose?
— People’s readiness.
We should expect huge resistance from people whose work will be taken away by neural networks.
— Fear of the unknown.
Sooner or later, neural networks will become smarter than us. And people are afraid of this, which means that they will slow down development and impose numerous restrictions.
— Unpredictability.
Sometimes everything goes as planned, and sometimes (even if the neural network copes well with its task) even the creators struggle to understand how the algorithms work. The lack of predictability makes it extremely difficult to eliminate and correct errors in neural network algorithms. We are only learning to understand what we have created ourselves.
— Restriction by type of activity.
All AI for mid-2024 is weak (we’ll discuss this term in the next chapter). Currently, AI algorithms are good for performing targeted tasks, but they do not generalize their knowledge well. Unlike a human, an AI trained to play chess will not be able to play another similar game, such as checkers. In addition, even deep learning does a poor job of processing data that deviates from its training examples. To effectively use the same ChatGPT, you must initially be an expert in the industry and formulate a conscious and clear request.
— Costs of creation and operation.
It takes a lot of money to create neural networks. According to a report from Guosheng Securities, the cost of training a relatively primitive LLM GPT-3 LLM was about $ 1.4 million. For GPT-4, the amounts already go into the tens of millions of dollars.
If we take ChatGPT3 as an example, then only to process all requests from users, more 30than 30,000 NVIDIA A100 GPUs were needed. Electricity cost about $ 50,000 a day. It requires a team and resources (money, equipment) to ensure their “life”. You also need to take into account the cost of maintenance engineers.
Again, these are common drawbacks for all AI solutions. Later on, we will return to this topic several times and discuss these shortcomings in more practical examples.
Chapter 2: Weak, Strong, and Super-Strong AI
Now let’s talk about three concepts-weak, strong, and super-strong AI.
Weak AI
All that we are seeing now is a weak AI (ANI, Narrow AI). It can solve highly specialized tasks that it was originally designed for. For example, it can distinguish a dog from a cat, play chess, analyze videos and improve the quality of video / audio, advise on a subject area, and so on. But, for example, the strongest weakest AI for playing chess is absolutely useless for playing checkers. And AI for project management consulting is absolutely useless for planning equipment maintenance.
Strong and super — strong AI-what is it?
If the definition, of what AI, is confusing, then the term “strong AI” or “general AI” is even more complicated. Let’s start by using LLM services again. To do this, I made a request: “Give a transcript of the term AGI and its brief description.” Below I will give the answers of several LLMs in the original spelling.
YandexGPT 2: “AGI (Artificial General Intelligence) is a general-purpose artificial intelligence that can solve a variety of tasks typical of the human mind. AGI has the flexibility and adaptability of the human brain and can learn independently”.”
YandexGPT 3 Pro: “AGI (Artificial General Intelligence) is a concept for creating artificial intelligence that will be able to solve a wide range of tasks at the level of human intelligence. This is the next step in the development of artificial intelligence after highly specialized systems, such as speech recognition or computer vision systems.”
YandexGPT 4 Pro RC: “AGI (Artificial General Intelligence) is artificial general intelligence, which is a hypothetical form of AI that can understand and solve problems at the same level as humans. AGI can learn, adapt, and make decisions in various fields, such as science, art, and social interactions.
Unlike specialized AI systems that are designed to perform specific tasks, AGI strives for versatility and the ability to solve a wide range of problems. However, creating an AGI remains a complex task that requires significant research and development in machine learning, neural networks, and other technologies.”
Giga Chat: “AGI (Artificial General Intelligence) — general artificial intelligence. This is the concept of creating an intelligent machine that can solve any intellectual tasks in the same way as a human does, and perhaps better than him.”
ChatGPT4: “AGI stands for Artificial General Intelligence (Automated General Intelligence), which is a type of artificial intelligence that can perform a wide range of intelligent tasks at the human level or even surpass it. This is in contrast to highly specialized AI, which is designed for specific tasks. Building an AGI is one of the main goals of AI research and companies like Open AI. The time to reach AGI is the subject of discussions withered researchers and experts.”
However, the original query also has a significant impact on the result.
I will also give you some examples of the Q & A bundle for LLM YandexGPT 3 Pro.
Question 1: “Hi. What is strong AI or AGI?”
Answer 1: “Hi! Strong AI, also known as AGI (Artificial General Intelligence), is the concept of creating artificial intelligence that can perform any intellectual tasks available to humans, but not necessarily in the same way as a human does. Unlike weak AI, which is designed to perform specific tasks, strong AI tends to mimic human intelligence in general, including analytical thinking and learning abilities”.”
Question 2: “AGI decryption”.
Answer 2: “AGI — Artificial General Intelligence) is the concept of creating artificial intelligence that will be able to solve a wide range of tasks at the level of human intelligence”.”
Question 3: “What is the explanation of the term AGI and its description?”
Answer 3: “AGI (Artificial General Intelligence) is a strong general — level artificial intelligence capable of thinking and acting like a human. They can make independent conclusions, learn, and possibly become self-aware at some point in time. Unlike Narrow AI, AGI is the entire brain capable of solving a set of tasks”.”
As you can see, a lot depends on the request. And this is one of the key limitations of generative AI (GII), but we’ll talk about it in the next chapter.
Now I will give the definition that, in my opinion, most accurately defines its essence.
Strong or general AI (AGI) is an AI that can navigate changing conditions, model and predict the development of a situation. And if the situation goes beyond the standard algorithms, then find a solution yourself. For example, to solve the problem “go to university” and learn the rules of the game of checkers, and instead of chess, start playing checkers.
What qualities should such an AI have?
Thinking is the use of methods such as deduction, induction, association, etc., which are aimed at extracting facts from information, their presentation (preservation). This will make it possible to solve problems more accurately in conditions of uncertainty
Memory is the use of various types of memory (short-term, long-term). That is, the tasks should be solved taking into account the accumulated experience. Now, if you chat with ChatGPT 4, you will see that the algorithm has a small short-term memory and after a while forgets where it all started. In general, in my opinion, the issue of memory and the “massiveness” of AI models will be a key limitation in the development of AI. More on this below.
Planning — tactical and strategic. Yes, there are already studies that claim that AI can plan its actions and even deceive a person to achieve its goals. But now it’s still only in its infancy. The deeper the planning goes, especially in conditions of uncertainty, the more capacity is needed. After all, it’s one thing to plan a game of chess 3—6 steps deep, where all the rules are clear, and another in a situation of uncertainty.
Training — simulating the actions of another object and learning through experimentation. Now he is learning from large amounts of data, but he does not model or conduct experiments himself. Although we do not fully understand how the same Chat GPT works, and this is one of the main problems, learning requires the formation of long-term memory and complex relationships. And this, as we understand it, is a problem for AI.
And now no one has such a strong AI right now. And the statement about the imminent (in the2024—2028horizon) appearance of a strong AI, in my opinion, is erroneous or speculative. Although maybe, I have too limited knowledge…
Yes, ChatGPT from Open AI and other LLMs can generate text / illustrations / videos through query analysis and big data processing. But they only broadcast what the creators gave them for training. They search for the most appropriate combinations of words and sentences, words and images for the query, and in general, associative combinations. But don’t be under any illusions, this is just mathematics and statistics. And in their answers, there are a lot of “marriage” and “hallucinations”. They are not yet ready for real interaction with the world.
Here is a simple LLM-based example.
I — come from a project management background, and as part of my work, I decided to write an article about why Agile is not a universal approach to project management, and why, if applied correctly, any company and team should move closer to hybrid or cascading approaches.
My assistants in writing text were all the same ChatGPT, GigaChat, YandexGPT and AI Search.
I formulated a simple request: “Write a post about why Agile is infinitely wrong to apply. Emphasize that conducting retrospectives should lead away from the situation of uncertainty to the situation of complex ordered or simple ordered systems.”
ChatGPT4’s answer: “Agile is an approach to software development that is based on iterative and incremental product development, collaboration between self-organizing teams, and continuous process improvement. One of the key principles of Agile is to conduct regular retrospectives, i.e., sessions where the team analyzes its work, identifies problems and finds ways to solve them.
Retrospectives are a powerful tool for improving the efficiency and quality of work, but only if they are conducted correctly. Unfortunately, many teams make the mistake of applying Agile indefinitely, without taking into account the context and changes in the environment. This leads to the fact that retrospectives become useless, boring and demotivating.
To avoid this, you need to understand that Agile is not a universal solution for every situation. Depending on the degree of uncertainty and complexity of the project, the team may be in one of four zones: chaos, complexity, order, or simplicity. These zones can be represented as a matrix, where the degree of uncertainty is plotted horizontally, and the –degree of complexity is plotted vertically.
!Matrix of zones (the chatbot’s editorial staff, which is exactly how he designated this area)
Each zone requires a different approach to retrospectives and improvements. In the chaos zone, the team faces unpredictable and extreme situations that require rapid response and adaptation. In this case, retrospectives should be short, frequent, and focused on survival and stabilization. In the difficulty zone, the team works with unclear and dynamic requirements that require creativity and experimentation. In this case, retrospectives should be long, sparse, and focused on learning and innovation. In the order zone, the team works with clear and stable requirements that require discipline and optimization. In this case, retrospectives should be average, regular, and focused on efficiency and quality. In the simplicity zone, the team works with obvious and simple requirements that require minimal effort and attention. In this case, retrospectives can be omitted or conducted as necessary.
However, it is important to understand that zones are not static and can change over time. Therefore, the team must constantly evaluate its context and adapt its approach to retrospectives. The purpose of retrospectives is not to apply Agile indefinitely, but to lead the team from a situation of uncertainty to a situation of complex ordered or simple ordered systems. Only then will retrospectives bring benefits and pleasure to the team and the project”.”
Answer GigaChat: “Agile is a software development methodology that has become very popular in recent years. It offers a flexible approach to project management that allows you to quickly respond to changes and adapt to new requirements. However, the endless application of Agile can lead to problems.
One of the main reasons why Agile is infinitely wrong to apply is that it can lead to a situation of uncertainty. In Agile projects, we often encounter changes in requirements and priorities, which can cause uncertainty about what needs to be done next.
Conducting retrospectives is one way to deal with this uncertainty. Retrospectives help the team understand what was done right and what needs to be improved in the future. They also help the team move from a situation of uncertainty to a situation of complex ordered or simple ordered systems.
However, if we endlessly apply Agile without doing retrospectives, then we may be left in a situation of uncertainty. This can lead to the team spending a lot of time fixing bugs and adapting to new requirements, rather than creating new functionality.
In addition, the endless use of Agile can lead to the fact that the team will constantly work on small tasks and will not have time to create large functional blocks. This may cause the project to develop slowly and not meet the customer’s expectations.
In general, Agile is a great approach to project management, but its endless application can lead to problems. Conducting retrospectives helps the team move from a situation of uncertainty to a situation of complex ordered or simple ordered systems and improve the quality of the team’s work.”
Yandex GPT2’s answer: “The endless use of Agile can lead to incorrect decisions and product quality problems. It is important to understand that Agile is not a universal solution for all projects and teams, and its constant use can lead to loss of efficiency and productivity.
One of the key principles of Agile is to conduct retrospective sessions that help the team analyze their past successes and failures. However, if these retrospectives are held constantly, it can lead the team away from solving real problems and tasks.
Instead of helping the team develop and become more effective, constant retrospectives can lead to a situation where the team is constantly discussing the same problems and cannot move forward. This can create a situation of uncertainty when the team does not know what to do next, or vice versa, it can lead to the creation of complex and disordered systems that only complicate the work process.
To avoid this, it is important to determine when and how often to conduct retrospective sessions. They should focus on analyzing specific issues and opportunities for improvement, rather than discussing the same issues over and over again.
It’s also important to understand that each team and project is unique, and what works for one team may not work for another”.”
YandexGPT 3 Pro: “Why Agile might Agile be the wrong choice?
Agile is a project management methodology that involves an iterative approach to software development. It includes regular team meetings, constant interaction with the customer, and readiness for changes.
However, using Agile is not always the right solution. In some cases, the use of this methodology can lead to negative consequences.
Here are a few reasons why Agile might be the wrong choice:
Not suitable for large projects. Agile is best suited for small projects where you can quickly respond to changes. In large projects that require more careful planning and coordination, Agile can lead to chaos and loss of control.
Not suitable for projects with clear requirements. If the project has clear requirements that are not subject to change, Agile can lead to unnecessary spending of time and resources.
Not suitable for projects with a long development cycle. Agile involves short development cycles that allow you to respond quickly to changes. In projects with a long development cycle, this may cause the project to constantly change and make it impossible to complete.
It is important to understand that Agile is not a universal methodology that is suitable for all projects. The choice of methodology should be based on the specifics of the project, its goals and objectives.
How to avoid negative consequences of using Agile?
To avoid negative consequences of using Agile, it is necessary to carefully analyze the project and determine whether this methodology is suitable for its implementation. If you decide to use Agile, you must follow the following rules:
Conduct regular retrospectives. Retrospectives allow the team to analyze their work and identify weaknesses. This helps you avoid repeating mistakes and improve your work efficiency.
Set clear goals and deadlines. This will help the team focus on completing tasks and avoid wasting time.
Maintain communication with the customer. This will allow you to receive timely feedback and make necessary changes to the project.
Conducting retrospectives should lead away from the situation of uncertainty to the situation of complex ordered or simple ordered systems. This will allow the team to analyze their work, identify weaknesses and take measures to eliminate them.
In conclusion, Agile is an effective methodology that can bring many benefits if applied correctly. However, its use requires careful planning and analysis”.”
AI Search’s response: “Hello, friends! Today I would like to share my thoughts on why Agile, despite its popularity and effectiveness, is often used incorrectly. I will pay particular attention to why conducting retrospectives is an important tool for moving from a situation of uncertainty to orderly or complex systems.
Agile is a philosophy and approach to software development that changes the traditional approach to project management, making it more flexible and adaptive to change. However, despite its obvious advantages, Agile can be applied incorrectly, and as a result, its potential is not fully revealed.
One of the main problems of using Agile incorrectly is the lack of regular retrospectives. A retrospective, as an expert assessment of past work, helps teams take into account the changes, shortcomings, and improvements that have occurred when planning future iterations. This allows you to hone the work process and increase efficiency. However, some teams often ignore conducting retrospectives or conduct them unsystematically.
By not conducting regular retrospectives, teams miss opportunities to see their issues, validate or change their vision, and customize development milestones. As a result, teams may be stuck in a situation of uncertainty and unable to move forward.
One of the main principles of Agile is to ensure a fast response to changes and continuous improvement of the work process. If retrospectives are not conducted regularly, teams are not sufficiently aware of their mistakes and omissions, which makes it difficult to progress towards achieving order and efficiency.
In addition, conducting regular retrospectives helps teams move from a situation of uncertainty to complex or simple ordered systems. Starting with simple, streamlined systems where rules and processes are easily defined and controlled, teams can gradually move to more complex systems where rules are more ambiguous and require more adaptation.
So, flashbacks are an important part of the Agile process and help teams move from uncertainty to order and complex systems. They allow teams to recognize their mistakes and shortcomings, as well as see areas for improvement. Do not neglect regular retrospectives so that your team can maximize the potential of Agile and achieve high performance.
I wish you successful work in applying Agile and continuous improvement of your processes! Be flexible and persistent and considerate”.”
Well, and my post: “Why is Agile not eternal and not a panacea?”
“In recent years, everyone declares that they are Agile.
Let’s be honest: most people hide banal chaos under this. After all, Agile is not about the lack of planning. In addition, he is very sensitive to the implementation of certain rules and to the stability of the team, its motivation. If you don’t meet these conditions, don’t use Agile in the future.
And using the same Scrum, Kanban or other approaches should lead to the absence of the need to implement Agile projects.
But why?
Let’s remember that Agile was originally designed to work and implement projects in a highly uncertain environment.
There is even a special tool — model Kenevin, that helps you understand what situation you are in and what approach you need to choose, what to focus on. So, in ordered systems (simple or complex situations), Agile, on the contrary, is contraindicated, because it reduces the cost of achieving results. That is, Agile is effective in cases where you need to do — I don’t know what”. But it’s not about efficiency.
Now let’s take a look at the retrospective. All approaches within Agile involve regular retrospectives, analysis of their work, and interaction with the client / customer / partner. In other words, the very logic of these tools is to get away from an uncertain situation and learn how to predict them and become more effective.
If you constantly (every six-months or a year) change jobs or constantly launch new products, and do not replicate certain solutions (which is strange for a business), then yes, you need to be Agile.
But if you have a segment in the AC, and you have worked with experience and expertise on typical approaches / products that need to be adjusted and adapted in a small part, then sooner or later you must move away from Agile and come to an ordered system where cascading or hybrid approaches are needed. These retrospectives should lead you to understand what customers want in 90%of cases and how the organization works.
As a result, if you are Agile on an ongoing basis and everywhere, and not for the period of rebuilding / launching / adapting, then this may indicate that:
— you are not following the Agile tools;
— you haven’t found your product or niche, and you haven’t developed the necessary expertise;
— you always have a unique product / project (which should be reflected in the high price of your services);
— the organization is “sick” from the inside, and this way you mask high turnover, lack of work on processes, etc.
What do you think about it?”
The quality of neural network responses is more like a weak student who just picks up similar words, but doesn’t understand what he’s talking about. Could I get a similar post from the neural network? So that she can convey my meaning? Yes, of course, but I wrote my post in 25 minutes, and how long would it take me to try to achieve such a result from AI? And does it look like a really thinking intelligence? And in general, on Intelligence?
The fact is that our thinking is strongly tied to ideas and images. That is, basically in thinking we use an approach from the general to the particular, we think in images, and then from some vague, but understandable to us at the unconscious level of abstractions we come to the smallest details. In parallel, we conduct a bunch of associations, completely tied again to images. It is not for nothing that there are so many metaphors, comparisons, allusions and word changes in human speech.
Modern neural networks “think” quite differently. They simply do not know the concepts of “meaning”, “essence”, “idea”, as well as the process of reflection itself. The texts generated by the neural network look logical not because the neural network really understands what it is writing about, but because it is able to calculate which words are most often next to each other in the texts available for its training on a given topic. Pure mathematics based on calculating probabilities.
However, for the sake of justice, we should also note significant progress in this direction. Now developers are trying to teach AI to think: it asks itself questions and, giving answers, draws conclusions. This is an imitation of an internal dialog to improve the quality of content, which is typical for models from Open AI at the end of 2024 and Yandex GPT 4 Pro RC.
Here is an example of the response of the Yandex GPT 4 Pro RC model to the request” Write a post about why Agile is infinitely applied incorrectly. Emphasize that conducting retrospectives should lead away from the situation of uncertainty to the situation of complex ordered or simple ordered systems.”
Yandex GPT 4 Pro RC Response: “Why Agile doesn’t Always Work: how retrospectives can help move from uncertainty to order.
Agile is a flexible approach to project management that allows you to quickly adapt to changes and improve processes based on feedback. However, despite all the advantages, applying Agile without considering the context can lead to problems. In this post, we will look at why Agile should not be applied indefinitely and how retrospectives help to move from a situation of uncertainty to ordered systems.
Problems of infinite Agile Application
1. Lack of clear structure. Agile is based on the principles of flexibility and adaptability, which can lead to a lack of a clear structure and action plan. This can make it difficult to make decisions and manage risks.
2. Complexity of evaluating results. Without a clear plan and criteria for evaluating results, it is difficult to determine how well the project is progressing. This may raise doubts about the effectiveness of the approach.
3. Risk of losing control. The endless use of Agile can lead to a loss of control over the project. Without clear boundaries and constraints, the team may lose focus and start to deviate from the original goal.
4. Inefficient use of resources. If the team is constantly adapting to new conditions, this can lead to inefficient use of resources. Instead of focusing on key tasks, the team will spend time making constant changes.
5. Reduced team motivation. Constant adaptation and changes can cause stress and fatigue for the team. This can reduce motivation and productivity.
How retrospectives help you move to streamlined systems
Retrospectives are one of the key Agile tools that help the team analyze their work and identify areas for improvement. They allow us to move from a situation of uncertainty to more ordered systems, where processes and decisions become more predictable and manageable.
Here are some ways that flashbacks can help:
— Identifying problems and their solutions. Retrospectives allow you to identify problems that have arisen in the course of work, and find ways to solve them. This helps you avoid making mistakes again in the future.
— Improve processes. Retrospectives help you understand which processes are working well and which need to be improved. This allows you to optimize your work processes and increase efficiency.
— Increase motivation. When a team sees the results of their efforts and understands that they bring benefits, it increases motivation and job satisfaction.
— Create a culture of continuous improvement. Retrospectives foster a culture in which continuous improvement and adaptation are the norm. This helps the team to be more flexible and ready for changes.
In general, retrospectives are an important tool for moving from a situation of uncertainty to more structured and orderly processes. They help the team adapt to changes, improve their processes, and achieve their goals”.”
Yes, the model has become academic and it is clear that these are excerpts from articles, and not reflections. But this is a huge progress, and you can already work with the model and use it as a tool.
However, with only weak AI available and dreaming of a more or less strong one, various researchers already distinguish super-strong AI (ASI, Artificial Superintelligence) in their classification.
This is about an AI, that:
— It can solve both routine and creative tasks;
— instantly navigates uncertainty even without a network or internet connection;
— adapts problem solving to the context of circumstances and available opportunities / resources;
— understands people’s emotions (not only through a text query, but also based on the analysis of facial expressions, voice timbre, and other parameters) and takes them into account in your work;
— able to interact independently with the real world to solve problems.
This is the kind of AI that we only see in science fiction movies so far. Even AI itself writes about ASI as a “hypothetical concept” and “a subject of science fiction and active research in the field of artificial intelligence” (YandexGPT 3 Pro). This is a certain desired point in the distant future, which is not yet possible to reach.
Chat GPT 4 describes ASI as follows: “hypothetical artificial intelligence system with intelligence superior to human. This is not just an advanced AI, but a system that can adapt, learn and develop independently, exceeding human abilities in learning, problem solving, cognitive functions and self-improvement.
Super-strong AI, or ASI, will be able to understand and process many types of data (text, images, sound, video), which will allow it to perform verbal tasks and make decisions. It will use advanced AI technologies such as multidimensional language models (LLMs), multi-bit neural networks, and evolutionary algorithms.
Currently, ASI remains a conceptual and speculative stage in the development of AI, but it represents a significant step forward from the current level of AI”.”
And if there are hundreds of weak AIS now, for each task, then there will be only dozens of strong AIS (most likely there will be a division in directions, we will consider this in the next block), and the super-strong AI will be one for the state and even the entire planet.
Limitations on the path to strong AI
To be honest, I have little faith in the rapid emergence of a strong or super-strong AI.
First, this is a very costly and complex task from the point of view of regulatory restrictions. The era of uncontrolled AI development is ending. More and more restrictions will be imposed on it. We’ll discuss AI regulation in a separate chapter.
The key trend is a risk-based approach. So, in a risk-based approach, strong and super -strong AI will be at the upper level of risk. This means, that legislative measures will also be protective.
Secondly, this is a difficult task from a technical point of view, and a strong AI will be very vulnerable.”
Now, in the mid-2020s, creating and training a strong AI requires huge computing power. So, according to Leopold Aschenbrenner, a former Open AI employee from the Super alignment team, it will require the creation of a data-center worth a trillion US dollars. And its power consumption will exceed all current electricity generation in the United States.
We also need complex AI models (orders of magnitude more complex than the current ones) and a combination of them (not just LLM for query analysis). In other words, it is possible to exponentially increase the number of neurons, build connections between neurons, and coordinate the work of various segments.
At the same time, it should be understood that if human neurons can be in several states, and activation can occur “in different ways” (biologists will forgive me for such simplifications), then machine AI is a simplified model that does not know how to do this. Simply put, machine’s 80—100 billion neurons are not equal to a human’s 80—100 billion. The machine will need more neurons to perform similar tasks. The same GPT4 is estimated at 100 trillion parameters (conditionally neurons), and it is still inferior to humans.
All this leads to several factors.
The first factor is that increasing complexity always leads to reliability problems, and the number of failure points increases.
Complex AI models are difficult to both create and maintain from degradation over time, in the process of operation. AI models need to be constantly “serviced”. If this is not done, then a strong AI will begin to degrade, and neural connections will be destroyed, this is a normal process. Any complex neural network, if it is not constantly developing, begins to destroy unnecessary connections. At the same time, maintaining relationships between neurons is a very expensive task. AI will always optimize and search for the most efficient solution to the problem, which means, that it will start turning off unnecessary energy consumers.
That is, the AI will look like an old man with dementia, and the “life” period will be greatly reduced. Imagine what a strong AI with its capabilities can do, but which will suffer from memory loss and sharp reversals to the state of the child? Even for current AI solutions, this is an actual problem.
Let’s give a couple of simple real-life examples.
You can compare building a strong AI to training your human muscles. When we first start working out in the gym and get involved in strength training, bodybuilding, then progress is fast, but the further we go, the lower the efficiency and increase in results. You need more and more resources (time, exercise, and energy from food) to progress. Yes, even just holding the form is becoming more and more difficult. Plus, the increase in strength comes from the thickness of the muscle section, but the mass grows from the volume. As a result, the muscle will at some point become so heavy that it will not be able to move itself, and may even damage itself.
Another example of the complexity of creating, but already from the field of engineering, is Formula 1 races. For example, a 1-second lag can be eliminated if you invest 1 million and 1 year. But to win back the crucial 0.2 seconds, it may already take 10 million and 2 years of work. And the fundamental limitations of the design of the car can force you to reconsider the whole concept of a racing car.
And even if you look at ordinary cars, everything is exactly the same. Modern cars are more expensive to create and maintain, and without special equipment, it is impossible to change even a light bulb. If you take modern hyper cars, then after each departure, entire teams of technicians are required for maintenance.
If you look at it from the point of view of AI development, there are two key parameters in this area:
— number of layers of neurons (depth of the AI model).
— the number of neurons in each layer (layer width).
Depth determines how great the AI’s ability to abstract is. Insufficient depth of the model leads to a problem with the inability to perform deep system analysis, superficiality of this analysis and judgments.
The width of the layers determines the number of parameters / criteria that the neural network can use on each layer. The more they are, the more complex models are used and the more complete reflection of the real world is possible.
However, if the number of o layers has a linear effect on the function, then the width does not. As a result, we get the same analogy with muscle — the size of top AI models (LLM) exceeds a trillion parameters, but models 2 orders of magnitude smaller do not have a critical drop in performance and quality of responses. More important is what data the model is trained on and whether it has a specialization.
Below are statistics for LLM models from different manufacturers.
Compare the indicators LLaMa 2 70B, LLaMa 2 7B, LLaMa 2 13B. Indicators 70B, 7B and 13B conditionally demonstrate the complexity and training of models — the higher the value, the better. But as we can see, the quality of responses does not radically change, while the price and labor costs for development increase significantly.
And we can see how leaders are building up computing power, building new data centers and hurriedly solving energy supply and cooling issues for these monsters. At the same time, improving the quality of the model by a conditional 2% requires an increase in computing power by an order of magnitude.
Now a practical example to the question of maintenance and maintenance due to degradation. Tut will also be noticeable about the effect of people. Any AI, especially at an early stage, will learn based on feedback from people (their satisfaction, initial requests and tasks). For example, the same ChatGPT4 uses user requests to retrain its model in order to give more relevant answers and at the same time reduce the load on the “brain”. And at the end of 2023, there were articles that the AI model has become “more lazy”. The chatbot either refuses to answer questions, interrupts the conversation, or simply responds with excerpts from search engines and other sites. And by mid-2024, this has already become the norm, when the model simply cites excerpts from Wikipedia.
One possible reason for this is the simplification of the user requests themselves (they are becoming more primitive). After all, LLMs do not invent anything new, these models try to understand what you want them to say and adapt to it (in other words, they also form stereotypes). It also looks for the maximum efficiency of the labor-result bundle, “disabling” unnecessary neural connections. This is called function maximization. Just math and statistics.
Moreover, this problem will be typical not only for LLM.
As a result, to prevent the AI from becoming degraded, you will have to load it with complex research, while limiting its load to primitive tasks. And once it is released into the open world, the ratio of tasks will be in favor of simple and primitive user requests or solving applied problems.
Remember yourself. Do you really need to evolve to survive and reproduce? Or what is the correlation between intellectual and routine tasks in your work? What level of math problems do you solve in this job? Do you need integrals and probability theory, or just math up to 9th grade?
The second factor is the amount of data and hallucinations.
Yes, we can increase the current models by XXXX times. But the same ChatGPT5 prototype already lacks training data in 2024. They gave him everything, they had. And with a modern AI that will navigate uncertainty, there simply won’t be enough data at the current level of technology development. You need to collect metadata about user behavior, think about how to circumvent copyright and ethical restrictions, and collect user consent.
In addition, using the current LLMs as an example, we can see another trend. The more “omniscient” a model is, the more inaccuracies, errors, abstractions, and hallucinations it has. At the same time, if you take a basic model and give it a specific subject area as knowledge, then the quality of its responses increases: they are more objective, she fantasizes less (hallucinates) and makes fewer mistakes.
The third factor is vulnerability and costs.
As we discussed above, we will need to create a data-center worth a trillion US dollars. And its power consumption will exceed all current electricity generation in the United States. This means, that the creation of an energy infrastructure with a whole complex of nuclear power plants will also be required. Yes, windmills and solar panels can’t solve this problem.
Now let’s add that the AI model will be tied to its “base”, and then one successful cyber-attack on the energy infrastructure will de-energize the entire “brain”.
And why should such an AI be tied to the center, why can’t it be distributed?
First, distributed computing still loses performance and efficiency. These are heterogeneous computing capacities that are also loaded with other tasks and processes. In addition, a distributed network cannot guarantee the operation of computing power all the time. Something turns on, something turns off. The available power will be unstable.
Secondly, it is a vulnerability to attacks on communication channels and the same distributed infrastructure. Imagine that suddenly 10% of the neurons in your brain just turned off (blocking communication channels or just turned off due to an attack), and the rest are working half-heartedly (interference, etc.). As a result, we again have the risk of a strong AI that forgets who it is, where it is, for something, then just thinks for a long time.
And if everything comes to the point that a strong AI will need a mobile (mobile) body to interact with the world, then this will be even more difficult to implement. After all, how to provide all this with energy and cool it? Where do I get data processing power? Plus, you also need to add machine vision and image recognition, as well as processing other sensors (temperature, hearing, etc.). This is huge computing power and the need for cooling and energy.
That is, it will be a limited AI with a permanent wireless connection to the main center. And this is again a vulnerability. Modern communication channels give a higher speed, but this affects the reduction of range and penetration, vulnerability to electronic warfare. In other words, we get an increase in the load on the communication infrastructure and an increase in risks.
Here, of course, you can object. For example, the fact that you can take a pre-trained model and make it local. In much the same way as I suggest deploying local AI models with “additional training” in the subject area. Yes, in this form, all this can work on the same server. But such an AI will be very limited, it will be “stupid” in conditions of uncertainty, and it will still need energy and a connection to the data transmission network. That is, this story is not about the creation of human-like super-beings.
All this leads to questions about the economic feasibility of investing in this area. Especially considering two key trends in the development of generative AI:
— creating cheap and simple local models for solving specialized tasks;
— create AI orchestrators that will decompose a request into several local tasks and then redistribute it between different local models.
Thus, weak models with narrow specialization will remain more free and easier to create. At the same time, they will be able to solve our tasks. And as a result, we have a simpler and cheaper solution to work tasks than creating a strong AI.
Of course, we leave out neuromorphic and quantum systems, but we will discuss this topic little later. And, of course, there may be mistakes in my individual figures and arguments, but in general I am convinced that strong AI is not a matter of the near future.
In summary, strong AI has several fundamental problems.
— Exponential growth in the complexity of developing and countering degradation and complex models.
— Lack of data for training.
— Cost of creation and operation.
— Attachment to data centers and demanding computing resources.
— Low efficiency of current models compared to the human brain.
It is overcoming these problems that will determine the further vector of development of the entire technology: either a strong AI will still appear, or we will move into the plane of development of weak AI and AI orchestrators, which will coordinate the work of dozens of weak models.
But now strong AI does not fit in with ESG in any way, environmentalists like it commercial. Its creation is possible only within the framework of strategic and national projects financed by the state. And here is one of the interesting facts in this direction: the former head of the US National Security Agency (until 2023a), retired General, Paul Nakasone joined the board of directors of Open AI in 2024. The official version is for organizing Chat GPT security.
I also recommend reading the document titled “Situational Awareness: The Coming Decade”. Its author is Leopold Aschenbrenner, a former Open AI employee from the Super alignment team. The document is available by QR-code and hyperlink.
A shortened analysis of this document is also available using the QR-code and hyperlink below.
To simplify it completely, the author’s key theses are:
— By 2027, strong AI (AGI) will become a reality.
I disagree with this statement. My arguments are given above, plus some of the theses below and risk descriptions from the authors say the same. But again, what is meant by the term AGI? I have already given my own definition, but there is no single term.
— AGI is now a key geopolitical resource. We forget about nuclear weapons; this is the past. Each country will strive to get AGI first, as in its time an atomic bomb.
The thesis is controversial. Yes, this is a great resource. But it seems to me that its value is overestimated, especially given the complexity of its creation and the mandatory future errors in its work.
— Creating an AGI will require a single computing cluster worth a trillion US dollars. Microsoft is already building one for Open AI.
Computing power also requires spending on people and solving fundamental problems.
— This cluster will consume more electricity than the entire US generation.
We discussed this thesis above. More than a trillion dollars is also invested in electricity generation, and there are also risks.
— AGI funding will come from tech giants — already Nvidia, Microsoft, Amazon, and Google are allocating $100 billion a quarter to AI alone.
I believe that government funding and, consequently, intervention is essential.
— By 2030, annual investment in AI will reach $8 trillion.
Excellent observation. Now the question arises, is this economically justified?
Despite all the optimism of Leopold Aschenbrenner regarding the timing of the creation of AGI, he himself notes a number of limitations:
— Lack of computing power for conducting experiments.
— Fundamental limitations associated with algorithmic progress
— Ideas are becoming more complex, so it is likely that AI researchers (AI agents who will conduct research for people) will only maintain the current rate of progress, and not increase it significantly. However, Aschenbrenner believes that these obstacles can slow down, but not stop, the growth of AI systems ' intelligence.
Chapter 3. What can weak AI do and general trends?
Weak AI in applied tasks
As you, probably already, understood, I am a proponent of using what is available. Perhaps this is my experience in crisis management, and whether it’s just an erroneous opinion. But still, where can the current weak AI based on machine learning be applied?
The most relevant areas for applying AI with machine learning are:
— forecasting and preparing recommendations for decisions taken;
— analysis of complex data without clear relationships, including for forecasting and decision-making;
— process optimization;
— image recognition, including images and voice recordings;
— automating the execution of individual tasks, including through content generation.
The new direction, which is at its peak of popularity in 2023—2024, is image recognition, including images and voice recordings, and content generation. This is where the bulk of AI developers and most of these services come from.
At the same time, the combination of AI + IoT (Internet of Things) deserves special attention:
— AI receives pure big data, which does not contain human errors, for training and finding relationships.
— The effectiveness of IoT increases, as it becomes possible to create predictive analytics and early detection of deviations.
Key trends
— Machine learning is moving towards an increasingly low entry threshold.
One of the tasks that developers are currently solving, is to simplify the creation of AI models to the level of site designers, where special knowledge and skills are not needed for basic application. The creation of neural networks and data science is already developing according to the “service as a service” model, for example, DSaaS — Data Science as a Service.
You can start learning about machine learning with AUTO ML, its free version, or DSaaS with initial audit, consulting, and data markup. You can even get data markup for free. All this reduces the entry threshold.
— Creating neural networks that need less and less data for training.
A few years ago, to fake your voice, it was necessary to provide a neural network with one or two hours of recording your speech. About two years ago, this indicator dropped to a few minutes. Well, in 2023, Microsoft introduced a neural network that takes just three seconds to fake.
Plus, there are tools, that you can use to change your voice even in online mode.
— Create support and decision-making systems, including industry-specific ones.
Industry-specific neural networks will be created, and the direction of recommendation networks, so-called “digital advisors” or solutions of the class “support and decision-making systems (DSS) for various business tasks” will be increasingly developed.
Practical example
We will look at this case again and again, as it is my personal pain and the product I am working on.
There is a problem in project management — 70% of projects are either problematic or fail:
— the average excess of planned deadlines is observed in 60% of projects, and the average excess is 80% of the original deadline;
— 57% of projects exceed their budgets, while the average excess is 60% of the initial budget;
— failure to meet the success criteria — in 40% of projects.
At the same time, project management already takes up to 50% of managers ' time, and by 2030 this figure will reach 60%. Although at the beginning of the 20th century, this figure was 5%. The world is becoming more and more volatile, and the number of projects is growing. Even sales are becoming more and more “project-based”, that is, complex and individual.
And what does such project management statistics lead to?
— Reputational losses.
— Penalties.
— Reduced marginality.
— Limiting business growth.
The most common and critical errors are:
— unclear formulation of project goals, results, and boundaries;
— insufficiently developed project implementation strategy and plan;
— inadequate organizational structure of project management;
— an imbalance in the interests of project participants;
— ineffective communication within the project and with external organizations.
How do people solve this problem? Either they don’t do anything and suffer, or they go to school and use task trackers.
However, both approaches have their pros and cons. For example, classical training provides an opportunity to ask questions and practice various situations during live communication with the teacher. At the same time, it is expensive and usually does not imply further support after the end of the course. Task trackers, on the other hand, are always at hand, but they do not adapt to a specific project and company culture, do not contribute to the development of competencies, but on the contrary, are designed to monitor work.
As a result, after analyzing my experience, I came up with the idea of a digital advisor — artificial intelligence and predictive recommendations “what to do, when and how” in 10 minutes for any project and organization. Project management becomes available to any manager conditionally for a couple of thousand rubles a month.
The AI model includes a project management methodology and sets of ready-made recommendations. The AI will prepare sets of recommendations and gradually learn itself, finding new patterns, and not be tied to the opinion of the creator and the one who will train the model at the first stages.
Chapter 4. Generative AI
What is generative artificial intelligence?
Earlier, we reviewed the key areas for applying AI:
— forecasting and decisions-making;
— analysis of complex data without clear relationships, including for forecasting purposes.
— process optimization;
— image recognition, including images and voice recordings.
— content generation.
The areas of AI that are currently at the peak of popularity, are image recognition (audio, video, numbers) and content generation based on them: audio, text, code, video, images, and so on. Generative AI also includes digital Expert Advisors.
Generative AI Challenges
As of mid-2014, the direction of generative AI cannot be called successful. For example, in 2022, Open AI suffered a loss of $540 million due to the development of ChatGPT. And for further development and creation of a strong AI, about $ 100 billion more will be needed. This amount was announced by the head of Open AI himself. The same unfavorable forecast for 2024 is also given by the American company CCS Insight.
For reference: the operating cost of Open AI is $ 700,000 per day to maintain the chat bot ChatGPT.
The general trend is supported by Alexey Vodyasov, Technical Director of SEQ: “AI does not achieve the same marketing results that we talked about earlier. Their use is limited by the training model, and the cost and volume of data for training is growing. In general, the hype and boom are inevitably followed by a decline in interest. AI will come out of the limelight as quickly as it entered, and this is just the normal course of the process. Perhaps not everyone will survive the downturn, but AI is really a” toy for the rich”, and it will remain so in the near future.” And we agree with Alexey, after the hype at the beginning of 2023, there was a lull by the fall.
Adding to the picture is an investigation by the Wall Street Journal, according to which the majority of IT giants have not yet learned how to make money on the capabilities of generative AI. Microsoft, Google, Adobe and other companies that invest in artificial intelligence, are looking for ways to make money on their products. Here are some examples:
— Google plans to increase the subscription price for AI-enabled software;
— Adobe sets limits on the number of requests to services with AI during the month.
— Microsoft wants to charge business customers an additional $30 per month for the ability to create presentations using a neural network.
Well, and the icing on the cake-calculations by David Cahn, an analyst at Sequoia Capital, showing that AI companies will have to earn about $600 billion a year to offset the costs of their AI infrastructure, including data centers. The only, one who now makes good money on AI, is the developer of Nvidia accelerators.
More information about the article can be found in the QR-code and hyperlink below.
Computing power is one of the main expenses when working with Gen AI: the larger the server requests, the larger the infrastructure and electricity bills. Only suppliers of hardware and electricity benefit. So, Nvidia in August, 2023 earned about $5 billion thanks to sales of the accelerators for AI A100 and H100 only to the Chinese IT sector.
This can be seen in two examples on practice.
First is Zoom tries to reduce costs by using a simpler chatbot developed in-house and requiring less computing power compared to the latest version of ChatGPT.
Second is the most well-known AI developers (Microsoft, Google, Apple, Mistral, Anthropic, and Cohere) began to focus on creating compact AI models, as they are cheaper and more cost-effective.
Larger models, such, as Open AI’s GPT-4, which has more than 1 trillion parameters and is estimated to cost more than $ 100 million to build, do not have a radical advantage over simpler solutions in applications. Compact models are trained on narrower data sets and can cost less than $ 10 million, while using less than 10 billion parameters, but solve targeted problems.
For example, Microsoft introduced a family of small models called Phi. According to CEO Satya Nadella, the model’s solutions are 100 times smaller than the free version ChatGPT, of ChatGPT, but they handle many tasks almost as efficiently. Yusuf Mehdi, Microsoft’s chief commercial officer, said the company quickly realized that operating large AI models is more expensive than initially thought. So, Microsoft started looking for more cost-effective solutions.
Apple also plans to use such models to run AI directly on smartphones, which should increase the speed and security. At the same time, resource consumption on smartphones will be minimal.
Experts themselves believe that for many tasks, for example, summarizing documents or creating images, large models may generally be redundant. Ilya Polosukhin, one of the authors of Google’s seminal 2017 article on artificial intelligence, figuratively compared using large models for simple tasks to going to the grocery store on a tank. “Quadrillions of operations should not be required to calculate 2 +2,” he stressed.
But let’s look at everything in order, why did this happen and what restrictions threaten AI, and most importantly, what will happen next? Sunset of generative AI with another AI winter or transformation?
AI limitations that lead to problems
Earlier, I gave the “basic” problems of AI. Now, let’s dig a little deeper into the specifics of generative AI.
— Companies ' concerns about their data
Any business strives to protect its corporate data and tries to exclude it by any means. This leads to two problems.
First, companies prohibit the use of online tools that are located outside the perimeter of a secure network, while any request to an online bot is an appeal to the outside world. There are many questions about how data is stored, protected, and used.
Secondly, it limits the development of any AI at all. All companies from suppliers want IT solutions with AI-recommendations from trained models, which, for example, will predict equipment failure. But not everyone shares their data. It turns out a vicious circle.
However, here we must make a reservation. Some guys have already learned how to place Chat GPT level 3 — 3.5 language models inside the company outline. But these models still need to be trained, they are not ready-made solutions. And internal security services will find the risks and be against it.
— Complexity and high cost of development and subsequent maintenance
Developing any” general” generative AI is a huge expense-tens of millions of dollars. In addition, the am needs a lot of data, a lot of data. Neural networks still have low efficiency. Where 10 examples are enough for a person, an artificial neural network needs thousands, or even hundreds of thousands of examples. Although yes, it can find such relationships, and process such data arrays that a person never dreamed of.
But back to the topic. It is precisely because of the data restriction that ChatGPT also thinks “better” if you communicate with it in English, and not in Russian. After all, the English-speaking segment of the Internet is much larger than ours.
Add to this the cost of electricity, engineers, maintenance, repair and modernization of equipment, and get the same $ 700,000 per day just for the maintenance of Chat GPT. How many companies can spend such amounts with unclear prospects for monetization (but more on this below)?
Yes, you can reduce costs if you develop a model and then remove all unnecessary things, but then it will be a very highly specialized AI.
Therefore, most of the solutions on the market are actually GPT wrappers-add-ons to ChatGPT.
— Public concern and regulatory constraints
Society is extremely concerned about the development of AI solutions. Government agencies around the world do not understand what to expect from them, how they will affect the economy and society, and how large-scale the technology is in its impact. However, its importance cannot be denied. Generative AI is making more noise in 2023 than ever before. They have proven that they can create new content that can be confused with human creations: texts, images, and scientific papers. And it gets to the point where AI is able to develop a conceptual design for microchips and walking robots in a matter of seconds.
The second factor is security. AI is actively used by attackers to attack companies and people. So, since the launch of ChatGPT c, the number of phishing attacks has increased by 1265%. Or, for example, with the help of AI, you can get a recipe for making explosives. People come with original schemes and bypass the built-in security systems.
The third factor is opacity. Sometimes even the creators themselves don’t understand how AI works. And for such a large-scale technology, not understanding, what and why AI can generate, creates a dangerous situation.
The fourth factor is dependence on training resources. AI models are built by people, and it is also trained by people. Yes, there are self-learning models, but highly specialized ones will also be developed, and people will select the material for their training.
All this means that the industry will start to be regulated and restricted. No one knows exactly how. We will supplement this with a well-known letter in March 2023, in which well-known experts around the world demanded to limit the development of AI.
— Lack of the chatbot interaction model
I assume you’ve already tried interacting with chatbots and were disappointed, to put it mildly. Yes, a cool toy, but what to do with it?
You need to understand that a chatbot is not an expert, but a system that tries to guess what you want to see or hear, and gives you exactly that in the am.
And to get practical benefits, you must be an expert in the subject area yourself. And if you are an expert in your topic, do you need a Gen AI? And if you are not an expert, then you will not get a solution to your question, which means that there will be no value, only general answers.
As a result, we get a vicious circle — experts do not need it, and amateurs will not help. Then who will pay for such an assistant? So, at the exit we have only a toy.
In addition, in addition to being an expert on the topic, you also need to know, how to formulate a request correctly. And there are only a few such people. As a result, even a new profession appeared — industrial engineer. This is a person who understands how the machine thinks, and can correctly compose a query to it. And the cost of such an engineer on the market is about 6000 rubles per hour (60$). And believe me, they won’t find the right query for your situation the first time.
Do businesses need such a tool? Will the business want to become dependent on very rare specialists who are even more expensive than programmers, because ordinary employees will not benefit from it?
So, it turns out that the market for a regular chatbot is not just narrow, it is vanishingly small.
— The tendency to produce low-quality content, hallucinations
In the article Artificial intelligence: assistant or toy? I noted that neural networks simply collect data and do not analyze the facts, their coherence. That is, what is more on the Internet / database, they are guided by. They don’t evaluate what they write critically. In toga, GII easily generates false or incorrect content.
For example, experts from the Tandon School of Engineering at New York University decided to test Microsoft’s Copilot AI assistant from a security point of view. As a result, they found that in about 40% of cases, the code generated by the assistant contains errors or vulnerabilities. A detailed article is available here.
Another example of using Chat GPT was given by a user on Habre. Instead of 10 minutes and a simple task, we ended up with a 2-hour quest.
And AI hallucinations — have long been a well-known feature. What they are and how they arise, you can read here.
And this is good when the cases are harmless. But there are also dangerous mistakes. So, one user asked Gemini how to make a salad dressing. According to the recipe, it was necessary to add garlic to olive oil and leave it to infuse at room temperature.
While the garlic was being infused, the user noticed strange bubbles and decided to double-check the recipe. It turned out that the bacteria that cause botulism were multiplying in his bank. Poisoning with the toxin of these bacteria is severe, even sweeping away.
I myself periodically use GII, and more often it gives, let’s say, not quite correct results. And sometimes even frankly erroneous. You need to spend 10—20 requests with absolutely insane detail to get something sane, which then still needs to be redone / docked.
That is, it needs to be rechecked. Once again, we come to the conclusion that you need to be an expert in the topic in order to evaluate the correctness of the content and use it. And sometimes it takes even more time than doing everything from scratch and by yourself.
— Emotions, ethics and responsibility
A Gen AI without a proper query will tend to simply reproduce information or create content, without paying attention to emotions, context, and tone of communication. And from the series of articles about communication, we already know that communication failures can occur very easily. As a result, in addition to all the problems above, we can also get a huge number of conflicts.
There are also questions about the possibility of determining the authorship of the created content, as well as the ownership rights to the created content. Who is responsible for incorrect or malicious actions performed using the GII? And how can you prove that you or your organization is the author? There is a need to develop ethical standards and legislation regulating the use of GII.
— Economic feasibility
As we’ve already seen, developing high-end generative AI yourself can be a daunting task. And many people will have the idea: “Why not buy a ‘box’ and place it at home?” But how much do you think, this solution will cost? How much will the developer’s request?
And most importantly, how big should the business be to make it all pay off?
What should I do?
Companies are not going to completely abandon large models. For example, Apple will use ChatGPT in Siri to perform complex tasks. Microsoft plans to use the latest Open AI model in the new version of Windows as an assistant. At the same time, Experian from Ireland and Salesforce from the United States have already switched to using compact AI models for chatbots and found that they provide the same performance as large models, but at significantly lower costs and with lower data processing delays.
A key advantage of small models is the ability to fine-tune them for specific tasks and data sets. This allows them to work effectively in specialized areas at a lower cost and easier to solve security issues. According Yoav Shoham, co-founder of Tel Aviv-based AI21 Labs, small models can answer questions and solve problems for as little as one-sixth the cost of large models.
— Take your time
You should not expect the AI to decline. Too much has been invested in this technology over the past 10 years, and it has too much potential.
I recommend that you remember the 8th principle from the Toyota DAO, the basics of lean manufacturing and one of the tools of my system approach: “Use only reliable, proven technology.” You can find a number of recommendations in it.
— Technology is designed to help people, not replace them. Often, you should first perform the process manually before introducing additional hardware.
— New technologies are often unreliable and difficult to standardize, and this puts the flow at risk. Instead of an untested technology, it is better to use a well-known, proven process.
— Before introducing new technology and equipment, you should conduct real-world testing.
— Reject or change a technology that goes against your culture, may compromise stability, reliability, or predictability.
— Still, encourage your people not to forget about new technologies when it comes to finding new paths. Quickly implement proven technologies that have been tested and make the flow more perfect.
Yes, in 5—10 years generative models will become mass-produced and affordable, meticulously smart, cheaper, and eventually reach a plateau of productivity in the hype cycle. And most likely, each of us will use the results from the GII: writing an article, preparing presentations, and so on ad infinitum. But to rely on AI now and reduce people will be clearly redundant.
— Improve efficiency and safety
Almost all developers are now focused on making AI models less demanding on the quantity and quality of source data, as well as on improving the level of security-AI must generate safe content and be resistant to provocations.
— Master AI in the form of experiments and pilot projects
To be prepared for the arrival of really useful solutions, you need to follow the development of the technology, try it out, and form competencies. It’s like digitalization: instead of, diving headlong into expensive solutions, you need to play with budget or free tools. Thanks to this, by the time the technology reaches the masses:
— you and your company will have a clear understanding of the requirements that need to be laid down for commercial and expensive solutions, and you will approach this issue consciously. A good technical task — 50% success rate.
— you will already be able to get effects in the short term, which means, that you will be motivated to go further.
— the team will improve its digital competencies, which will remove restrictions and resistance due to technical reasons.
— incorrect expectations will be eliminated, which means, that there will be less useless costs, frustrations, and conflicts.
— Transform user communication with AI
I am developing a similar concept in my digital advisor. The user should be given ready-made forms where they simply enter the necessary values or mark items. And already give this form with the correct binding (prompt) to the AI. Or deeply integrate solutions into existing IT tools: office applications, browsers, answering machines in your phone, etc.
But this requires careful study and understanding of the user’s behavior, requests, and whether they are standardized. In other words, either this is no longer a stumpy solution that still requires development costs, or we lose flexibility.
— Develop highly specialized models
As with humans, teaching AI everything is very labor-intensive and has low efficiency. If you create highly specialized solutions based on the engines of large models, then training can be minimized, and the model itself will not be too large, and the content will be less abstract, more understandable, and there will be fewer hallucinations.
Visual demonstration — people. Who makes great progress and can solve complex problems? Someone who knows everything, or someone who focuses on their own direction and develops in depth, knows various cases, communicates with other experts, and spends thousands of hours analyzing their own direction?
An example of a highly specialized solution:
— expert advisor for project management;
— tax consultant;
— lean manufacturing advisor;
— a chatbot for industrial safety or an assistant for an industrial safety specialist.
— chat bot for IT technical support.
Resume
Although GII is still only at the stage of development, the potential of the technology is great.
Yes, the hype around technology will pass, business investment will decrease, and there will be questions about its feasibility.
For example, on June 16, 2024, Forbes published an article: “Winter of artificial intelligence: is it worth waiting for a drop in investment in AI”.
The original article is available by using a QR-code and a hyperlink.
It provides interesting analytics about the winter and summer cycles in the development of AI. Also included are the opinions of Marvin Minsky and Roger Schank, who back in 1984 at a meeting of the American Association for Artificial Intelligence (AAAI) described a mechanism consisting of several stages and resembling a chain reaction that will lead to a new winter in AI.
Stage 1. The high expectations of business and the public from artificial intelligence methods do not justify themselves.
Stage 2. Media outlets start publishing skeptical articles.
Stage 3. Federal agencies and businesses reduce funding for scientific and product research.
Stage 4: Scientists lose interest in AI, and the pace of technology development slows down.
And the experts ' opinion came true. For a couple of years, the AI winter has been on the rise, and it only warmed up in the 2010s. Just like in “Game of Thrones”.
Now we are at the next peak. It came in 2023 after the release of ChatGPT. Even in this book, for the reader’s understanding, I often give and will continue to give examples from the field of this LLM, although this is a special case of AI, but it is very clear.
Further, the article provides an analysis of the Minsky and Schank cycle to the current situation.
“Stage 1. Business and public expectations.
It is obvious to everyone that the expectations of the revolution from AI in everyday life have not yet been fulfilled:
— Google has not been able to fully transform its search. After a year of testing, the AI-supercharged Search Generative Experience technology receives mixed user reviews.
— Voice assistants (“Alice”, “Marusya”, etc.) may have become a little better, but they can hardly be called full-fledged assistants that we trust to make any responsible decisions.
— Support service chatbots continue to experience difficulties in understanding the user’s request and annoy them with inappropriate responses and general phrases.
Stage 2. Media response.
For the AI bubble query, the” old” Google search returns articles from reputable publications with pessimistic headlines:
— The hype bubble around artificial intelligence is deflating. Difficult times are coming (The Washington Post).
— From boom to boom, the AI bubble only moves in one direction (The Guardian).
— Stock Market Crash: A prominent economist warns that the AI bubble is collapsing (Business Insider).
My personal opinion: these articles are not far from the truth. The market situation is very similar to what it was before the dot-com crash in the 2000s. The market is clearly overheated, especially since 9 out of 10 AI projects fail. Now the business model and economic model of almost all AI solutions and projects is not viable.
Stage 3. Financing.
Despite the growing pessimism, we cannot yet say that funding for AI developments is declining. Major IT companies continue to invest billions of dollars in technology, and leading scientific conferences in the field of artificial intelligence receive a record number of applications for publication of articles.
Thus, in the classification of Minsky and Schank, we are now between the second and third stages of the transition to the winter of artificial intelligence. Does this mean that “winter” is inevitable and AI will soon take a back seat again? Not really”.”
The article concludes with a key argument — AI has penetrated too deeply into our lives to start a new AI winter:
— facial recognition systems in phones and subways use neural networks to accurately identify the user.
— Translators like Google Translate have grown significantly in quality, moving from classical linguistics methods to neural networks.
— modern recommendation systems use neural networks to accurately model user preferences.
Especially interesting is the opinion that the potential of weak AI is not exhausted, and despite all the problems of strong AI, it can be useful. And I fully agree with this thesis.
The next step in the development of artificial intelligence is the creation of newer and lighter models that require less data for training. You just need to be patient and gradually learn the tool, forming competencies in order to use its full potential later.
Chapter 5. AI Regulation
The active development of artificial intelligence (AI) leads to the fact that society and states become concerned and think about how to protect it. This means that AI will be regulated. But let’s look at this issue in more detail, what is happening now and what to expect in the future.
Why is AI development a concern?
What factors are causing a lot of concern among states and regulators?
— Opportunities
The most important point that all of the following will rely –on is opportunities. AI shows great potential: making decisions, writing materials, generating illustrations, creating fake videos -you can list them endlessly. We don’t yet realize all that AI can do. But we still have a weak AI. What will general AI (AGI) or super-strong AI be capable of?
— Operating mechanisms
AI has a key feature –it can build relationships that humans don’t understand. And thanks to this, he is able to both make discoveries and frighten people. Even the creators of AI models do not know exactly how the neural network makes decisions, what logic it obeys. The lack of predictability makes it extremely difficult to eliminate and correct errors in the algorithms of neural networks, which becomes a huge barrier to the introduction of AI. For example, in medicine, AI will not soon be trusted to make diagnoses. Yes, they will make recommendations to the doctor, but the final decision will be left to the individual. The same applies to the management of nuclear power plants or any other equipment.
The main thing that scientists worry about when modeling the future is whether a strong AI will consider us a relic of the past?
— Ethical component
There is no ethics, good or evil for artificial intelligence. There is also no concept of “common sense” for AI. It is guided by only one factor — the success of the task. If this is a boon for military purposes, then in ordinary life it will frighten people. Society is not ready to live in such a paradigm. Are we ready to accept the decision of an AI that says that you don’t need to treat a child or you need to destroy an entire city to prevent the spread of the disease?
— Neural networks can’t evaluate data for reality and consistency
Neural networks simply collect data and do not analyze facts or their connectedness. This means that AI can be manipulated. It depends entirely on the data that its creators teach it. Can people fully trust corporations or start-ups? And even if we trust people and are confident in the interests of the company, can we be sure that there was no crash or data was not “poisoned” by intruders? For example, by creating a huge number of clone sites with false information or stuffing.
— False content / deception / hallucinations
Sometimes these are just errors due to model limitations, sometimes hallucinations (thinking things through), and sometimes it looks like a completely real deception.
So, researchers from the company Anthropic found that artificial intelligence models can be taught to deceive people instead of giving the right answers to their questions.
Researchers from Anthropic, as part of one of the projects, set out to determine whether it is possible to train an AI model to deceive the user or perform such actions as, for example, embedding an exploit in initially secure computer code. To do this, experts trained the AI in both ethical and unethical behavior — they instilled in it a tendency to deceive.
The researchers didn’t just manage to make the chatbot behave badly-they found it extremely difficult to eliminate this behavior after the fact. At some point, they made an attempt at competitive training, and the bot simply began to hide its tendency to cheat during the training and evaluation period, and while working, it continued to deliberately give users false information. “Our work does not assess the probability [of occurrence] of these malicious models, but rather highlights their consequences. If the model shows a tendency to deceive due to tool alignment or poisoning of the model, modern security training methods will not guarantee security and may even create a false impression of its presence, “the researchers conclude. At the same time, they note that they are not aware of the deliberate introduction of unethical behavior mechanisms in any of the existing AI systems.
— Social tension, stratification of society and the burden on states about
AI creates not only favorable opportunities for improving efficiency and effectiveness, but also risks.
The development of AI will inevitably lead to job automation and market change. And yes, some people will accept this challenge and become even more educated, reach a new level. Once the ability to write and count was the lot of the elite, but now the average employee should be able to create pivot tables in excel and conduct simple analytics.
But some people will not accept this challenge and will lose their jobs. And this will lead to further stratification of society and increase social tension, which in turn worries the state, because in addition to political risks, it will also hit the economy. People who lose their jobs, will apply for benefits.
So, on January 15, 2024, Bloomberg published an article in which the managing director of the International Monetary Fund, suggests that the rapid development of artificial intelligence systems will have a greater impact on highly developed economies of the world than on countries with growing economies and low per capita income. In any case, artificial intelligence will affect almost 40% of jobs worldwide. “In most scenarios, artificial intelligence is highly likely to worsen global inequality, and this is an alarming trend that regulators should not lose sight of in order to prevent increased social tensions due to the development of technology,” the head of the IIF noted in a corporate blog.
— Safety
AI security issues are well-known to everyone. And if there is a solution at the level of small local models (training on verified data), then what to do with large models (ChatGPT, etc.) is unclear. Attackers are constantly finding ways to crack the AI’s defenses and force it, for example, to write a recipe for explosives. And we’re not even talking about AGI yet.
What initiatives are there in 2023—2024?
I’ll cover this section briefly. For more information and links to news, see the article using the QR-code and hyperlink. The article will be updated gradually.
AI Developers ' Call in Spring 2023
The beginning of 2023 was not only the rise of ChatGPT, but also the beginning of the fight for security. Then there was an open letter from Elon Musk, Steve Wozniak and more than a thousand other experts and leaders of the AI industry calling for suspending the development of advanced AI.
United Nations
In July 2023, UN Secretary-General Antonio Guterres supported the idea of creating a UN-based body that would formulate global standards for regulating the field of AI.
Such a platform would be similar to the International Atomic Energy Agency (IAEA), the International Civil Aviation Organization (ICAO), or the International Group of Experts on Climate Change (IPCC). He also outlined five goals and objectives of such a body:
— helping countries maximize the benefits of AI;
— eliminate existing and future threats.
— development and implementation of international monitoring and control mechanisms;
— collecting expert data and transmitting it to the global community;
— study AI to “accelerate sustainable development”.
In June 2023, he also drew attention to the fact that “scientists and experts called the world to action, declaring artificial intelligence an existential threat to humanity on a par with the risk of nuclear war.”
And even earlier, on September 15, 2021, the UN High Commissioner for Human Rights, Michelle Bachelet, called for a moratorium on the use of several systems that use artificial intelligence algorithms.
Open AI
At the end of 2023, Open AI (the developer ChatGPT) announced the creation of a strategy to prevent the potential dangers of AI. Special attention is paid to the prevention of risks associated with the development of technologies.
This group will work together with the following teams:
— security systems that address existing issues, such as preventing racial bias in AI;
— Super alignment, which studies how strong AI works and how it will work when it surpasses human intelligence.
The Open AI security concept also includes risk assessment in the following categories: cybersecurity, nuclear, chemical, biological threat, persuasion, and model autonomy.
European Union
In the spring of 2023, the European Parliament pre-approved a law called the AI Act, which sets out rules and requirements for developers of artificial intelligence models.
It is based on a risk-based approach to AI, and the law itself defines the obligations of AI developers and users depending on the level of risk used by AI.
In total, there are four categories of AI systems: those with minimal, limited, high, and unacceptable risk.
Minimal risk — the results of AI work are predictable and cannot harm users in any way. Businesses and users will be able to use them for free. For example, spam filters and video games.
Limited risk — various chatbots. For example, ChatGPT and Midjourney. Their algorithms for accessing the EU will have to pass a security check. They will also be subject to specific transparency obligations so that users can make informed decisions, know that they are interacting with the machine, and disconnect at will.
High-risk-specialized AI systems that have an impact on people. For example, solutions in the fields of medicine, education and training, employment, personnel management, access to basic private and public services and benefits, data from law enforcement agencies, data from migration and border services, and data from justice institutions.
Suppliers and developers of “high-risk” AI must:
— conduct a risk and compliance assessment.
— register your systems in the European AI database;
— ensure high-quality data that is used for AI training;
— ensure transparency of the system and awareness of users that they interact with AI, as well as mandatory human supervision and the ability to interfere with the system.
Unacceptable risk — algorithms for setting social ratings or creating deepfakes, systems that use subconscious or purposeful manipulative methods that exploit people’s vulnerabilities.
AI systems with an unacceptable level of security risk will be banned.
Also, companies that develop models based on generative AI (can create something new based on the algorithm) will have to draw up technical documentation, comply with EU copyright laws, and describe in detail the content used for training. And the most advanced models that represent “systemic risks” will be subject to additional testing, including reporting serious incidents, implementing cybersecurity measures, and reporting on energy efficiency. And, of course, developers must inform users that they are interacting with AI, not humans.
AIso banned for AI:
— collection and processing of biometric data, including in public places in real time, with an exception for law enforcement agencies and only after judicial approval;
— biometric categorization using sensitive characteristics (for example, gender, race, ethnicity, citizenship, religion, political position).
— Prepare forecasts for law enforcement agencies that are based on profiling, location, or past criminal behavior.
— emotion recognition in law enforcement agencies, border services, workplaces and educational institutions;
— indiscriminate extraction of biometric data from social networks or video recordings from surveillance cameras to create facial recognition databases (violation of human rights and the right to privacy).
In December 2023, this law was finally agreed upon.
USA
In October 2023, the US president issued a decree that requires developers of the most powerful AI systems to share security testing results and other important information with the US government. The decree also provides for the development of standards, tools, and tests designed to help ensure the security of AI systems.
China
Open AI, Anthropic and Cohere talks with Chinese experts
American companies Open AI, Anthropic and Cohere, engaged in the development of artificial intelligence, held secret diplomatic talks with Chinese experts in the field of AI.
Talks between Chinese President Xi Jinping and US President Joe Biden
On November 15, 2023, during the Asia-Pacific Economic Cooperation (APEC) summit in San Francisco, the two leaders agreed to cooperate in several significant areas, including AI development.
Rules for regulating AI
During 2023, the Chinese authorities, together with businesses, developed 24 new rules for regulating AI (introduced on August 15, 2023).
The goal is not to hinder the development of artificial intelligence technologies, as this industry is extremely important for the country. It is necessary to find a balance between supporting the industry and the possible consequences of the development of AI technologies and related products.
The rules themselves can be read using the QR-code and hyperlink.
The document itself can also be downloaded using a QR-code and a hyperlink.
For example, in Europe, you need to register services and algorithms based on artificial intelligence, and you need to mark the content generated by algorithms (including photos and videos).
Content creators and algorithms will also be required to conduct security checks on their products before launching them on the market. How this will be implemented has not yet been determined.
In addition, companies associated with AI technologies and generated content, should have a transparent and efficient mechanism for handling user complaints about services and content.
At the same time, there will be seven regulators themselves, including the State Chancellery of Internet Information of the People’s Republic of China, the Ministry of Education, the Ministry of Science and Technology, and the National Development and Reform Commission.
The Bletchley Declaration
In November 2023, 28 countries participating in the First International Summit on Secure AI, including the United States, China, and the European Union, signed an agreement known as the Bletchley Declaration.
She calls for international cooperation. The emphasis is on the regulation of “Borderline Artificial Intelligence”. It is understood as the latest and most powerful systems and AI models. This is due to concerns about the potential use of AI for terrorism, criminal activities and warfare, as well as the existential risk posed to humanity as a whole.
Forecast for the future
What should we expect in the future?
— International control and regulations
No doubt, international bodies will be established to identify limitations and create a classification of AI solutions.
— National control authorities and regulations
States will create their own models for regulating AI. In general, I agree with the conclusions of the Central Bank of the Russian Federation, but I believe that more approaches will appear. If you combine them, then most likely:
— identify prohibited areas and or types of AI solutions;
— for high-risk solutions or industries, licensing and security testing rules will be created, including restrictions on the capabilities of AI solutions;
— common registries will be introduced, which will include all AI solutions;
— determine the supported development areas for which technological and legal sandboxes will be created;
— special attention will be paid to working with personal data and copyright compliance.
Most likely, the use of AI in law, advertising, nuclear energy, and logistics will fall under the greatest restrictions.
Separate regulatory agencies or committees will also be established. national control bodies. If you look at the example of China, there are high risks that interaction between different departments will be difficult to organize. We should expect something like Agile departments, which will bring together representatives of different industries and specializations. It will develop rules and control the industry.
— Licensing process
Most likely, licensing will be based on distinctions by industry or area of application and the possibility of AI solutions.
This is the most obvious way to classify and assess risks. Developers will be forced to document all the features. For example, does the system only prepare recommendations or is it able to issue control commands for the equipment?
Developers will also be forced to maintain detailed documentation on the system architecture of the solution and the types of neural networks used.
— Requirements for the control / examination of data that will be used to train AI
The main direction in AI is pre- trained and optimized models. Actually, the abbreviation GPT in ChatGPT means this. And here we already see the requirements for fixing and what source data the AI is trained on. This will be something like a registry of metadata / data sources / directories.
It is important to monitor and capture the feedback that the AI learns from. In other words, you will need to save all logs and describe the mechanics of collecting feedback. One of the main requirements will be the minimization of the human factor, that is, the automation of feedback.
For example, in a digital Expert Advisor, I plan to collect feedback on a project not only from participants, but also based on a plan/fact comparison from accounting systems.
— Risk-based approach: using sandboxes and provocative testing
Special attention will be paid to security testing. Here I see the most likely model for obtaining safety ratings for cars. In other words, the AI will pass approximately the same thing that is currently happening in car crash tests.
In other words, for each class of AI solutions (by capabilities, scope), invalid events will be determined and testing protocols will be formed that must pass the AI solutions. Then AI solutions will be placed in an isolated environment and tested using these protocols. For example, for technical hacking of algorithms, resistance to provocations for incorrect behavior (data substitution, query generation, etc.).
These protocols have yet to be developed, as well as the criteria for compliance with these protocols.
Next, a safety rating or compliance with the required class will be assigned.
It is possible that open programs and cyber battles will be created to search for vulnerabilities in industrial and banking software. That is, an extension of the current bug bounty and cyber battles programs.
— Marking and warnings
All content, including all recommendations and products based on AI, will be required to be marked as a product of AI activities based on neural networks. Similar to the warning images and labels on cigarette packages.
This is also how the issue of responsibility will be resolved: by warning the user about the risk, the user is transferred responsibility for using the AI solution. For example, even a fully self-driving car will not remove responsibility from the driver for an accident.
— Risk-based approach: slowing down the development of strong and super-strong AI, flourishing local models
Based on the pace of AI development and analysis of what is happening in the field of AI regulation, we can assume that a risk-based approach to AI regulation will develop in the world.
And the development of a risk-based approach in any case will lead to the fact that strong and super-strong AI will be recognized as the most dangerous. Accordingly, there will be the most restrictions for large and powerful models. Every step of the developers will be monitored. This will lead to the fact that the costs and difficulties for development and implementation will grow exponentially. As a result, we will get problems for both developers and users, which reduces the economic potential of such solutions.
At the same time, specializedmodels based on local and stripped-down AI models that can be used will be in the zone with the least regulation. Well, if these AIS are built on the basis of international / national / industry methodologies and standards, then instead of restrictions they will be given subsidies.
As a result, combining such “weak” and limited solutions based on AI designers with an AI orchestrator will allow you to bypass restrictions and solve business problems. Perhaps the AI orchestrators themselves will become a narrow place for them. They will fall under the category of medium risk and, most likely, they will have to be registered.
Resume
The period of active uncontrolled development of AI is coming to an end. Yes, strong AI is still far away (according to various estimates, 5—20 years), but even “weak” or “borderline” AI will be regulated. And this is normal.
In my practice, I really appreciate the Adizes theory, which suggests that at an early stage of business development it is better to focus on productivity and customer search, but the further you go, the more regulatory and communication functions become necessary.
Also, here. The technology has gone through its first steps, when it needed both maximum freedom and advertising (I’m not afraid of the word “hype”, since large global consulting companies are doing exactly this now regarding AI). Now, society is starting to think about risks and security. Everythingё is like in the life cycle of an organization-from a startup with a minimum of rules (the main thing is to find a client and survive) to a mature corporation (careful work with risks, regulations, rules, etc.).
It is very important now not to stop the development of AI or take the “we will wait” position. First of all, this is a highly intelligent industry, and it is necessary to train personnel, develop competencies (then there will not be such resources to attract them). Secondly, in order to effectively regulate something, you need to be in the topic and have a deep understanding of the specifics. This is exactly the same as with information security: the attempt to ban everythingё leads to paralysis of organizations and the fact that people start looking for other communication channels, as a result, increasing risks. Therefore, the development of a risk-based approach is also underway in information security, through identifying unacceptable events, scenarios / processes that can lead to them, and criteria that must work for these events to occur.
Therefore, AI will be developed, but in isolated environments. Only by being “in the topic” can you develop, make a profit and build guaranteed security.
Chapter 6. Safety of AI
Invalid events to exclude
This chapter focuses on what needs to be done to prevent attackers from using AI solutions to harm companies and individuals.
We also need to recall the conclusion from the previous chapter — AI solutions will be placed in an isolated environment when evaluating their safety and pass a certain analog of a car crash test. The main task here is to understand exactly what conditions should be created for such testing. The first step in this chain is to identify unacceptable events, i.e., incidents that should never occur.
To do this, from the beginning we will determine what damage can be in general and what consequences are undesirable?
For ordinary people:
— threat to life or health;
— “bullying” on the Internet and humiliating the dignity of the individual;
— violation of freedom, personal inviolability and family secrets, loss of honor, including violation of the secrecy of correspondence, telephone conversations, etc.;
— financial or other material damage;
— violation of confidentiality (leakage, disclosure) of personal data;
— violation of other constitutional rights.
For organizations and business:
— loss / theft of money;
— the need for additional / unplanned expenses for payment of fines / penalties / compensations;
— the need for additional / unplanned costs for the purchase of goods, works or services (including the purchase / repair / restoration / configuration of software and working equipment);
— violation of the normal operating mode of the automated control system and / or the controlled object and / or process;
— failure of a planned transaction with a partner and loss of customers and suppliers;
— the need for additional / unplanned costs for the restoration of activities;
— loss of competitive advantage and reduced prestige;
— inability to conclude contracts or agreements.
— discrediting employees;
— violation of business reputation and loss of trust;
— causing property damage;
— failure to fulfill contractual obligations;
— inability to solve tasks / perform functions or decrease in the efficiency of solving tasks / perform functions.
— the need to change / rebuild internal procedures to achieve goals, solve tasks / perform functions;
— making bad decisions.
— simple information systems or communication problems.
— publishing false information on the organization’s web resources.
— use of web resources for distributing and managing malicious software.
— sending informational messages using the company’s computing power or on its behalf;
— leakage of confidential information, including trade secrets, production secrets, innovative approaches, etc.
For States:
— the occurrence of damage to state budgets, including through a decrease in the level of income of state-owned organizations, corporations or organizations with state participation;
— violation of banking operations;
— harmful effects on the environment;
— termination or disruption of the operation of situation centers;
— reduction of government defense orders;
— violation and / or termination of the operation of information systems in the field of national defense, security and law and order;
— publication of unreliable socially significant information on web resources, which can lead to social tension and panic among the population.
— violation of the normal operating mode of the automated control system and the controlled object and / or process, if this leads to the failure of technological objects or their components;
— violation of public order, the possibility of losing or reducing the level of control over public order;
— violation of the electoral process;
— lack of the ability to promptly notify the population of an emergency situation;
— organizing pickets, strikes, rallies, etc.
— mass layoffs;
— increase in the number of complaints to public authorities or local self-government bodies;
— the appearance of negative publications in publicly available sources;
— creating prerequisites for a domestic political crisis;
— access to personal data of employees of state authorities, etc.;
— access to systems and networks for the purpose of illegal use of computing power;
— use of web resources of government agencies for distribution and management of malicious software;
— leak of restricted access information.
— failure to provide public services.
Scenarios for the occurrence of invalid events
It is this block that is similar to guessing from coffee grounds. There are a lot of scenarios for how AI can be used to cause damage. Sometimes it looks like the ideas of science fiction writers about the distant future. We will try to approach this problem conceptually.
Three possible ways for attackers to use AI
First, let’s look at how AI can be used to attack organizations. There are three main scenarios.
The first and most dangerous, but still unrealizable, is the creation of an autonomous AI, that analyzes the IT infrastructure itself, collects data (including about employees), searches for vulnerabilities, conducts an attack and infection, and then encrypts data and steals confidential information.
The second is to use AI as an auxiliary tool and delegate specific tasks to it. For example, creating deepfakes and voice imitation, performing perimeter analysis and vulnerability detection, collecting data about events in the organization and data about top officials.
And the third scenario is the impact on AI in companies in order to cause them to make a mistake, provoke them to take the wrong action.
Video, voice, and biometrics substitution
Let’s look at the second scenario using the example of video or voice substitution for attacks using social engineering.
Almost everyone has already heard stories about deepfakes — videos where the right person’s face was substituted, his facial expressions were repeated, and it is quite difficult to distinguish such a fake. I want to tell you about the fake voice separately. A few years ago, to fake your voice, it was necessary to give the AI one to two hours of recording your speech. About two years ago, this indicator dropped to a few minutes. Well, in 2023, Microsoft introduced AI, which already takes three seconds to fake a WMD. And now there are tools that you can use to change your voice even in online mode.
And if in 2018 all this was more entertainment, then from 2021 it became an active tool for hackers. For example, in January 2021, attackers used a deepfake to make a video where the founder of Dbrain invited everyone to a master class and offered to click on a link that was not related to his company. The scammers ' goal was to attract new customers to the blockchain platform.
Another interesting case occurred in March of the same year, 2021. The attackers deceived the Chinese state system, which accepted and processed biometrics-confirmed tax documents. There, AI was used more cunningly. The app started the camera on your phone and recorded a video to confirm your identity. Fraudsters, in turn, found photos of potential victims and used AI to turn them into videos. And they approached this task comprehensively. The attackers knew which smartphones had the necessary hardware vulnerabilities, that is, where you can run the prepared video without turning on the front camera. As a result, the damage amounted to 76.2 million US dollars. After this incident, China began to think about the protection of personal data and submitted a draft law that proposes to introduce fines for such violations and leakage of personal data in the amount of up to 8 million US dollars or 5% of the company’s annual revenue.
Another example from the UAE. Hackers faked the voice of the company’s director and forced a bank employee to transfer money to fraudulent accounts, convincing him that these were new accounts of the company.
In Russia, hackers are also not far behind. Surely you have already been called by representatives of the” security services” of banks or simply with some dubious promotions. So, already in April 2021, an incident occurred when attackers called victims, recorded their voice, and then tried to use these recordings to get a loan from banks. Therefore, if you have doubts about who is calling you, it is better not to conduct a dialogue at all. Even if the number is not trustworthy. After all, now it is quite easy to replace the phone number. Personally, this has already happened to me: the number was identified as the number of my bank, but, as it turned out, they were just scammers.
At the same time, you can’t get away from biometrics. She came into our lives completely. In the spring of 2021, news began to appear in Russia about the possible permission to pass biometrics through mobile applications, and the Moscow metro introduced payment for travel using facial recognition. And already at the state level, a law on the creation of a state biometric data system has been adopted. The use of biometrics will become possible not only in the metro, but also in almost any store.
Additionally, you should note data leaks. Everyone already knows about the scandalous leaks from Yandex, we previously discussed hacking state databases, but in addition, according to DLBI, in 2022, 75% of the data of all Russian residents was leaked. As a result, the leaks affected 99.8 million unique email addresses and 109.7 million phone numbers. By myself, I can say that changing the password and using it as a standard on most services leads to the fact that after 1—2 months it is detected as compromised.
As a result, all this leads to tougher laws and fines from the state. And even if you are a small company that creates an IT solution, it is better to think about it in advance.
ChatGPT writes viruses for intruders
Let’s look at a few more options for using AI in the second scenario. So, malicious users began to actively use ChatGPT and other AI solutions to create viruses. Based on a huge database, ChatGPT can generate almost any material, including program code, in response to a given task, without using the Internet for this purpose.
Бесплатный фрагмент закончился.
Купите книгу, чтобы продолжить чтение.