Artificial Intelligence trends for 2022 - Sia Partners
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
0 0 Summary • • 4 Words from CEO Matthieu Courtecuisse 6 1. The democratization of AI 6 An accelerated development through several phases 6 Challenging journey of scaling up POCs 6 Data, at the essence of corporate strategies • 6 Data-driven organization : a challenge of acculturation 8 2. T he diversification of sectoral applications 8 Energy & Utilities: a matter of data capture 8 Insurance, a sector with a head start in AI • 9 Health: the eldorado of AI innovation 10 3. AI and R&D intertwined 10 A theoretical corpus that keeps on developing • 10 AI challenges conventional research 11 4. The academic rise of AI. 11 The effervescence of online education 11 Communities of data scientists as catalysts of this democratization 11 The massive integration of AI in academic curriculum • 11 Companies, key players in training and academic projects 13 5. An ever evolving business-skill reference system 13 The end of “Data Scientists” ? • 13 Adaptation of recruitment processes 14 6. Continuous innovation of technologies 14 The new generation of algorithms: continuity rather than disruption 15 The new paradigms of AI: Edge computing and hybrid models 16 Low-code: evolution or revolution ? • 16 Beyond algorithmic challenges 17 7. Public policies and competitiveness around AI 17 The race for AI: who are the champions today? 17 The proliferation of national AI programs • 17 Zoom in on the french strategy 19 8. Ethical and societal issues : concerns, controversies and opportunities 19 AI, an asset for ecological issues 19 Ethics and AI, an urgent societal issue • 20 AI in the collective imagination, the new normal ? 21 9. AI regulation is maturing but still in its early stages. 21 Towards a «Europeanization» of AI regulation 22 Growing legislative pragmatism with similarities to cybersecurity • 22 A disadvantage to international competitiveness • 23 10. The closing word by David Martineau. 24 Glossary
0 0 Words from CEO Matthieu Courtecuisse. COVID has pushed the progression of AI across all sectors. The health crisis and the automation trend significantly challenged business operating models. We are also witnessing an acceleration of investments in these areas from our clients. From individuals to states and throughout companies, artificial intelligence is now the primary fuel for technological disruptions at all scales. Far from the fantasies accompanying its first steps, AI contributes to progress throughout all sectors. Without AI and digitalization, lock-down would have been even more damaging to our lives and economies. For AI to continue its fast deve- lopment, we must encourage education in science and mathematics, a regulatory framework that promotes competition and innovation, and finally aim for carbon neutrality in digital technology within 20 years. This political, social and economic commitment will be decisive for the competitiveness of companies and states. This is the sine qua non condition to catalyze the development of AI and the proliferation of its use cases. For the past few years, the trend in all sectors has been to integrate toolkits powe- red by AI, RPA, or blockchain to develop new use cases: forecasting needs, optimi- zing to increase productivity, improving data quality, helping with decision making, improving customer knowledge, etc. For us, this takes the form of Augmented Consulting: business experts who intervene using technical tools to accelerate their intervention and bring more value to their customers. The main difficulty for our clients today in seizing the opportunities offered by AI is to identify the right use cases and industrialize them. AI projects are often confronted with the reality of business issues such as large volumes of data, IT legacy, existing business processes, etc. If we exclude the tech giants, which are natively involved in these subjects and which are designed to integrate AI into all of their activities, the big challenge for companies in 2022-2023 will be the ability to industrialize the production of AI use cases at a large scale, by providing a favorable human and technological ecosystem. This publication by Sia Partners aims to provide an overview of the major dyna- mics around AI and what the year 2022-2023 has in store for us in this area. This exercise was based on our knowledge and expertise in Data & IA, on the business expertise of our sectoral directors. They support our clients daily in their Data & IA transformation and in numerous consultations with experts in Business & Tech.
0 0 1. The democratization of AI. In 2021, the technology sector had the largest number of IPOs (initial public offe- Data, at the essence of rings) with 611 - 26% of IPOs - for a total amount of $147.5 billion. GAFAMs, which are corporate strategies continuing to conquer the world of the future by increasing their hold on data, have Last year, 69% of French data-driven seen their stock market valuation double between January 2019 and July 2020. In companies maintained or even in- companies, data teams, data factories and CDOs are on the rise. creased their investment in data. This position was reinforced by the pande- mic, with more than 80% of business An accelerated also adopt and design data driven use leaders claiming to have gained deci- development through cases. sive advantages thanks to data during several phases this period. Furthermore, the startup The multiplicity of data collection chan- Challenging journey of community is increasingly focusing on nels and the technical breakthroughs in scaling up POCs data. In 2021, BPI France listed 502 French startups dedicated to AI in all the ability to manipulate large volumes Moving from the prototyping phase to areas (industry, transportation, human of data have been the main drivers of the industrialization phase, remains the resources, etc.). the Big Data boom, leading to a 20- reason for many discarded prototypes. fold increase in the volume of digital In addition to start-ups, most compa- data created worldwide over the past The first prerequisite for industrializa- nies, large groups, SMEs and ETIs, are decade. This significant volume has tion is the availability of an infrastruc- embracing data by creating data teams been combined with lower storage and ture that allows the implementation of dedicated to data recovery and mana- processing costs, making Big Data an centralized data models and the crea- gement. This talent pooling often goes attractive and accessible market. tion of DataLakes. These facilitate the hand in hand with the designation of a The last few years have also been encapsulation and harmonization of Chief Data Officer (CDO). This position, marked by the emergence of more POCs into a single robust platform. The which only existed in 12% of companies powerful algorithms. For example, in 2012, is now present in 65% of them. second prerequisite is the compatibility image processing by neural networks has reached such a maturity that it now of AI platforms with existing IS systems Despite the swelling ranks, the scope challenges the most advanced human to enable the deployment of AI-powe- and the hierarchical structure of CDOs diagnosis of lung cancer. NLP tech- red tools in real-world scenarios. Last still varies from one company to ano- niques, on the other hand, enable all but not least, one major prerequisite is ther. A sign that a winning combination sorts of treatments and analyses on all the translation of algorithmic models has not yet been found ? types of textual data. into business interfaces adapted to the Furthermore, according to the same users’ processes. The industrialization survey only half of CDOs (49.5%) have At the dawn of the big bang of data, the of AI models calls for a revision of the led supervisory authority over data in need for data tools and the numerous algorithm (data volume, processes ca- their organizations, and only a third use cases have given free rein to the pacity etc.) in order to robustly expand rate the CDO role as «successful and imaginations of statisticians to design the scope of the POC to a larger scope established”. forecasting or modeling algorithms of application. In addition, as the com- that, within a limited scope, have de- plexity of the model increases, we are Data-driven organization : monstrated their value within compa- witnessing the emergence of low-code a challenge of acculturation nies. Two crucial challenges remain: to solutions for non-specialists, enabling move forward to the industrialization the manipulation of databases in a mat- While it is still difficult for many to mea- stage and to ensure that, in addition ter of a few clicks. sure the business value of chief data to statisticians, business units can officers, their role as coordinators and (1) https://www.decideo.fr/80-des-entreprises-europeennes-data-driven-affirment-beneficier-d-un-avantage-decisif-alors-meme-qu-elles-continuent-a_a12372.html
0 0 catalysts is essential, particularly in terms of advancing the adoption of new technologies and facilitating change management. Gartner mentions the difficulties of effectively implementing intelligent technologies in everyday life, referring to a plateau of AI produc- tivity (Gartner AI hype cycle). Industria- lization and data acculturation will be two major aspects of the AI maturation process in the upcoming years. image
0 0 2. Data-driven organization: a challenge of acculturation. The application of AI capabilities follows different dynamics depending on the AI has boosted the advent of self-ser- sector. These dynamics can be catalyzed by economic circumstances such as the vice (especially in insurance underwri- sanitary crisis for the health sector, but more fundamentally by the quality of the ting and simple claims) via the automa- technological ecosystem that drives each sector. While the advent of AI has made tion of management actions, in order it possible to boost business models of various sectors such as energy or insurance, to refocus insurers’ efforts on actions it has been transforming other sectors, such as health, in a more disruptive way. with high added value. The mana- These transformations are already giving rise to major use cases in these sectors. gement of complex or bodily claims The sectors presented in this article intend to provide a non-exhaustive illustration becomes more efficient, leading to a of sectoral trends in AI by 2022-2023. better customer experience. Similarly, automating the processing of incoming documents from policyholders or third Energy & Utilities: a matter dynamic profiling of segments, appro- parties brings significant competitive of data capture priate definition of pricing policies or advantage. marketing campaigns improvement. Since 2010, the massive deployment of smart metering devices, on the AI also contributes to a better un- “More than 203 million smart meters electricity, gas and water networks, derstanding of clients and therefore will be installed worldwide in 2024, or has made it possible to address new to a deepening of their segmentation as many points of data collection”. use cases where AI allows large-scale thanks to clustering methods. This ap- Beyond the proven short-term return achievements: detection of leaks in wa- plies to several areas, including the de- on investment of these use cases (in ter, detection of electricity consumer termination of tailor-made pricing poli- particular in fraud recovery for exa- fraud, predictive maintenance or even cies according to the insured’s profile, mple), these algorithms also make it enhanced understanding of consumer the adoption of preventive anti-churn possible to plan for the medium term behavior. measures thanks to the detection of «Smart meters electricity consump- and are key tools used to plan the clients likely to leave the portfolio or tion data is a major source of informa- design of tomorrow’s networks in the finally the detection of insurance fraud, tion for AI algorithms in the Energy & context of the energy and ecological which remains a major challenge for Utilities sector» transition! the sector, thanks to anomaly detection Based on machine learning, managers mechanisms. of distribution grids can accurately pre- Insurance, a sector with a dict flows passing through all the links head start in AI While claims management remains a in their networks and thus anticipate major element of the economic profi- corrective actions to be implemented: The insurance sector, which is in- tability of an insurer, the progressive intelligent network monitoring via trinsically quantitative in nature, has application of AI use cases affects the constraint management, activation of benefited from a momentum already entire value chain. This is especially necessary flexibilities, better integra- underway regarding the integration of true regarding the pillars of customer tion of renewable energies, detection AI use cases into its business model. AI experience, namely servicing and cus- of anomalies and finally, appropriate has indeed greatly contributed to the tomer relations, which are key to im- maintenance of the most sensitive operational excellence of insurance, as proving the Net Promoter Score (NPS) equipment. we observe an acceleration of proces- or voice analytics which allows the On the supplier side, the use of this sing times on many components of its «augmented» manager to better adapt data allows a better knowledge of value chain. to the intention and psychological dis- consumers: portfolio view optimization, position of the caller. (2) https://www.openaccessgovernment.org/how-automated-meter-readers-are-transforming-the-water-utilities-industry/117799/ (3) https://iot-analytics.com/smart-meter-market-2019-global-penetration-reached-14-percent/
0 0 image Health: the eldorado of AI Nevertheless, the democratization of AI Doctors will benefit from AI applica- innovation faces major obstacles. The first being tions trained on a large number of the absence of collaborative databases medical cases and available on a col- Health is a popular sector for AI appli- that would consolidate medical data at laborative exchange platform, similar to cations, and the sanitary crisis has fur- a large scale. This is a key prerequisite the business models of certain major ther strengthened this dynamic. Today, for the training of algorithms but also online marketplaces. one in five AI start-ups choose to invest for the proper processing of patients’ in e-health. medical histories. Whether in the field of radiology, der- The second obstacle or safeguard de- matology or oncology, computer vision pending on one’s point of view, is that techniques provide specialists with in many countries, the management of tools to improve the reliability of their medical data faces issues such as the diagnoses. Some AI algorithms enable administrative complexity, the confi- early prediction of kidney cancer me- dentiality of medical records or the tastases, thus increasing the chances reluctance of patients to share their of patient survival between 20% and medical data. 80%. AI use cases are proliferating in medical biology, ophthalmology or In the near future, the emergence of neurology, while in the field of epide- collaborative health platforms will uni- miology, AI makes it possible to predict versalize medical profiles of consenting the outbreak, amplitude and spread patients, and will make them available patterns of pandemics. to their doctors. 3. L’IA et la R&D 3. AI (4) Net Promoter Score: Indicateur mesurant l’intention de recommandation des clients (5) https://www.alliancy.fr/france-digitale-cartographie-startups-ia-france
1 1 3. AI and R&D intertwined. Data and statistics are the two cornerstones of scientific research. Today AI ca- in particular in the field of drug design pabilities are disrupting research methods, whereas research activities originated for which the understanding and mo- the theoretical foundations of AI since Alan Turing’s work on the idea of machine deling of 3D protein structure is crucial. intelligence in the 1940s. Seventy years later, AI revolutionizes research in many fields by providing researchers with world-class analytical capabilities while redu- cing the cost, time and complexity of research. A theoretical corpus that Credit. The country now has nearly 81 keeps on developing R&D laboratories and research centers and more than 13,000 researchers wor- Nowadays, the immediate availability king on AI topics. of very large volumes of data asso- ciated with new algorithmic techniques and computer power has led to major AI challenges conventional technological breakthroughs: Convolu- research tional Neural Networks, Reinforcement AI allows the consideration of less Learning, Transfer Learning, Genera- conventional perspectives: probabilis- tive Adversarial Networks (GANs), etc. tic programming allows for the gene- These breakthroughs act as catalysts ration of hypothetical data that can be for scientific discoveries. compared to experimental data thus On the other hand, the work on the broadening the spectrum of data analy- explicability and interpretability of al- zed. In the field of molecular modeling, gorithms facilitates and accelerates the AI disrupted conventional methods of popularization of their use to research modeling proteins and molecules. This communities for which traceability of is what enabled AlphaFold to manage results is a major concern. and process significant volumes of data, with over 100 million proteins, This enthusiasm in the scientific com- to understand protein folding mecha- munity is reflected in the exponential nisms on a large scale, which resear- growth in the number of publications chers had previously been unable to related to artificial intelligence on the achieve. The autonomy of learning arXiv : multiplied by 6 between 2015 algorithms is remarkable, enabling the and 2020. Similarly, open archives, shift from deterministic modeling go- conferences, and publications dea- verned by molecular dynamics equa- ling with AI are multiplying, with an tions to ML capable of deducing the be- increase of nearly 14% in the number havior of proteins from past behaviors. of publications between 2019 and Therefore, all experimental and ana- 2020. This craze has been boosted by lytical chains are impacted by AI. The international competitiveness. In 2020, scaling achieved by AlphaFold would China surpassed the United States for have required decades of research the first time in the number of AI-re- to obtain the same results without AI. lated citations. France is also a fertile Beyond the economic advantages, the ground, with AI-research incentive gains in time and precision described mechanisms, such as the Research Tax above have benefits on a medical level, (6) arXiv : archive ouverte de prépublications électroniques d’articles scientifiques (7) https://www.intelligence-artificielle.gouv.fr/fr/strategie-nationale/la-strategie-nationale-pour-l-ia (8) https://www.journaldunet.com/solutions/reseau-social-d-entreprise/1192757-carte-de-france-des-laboratoires-d-intelligence-artificielle/
1 1 4. The academic rise of AI. With a progressively positive public perception, AI has become an important source of economic business and is increasingly implemented within companies. Many people have developed an interest in the applications of AI and are eager to ac- quire the technical skills to tackle them. The effervescence of repository or PyTorch, play an impor- online education tant role in supporting the developer community in creating AI solutions. «With more than three million students enrolled since the course launched in October 2011, the most followed MOOC The massive integration of in the history of online education is AI in academic curriculum Machine Learning,» from Stanford Uni- The proliferation of data use cases versity on the platform Coursera. The and the emergence of new statis- success of this online course, taught tical models have been supported by Andrew Ng, is not trivial. It demons- by the development of numerous trates the strong appeal of AI and the relevant academic courses. These new forms of learning that promote and allow specialization in specific fields democratize it. such as word processing or reinfor- cement learning. «The number of AI Communities of data courses has increased by 103% at the scientists as catalysts of undergraduate level and 47% at the this democratization master’s level over the past 5 years». A specificity of digital revolutions, France now has nearly 18 masters’ de- particularly that of AI, is the conside- grees specialized in AI and more than rable influence of online communi- 1,000 theses on these subjects. These ties contributing to their growth. As a figures are expected to increase: Em- data scientist, it is essential to remain manuel Macron has pledged to «triple abreast of technological evolutions the number of people trained in AI and the best development practices. within three years» in France. There are many online communities, such as Stack Overflow, one of the Companies, key players largest developer communities with 14 in training and academic million users, where AI enthusiasts can projects share their code, knowledge, ideas and challenges. Github is another perfect Increasingly, schools are collaborating example. As a leading code hosting with companies to create content for and versioning platform (with 40 mil- their academic programs. Some artifi- lion users and 190 million projects), it cial intelligence courses are designed has become a de facto clearinghouse and taught entirely by corporate spea- for all open source projects, including kers. These partnerships between libraries used in the data science com- schools and companies are accelera- munity. «More than 73 percent of data ting the professionalization of students scientists use open source libraries’’. and facilitating the recruitment of talent Popular AI-specific libraries, such as trained in AI. TensorFlow with 85,500 clones of the (9) https://aiindex.stanford.edu/wp-content/uploads/2021/03/2021-AI-Index-Report-_Chapter-4.pdf
1 1 image
1 1 5. An ever evolving business-skill reference system. As roles in the data value chain multiply, the data scientist’s job is to maintain focus on their core function: modeling. These evolutions lead to many conceptual changes in the recruitment processes of companies.. The end of “Data cessitates astute project management. Scientists” ? This implies a strong capacity for colla- boration and communication between In 2017, the Harvard Business review business and data experts, as well as classified data scientist as the «sexiest pedagogy on both sides. To be suc- job in the world». The definition of data cessful, data science teams need to be scientist given at that time is now being supervised and supported by roles such challenged! Previously a Jack of all as chief data officer, data quality officer, trades, the job of data scientist is now product owner, or data protection offi- re-focusing on the heart of the trade, cer. The exact roles and responsibilities which is modeling. The roles of data of these professions are still being de- specialists are multiplying in the value bated, but there is a consensus that they chain, upstream and downstream of are necessary. modelisation tasks. Beyond the deve- lopment of POCs, the complete produc- Adaptation of recruitment tion of an artificial intelligence project processes requires specific technical skills such as mass data manipulation, dynamic visua- In the context of fierce competition lization on a web interface or the use of between companies for AI skills, com- cloud technologies. An AI solution «pro- panies are shifting their recruitment duct» team is now made up of a wide processes towards unconventional me- variety of profiles: data scientist, analyst thods, in order to increase attractive- or engineer, machine learning engineer, ness to highly sought-after profiles. In cloud specialist, cloud archi- tect, NLP addition to traditional recruitment chan- scientists, or even frontend developer. nels, monitoring both specialized coder forums and data science competitions is becoming a competitive advantage This evolution is a result of specific scien- for recruiting the best professionals. A tific and technical requirements, but good ranking in such competitions, or also of managerial ones. Algorithms are significant activity on forums, are now essential, but to succeed in an artificial key performance indicators in the eyes intelligence project, business stakehol- of recruiters. Decisive weight is now ders must be involved throughout the given to these elements in the recruit- design process, an integration that ne- ment process..
1 1 image 6. Continuous innovation of technologies. The field of AI has seen one of the fastest evolutions in the market. Innovations and data storage to cloud providers in terms of algorithms and modeling as well as in terms of technologies are acce- has many advantages for companies lerating. Nevertheless, while the last few years have seen the democratization of , even in cases of sensitive data, for deep learning and cloud computing, the next few years could bring algorithmic which cloud providers have specific and technological disruptions, with the emergence of new paradigms (edge AI, offerings. For example, a GPU (Gra- low-code, etc.) phic Processing Unit) must be running at least 70% of the time to be profitable, The new generation of However, deep learning algorithms ge- which is rarely the case for a single algorithms: continuity nerally have anywhere from thousands company. Furthermore, a complete and rather than disruption to millions more parameters than tra- expensive technological ecosystem is ditional machine learning algorithms, required to take full advantage of a The new generations of modeling algo- which makes them more demanding GPU, which is very costly. Transitioning rithms are not disruptive innovations, in terms of the data and machine re- to cloud resources, where companies but rather improvements of existing sources required to train them. This only pay-what-they-use and don’t have models, the most powerful of which are growing need for computational power to manage the hardware, is therefore based on deep learning. The latest ite- and storage capacity puts cloud provi- a strategic move for many companies. ration of deep learning models, trans- ders in a prime position to offer cost-ef- formers, offer great performances and fective solutions to businesses, thanks Cloud providers are therefore bene- are becoming increasingly prevalent to economies of scale on massive fiting from a favorable dynamic in a in all areas of AI, from computer vision infrastructures. rapidly growing market. Nevertheless, (image processing) to NLP (natural competition remains strong, firstly language processing) and time series. Outsourcing computational processing from services dedicated to artificial
1 1 intelligence, but also from data lakes vernment is seeking to give preference The new paradigms of and data warehouses. The latter are to French and European operators and AI: Edge computing and beginning to merge to form a single is implementing a sovereign cloud hybrid models system, accompanied by computation strategy with the «trusted cloud» label. capacities and business intelligence This new situation in France has led to While deep learning is now one of the tools. The field of cloud services has the emergence of new partnerships, most widespread trends in artificial in- recently seen a number of newcomers such as the one between Thales and telligence, other approaches are retur- snatch significant market shares. One Google. ning to the forefront. For example, this of these new players, Snowflake, made is the case for probabilistic approaches a huge IPO in 2021, worth USD 12.4 bil- With a trend of service optimization and expert systems, which were very lion, illustrating the scale of the pheno- (less networking, more simplicity), the widespread until the mid-2000s. These menon. This represented not only the next few years could see a conver- different trends are becoming more and biggest IPO of the year on the US stock gence between cloud providers and more intertwined and operate less and exchange, but also the biggest IPO in SaaS cloud platforms (Snowflake, less in silos, and it is probably from a history for a software company. Today Databricks, etc.), leading to the emer- hybrid approach that the next break- in France, 70% of the data hosting mar- gence of new offerings from cloud pro- through innovation will emerge. Such ket is held by Amazon, Microsoft and viders to fill this gap in their services. algorithms will undoubtedly be much Google. In response, the French go- less resource- and data-intensive, with equivalent or better performance. Re- placing big data with smart data, and memory with the ability to abstract concepts, would greatly reduce storage and computing capacity requirements. Other emerging technologies such as edge AI or federated learning also reduce storage and communication re- quirements while offering better data security. What these two practices have in common is a profound paradigm shift in the way we use data and train AI models. Traditionally, data from many sources such as sensors or cameras is centralized in a data center, before being processed and used to train a single model. In the case of edge AI or federated learning, the data does not leave the source, but rather the algo- rithms and models are placed as close as possible to the data source. This pa- radigm shift offers many advantages, image particularly in terms of network latency and data protection, since data never travels through a network.
1 1 Low-code: evolution or being. It is indeed essential that these error rate is lower than that of existing revolution ? new applications can be integrated into processes. The acculturation of the data the existing ecosystem, have access to within companies plays a decisive role One of the main obstacles to wides- proper data and be maintained in case in the responsibility of CDOs (chief data pread AI adoption is the shortage of of malfunction. The development of this officers). data scientists capable of creating the type of application has the merit of rai- necessary tools and algorithms. No- sing and clarifying data governance is- code and low-code solutions aim to sues, previously implicit and underlying overcome this problem by providing in AI products. While no-code offers simple interfaces that can be used, in limited functionality and little flexibility theory, to build increasingly complex AI in the design of algorithms, low-code systems. Just as no-code web design facilitates the creation of powerful and and user interface tools allow users to more responsive applications, while create web pages and other interactive leaving a certain amount of flexibility to systems by simply dragging and drop- developers. ping graphical elements, low-code AI systems will allow intelligent programs Beyond algorithmic to be created by plugging in different challenges pre-built modules and feeding them with specific data. All this will play a key If algorithmic performance and techno- role in the on-going «democratization» logies used are key success factors in of AI and data technologies. the AI field, the elements upstream and downstream of algorithmic modeling, This paradigm shift in development i.e. data collection and processing on tools is reflected in the growing market the one hand, and the industrialization share of no-code and low-code in the of use cases on the other, are just as highly competitive SaaS (Software as important. However, most POCS are a Service) market. These no-code and never deployed on a large scale, due low-code tools are emerging as very to a lack of production capacity of com- promising solutions. In 2020, they re- panies, which requires the synergistic presented a market share of 15 billion efforts of many departments (project dollars and is expected to reach 112 management, adapted technological billion dollars in 2026. This new type of infrastructure, integration with existing platform flips the development process IT systems, etc.). of an application or service, allowing business experts to be the architects A second challenge lies in the quality of of their own AI solutions. The advan- the data. In many companies today, data tages of this type of tool are multiple, is still too disparate and insufficiently including more permanent added va- standardized to allow industrialization lue to the business with faster design of AI models. In the short and medium times prototyping and deployment. But term, the greatest value of AI likely lies their main advantage lies in facilitating in the standardization and automated change management around AI pro- enrichment of data. jects, a major challenge for companies in the upcoming years. Resistance to change is an obstacle to the development of many artificial In practice, the shift towards these new intelligence projects. For example, the types of platforms remains a little bit automatic modification of data by an al- complicated and the support of data gorithm with results that are difficult to experts is still necessary for the time explain is a disturbing idea, even if the (10) D’après les analyses de Forrester Research
1 1 7. Public euros invested in AI in 2017 (following the submission of the report by MP Cédric Villani), at the end of 2021, the French government announced a new policies and envelope of 2 billion euros, of which a large part, nearly 50%, will be de- dicated to skill development through future institutions of excellence and AI competitiveness training programs. The emphasis on training aims to strengthen France’s competitiveness. As part of the natio- nal AIForHumanity policy, interdiscipli- around AI. nary AI institutes (3IA) have emerged in different cities such as Toulouse, Nice and Grenoble. While the R&D centers of Facebook and Google rea- dily recognize the skill level of French talents, French players are concerned with the difficulties in recruiting. Other priorities of this new investment plan: As AI becomes a lever for transforming society, many governments have adop- the government wants its AI industry ted proactive policies to support an AI-ecosystem. More than 30 countries have to be positioned in emerging markets, such as embedded AI, edge, trusted already created national AI strategies to improve their prospects. Nations that take technologies or even frugal AI. the lead in the development and use of AI will shape the future of this technology and significantly improve their economic competitiveness. To increase the competitiveness of companies, it is not only a matter The race for AI: who are the strategies vary. France sets a 4-year of transforming research work into champions today? horizon with the National AI Research «economic opportunities» via the Program, which precisely lays out the emergence of AI champions, but also So far, according to the ranking com- actions to be taken in this period. Other the dissemination of concrete uses of piled by the Center for Data Innovation countries are planning their strategies these technologies within companies. based on 30 metrics and 6 catego- for the next decade, such as the United The Île-de-France region implemented ries (talent, research, development, Arab Emirates with the AI 2031 strate- the Pack AI system in 2021, with the aim hardware, adoption and data), the gy, which is presented according to of integrating data into the ecosystems United States has established itself certain rankings as being one of the of companies where the complexities as a pioneer of AI, but China conti- most ambitious national AI programs of implementation and appropriation nues to close the gap in certain major in the world in this area. are significant. The “direct return on areas (data quality, computer power, investment” of AI public policies is dif- research, etc.). Without significant po- The strategies for implementing AI po- ficult to evaluate. However, the govern- licy changes in the EU (still recovering licies are critical to their success and ment has mapped out, in its 2021 in- from the UK’s departure), such as the reflect the ambitions of nations. The ventory analysis of the AI strategic plan EU changing its regulatory system to emerging trend is to establish dedi- first phase, 81 AI laboratories in 2021 be more innovation-friendly, it is likely cated administrations or government – the highest number among European that the EU will remain behind the US. institutions specially designed to ma- countries -, 502 start-ups specializing Similarly, without the U.S. developing nage and implement AI strategies and in AI – an increase of 11% compared and funding a more proactive national programs, such as the Ministry of Artifi- to 2020 -, 13,459 people working in AI AI strategy, China will dethrone the U.S. cial Intelligence in the United Arab Emi- start-ups - for 70,000 jobs indirectly ge- rates. In other cases, AI mandates are nerated. With a hint of patriotism, we The proliferation of assigned to existing interdepartmental observe the rise of 6 French unicorns national AI programs bodies or committees or to public-pri- associated with AI, such as ContentS- vate partnerships. quare, Shift Technology and Mirakl, al- Canada was the first country to launch though the capital raised is mostly from a national Artificial Intelligence strate- foreign private investors such as Soft Zoom in on the french gy in 2017 before being followed by Bank or Advent International. strategy more than thirty countries. The time horizons of the different national AI After the announcement of 1.5 billion Finally, in order to further strengthen (11) https://www.tortoisemedia.com/intelligence/global-ai
1 1 the data culture at a regional level, se- veral Hackathons are being organized on a regular basis. The organization of such events show that participants are part of a knowledge management and collaborative work approach. France ranks 6th in terms of organized hacka- thons with 195 per year in 2019. (12) https://www.economie.gouv.fr/la-strategie-nationale-pour-lintelligence-artificielle (13) https://www.maddyness.com/2019/01/15/le-succes-des-hackathons-ne-flechit-pas/
1 1 8. Ethical and societal issues : concerns, controversies and opportunities. AI raises questions. In fact, it relies on very power-hungry structures and contributes AI algorithms can help reduce the en- significantly to digital pollution. The algorithms race raises concerns on the mea- vironmental impact. This involves first ning of human control in light of the «black box» effect of certain algorithms. On a choosing well-scaled algorithms and more positive note, AI has the potential to accelerate global efforts to protect the optimizing computation costs when environment and conserve resources by detecting reductions in energy emissions beginning the design process. The and promoting the development of specific use cases. It could be a key asset in the choice of hardware on which the algo- coming years to address the most pressing environmental and societal challenges. rithm is trained, as well as the choice to use «green» datacenters powered by renewable energy, are other levers that AI, an asset for ecological enable the implementation of AI with a issues In addition to the numerous AI use low carbon footprint. In the long term, cases in the environmental sector, The pressure on companies to respond the use of AI will certainly contribute it is clear that the computing power to climate change and reduce their to the fight against climate change, required for AI algorithms has signifi- carbon footprint is increasing every provided that companies developing cantly increased in recent years. For year. AI is proving to be a real asset algorithms account for environmental instance, DeepMind’s AlphaZero Go as it offers the ability for companies impact and avoid the oversizing of in- game has doubled computing power to better leverage their data to obtain frastructure and computational power roughly every 3.4 months between key indicators on different aspects of required for the algorithms. 2012 to 2018 –increasing 300,000 their carbon footprints. Other AI use times– according to an OpenAI study. cases with high added value for civil Let’s recall that, according to Moore’s Ethics and AI, an urgent protection and the environment are empirical Law, computing power societal issue bound to proliferate in the years to come. For example, Computer Vision doubles every two years. As a result, Ethical issues surrounding AI are gai- techniques can improve the monitoring machine learning systems are beco- ning more and more public attention, of deforestation and offer the ability ming more resource-intensive and with research publications on ethics in for civil protection services to detect generating a significant amount of this area more than doubling between forest fires, which are very harmful to carbon emissions. A 2019 study found 2015 and 2020. The drive to create the environment. Though the fires of that the training process of an NLP mo- increasingly efficient models has led 2021 were among the largest ever ob- del (transformer with neural architec- many companies to prioritize opera- served, caused by increasingly extre- ture) can emit more than 285,000 kg tional gains in recent years, at the ex- me weather conditions, with recurring of CO2, which is nearly five times the pense of interpretability and explaina- droughts, AI applications demonstrate lifetime emissions of an average car, bility. Moreover, the use cases related strong operational potential in the fight or the equivalent of about 300 round- to AI are multiple and diversified but against global warming. For example, trip flights between New York and San the data used to train models is not drones equipped with computer vision Francisco. always as diversified; many groups algorithms could locate «hotspots» of are underrepresented or absent from fires, allowing firefighters to more ef- Planning and assessing the environ- landmark datasets. The proliferation fectively target their efforts. mental impact before implementing of AI solutions raises many concerns (14) https://arxiv.org/pdf/1906.02243.pdf
2 2 about their potential consequences tial to reduce these biases by building of AI will always depend on the purpose such as privacy intrusion, discrimination, adversarial models to highlight dispari- of the use cases. and especially the opacity and racism ties. It has never been more important to inherent in various AI decision-making address existing ethical challenges and processes. Automated decision-making develop responsible and equitable AI tools often reproduce human biases innovations before deployment. based on the race, gender or social class of an individual. AI in the collective As our societies become more and more imagination, the new dependent on decisions made by algo- normal ? rithms, companies must become more The enthusiasm generated by subjects transparent about this decision-making related to artificial intelligence induces process and find solutions to counter the a split among the general public, like black box effect of models and data col- any disruptive innovation. Between lection processes. This is the case, for threat and opportunity, the technicality example, with neural networks, which and complexity of the field of artificial are proving to be very effective for intelligence often leads to an erroneous voice and image recognition, but whose understanding of the stakes, making functioning and results remain difficult widespread adoption of AI difficult. The to explain. In his report, Cédric Villani feelings of loss of control over the un- recommends in-depth studies to make derlying computing processes, as well these algorithms interpretable for regu- as the media coverage of certain use latory purposes as well as for ethical and cases that infringe on privacy, are all performance considerations. Finally, AI elements that will require pedagogical can theoretically reduce discriminatory efforts to better understand how these biases in decision-making processes algorithms operate. The developing within companies but only if significant regulatory framework via the General changes to the status quo are made. Data Protection Regulation (GDPR) and Examples of such decision-making the AI Act contribute to improving the re- processes include the granting of loans putation of AI. The AI Index 2018 reveals and financing in the banking sector, and that media articles on AI have become hiring in the recruitment sector. Artificial 2.5 times more positive between 2016 intelligence techniques have the poten- and 2018. Overall, the public impression
2 2 9. AI regulation is maturing but still in its early stages. The regulation of artificial intelligence is gradually being put in place with the upco- Towards a ming arrival of the Artificial Intelligence Act, which is in line with the first European «Europeanization» of AI regulation around data, the GDPR of May 2018. The AI Act, which is currently being regulation drafted, will be a worldwide reference for a legislative framework for AI, both tech- After having positioned itself as a pioneer nically and in terms of its intended purposes. This regulation, which is supposed to in the regulation of personal data with regulate and not penalize the development of AI, is designed in consultation with the GDPR, the European Union wants private and public actors in the field of AI. to tackle the whole chain, extending regulation to the processing of all types of data and AI use. This is the aim of the draft regulation called the Artificial Intel- ligence Act, first published in April 2021. Currently under review of companies, the draft will be submitted to the Euro- pean Parliament for a vote before being implemented a year and a half later. Gi- ven the massive deployment of artificial intelligence in the European ecosystem, this act aims to establish a regulatory and ethical framework for the use of AI, by placing constraints on its techniques and applications. Thus, any implementation of artificial intelligence in the European Union that falls within the material scope of the AI Act will be subject to this legis- lative framework, regardless of where it image is hosted. The legislative grid chosen for this founding act, namely the European Union, is in line with the desire to deve- lop a common policy for the promotion of AI, which is already reflected in a whole series of initiatives. An example of such an initiative is the European Data Portal, which includes more than a million data- sets shared between the member coun- tries of the European Union15. This act will enable the pooling of access to data and capitalize on synergies.
2 2 A disadvantage to interna- tional competitiveness However, given the international com- petitiveness of the AI market, the aim image of the text is sustainable development, not to disadvantage European compa- nies relative to unconstrained interna- tional companies. Thus, relaxation of some restrictions are planned in order to avoid penalizing innovations. The European Union plans to set up «regu- latory sandboxes’’ that will establish a controlled environment for testing in- novative technologies for a limited pe- riod of time. Companies will be able to access these sandboxes after showing a test plan to the relevant authorities. The European Union wants to promote access to this facility while giving prio- rity to SMEs and start-ups. On the other hand, the regulation is accompanied by a strong budgetary policy to support AI: the Digital Europe and Horizon Eu- Growing legislative at all levels from system-design to rope schemes will devote 1 billion euros pragmatism with development. Requirements, such as each year to Artificial Intelligence pro- similarities to the systematic drafting of technical jects. In addition, 20% of the European cybersecurity documentation serving as references, recovery plan must be devoted to the activity registers or user accessibi- digital transition and AI projects. The AI Act aims to regulate and cate- lity requirements, will help prove the gorize AI applications according to compliance of internal systems and their level of risk. This allows the pro- devices, particularly in the event of an tection of citizens against fraudulent external IT audit. In addition, the IA Act practices performed on sensitive data, sets requirements at the supplier, dis- such as certain real-time biometric tributor and end-user levels, in order to identification systems or protection hold all stakeholders accountable. The against the unconscious manipulation inclusion of all actors in the drafting of of behaviors, for example by promoting the text and the specification of roles scandalous and fake news on social and responsibilities, allows expanded media. The risk-based approach has risk coverage. already been adopted in areas such as cybersecurity, with the concept of CCS The text gives detailed classifications of (Cyber Security Culture) implemented different AI systems and the associated by ENISA (European Network and provisions, which facilitates the proper Information Security Agency). To mi- implementation by the stakeholders by nimize cyber risk, companies have im- reducing possible misinterpretations. plemented awareness campaigns and The AI Act will also allow compliant other means to promote a culture of IT companies to benefit from a better mar- security internally and shift behaviors ket image and more visibility. This risk and beliefs towards cyber security. classification is in line with the current trajectory of the risk-based approach, Similarly, based on risk, the AI Act aims which allows for the adaptability of to create an AI culture. Each company constraints to operational realities. If involved will have to implement go- violations occur, they can lead to re- vernance and processes in order to cord fines ranging from 2% to 6% of prove the mastery of its technologies the company’s global annual turnover. (15) https://data.europa.eu/en
2 2 10. The closing word by David Martineau. For Sia Partners, AI has fundamentally bed our algorithmic cores in industrial Over the next 18 months, we plan to changed the game. The needs of our AI solutions for business users. Today, double the size of the team. This will clients, who now have access to te- our team includes more than 200 data allow us to address clients in new geo- rabytes of data thanks to the Internet experts, in 5 Centers of Excellence in graphies (Middle East, Asia, etc.) and and the IoT, have evolved. We have Europe and North America, with an to offer an even wider range of pro- chosen to put ourselves in a position annual growth rate of nearly 100%. ducts and services (quantitative and to accompany them in the management actuarial services, etc.). We will also and exploitation of their data on two To cultivate our attractiveness and en- invest in the development of emerging fronts : sure the retention of talent, we invest technologies: Blockchain, Cryptos, considerably in R&D. This takes the Quantum Computing, etc. ● Strategy and management consul- form of AI accelerators - elementary ting, with AI roadmap projects, go- algorithmic bricks- that we are deve- vernance projects and data quality loping through thematic Labs, such as approaches on all sectors covered NLP, Time Series, Computer Vision, by Sia Partners. This component has etc. They enable our experts to be developed quite naturally through at the cutting edge of each of these the crossroads of our consulting rapidly evolving AI technologies, and teams and our Data teams. to be trained on a continuous basis. ● T he technological component is part of a stronger break within our For 20 years, the DNA of Sia Partners’ traditional consulting activities. It Business Consulting interventions has is through Heka, our Data and AI been «problem-solving» thanks to the ecosystem, that we have chosen sector expertise of our consultants. to reinvent ourselves in order to We use AI to address these challen- attract and grow talent around Data ges differently. This set of expertise and develop AI software solutions. and solutions allows us to intervene Heka allows us to have a dual value on very diverse issues. We support a proposition: Data experts to support major insurance company on its entire our clients’ projects at the heart of AI strategy and roadmap, just as we the Data Labs, as well as industrial work with ski area operators to opti- AI solutions to further accelerate mize snowmaking using predictive our interventions and respond to algorithms and intelligent sensors. the common problems of many of our clients. To ensure the proper deployment of these solutions to the standards of our We have built our AI team around a clients’ IT teams, we have equipped core workforce of 75% Data Scien- ourselves with an industrial platform. tists, whereas more traditional firms It ensures rapid and sound develop- have invested more in Data Analysts. ments and guarantees the scaling of Heka’s positioning is therefore closer our solutions. It is the main tool of our to the pure players in AI. We have teams, from POC to industrialization. completed this Data Science exper- Today it hosts nearly 100 customer tise with profiles of Data Engineers, platforms, over 1,000 projects and Devops Engineers, Web Developers, 125,000 pipelines. UI/UX Designers. They allow us to em-
2 2 Glossary (1/3). Big Data : The definition of Big Data is as follows: more varied data, arriving in increasingly large volumes and with greater volumes and at a higher speed. This definition is also known as the three V’s. (https://www.oracle.com/fr/big-data/what-is-big-data/) Cloud : «The cloud» refers to servers that are accessible via the Internet, as well as the software and databases that run on these servers. Cloud servers are located in data centers around the world. (https://www.cloudflare.com/fr-fr/learning/cloud/what-is-the-cloud/) Clustering : Clustering is a machine learning method that consists of grouping data points by similarity or distance. It is an unsupervised learning method and a popular technique for statistical analysis of data. (https://analyticsinsights.io/le-clustering-definition-et-implementations/) Data-driven : Data Driven, also known as Data-Driven Marketing, is based on an approach that consists of making strategic decisions based on data analysis and interpretation. (https://www.atinternet.com/glossaire/data-driven/) Data Lake : A data lake contains data in an unstructured way. There is no hierarchy or organization between data elements. The data is kept in its most raw form and is not processed or analyzed. A data lake accepts and stores all data from different sources and supports all data types. (https://www.oracle.com/fr/database/data-lake-definition.html) Data Warehouse : A Data Warehouse is a relational database hosted on a server in a Data Center or in the Cloud. It collects data from various and heterogeneous sources with the main purpose of supporting the analysis and facilitating the decision making process. (https://www.oracle.com/fr/database/data-warehouse-definition.html) Deep Learning : Deep learning is a sub-field of machine learning. It consists of algorithms capable of mimicking the human brain through a large network of artificial neurons. The systems are therefore able to learn, predict and decide autonomously. (https://blog.hubspot.fr/marketing/deep-learning) Edge Computing : Edge computing is a method of optimization used in cloud computing that consists of processing data at the edge of the network, near the source of the data. (https://fr.wikipedia.org/wiki/Edge_computing) Federated learning : Federated learning is a machine learning technique that trains an algorithm on multiple decentralized servers, containing local data samples, without exchanging their data samples.. (https://24pm.com/117-definitions/359-apprentissage-federe) Hackatons : A hackathon, programming marathon or programmathon is an event during which groups of volunteer developers come together for a given period of time to work on computer programming projects in a collaborative manner. (https://fr.wikipedia.org/wiki/Hackathon) Hyper-segmentation : A marketing strategy that consists of developing offers in such a way as to best respond to the needs of the consumer, or even to create custom-made products. (https://academy.visiplus.com/ressources/definition/hypersegmentation) AI : The term «Artificial Intelligence», created by John McCarthy is «the set of theories and techniques implemented to create machines capable of simulating human intelligence» according to Larousse dictionary. It is defined by one of its creators, Marvin Lee Minsky, as «the construction of computer programs that perform tasks that are, for the time being, more satisfac- torily accomplished by human beings because they require high-level mental processes such as perceptual learning, memory organization and critical reasoning». (https://fr.wikipedia.org/wiki/Intelligence_artificielle#D%C3%A9finition)
You can also read