Connect with us


DeepMind, the Artificial Intelligence division of Google, does not achieve success in computer programming

Avatar of Thomas Grimm



deepmind 1000x600 jpeg

For DeepMind there were no limits. He has been able to solve such disparate problems that go from the StarCraft video game saga to protein analysis, but now he has found a serious problem; the computer programming.

Google’s AI division is not capable of generate more elaborate codealthough in itself it has merit that it can work without having received basic information about algorithms and programming languages.

Computer programming is highly internalized among humans, and therefore should be a simple operation for an efficient AI system. But it has been causing problems in this regard, and the DeepMind team has not been able to find the key to the error. At first they associated it with a language problemthat is, the challenge description is an expression of what the algorithm should do, but the code is an identical expression albeit in a different language.

Thus, DeepMind AI it has been trained to acquire the description and convert it into an internal representation that would later generate a functional code. At first I was asked to process some of the material on GitHub, some 700GB of codeto which if you add the plain text would represent many more lines per gigabyte, a natural language that can be confusing.

DeepMind fed the results into the system very well organized: problem description, working code, failed code, and test cases used.

a latent problem

More than 40% of the solutions offered by Google’s AI system tended to run out of memory or simply failed to find an efficient solution in a reasonable amount of time, resulting in poor code. The initial idea was to see if any program AlphaCode could pass that initial test, but only 1% did.

The solutions that worked for DeepMind were similar to each other, with the entire set of incorrect answers being randomly distributed. The system identified the ten largest code groups and chose a representative from each group.

But even so, DeepMind was not sufficiently prepared, and the results denote that more than 54% of the computer programmers who would have faced the AI ​​would have defeated it, since it has the level of a programmer with little training and experience.

The training of the AI ​​system was based on more than 2,000 petaflops and took with him 16 times the annual energy budget of an average home. By increasing the number of solutions, the system was able to generate a greater number of correct solutions, following a factor of 10. Thus, we determined that AlphaCode works better, but is much more expensive.

However, and despite its limitations, the system of DeepMind it is working as expected and is doing internal translations between the problem and the solutions, not just throwing out chunked code. AlphaCode will continue to send you data constantly, allowing the new version to improve considerably.

The system devised by DeepMind could generate shorter code snippets that handle specific requirements. Similarly, it is curious that AlphaCode was never given an indication of what constitutes an algorithm. However, it generated functional code using an AI structure similar to that used for language translation.

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *


Microsoft increases the staff of its artificial intelligence center in Barcelona by a hundred

Avatar of Thomas Grimm



Microsoft aumenta en un centenar la plantilla de su centro de inteligencia artificial en Barcelona

After the company announced a total of 10,000 layoffs worldwide, representing 5% of its workforce, Microsoft is now announcing the hiring of 100 new employees for its artificial intelligence (AI) center in Barcelona. This new wave of hiring, which contrasts with the economic recession and job cuts in the technology sector, adds to the 80 people who have joined the center in recent months.

That Microsoft hires 100 new employees for its AI center is not trivial, since since its opening in September 2021 its activity and productivity based on the development of models on artificial intelligence have not stopped growing. And it is that this space, attached to the division WebXT (Web Experiences Team), works to improve services such as Azure, Bing or Windows.

Microsoft seeks to strengthen its technology hub in Barcelona by attracting talent in new vacancies for specialists in Artificial Intelligence, Machine Learning, Software Development and Data Science.

With implementations in software engineering, data science, and machine learning, new AI models and innovative web experiences can be developed. The goal is to create a diverse work team in gender, culture and nationalities.

The president of Microsoft Spain, Alberto Granadoshas appeared during an event organized by the New Economy Forum and has stated that the employment adjustment plan will not affect the data center projects that exist in our country, although they will give priority to certain areas in order to face new challenges and projects .

However, he values ​​the global layoffs as a conservative and positive lineafter having increased its workforce by more than 50% in recent years.

Granados has assured that the layoffs from Microsoft will neither reach Spain nor will they affect the construction of data centers in Madridas well as the launch of his cloud region. Similarly, it ensures that the data centers of Meco, Algete and San Sebastián de los Reyes will have a impact of 15,000 million euros in the national economy and will generate 50,000 jobs.

A clear commitment to AI

The AI ​​center in Barcelona is the fourth that Microsoft has opened in Europe, after those in London, Paris and Munich, the eighth in the world and the first in Spain. It must be taken into account that AI does not stop growing and that, in addition to having been declared ‘the word of the year’ by the FundéuRAE, it is expected to that ends up being fully integrated into the business structure and assuming a brutal transformative effect on the economy.

Since Microsoft announced its investment in Open AI, its commitment to the integration of all Technologys (such as ChatGPT) in its products is complete. Thus, it is sought that it can be applied in Azure or Power Point so that users could generate texts or images organically. The objective of the company in Spain, in the words of Granados himself, is to Mark new opportunities with various Ibex 35 companies.

The WebXT organization It works as a start-up environment where you work on challenges that allow you to improve the consumer experience through optimal products and services such as Bing, Edge and MS News.

Spain, core of digital talent

To speak of Spain is to do it one of the ten countries that most attract and retain digital talent, being the fourth country in Europe, after the United Kingdom, Germany and France, in number of professionals and developers in the cloud. All this represents 10% of professionals certified in advanced Technologys in Europe.

With its new AI center in Barcelona, ​​Microsoft seeks to make Spain and Catalonia a reference in the sector. His plan is to continue working in partnership with universities, research centers and technology companies to promote training in AI and Machine Learning Technologys.

Continue Reading


OpenAI launches a tool to catch those who use ChatGPT and other AIs to generate texts

Avatar of Thomas Grimm



openai herramienta pillar textos chatgpt ias

OpenAI has announced the launch of a tool to distinguish between human-generated texts and the created by Artificial Intelligence systems, such as ChatGPT and GPT-3. Is named OpenAI AI Text Classifier, and for now it is not excessively accurate, since it has a success rate of 26%. However, according to OpenAI, if used in combination with other systems, it can be useful in preventing users of text-generating AI systems from abusing them.

Until now, mostly because students were starting to cast texts created with ChatGPT as part of their homework. They have also used them in other areas, which has led various institutions to prohibit the use of ChatGPT on their networks and equipment. This is the case of several schools in the United States. They have done so out of fears of the impact they may have on student learning.

Among the entities that have banned ChatGPT is Stack Overflow, whose managers have decided that their users cannot share content generated with ChatGPT, alleging that Artificial Intelligence makes it very easy for users to flood their discussion forums with more than dubious answers to the questions posed.

In other words, their mission will be to help discover those who use texts generated by Artificial Intelligence, ensuring that they have written them, or any other human. Of course, as the company points out, it is best to use it as a complement to other systems for identifying the text source, instead of using it as a primary tool to identify if it is written by humans.

In addition, it is a classification tool considered initial by OpenAI, although the company has decided to launch it to obtain information from its users about its utility. In OpenAI they also assure that they hope to share improved systems in the future, although without specifying.

As with ChatGPT, OpenAI AI Text Classifier is an Artificial Intelligence language model trained with a multitude of publicly available text examples on the web. But unlike said tool, it is tuned to predict the probability that the text it analyzes is generated by any Artificial Intelligence model capable of generating text.

Specifically, OpenAI has trained this tool with text from 34 text generation systems from five different organizations, including OpenAI. This text was matched with somewhat similar human-written text on Wikipedia, websites pulled from links shared on Reddit, and a set of human-written samples compiled by an older OpenAI text generation system.

Text Classifier does not work, yes, with any type of text. You need me to have a minimum of 1,000 characters, or between 150 and 200 words. In addition, it does not detect plagiarism, and according to the company, it is more likely to fail to detect texts written by an AI if the text is written by children or in a language other than English. The latter is mainly due to the fact that the data with which the tool has been trained is fundamentally in English.

The tool, when analyzing a text, displays a response, which determines the probability that the text is or is not generated by Artificial Intelligence. Thus, it will classify it as “very unlikely” that it is generated by an AI when it believes that this probability is less than 10%; “unlikely”, when the probability of this being the case is between 10% and 45%; “not clear” if the probability that it was not written by a human is between 45% and 90%; “possibly” generated by AI if it has between 90% and 98% probability that it is so, and “probably” generated by an AI if the percentage probability exceeds 98%.

Continue Reading


Did you miss the webinar? Find out how to reduce your environmental impact here

Avatar of Thomas Grimm



amazon proyectos energia renovable centros datos

Sustainability in IT operations and data centers has become one of the main challenges that companies face. Being sustainable makes it easier for companies to minimize their carbon footprint, reduce their costs, increase their productivity and improve their brand image among consumers who are increasingly concerned about the future of the planet. However, to be sustainable in an efficient and sustained way over time, it is essential to have the right partner.

A few days ago we held our webinar “How VMware helps you reduce your environmental impact” in which experts from the American multinational told us why it is important for all companies to bet on technological sustainability, what initiatives and strategies the company is implementing. company for reduce your environmental impact and how organizations can use VMware solutions to reduce their carbon footprint

Throughout this virtual meeting we have the participation of Lluis Altés, Senior Business Solutions Strategist at VMware, Clara Arbide
Iberia Regional Business Manager-Cloud Management at VMware and Carlos CarballeiraCloud Management Lead Solution Specialist at VMware, who told us about initiatives like “VMware Zero Carbon Committed” a program to catalyze and accelerate the transition to zero carbon clouds through the company’s associated data centers that are powered by renewable energy sources; or how VMware Aria helps companies become more sustainable, thanks to a cloud management solution that unifies applications, infrastructure and services in private, public and hybrid clouds, from a single platform, with a single data model.

If you missed the webinar or if you want to review some of the most interesting moments of this technical session, you can now do so by clicking on the following link. Do not think twice and discover how you can advance on the path of sustainability.

Continue Reading