Personalized GPTS are one of Openai’s great inventions. Now Google has just released yours in Gemini

One of the most interesting functions that chatgpt has are GPTS. In a nutshell, they are Chatgpt personalized versions created for specific purposes. We could have a GPT focused on correcting our texts, to solve mathematical problems or plan trips. It is a really useful function, but for now they can only be created by payment users. Anyone can use them, but only premium users can create them. Well Google has decided to opt for a different path with Gemini and its Gems. And yes, the user win. Gems? That is the name that the equivalents of GPTs receive on Google Gemini. To all purposes, they are exactly the same. Instead of using Gemini’s “general” version, Gems allow us to use a specialized version in certain tasks. It is a function that, having dedicated the necessary time and care, can be very useful. 10 Google applications that could have triumphed For everyone. So far, only payment users could create and use Gems. That is, the only way to access this function was paying the 21.99 euros per month that costs access to Gemini Advanced. That is over. As plannedGoogle has released access to gems and, from today, creating them and using them is completely free. Gems creator | Capture: Xataka Options. Google gives us five predetermined gems focused on brainstorming, professional orientation, programming, learning and writing review. Grace, however, is to create ours. To do this, you just have to go to the Gems manager and start the process (or you can do it by clicking directly in This linkwhich is direct access). Important: Although gems can be used from mobile app (and deployment is being progressive), they can only be created in the web version. Some keys. When creating a GEM it is important to be clear, concise and descriptive. Here are Some tricks to get the perfect prompt. For example, if we want our Gem to serve us to correct texts in English, something like this should be put: “You are a text reviewer in English and aid to people to detect and correct failures in their writings. Your work is to analyze the texts, find all errors, explain to the user why he is poorly written and suggest improvements. Use a friendly tone. Use Spanish to give explanations. Be patient.” The result will be something similar to this: when giving the badly written phrase “I are not feeling lots well”, the GEM returns the following answer: Example of use of a Gem created by us | Capture: Xataka Models. In our Gems we can use the models we have access to Google Gemini. In the free version we can use Gemini Flash 2.0 and Gemini Flash 2.0 Flash Thinkingwhich is experimental. If we had the payment version we could use the most advanced models. Using the reasoning model can be really useful if we create a very specific Gem focused on answers that need precision. Limitations. Gems are very useful, but they have an important limitation: they do not admit the rise in documents, at least in the Spanish version and for now. In the English version they seem to admit them. Being able to upload documents is a very interesting function to consult bibliography, interact with a PDF, with an Excel sheet, etc. Let’s think about the potential that this has to analyze data, extract trends or digest a lot of information more easily. The problem is that, for the moment, we do not have it available. Cover image | Xataka In Xataka | Google’s results with generative come to Spain. And with them, an elephant in the media room

Openai’s new voice models already speak as customer service agents. His next destination: the call centers

Since the beginning of the year, the objective of great technological ones has been clear: that we talk to artificial intelligence (ia). Openai, Microsoft, Google and Meta have added voice functions to their assistants. But this seems to be just the beginning. The industry advances at a frantic pace and the way we interact with these tools continues to evolve. Tell the voice agents ‘hello’. Sam Altman’s company has been betting on text agents with tools such as Operator either Computer-Useing agents. However, Openai already has it ready if next great movement to continue highlighting in the race for the development of AI: to promote a new and powerful generation of voice agents. New models on stage. OpenAI has announced The launch of new audio models to turn voice into text and vice versa. They are not in chatgpt, but in the APIwhere developers can use them to create voice agents. The important thing? They aim to be much more precise and to bring customization to the next level. The new OpenAI models, built on GPT-4O and GPT-4O-minipromise to improve Whisper Already its previous text to voice tools, which will also remain active through the API. But it is not just a matter of performance: now they can also modulate their tone to sound, for example, “as an empathic customer service agent.” Destination: the call centers. Openai makes it clear where they point with this launch. He assures that “for the first time, developers can tell the model not only to say, but also how to say it, which allows more personalized experiences for use cases ranging from customer service to creative narrative.” According to Openai, this technology will allow creating much richer “conversational experiences.” If we take into account that Chatgptpowered by GPT-3.5arrived in November 2022, it is evident that the progress has been vertiginous. And everything indicates that these models will end up arriving at the call centers. We might think that at first the interactions will be somewhat limited, but well above the current voice systems. They will move away from traditional automated assistants and will be much more natural. Over time, the line between a conversation with a person and an AI could become almost imperceptible. Images | Charanjeet Dhiman | OpenAI In Xataka | We have tried Sesame’s conversational. It is the experience closest to a “human voice” that we have seen In Xataka | China has found an unusual strategy to avoid US mosquadillas with AI: bet on the Open Source

Musk is trying to block Openai’s transition to “for-profit.” A judge just put it very difficult

Sam Altman wants to make Openai a company with profit (“for-profit”), but the process was notably complicated when Musk offered to buy it for 97.4 billion dollars. The tycoon did not stop there and He also tried to sue A Openai to try to block that transition to For-Profit, but just lost that legal battle. What happened. A federal judge in California has blocked Musk’s legal effort to stop Openai’s transformation to an entity of profit. As indicated In BloombergJudge Yvonne Gonzalez Rogers determined that the lawsuit “have not satisfied the probative charge” that would have needed for that demand to prosper. A case that was entangled. In March 2024 Musk He sued Openai for violating their contracts and fiduciary duties. The case He retired of the State Court and was activated in the Federal Court, and that was when Musk expanded the demand indicating that Microsoft and Openai had violated antitrust laws. Openai defended himself from these allegations publishing internal messages of the billionaire, and now the federal judge has made it clear that the evidence and arguments presented by Musk are not enough to avoid that transition sought by Altman. But not everything is lost. The demand contained other Musk requests with respect to OpenAI, although it is not detailed which. Even so, the judge has indicated that these other components of the lawsuit may remain active in the legal process. Sam Altman has it easier now. The Openai Directorate Council declared a few weeks ago that “the hundreds of billions of dollars that large companies are now investing in the development of AI show what is really needed so that Openai continues to pursue the mission.” With the Form-Profit structure, the company can avoid the limitations of investments in your company. Thus, Openai will be able to eliminate that benefit to investors, which can attract even more money for the company. Quick judgment. Rogers added in his sentence that Musk’s complaints are resolved as soon as possible “the public interest in play and potential damage if a transition contrary to the law occurs.” Thus, the judge indicated that he will hold an accelerated trial focusing on the main demand that the Openai conversion plan is illegal and “potentially the interrelated demands based on contracts”. The war between Musk and Altman continues. Openai’s lawyers stood out as Musk’s demand is basically a demand for the competition with the company. “Elon’s own emails,” they explained, “they show that they wanted to merge an openai with profit with Tesla. That would have been great for their personal benefit, but not for our mission or the interests of the US.” Image | Ted | Techcrunch In Xataka | Elon Musk’s continued criticism A OpenAi have a simple explanation: it went too soon

Deepseek does the same as Openai’s most advanced models with much less resources. The key: “Reinforcement Learning”

The entire world is wondering how it is possible that the models of AI of Deepseek They have become overnight the great protagonists of today in the field of artificial intelligence. The answer is relatively simple. These models have managed to demonstrate that You can do more with much less. Both Deepseek V3 and Deepseek-R1 are comparable to GPT-4 or O1 OPENAI respectively, but it is estimated that their training has been much less expensive and its inference, of course, is: the prices of the Deepseek API are up to 35 sometimes lower than those of OpenAi, but that makes one wonder how it is possible. The answer is clear, and it is because we have at our disposal the technical reports of these AI models. Precisely his study has allowed us to clarify What are the techniques that this Chinese R&D laboratory has used to develop these models so efficient and capable. Many techniques, a single objective: efficiency There are several differences that make Deepseek’s new model especially efficient. Its creators explain in detail in the detailed Technical Report that is publicly available. Here are the most relevant: Deepseekmoe (“Mixture of experts”): In models such as GPT-3.5 the entire model was activated in both training and inference (when we use it). However, not all model components are necessary for our requests. The MOE technique – already introving with Deepseek V2 – precisely divides the model into multiple “experts” and only activates those that are necessary according to the request. GPT-4 is already a MOE model. But as we said, Depseekmoe even went further and differentiated between even more specialized experts, in addition to using some somewhat more generalist experts that could contribute value in certain requests. Managing all those specialized or generalist experts not only benefits inference, but also the training phase, making it more efficient. This technique is similar to the so -called “Time Scaling test” that also adjusts the size or complexity of a model during efficiency. Deepseekmla (Multi-Head Latent attention): It is another substantial improvement-even more than the previous one, and also introduced with Deepseek V2-that affects the way in which memory is managed in these models. Normally it is necessary to load both the model and the entire context window – the one that allows us to write prompts and include long texts, for example. Context windows are especially expensive because each token requires both a key and their corresponding value. With the improvement introduced with this technique, what was made possible was to compress that warehouse of keys and values, dramatically reducing memory use during inference. Auxiliary -los-Free Load Balancing: If we imagine a model like a great orchestra, each musician is an “expert” within the model. To play a complex piece, not all musicians are necessary all the time. Traditionally the so -called “auxiliary losses” were used to make sure that all musicians played enough, but these losses could interfere with that interpretation of the musical piece (model training), which could degrade general performance. With Deepseek V3 the model is able to balance the work of each expert dynamically. That does the simplest, direct and efficient training by eliminating “auxiliary losses.” In addition, the elimination of interference allows the model to learn better and with less resources … and get better results. Multi-Token Prediction Training Objective: Often predicting the following word depends on several previous words or context. With this technique instead of predicting only the following word, the model learns to predict several words at the same time. That makes more natural and understandable and less ambiguous texts generate, but also accelerates training by reducing the number of steps necessary to generate the complete text sequence. FP8 Mixed Precision Training: The use of Numbers FP8 allows significantly reducing memory consumption and accelerates calculations. Some critical parts of the model continue to use FP32 training to guarantee precision, but there is another additional benefit of FP8: the size of the models is reduced. Other models use techniques such as quantization or parameter pruning. Although Openai does not give data on GPT-4 in this section, the assumption is that it works with BF16, more expensive in terms of memory. Although FP8 theoretically leads to less precise models, other complementary techniques such as fine-grained quantization are used to reduce the negative impact of values ​​that come out of the common, which makes a stable training possible. Cross-Node All-to-Lall Communication: During training it is necessary to constantly exchange information between all nodes (computers) connected in training data centers. That can become a bottleneck, but these new Deepseek V3 techniques include efficient communication protocols, data traffic reduction and efficient synchronization to accelerate training and, once again, reduce the costs of that process. Reinforcement and “distillation” learning as keys But in addition to all these techniques, those responsible for Deepseek V3 explain how they pressed it with 14.8 billion tokens, a process to which a supervised adjustment followed (Superved Fine-Tuning, SFT) and several stages of Reinforcement Learning (Reinforcement Learning, RL). The SFT phase-which is mentioned in the Deepseek V3 report-was completely omitted in the case of Deepseek-R1. However, learning by reinforcement is an absolute protagonist in the development of both models, especially in R1. The technique is well known in the field of artificial intelligence, and it is as if we trained a dog with prizes and punishments. The model learns to respond better by giving rewards if you do well. Over time, the model learns to take actions that maximize long -term reward. In Deepseek, learning for reinforcement is used to break down complex problems in smaller steps. In it Deepseek R1 technical report It also indicates how this model makes use of RL techniques directly on the base model, without the need for supervised training. That saves computing resources. The call also comes into play here Thought chain (chain-of-though)also mentioned in the technical report. This refers to the ability of a language model to show the intermediate steps of its reasoning. The model not only … Read more

OpenAI’s rival company is taking steps to address the problem

The latest innovation in the field of artificial intelligence (AI) It comes from China and is called DeepSeek. This chatbot, which competes directly with OpenAI’s ChatGPT, has gained notable popularity in recent hours. Many users highlight its capabilities, and some even wonder if it is worth investing in similar tools. However, it is not all good news. DeepSeek, the company behind the tool of the same name, says it has suffered a cyberattack. According to the interface itself, these are “large-scale malicious attacks” that have impacted its services. The nature of the incident is still unclear, although there are indications that point to a distributed denial of service (DDoS) attack. The measure the company has taken to address the problem has been to temporarily limit registration of new users. This means that people who want to use the AI ​​chatbot for the first time will likely not be able to register, at least on the first try. DeepSeek itself recommends trying again if you are not successful. Users who were already registered on the platform should have no problems accessing it. However, in our tests we observed that the chatbot took longer than usual to respond. It should be noted that, in the last few hours, the service has remained inaccessible for some users, as indicated by the DeepSeek status page. In development. Images | DeepSeek In Xataka | NVIDIA has lost 400 billion in market value. The finishing touch has been given to the Chinese AI DeepSeek

Log In

Forgot password?

Forgot password?

Enter your account data and we will send you a link to reset your password.

Your password reset link appears to be invalid or expired.

Log in

Privacy Policy

Add to Collection

No Collections

Here you'll find all collections you've created before.