Leveraging AI
Dive into the world of artificial intelligence with 'Leveraging AI,' a podcast tailored for forward-thinking business professionals. Each episode brings insightful discussions on how AI can ethically transform business practices, offering practical solutions to day-to-day business challenges.
Join our host Isar Meitis (4 time CEO), and expert guests as they turn AI's complexities into actionable insights, and explore its ethical implications in the business world. Whether you are an AI novice or a seasoned professional, 'Leveraging AI' equips you with the knowledge and tools to harness AI's power responsibly and effectively. Tune in weekly for inspiring conversations and real-world applications. Subscribe now and unlock the potential of AI in your business.
Leveraging AI
147 | OpenAI’s 12 Days of Releases, Google’s Game-Changing Video Generation Leap, and Amazon’s New Models (Nova) and more AI news for the week ending on Dec 7, 2024
Are we living through the fastest AI innovation sprint in history?
This week, the AI world was set ablaze with groundbreaking releases from OpenAI, Google, and Amazon, among others. From OpenAI’s ambitious “12 Days of OpenAI” campaign to Google’s jaw-dropping Veo video generation capabilities, and Amazon unveiling the Nova series, innovation is moving at lightning speed.
What do these advancements mean for businesses, creators, and technology leaders? How will the introduction of multimodal agents, advanced reasoning tools, and video AI redefine the future of work and creativity? This episode dives into the transformative announcements and their implications.
If you’re looking to stay ahead of the curve, you can't miss this episode:
In this session, you’ll discover:
- OpenAI’s O1 Pro model: what it is, why it matters, and how it’s setting new benchmarks in STEM reasoning.
- Google’s Veo: a 1080p video generation model designed to disrupt content creation.
- Amazon’s Nova family: affordable and advanced AI models for text, images, and video.
- The rapid evolution of multi-agent systems and their potential to redefine enterprise operations.
- How OpenAI’s new pricing tiers could shape access to advanced AI tools.
- Why video AI could soon surpass human-generated content in quality and speed.
About Leveraging AI
- The Ultimate AI Course for Business People: https://multiplai.ai/ai-course/
- YouTube Full Episodes: https://www.youtube.com/@Multiplai_AI/
- Connect with Isar Meitis: https://www.linkedin.com/in/isarmeitis/
- Free AI Consultation: https://multiplai.ai/book-a-call/
- Join our Live Sessions, AI Hangouts and newsletter: https://services.multiplai.ai/events
If you’ve enjoyed or benefited from some of the insights of this episode, leave us a five-star review on your favorite podcast platform, and let us know what you learned, found helpful, or liked most about this show!
hello and welcome to another Weekend News episode of the Leveraging AI podcast, the podcast that shares practical, ethical ways to leverage AI to improve efficiency, grow your business and advance your career. This is Isar Mehti, your host, and we have a jam packed week. It seems that there has been more new feature and model releases this past week than ever Any week in history. So while we usually dive into several large topics and then rapid fire, this is going to feel more like rapid fire beginning to end. But the entire first segment of the show is going to talk about new models and new features that have been released or just about to be released by more or less anybody who has anything to do with AI on the planet. So we have a lot to talk about. So let's get this started. If you've been listening to the show for a while, you know, there've been huge anticipation to see what open AI is going to do towards the end of the year, or more specifically towards the two year anniversary or birthday of ChatGPT. Well, nothing actually happened on the ChatGPT day itself, but what they have done is they just announced 12 days of open AI, which basically are 12 days in which every single day they're going to release a new capability. Some are big and some are small, or how they labeled it, big ones, and some stocking stuffers. So the first one that we got is the full O one model. So we had access to all one preview and O one mini, both were groundbreaking, new kind of family of models that can think and analyze things in a much deeper way than before. This took the world by craze, and now basically everybody's chasing the same thing. And we saw multiple releases from Chinese companies, which we're going to talk about as well in the show that are trying to do the same thing. Well, now we finally got the full O1 model, and that was the first release in the first days of the 12 days of OpenAI. Just as a quick reminder, we talked about this when this was announced, but oh, one is achieving incredible solutions in problem solving, and it scored an 83 percent success rate on the international mathematics Olympiad qualifying exam, which is a very big spread than GPT for Oh, who scored 13%. So that's a huge spike. also big, huge decline in error, a significant improvement in error reduction and a very high success on anything that has to do, that requires deeper reasoning, as well as STEM capabilities. Now, in addition, they announced a new ChatGPT Pro subscription that is going to be 200 per month. So that's a pretty big spike from the 20 bucks a month. Most of us are paying of the 25 or 30 for the team's versions, depending exactly which plan you pick. And then there's obviously the enterprise versions that are priced very differently. So right now there is going to be one more tier. There's a free version. There's the plus version, which most of us use. If you're just a regular user, which is 20 bucks a month. And then it's going to be a 200 a month, for the pro version. And I'm reading for their website. It's going to have everything in plus and. Unlimited access to GPT 4. 0 and GPT 4. 1 and access to advanced voice and access to 01 pro mode, which uses more compute for the best answers to the hardest questions, which basically tells us that there is going to be limited access to 01 if you're not on that mode and standard and advanced voice mode accessible, but with some limits if you're just on the 20 bucks a month plan. Basically what this is it's targeting people who are heavy users who are going to use this all the time and needs more bandwidth. And for them, it will probably make sense to pay 200 a month. It will be very interesting to see if that actually works because that's a very, very big spike. That being said, people who actually use this at that capacity probably see the value in it and probably will pay the 200. This will open the door for many other providers to do similar, significantly higher early to follow a similar approach with significantly higher monthly rates to use various advanced services. On day two, which was on Friday, December 6th, they released the capability to allow people to do their own reinforcement fine tuning. So you can now apply And now I'm reading from their website again, applied to the reinforcement, fine tuning research program, or expanding alpha access to reinforcement, fine tuning and inviting researchers, universities and enterprises with complex tasks to apply, spots are limited. So that was the announcement of day two on the day this is released. It's going to be day three and so on. So if you're listening, they're releasing one of those at 10 a. m. every single day. Pacific time every single day. And you can chime in and see exactly what the release is, but there's going to be a lot of exciting stuff. What might be some of that exciting stuff I really hope, and many other people hope that Sora will be released as part of those capabilities, which is the really incredible, presumably video generation model that they have announced and showed off in February of this year and haven't released to the yet. And now a lot of the other companies are catching up to it. And so that's a highly anticipated one. And they may or may not, release that and maybe glimpse or segments or previews of GPT five or whatever they're going to end up calling it. We may get that, maybe more advanced voice capabilities and stuff like that. So there's a lot that they may release, but we don't really know. And like I said, some of them announced that are going to be big. Some are going to be small, but definitely pay attention because every single day in the next few days, there's going to be another announcement from open AI. Now staying on the topic of interesting releases. A Chinese AI firm called oh one AI has achieved an really interesting release. So they released an AI model that is at the level or close to the level of GPT-4. The company is called Y Lightning, and their model is currently rank number six on the LM sys chatbot arena, and they have trained the model with an investment of 3 million to put things in perspective chat GPT four. Oh, was trained at an 80 to 100 million dollar investment. They've done this with only 2000 GPUs, despite the U S restrictions. That's what they were able to get their hands on. And they were able to achieve that by just solving for problems and being more innovative than what the big companies are just have the money and the brute force to put into this. And they solved things like reduced computational bottlenecks and multi layer caching implementation, specialized in inference engine that they developed specifically for their needs and optimized memory usage. So when we hear people like Ilya Saskover saying that it's not about just old scaling laws, but we need to think out of the box. This is a very good example in my eyes. So this model is now available. It's open source. You can use it. And it's significantly cheaper on inference as well than the big models while it's generating results that, as I said, place it six out of hundreds of models in the world right now. Another big announcement came from Google this week. So they're rolling out a major update to Gemini on pixel phones and pixel watches. And what it allows you to do is that allows you to use multiple extensions to connect and control and work with multiple apps on your phone. The biggest ones that they just released are Spotify, messaging, calling, and smart home. The biggest gripe that people had with the new Gemini that they're trying to replace just a Google assistant is that it didn't control most of the important apps on your phone. And now this feature has been released that allows these extensions to talk to the other apps. As I mentioned, a few has already been released and I'm sure we're going to see more and more of them, which will most likely make Gemini the new tool for most people use Android phones, like myself to engage with your phone and with different applications. But that's not the biggest announcement from Google. Google has launched Veo, Which is it's AI video generation model that was in private release for a very long time and now is available for preview via Vertex AI platform, which anybody can have access into if you're just creating an account and logging into that, and it is available. Incredible. So it generates 10 ADP resolution videos from either text or images. It is very good at following content in various visual cinematic styles. So you can book a specific style and it sticks with that style through the entire video generation. It can produce videos beyond a minute per them. I haven't tested it yet. And it includes built in safeguards against harmful content. And so people cannot generate violence and otherwise problematic content. The other additional interesting thing is that it also includes deep minds, synth ID watermarking technology, meaning it's nothing you can see in the video, but it allows Google to detect that these videos were generated. A synth ID was released as open source. Technically, now anybody will be able to find out which videos are created with Veo. This should be a very serious catalyst for open AI to release Sora, because this is now Sora domain capabilities. 1080p one minute long is not something we have from any other supplier. It's not surprising to me that Google are able to make that jump because they can train on any domain. Every piece of video from YouTube, which is the largest video repository on the planet. And they obviously have the compute and the human resources to develop this kind of tool. But I'm personally very curious to start playing with this and see what it can produce. Now, in addition to that, they are planning to also release a new version of Imagine 3, their text to image generator to all Google Cloud customers. And the combination, obviously, will allow you to create very specific images, high resolution and detailed to present exactly what you want, and then use that image as a feed into the new VO engine to create videos. Again, very exciting to any creator who wants an additional tool at their fingertips. I want to share with you some exciting news from multiply my company. We just opened the registration for our January AI business transformation course. The current course was sold out. The current course that started in November and is ending this week was sold out. We're not going to launch one in December because of the holidays, but there's another cohort that's opening on January 20th of 2025. This course has transformed hundreds of companies and business leaders that have taken the course since its inception in April of 2023. So if you're looking for ways to start 2025 with the right foot forward, whether for your personal career or your knowledge. Or for the sake of the success of your company, business team, organization, et cetera, come join us. There will be a link in the show notes, so you can open your phone right now or your computer and go and see exactly all the details over there. And now back to the episode. Now, staying on the same topic, but from a different company, Runway, which is one of the leading providers of video capabilities, just launched Frames. So Frames is an image generator built into Runway. They had one before. But this is a completely new model that they've developed from the ground up and it prioritizes stylistic control and visual fidelity. What it basically means, it means that you can generate multiple images with consistent artistic styles, which can be used as the starting point to create videos, which is really important when you're trying to create a consistent video that is generated from multiple images. So right now, the videos that runway can generate are significantly shorter and there are a few seconds each. So in order to create a longer video, you need to start with multiple images and you need these images to look consistent. Otherwise, your video will not look consistent. And this is exactly the capability they have just provided. Now, it's already rolling out gradually to anybody who has access to Gen 3 Alpha. Basically, any paying user of Runway and it will be available through the API as well. Staying on the topic of video generation, Kling has released motion brush. So it allows you to select up to six elements in a single image, paint them with a little brush and point an arrow where you want their motion to go. They also provided camera movements with six types of cameras with camera motion, horizontal pan, vertical pan, zoom, tilt, and roll, which means you can now control both the motion of the camera, as well as motion of elements within the scene. And they released two modes, standard mode with 720p video iteration, that is faster and more cost effective and professional mode that allows you to release 1080p video. HD capabilities. Now I've got to go back to a prediction I made in Q1 this year, and you can go back and check those episodes. But I was saying that by the end of 2025, we will, by the end of 2024, we will be able to create videos that will be indistinguishable between professionally generated videos of real life or cartoons or any other style we would want. And by 2025, we will have full control of our content. Over the cameras, the scenes, the view, the story and everything else in the video. And the first half of the prediction is already correct. Now, these tools are not perfect yet and they still have some morphing and there's still some issues in consistency, but it's getting better and better every day. And now with the release of Veo and potentially Sora as well, before the end of this year, we will have even more advanced capabilities with more consistency. And I think the consistency issues are going to be mostly resolved by the In the near future. And then the full control over camera and scene and action will happen in 2025. The other thing that I think will happen in 2025 is AI editing, meaning you'll be able to go to an existing video, whether a real one that was shot with a camera or an AI generated one and go back and request edits, like take this section out, change the camera pan from this to this and do whatever you want in order to actually edit the video that you already have created to have a lot more control. And if I have to bet, we will have those capabilities before next year is over, which will allow, and we'll probably start seeing full videos created completely with AI. Right now, people are kind of like playing with it, creating short commercials and, 30 seconds videos. But I assume, and I bet that by the end of next year, there's going to be a full featured film created with AI or at least an episode of like 30 minutes of something that would be like a series. What I just said, obviously have profound implications on TV and Hollywood. And it will be very interesting to see where that goes. Another company that made some big announcements in the past few days is Microsoft,. Microsoft launch magentic one, which is an open source multi agent system that is featuring five specialized agents working in concerts. So the idea is to create a multi layer approach where you have an orchestrator, which is the lead agent that's coordinating all the other operations, a web surfer that handles. Browsing to navigate the web and get information from the web file surfer that manages document and file operations, coder, which writes and analyzes code solutions and computer terminal, which executes the codes and provides system operations. So with all of these working together, you can create magical, really advanced, complex processes that are auto run and executed through the orchestrator agent that is going to run everything. Now, they're not the first ones to release something like this. AWS has already released something like this. IBM released B Agent. OpenAI has Swarm. So this multi layer agent control system is something that we're going to see more and more of, and it will allow real advanced development of sophisticated processes that will become easier and easier as these systems will learn our needs, and with simple prompts and instructions, we'll be able to complete very complex tasks. Now, Microsoft also is launching its screen reading AI that they're calling CoPilot Vision. It's something that they demoed before, and it can analyze text and images on web pages in real time. And then you can ask it to provide summaries and translations, and it can help in product discovery in online catalogs. And it offers gaming assistant if you're in the middle of playing a game while you're on a browser, et cetera, et cetera. Basically, it's going to be your a I layer integrated into the edge browser that can see everything in the browser and provide information and assistant to everything that's on that web page right now. It's going to be released in the U. S. Only. And it requires you to be a paid member of copilot with a 20 bro subscription. Now, the good news is that it only works with pre approved popular websites and it cannot do the same thing with like your bank account or stuff like that. So there's some guardrails that has been put in place out of the box for us to use it. the other safety guards is that it deletes all the data that it captured at the end of each session. And there is no storage or model training using the process content that was generated in each of those sessions. So in theory you can use it safely to get assistance on every web page you're on. If you are using the edge browser, Google has already shared that they're releasing something like this in the very near future, probably in the beginning of Q1 of next year for Chrome. This will become a standard thing, regardless of which browser you're using to be able to ask AI about what's on the screen right now. The next big and interesting feature comes from Anthropic. They just added Google Docs support. So You can connect multiple documents in a single chat. But the coolest feature in this is that there's an automatic sync, meaning if you are changing the version of the document, Claude will already know that. That to me was one of the biggest benefits of Gemini. So Gemini allows you to basically look into a folder and use that as the data set for a Gemini conversation without having to actually connect the files or upload them into the chat. This is still the case, meaning. It's still a unique feature of Gemini that allows you to look into a folder, but this new functionality by Claude closes the gap somewhat that allows you to at least see updates when there's an update to a document and re ask questions and get updates without having to re upload the file into Claude. The disadvantage of this tool is that it cannot process images or comments in Google Docs. So it doesn't know how to process that. I would say it doesn't know how to process that yet. Another big news that came from Anthropic this week is a game changing MCP protocol and MCP stands for model context protocol, which is a technical standard that enables seamless connection between AI assistants. And data sources aim tools. So if you think about to create successful, working agents, which is the future of everything AI, you need three things. You need the model itself, the thinking entity, you need access to data. So whatever the sources, the data are, and you need tools, you need access to the internet, a browser, an application, and so on to actually execute the things you need to execute in order to complete the task. And this MCP protocol that Anthropic just open source it's supposed to be an infrastructure that will enable all of that. And because it's now open sourced, anybody can use it and more and more people can contribute. And I actually see that as a very solid step in the right direction of sharing this kind of information across different companies and organizations. What it allows you to do is it allows to connect AI applications with databases and knowledge bases and business intelligence graph sources and then link chatbots to development tools and development environments and so on. So it's basically allows you to connect the three main components seamlessly through this interface. Now, they already have several connectors built into it, like Google Drive and BraveSearch and Slack, and they're planning to connect this to any other known organization data sources. The idea is obviously to break the silos of information. so far, to do stuff like that, you needed to Deploy data lakes and data lake houses, which is in many cases, a very complex and expensive process, and this may or may not replace data lakes, but it definitely provides a smaller footprint, significantly cheaper and less complicated solution for at least some of the company needs that can be deployed very fast using now an open source architecture. Now we both know that Anthropic just recently got another 4 billion investment from Amazon and AWS to, for a total of 8 billion so far. But amazon are not just counting on Anthropic. They're actually developing their own models and they just released a whole family of new models called Nova. And that series of model is going to be available on AWS Bedrock, which is their infrastructure for everything AI. And now, and they currently release three models, Nova micro, which is fast cost effective text model. Then there's Nova light, which is a low cost multimodal model for image, video, and text. And then there's Nova pro, which is advanced multimodal model. And in 2025, they're planning to release Nova premiere, which is advanced reasoning model, basically like GPT 01, Nova canvas, which is image generation tool and Nova real, which will be video generation model with watermarking and they're planning speech to speech and native multi modal models in 2025 as well. So as if we didn't have enough models in the mix, now we have a whole set of new models available directly from Amazon that are available on their bedrock platform, I assume they will make it highly competitive rates for anybody on their platform to give some benefit to people to use their models over other models on bedrock. Now, staying on Amazon and AWS, they released three big tools. In the AWS environment, one is automated reasoning check. So it's a new tool that allows you to check the data to reduce or maybe eliminate hallucinations. It can cross reference both public data as well as customer supplied information that resides in AWS, and there's and it's similar to such offerings that already exists from Microsoft and Google. They also added. Model distillation, which basically allows to take information that was put into one large model and move it into smaller models. The idea is obviously to get higher efficiency and lower cost and higher speed on tasks that do not require the large language model. There's a lot of discussions about this and you heard me talk about this. That's the future where everything is going right. And we talked about this before with delegation and multi agent collaboration, which is the third thing that they have released. So they released a multi agent collaboration tool that enables AI tasks to be distributed, With supervisor agents and allows parallel processing of complex tasks by multiple smaller agents. So this idea of multi layered multi tiered approach is something that's going to become the norm across all the different platforms and we'll allow the main tool to understand what the task and what the needs are, and then orchestrate it across multiple smaller agents that will work in parallel to complete the task faster, but also with more specifically oriented models that will be tailored for specific tasks. Another company that we talked about last week, which is Cohere, Cohere is focusing on enterprise solutions. They released 3. 5, which sets new benchmarks on enterprise search capabilities, its accuracy and speed. So It sees 23. 4 percent improvement over existing hybrid search systems. It's doing 30. 8 percent better than traditional search algorithms on financial data sets, et cetera. And it supports a hundred languages on the data sets, including Arabic, Japanese, and Korean, which were considered to be more complex to query through. So significant benchmark achieved by Cohere in this particular solution. Cohere has selected to not compete in the crazy race for the most advanced models and instead are developing tools that are geared and tailored for enterprise benefits. And this new release is just going to push them even to be more competitive in that particular field. Another company that has made a release. And I told you lots of releases this past few days, 11 labs, which is a company that has been around for a while doing really advanced voice models has made two releases. One of them is a customizable conversational AI agent builder. So what they released is a complete conversational agent building platform. It has multiple LLM options behind the scenes. So you can use Gemini, GPT, and Claude, and you can customize the voice parameters and response characteristics, and you connect it to multiple databases that they already have integrations with. So it comes with a full SDK for Python and JavaScript and React and Swift. So whichever platform you're using, you can connect it to that. And it has the capability to have multiple customizable variables for tone and response length. So you can adjust parameters such as the language selection, the response temperature, the token usage limit. So you can limit the length of each answer, the latency and stability of the model. So you can control how fast it's going to respond and take that into consideration when creating the answers conversation length, et cetera, et cetera. So multiple controllable parameters in there, and they're obviously coming to compete with OpenAI real time conversation API that has been released, in the past few months, basically jumping into their field. So they've been the leader in the voice field for a while. and now both OpenAI and Google has stepped into their field. So this is just them upping their game and providing a complete SDK and controllable capability for agent development. From an availability perspective, this is awesome for anyone who wants to develop any kind of voice agents, whether for internal or external usage, whether for customer service, employee support, whatever you want, these tools will become more and more capable, and there's going to be a whole universe of applications that are going to be developed on top of these capabilities. In addition, 11 Labs introduced an app that literally just competes with the podcast capability of Notebook LM that we showed you several times in the past. So you can now upload PDFs and articles and eBooks and other types of data into it. And it will generate a conversational style podcast between two AI generated hosts. Again, literally directly competing with Notebook LM's podcast capabilities. And the last big release for this week, which is not a release, but more an announcement. Apple is planning to do a major overhaul for Siri to be LLM driven, which will allow it to be a lot more conversational. So that's the exciting news for anybody in the Apple universe that not so exciting news is that they're planning to release it in 2026. So that's at least a year out. more likely more. And that's really disappointing from Apple's perspective. They have been very late to the AI game. What they released so far has been nothing impressive. and even that has been released later than they have suggested initially. I don't really know what's happening with Apple and their AI capabilities, but so far what they're doing is far from impressive and very late behind everybody else. And this is just another example of that. On the other hand, their competitors, Google are releasing more and more advanced Gemini capabilities into Android all ready. Now we're going to stay on rapid fire and now talk about stuff that is not new releases. And the biggest one for this week is that a few very, very interesting individuals are starting a company to create an operating system for agents. So the company has the weirdest name ever, and it's basically called forward slash dev forward slash agents. And they secured a 56 million seed round at a 500 million valuation led by some of the biggest VCs in the world. So you're asking, how is that possible? How can a company that just got founded gets this level of funding? the reason is the level of people and their background. So the CEO is David Singleton. He is the former Stripe CTO and was the Android Wear lead. The CPO is Hugo Barra, the former Android VP and Meta Oculus leader. And the team also includes former executives from Google, Meta, Dropbox and Figma. Now the goal is to create an operating system that will enable a I programs to collaborate on complex multi step tasks based on what I mentioned earlier in this episode. You understand that's the Holy Grail. And if they can do this across multiple tools, multiple platforms and not within a single platform. Platform. So it's not just on AWS. It's not just on Google's tools, et cetera. It can work across all of them. It's obviously very powerful and will provide an infrastructure for basically the future. So when you have people who have created Android, they're Decane, the operating system for most of the phones in the world, saying that they're going to create a new operating system for AI agents you understand the excitement. Now, even the investors have some really big names like Andrej Karpathy, which was one of open AI founders and worked for many years in Tesla, and scale AI CEO, Alexander Wong and Palo Alto networks, CEO, Nickish Aurora. So really big names are in the investor list as well. Definitely a company that we need to pay attention to. And I will keep updating you as I learn new stuff right now. There's very little to know their website basically doesn't say anything. It's very generic and they're in quasi stealth mode, other than saying what they're going to develop. The interesting news is that they're going to deploy it. They're saying early 2025. So the first versions of this are coming very quickly. There's been some big poaching of talent this past week. So open AI just recruited three senior engineers from Google deep mind. And all three are going to work on multimodal development at open AI Zurich's office. As I mentioned earlier, OpenAI is in a crazy race right now between Sora and Dali and image generation and voice and audio that is coming and that is being attacked from all different angles and definitely having three leading engineers from DeepMind will help them stay ahead of the curve. Staying on the topic of OpenAI, to block OpenAI's for profit transition. So we talked about this a lot in the past few weeks. OpenAI are in the process Of switching from a non profit structure to a for profit structure. Elon Musk has donated the first 44 million dollars to get OpenAI off the ground in its early days. And there's a whole history in there that I'm not going to dive into in this episode because it will take another 30 minutes. But what Elon is targeting is OpenAI, Sam Brockman, Microsoft, Reid Hoffman, and d Templeton, and he basically trying to prevent this move. Now, this may have very significant implications on OpenAI because they just raised a huge amount of money and a huge amount of debt for a total of over$10 billion all tied into a promise that they will be able to make that transition. Now, that transition is not easy, period, regardless of being sued by somebody. Because the whole point behind collecting money and building value in a non profit is keeping it in the non profit realm. And I talked about the different implications two weeks ago on the episodes. If you want to dive into that, you can go and check it out. But right now, this will definitely put another roadblock in their path to success. Combine that with the fact that Elon Musk is now buddy With elect president Trump, and he's going to play a role and have the president's ear. And there's a lot of conversations talking about potentially nominating an AI czar. And I assume that it will have at least some influence on the decision on who that is going to be. And the decisions that the new president and the new administration are going to put in place. So this definitely doesn't sound like promising news for OpenAI. Now, obviously, Sam Altman did not stay quiet on this, and he called the potential political interference of OpenAI's growth as profoundly un American. Now, I tend to agree with him, but that being said, when you pick a fight with somebody like Elon Musk, and you are growing the most influential startup in the past decade, but maybe in history, and you want to change it from a nonprofit to a for profit who should expect, that this was coming. It will be very interesting to see how that evolves while Elon is a big name. And now, again, he has the new elect president's ear. There are a lot of people in the open AI corner as well that are big and influential and have very deep pockets. So it will be very interesting to see how this evolves and I'll obviously keep you updated as this moves forward. Now there's some new rumors that open AI may start adding advertising as another way to generate revenue. While this was very clearly denied by Sam Altman in the past and still officially is the stance of the company that no active plans to pursue advertising. CFO Sarah Fryer confirms that OpenAI are exploring advertising possibilities. So that might show up on whatever way in the new AI universe of OpenAI. Open AI also had a new partnership deal this week, this time with future PLC, which is the company behind some of the biggest publications in the world, like Mary Claire, PC gamer, tech radar, Tom's guide, and many other names. There's 200 plus media brands under that umbrella. And there's been a relationship between these two companies before where future has been using more and more of AI's technology, but now it's going to be a two way relationship where OpenAI's GPT platforms will be able to use the data and the information coming from these platforms and to show it as part of the results in OpenAI. That's just another licensing deal that they have in place, like many others that they put in place before. And this is good news for anybody who's using chatGPT to get real time information across multiple domains. An interesting piece of news that is relevant to anyone who is trying to implement AI in their company actually comes from an interview with AWS CEO, That is talking about the shift in enterprise and companies implementation of AI on NWS. So Matt German, the CEO of AWS have shared in the interview that he's seeing a big shift of companies from broad experimentation to focused implementation. And he sees a dramatic change from companies trying hundreds of proof of concepts or fewer applications that drive the highest ROI and just focusing on those as a step one. When I work with my clients, we actually try to do kind of like both in parallel, but we try to guess what are going to be the things that are going to drive the highest ROI to begin with and focus on those while allowing employees to experiment and develop small proof of concepts within very well defined guardrails. And that's actually proves to be very, very Very helpful and very useful to most companies where they can benefit from just a few bigger projects, but also benefit from a lot of small efficiencies that are generated by employees after they get properly trained and provided the right tools and the right guardrails in order to move forward improve their own day to day work. We spoke earlier about competition to notebook LM that now comes from several different directions. Well, notebook LM leadership team has left Google to start their own stealth startup. Now it's unclear what the startup is, but the notebook LM team lead, the notebook LM designer, and one of the top engineers, All three of them have left to have a stealth startup. They haven't said anything yet about what it is other than the plans to focus on consumer facing AI products. That's obviously riding on the success on the work that they've done in Google. Now, we mentioned Elon Musk in several different segments in this episode, but we told you before that XAI raised a lot of money and that they built the largest single location training computer on the planet, which is called Colossus, which is in Memphis, Tennessee, and it has 100, 000 GPUs. We also told you that there's a short term plan to make it into 200, 000. Thousand GPUs from the latest version as well. So they're not just adding 100, 000 GPUs. They're going to be the stronger, newer version of GPUs from NVIDIA, but apparently there's a plan to bring it to 1 million GPUs in total. And that's what they're working on right now as a longer term project. That's the plan. Not something that's easy to achieve from any perspective. Both it means of power supply as well as many things don't scale linearly. So the fact that we're able to build a hundred thousand doesn't mean they can be 200,000. It definitely doesn't mean you can grow it up to a million, but that's the plan they have set themselves to pursue. And just to put things in perspective, when they built the current Colossus computer, which has a hundred thousand GPUs. As I mentioned, It's the largest GPU cluster in the world. They've done this in 122 days, similar, smaller projects have taken nine to 15 months. And so if anybody can pull this off, it's Elon Musk and XAI and the people that they've been working with to create this new, incredibly powerful and incredibly power consuming, new computer. That will definitely overshadow any other computer on the planet from an AI perspective. Staying on the Elon universe, Tesla just unveils their Gen 3 Tesla bot with advanced dexterity capabilities. So they significantly improved the dexterity https: And the eye hand coordination capabilities, and they've demonstrated some very incredible things. The most incredible one was that the robot can catch a tennis ball in midair. So catching something that's moving fast in midair is a very, very complex task to do. And now the robots can do that. But obviously the idea is not that is the fact that it has very high dexterity, which allows it to. Move stuff around, grab different objects in different ways and perform much more delicate tasks than most robots can perform today, which will allow it to be more useful in a lot more tasks, definitely in industrial manufacturing, but also in health care, et cetera, and later on also in household activities. And with a very interesting final, happy and positive piece of news, that stays on the robot realm, John Hopkins and Stanford researchers were able to build a training system for a surgical robot using videos. So they're using the da Vinci surgical system platform, which has been around for many years, but it was always operated by humans to do multiple types of surgery faster. And what they were able to do is they were able to build a training platform that will allow it to watch videos from hundreds of other surgeries. And the robots basically learn from watching these videos. They're now performing very successfully and very accurately. Three different critical surgical procedures. They do needle manipulation, tissue lifting, and suturing all on their own and all without any human intervention and all just by watching videos. So this will obviously allow surgical robots to do a lot of surgeries in a highly accurate way, and they're claiming that they can train new procedures in just a couple of days by allowing the system to watch these videos. This could be transformative to the surgical world if this can be done at scale and I don't see a reason why it couldn't. This could dramatically reduce the cost of surgery, allows to do more surgeries per day with doctors that don't get tired and can literally do this 24 seven without issues across multiple types of surgeries. And over time, this will be probably any kind of surgery, or at least most, which I find as a really important piece of news. That's it for today, as I mentioned, drinking from a fire hose with new capabilities and new features from more or less everybody in the AI industry. Stay tuned for the daily release on 10 a. m. Pacific time for everything OpenAI is going to release this week. If not, just come back next week and I will share with you everything that they have released. Until then, we will be back on Tuesday with really fascinating episode on how to, we're going to compare different leading large language models, but more importantly, we're going to show you how to do this with hunch. Hunch is an incredible AI platform that is flexible and allows you to do magical things, connecting multiple AI platforms together, but in a very easy to use user interface, and we're doing this with the help of hunches CEO. Don't miss this Tuesday's episode and my last thing, as I request every week, if you're a regular listener to this podcast and you find value in it, please open your phone right now, share the podcast with people that you know that can benefit from it and rate and rank us on your favorite platform, whether it's Apple podcast or Spotify, I would really appreciate that. And until next time, have an awesome weekend. Enjoy your time with family and friends, keep on experimenting with AI, share what you learn with the world, because we all have more to learn about it, so we can have the AI revolution help us while reducing the risks as much as possible.