Leveraging AI

63 | ChatGPT introduces a long term memory and insane video creation capabilities, Gemini 1.5 has 1M tokens content window, $349 stylish AI glasses, and many more AI news from the week ending on Feb 16, 2024

February 17, 2024 Isar Meitis Season 1 Episode 63
Leveraging AI
63 | ChatGPT introduces a long term memory and insane video creation capabilities, Gemini 1.5 has 1M tokens content window, $349 stylish AI glasses, and many more AI news from the week ending on Feb 16, 2024
Show Notes Transcript

🎧 What can you do with AI Chats with 5X more memory capabilities? 

In this weekend news special episode, host Isar Metis shares the latest explosive updates across key AI players like OpenAI (ChatGPT), Google, NVIDIA and more.

Topics we discussed:

👥 OpenAI’s ChatGPT to get long-term memory
🔒 Privacy concerns and solutions
🎥 OpenAI reveals Sora - text to ultra-realistic video
💻 ChatGPT suffers major outage
🔮 Andrej Karpathy departs OpenAI
🌟Google’s new Gemini 1.5 model revealed
🤯 Gemini 1.5 matches paid Ultra performance
👓 NVIDIA’s RTX AI chatbot runs locally
😎 Brilliant Labs’ AI glasses look like cool shades

About Leveraging AI

If you’ve enjoyed or benefited from some of the insights of this episode, leave us a five-star review on your favorite podcast platform, and let us know what you learned, found helpful, or liked most about this show!

Isar Meitis:

Hello and welcome to Leveraging ai. Podcast that shares practical, ethical ways to improve efficiency, grow your business, and advance your career with ai. This Isar Metis, your host, and this is a short weekend news episode of the podcast. This episode is brought to you by the AI Business Transformation Course. We've been teaching these courses twice a month. Some of them are open to the public and some of them are booked by specific organizations So if you're an organization that wants to train your people on ai, or if you're an individual that wants to take a course, that will dramatically change your understanding of AI and how to use it within a business context. You can check out the information in the link in the show notes. The courses have been sold out all the way through the end of March, so the next open course starts in the beginning of April, and we would love to see you join us in those courses. Like every week a lot of stuff happened, but this week has really been explosive with really incredible news. So the first big news comes from Open AI and ChatGPT. OpenAI just announced that ChatGPT will have a quote-unquote long-term memory. What this means is that it will remember things in between chats. so far the way ChatGPT works, every chat was a standalone universe where it remembered what happened through the chat information you gave it in documents you've uploaded, etc. But it did not know anything across chats other than the stuff you put in the custom instructions, which you have uploaded and created yourself. If you haven't, then you can go and explore that on how to add custom instructions, then carry across all the different chats. But this new feature will basically learn you. It will start saving different components of things that you write and different responses such as your tone and information about your personal life in your company and the style you wanna write in, and the type of outputs you like and don't like, and so on, and build a profile of you. In order to provide you with better personalized, customized results to everything that you're doing. Now, on one hand, this is amazing, right? It means that every one of us will have its own ChatGPT that is tailored to our needs. The concern is obviously privacy, and what data does it store, and how long does it store it, and so on. To resolve that, OpenAI introduced two different things. One is you can see. Everything that it is saving to its memories, and you can control that. You can actually actively ask it to save stuff, and you can go to the list and then delete anything you don't want it to remember. So that's number one. Number two, they're also introducing a feature that's called Private Chat that is going to be like an incognito window in a browser, meaning it'll not remember anything from that chat. So if you want to have a completely private conversation where Open AI is not gonna save any of your information, you'll be able to do that as well. on the bigger picture, what this tells us is that these AI systems are going to be more and more personalized and customized to our needs, whether it's company needs, tone, style, brand, target, persona, audience, we're writing to type of content we like to create and so on and so forth. This is obviously great news and that comes with obviously some baggage of giving up on personal information. Just like any other platform, you give up information to Google, to Facebook and so on, and it gives you more personalized content. The same thing is going to happen on probably any AI platform, and this is just the first step in that direction. This feature is currently not available to everybody. It's currently rolled out to some free users and some paid users, and only then it's gonna roll out to everybody else and then businesses. But I assume in the near future we are gonna have access to that another huge piece of news from OpenAI this week came with introduction of Sora, which is their text two video AI platform, and it's absolutely mind-blowing. If you are on any social media platform and you haven't seen 20 different videos of examples of this, then you're not following any AI titles in that platform because this has exploded everywhere. It is absolutely incredible. It literally looks the most realistic video. It runs up to 60 seconds and with very good consistency and very high resolution. It's really like nothing I've ever seen before. The most impressive video creation platform I've seen so far, which is not even available to the public is Google's Lumiere, but that, is a research paper and it showed. Good improvement over what we know today from platforms like Runway and Pica, but it's not available. On the other hand, Sora from OpenAI is started to being available to beta users, which again means probably within the next couple of months we are all gonna get access to it. And it's even better than Google's platform. If you haven't seen any of their demos, I will put a link in the show notes to the actual announcement page and you can go and check out their demo videos. It's absolutely amazing. I said that several times on this podcast. I. 2024 is going to be the year of video and audio on ai. It's already pretty good, but by the end of this year, there's zero.in my mind. We'll be able to create videos that will be super realistic, high resolution in probably any length we want. Definitely we stitch them together in a video editing software. And that has amazing benefits from anybody who's a content creator. And it has a lot of question marks when it comes to safety, knowing what's the truth is, or even people in the video creation industry, like people who are photographers, videographers, camera people, light people, actors, etc. Because you'll be able to create videos at a very high resolution without any of that equipment and any of these people. Now, I dunno if those two huge pieces of news from ChatGPT about the memory and Sora is what broke the internet, but it definitely broke ChatGPT. I'm just kidding. I dunno if it's related. But ChatGPT experienced massive outages on February 14th and they lasted for a few hours. There were tens of thousands of people who shared outages across multiple platforms, whether it's on X or on down detector platform. In those 24 hours, OpenAI did not comment on this to explain exactly what happened, but what it shows, it's showing that there's a huge demand for that platform or similar platforms and that is obviously not bulletproof from having issues. So the problem that I see with this is that we will become more and more dependent on these AI solutions, and if they are down, it's going to be basically like the internet is down, because many of our core processes will depend on that. I think companies that start building business solutions around ai. Needs to take this thing into consideration and have some kind of a backup plan that will allow you to continue working if these are down. As an example, don't use just one model, so know how to run on ChatGPT and Claude and Gemini, and potentially an open source model like Mistral, and then you are covered. If your main path is not working, you can fall back to your second or third different options. The last piece of news about open AI on this coming Tuesday, we are releasing an episode about convergence of technologies which one of the main topics that we're going to talk about is AI agents, and one of the most interesting people in the world when it comes to talk about AI agent is Andrej Karpathy. He is a researcher in that field. He's one of the dominant researchers in that field in the world. He is one of the founding members of OpenAI. He then left OpenAI to join Tesla as their director of ai, and he came back to OpenAI in early 2023 to lead. Exactly that field of agent development. And he just announced that he's departing the company immediately. And he said it's, there's no hard feelings. It's all a good departure and that he's going to be pursuing personal projects and see what that develops. So it's not great news for the development of agents with it opening I, even though I'm sure they have a lot of other. Really talented people that can continue his path, but I'm really curious to see what he comes up with, because I'm sure it's gonna be mind-blowing if it's something that is more than what he can do at OpenAI. And now from the. Insane amount of news that came from Open AI to the really exciting news that's coming out of Google. Just last week, Google announced their new advanced model called Gemini Advanced, which is running on their model that's called Gemini Ultra. But let's roll back for a second. At the end of last year, they finally released their new model called Gemini. Which is going to be the infrastructure for everything. AI within Google, they released three different levels. Nano for Mobile Pro that back then was the back end of Bard and they said that they will come up with a new model that is going to be Gemini Ultra. That is going to be their most advanced model that, as I mentioned, they released last week. Well, not a week has passed and Google has announced that they're releasing Gemini 1.5. So everything else that I said so far was version one. Now it's version 1.5, and the first platform that's gonna get this is Gemini. Pro, which is the thing that used to run Bard. But last week they've announced there's no Bard anymore. So let's try to create some order in everything I just said, they still run their Gemini nano on their advanced phones like Pixel eight Pro. There is no more Bard. It's now everything is called Gemini and Gemini. Browser that used to be called Bard, is now running Gemini 1.5 and there's still Gemini Ultra. That is the advanced model that you can pay for 20 bucks a month and be able to use that is their competitor of GPT-4 Turbo. That still runs version one. Now, what is Gemini 1.5 and why is it exciting? First of all, they're saying that Gemini 1.5, which is again the second tier version, it's not their top version. They're saying it's now as good and in some cases better than the Ultra version one. So Gemini Pro 1.5 is in par with Gemini Ultra version one. Using a lot less computing power, meaning it's more scalable and it's gonna cost less money and create less damage to the environment while getting to the same results. The other interesting part about Gemini Pro 1.5 is that it can run a context window of up to 1 million tokens. Now, those of you who don't know what tokens are, the whole large language model concepts run on tokens. It's basically fractions of words. It's about 0.75 words is one token. So a million tokens is about 700 to 750,000 words. To put things in perspective, the largest context windows we have right now is 128,000 tokens in GPT for Turbo and 200,000 tokens in Claude two. So this is five x. The second best context window, but more importantly, if you tried these large context window in either Claude or ChatGPT, you know it's not working great. The more data you put into the context window, the less accurate the results become. And what Google are claiming is that they can run these really, really large context between those with very little reduction in quality. So this is obviously really exciting to anybody who wants to run on really large volumes of data. This is still not available to everybody. I mean, this the large context window and it's only going to be available in their AI studio and Vertex ai. Vertex AI is their API connection and AI studio is their playground, sandbox development environment in which you can test stuff and then turn it into things. You can connect through the API, but. Still, if you wanna be able to experiment with that, you'll have to wait probably a little longer. And then I assume everybody will have access to it through AI Studio, which is a really cool tool that you can go and create and test different things. You can even create stuff like GPTs that exist in open ai. You can create similar things that they're calling them chat prompts that you can build within Google's AI studio. One more thing I wanna say about this is If the free version, which is Gemini Pro 1.5, is as good as their paid version, Gemini Ultra version one. There is zero doubt in my mind that Gemini Ultra 1.5 is coming out in the very near future, otherwise there would be no point for people to pay for the Ultra version. So this is obviously just another step in this crazy race between these most advanced model companies on what's coming next. I can't wait to see what's gonna be coming out of Open AI with GPT five or 4.5. There's rumors about both, but it will be very interesting to see how good that model is. Speaking of GPT-Five. Another just interesting geeky fact, one of the Microsoft executives that has been testing GPT-Five is saying that there's a very high likelihood it will be able to do things that we were not able to do before as humanity. One of the cool examples that he gave is that there's a very. Old written language called Linear A, which has been Undecoded since 3000 years ago. We have examples of this language, but nobody was able to identify it, and he's claiming that GPT-Five has a very good shot of actually understanding how this language works because it's really amazing capability to make sense in sequences and codes and so on. While this is a very geeky, maybe not mainstream usage. It just shows you the power of these tools to make sense in data that we couldn't make sense in before that is available to literally anybody in the world. So it's the ultimate democratization of amazing power. Another very interesting piece of news this week came from another huge player in the AI world, which is Nvidia and Nvidia. We all know them as a company that makes the GPUs. Run basically everything AI right now. That's why their valuation went through the sky in the last two years because everybody's buying their GPUs. This week NVIDIA announced that, they're releasing what they call RTX personal AI chatbot. It's a chatbot that runs locally on RTX 30 and 40 GPUs. So if you have a computer, a PC, and you can buy an RTX GPU, install it, and then you can run. A local chatbot that will have access to all your documents, as well as different links, like YouTube videos that you're watching, and you can quickly search and summarize information across all the information you have on your computer. The huge benefit of that is obviously that you're running it locally and you're not relying on cloud services, which means, A, it's gonna be faster, and B, you're not sharing data with anybody in order to be able to do that. That being said, it still has a lot of bugs from people who started using it. It takes a huge amount of ram and it takes even a bigger amount of storage on your hard drive. But if you have enough RAM and you have enough storage, and you're running a PC and you can run these GPUs, it's an interesting tool to try out. That being said, it does not have any. Context, meaning it does not remember anything between questions, and it has still some accuracy problems, but I think the direction is very, very clear. I don't see that as a long-term viable solution. I have very little doubt that the next operating system, both from Microsoft and from Apple. We'll integrate AI capabilities into the actual operating system of our computers. Meaning external tools like this are not going to be necessary because the operating system itself will most likely be able to know everything that's on the computer and respond to that through some kind of an AI chat, or even beyond that may be voice and video and so on. So, why is this important? It's important because it's showing the transition of NVIDIA from just being a hardware provider to being AI software provider, meaning, drifting into the worlds of Microsoft and Google, and OpenAI, and Anthropic and so on. By the way, at the same time, the opposite process is happening. So Microsoft already announced that they're developing AI chips. Google has announced that they're developing AI chips for their phones. Apple announced the same thing that the next set of iPhones is gonna have AI chips built on board. And obviously there's a huge announcement of Sam Altman that's saying he's trying to raise$7 trillion in order to build AI chips. They'll be able to run the AI of the rest of the world. As a follow up to that, by the way, after Sam came up with his announcement of trying to raise 7 trillion to build these chips, Nvidia, CEO Jensen hung replied by saying that this 7 trillion could buy basically every GPU in the future of the planet. But he's saying that because of advancement in that world and because of economies of scale, he thinks that the world will not need more than 1 trillion to maybe 2 trillion to provide it all the AI computing power it needs. Now, he obviously knows one of two things about developing AI related hardware because they're the leading manufacturers by a very big spread. I would assume he's probably more accurate with his number than Sam Altman, but there's definitely a race going on of hardware manufacturers going into the software world and vice versa. So from all of these huge piece of news to one last cool and fun piece of news. There's been a lot of hype in the last couple of weeks from Apple's Vision Pro, everybody's talking about them. Everybody who has enough money or is a real geek, has bought one and is sharing what he's finding and everybody finds it amazing. I assume it's incredible, from everything I've seen, it's a game changer. I. The only problem with it is heavy and it's big ski goggles that you're probably not gonna wear walking out in the street. Well, a startup called Brilliant Labs just came out with what they call frame glasses. These are AI-driven, really cool-looking regular glasses that you can wear anywhere, and they can project on a relatively small field of view, but still project different information. AI-based on the glasses. it has a microphone that it can listen to getting different commands through OpenAI. It can identify objects through the camera and it provide various information and you can communicate with this through your voice and get voice answers and information projected on the screen. These glasses look like pretty cool glasses. Their round looking looked like John Lennon kind of glasses, and so they're stylish and small and you can wear them anywhere and you can use them to connect with AI capabilities through. Normal day-to-day interface, which I find really attractive. They weigh only 40 grams and they cost only$349. So this is not much more expensive than just high-end sunglasses. It's obviously not similar in any way to. Apple's Vision Pro, but it shows you where this world is going, meaning definitely in the future, I see small glasses with a form factor of normal sunglasses that will be able to provide us similar capabilities to what Vision Pro can do right now. And this is definitely something that I see more and more people wearing in the next few years. If you're enjoying this podcast, please subscribe. Rate the podcast, give us whatever star rating you feel we deserve in your favorite podcasting platform, and share it with other people that you think can benefit from this. This is your way to help additional people learn more about AI and it also helps us and I would really appreciate that. That's it for this week. As I mentioned on Tuesday, we're coming up with a unique episode that is going to talk about the convergence of various technologies, including AI and where is that leading the world we live in, and how much it is going to change that's it for now and have an amazing weekend.