Leveraging AI

81 | AI summit at MIT - I took notes for you at Imagination in Action this week and more AI news for week April 20

April 20, 2024 Isar Meitis Season 1 Episode 81
Leveraging AI
81 | AI summit at MIT - I took notes for you at Imagination in Action this week and more AI news for week April 20
Show Notes Transcript

Last week, I had the privilege of attending the Imagination in Action summit at MIT, an experience that was nothing short of exhilarating. As leaders from across the globe presented their cutting-edge advancements, the leap from theoretical AI to real-world applications was not just visible; it was solid.

The summit underscored how essential human collaboration is, despite—or perhaps because of—our deep dive into the digital age.

Meanwhile, here are some of the news highlights of the week:

  • Meta's Llama 3 Model Release: Meta has launched two new versions of its Llama model, boasting significant improvements in reasoning and code generation capabilities. The standout feature is that the 8 billion parameters model surpasses its predecessor's 70 billion model in performance.
  • Open Source AI Tools: These new models are available on major platforms like AWS, Databricks, Google Cloud, and more, emphasizing Meta's commitment to open-source development.
  • Enhanced AI Safety Features: With tools like Llama Guard, Code Shield, and CyberSec Eval, Meta is enhancing the security framework around its AI models.
  • Upcoming Enhancements: Meta plans to introduce features like a longer context window and a multi-modal version with over 400 billion parameters in the coming months.

Let's continue to focus on how AI can serve humanity, pushing the boundaries of what's possible while grounding our efforts in ethical practices. The future is here, and it's a collaborative effort.

#AI #Innovation #MITSummit #TechnologyLeadership #HumanCentricAI


About Leveraging AI

If you’ve enjoyed or benefited from some of the insights of this episode, leave us a five-star review on your favorite podcast platform, and let us know what you learned, found helpful, or liked most about this show!

Hello, and welcome to Leveraging AI, the podcast that shares practical ethical ways to improve efficiency, grow your business and advance your career with AI. This is Isar Metis, your host, and this is a short weekend news episode of the podcast, or it is because this is a very unique episode. I'm recording this in a hotel room in Boston, Massachusetts, after attending the Imagination in Action summit put together by M. I. T. yesterday and the day before the summit was absolutely amazing. The top leaders in A. I. across everything from investment to research to company leadership to technological advancements. Literally the smartest people in AI in the world were going back to back on stage, either as individuals or in panels in several different channels. And I really want to share with you what I've learned in that summit. So we're going to focus on that, but let's start with the biggest news of the week that I just cannot skip. And then we will dive into what happened at the summit. Let's start with the biggest news of the week or of the day. Yesterday, Meta has released Llama 3. It was highly anticipated and it was known that it's coming in the next few weeks. I spoke about this in the last couple of episodes. So the model was released yesterday. Yesterday on thursday while we're actually speaking to Yan Lecun who is the head of ai at meta And he shared with us some of the details, but the details are this they released two models so far one with 8 billion parameters. The other was 70 billion parameters. They are claiming that the 8 billion parameters model Is better than the 70 billion model of Llama two, which is obviously a huge improvement in performance that only gets you to think what does the 70 billion parameter model can do it offers improved capabilities in things such as reasoning and code generation, and it is available basically anywhere you want, they released it as open source as they have with all their previous models, and it's available on AWS and Databricks and Google Cloud and Hugging Face and Kaggle and IBM Watson and Microsoft Azure and NVIDIA NIMH and basically anywhere that you can get access to models. It is available and it is supported on all the major chip manufacturers. Meta also announced that they're prioritizing responsible development. As part of Llama three and they're introducing new ways to keep those safeguards and it includes tools such as Llama guard to code shield and cybersec eval to multiple aspects of layers that is supposed to make the usage of Llama three safer for all of us. Now, if that's not enough, Meta share that they're planning to introduce additional capabilities such as longer context window and an enhanced performance in the next three months. And they're also planning to release a multi modal version with over 400 billion parameters, again, different than the 8 billion and 70 billion that are released right now, and they are all coming in the next few months. So this is the biggest piece of news of this week. Llama has been one of the best open source models so far, and there's no doubt in my mind that Llama3 will take the top off the list in at least the near future. As far as the open source models, it will be very interesting to see its benchmark in the real live world against the top models right now from OpenAI and Cloud. And OpenAI just released GPT 4 Turbo into two all paid ChatGPT users. So they've announced that back in November, ChatGPT 4 Turbo allows to run a much larger context window. So 128, 000 tokens versus 32, 000 tokens of GPT 4. This is the biggest difference. The other differences are significantly lower need for compute. And power to run these models. So basically everybody wins. We get a faster model that is also running a much larger context window. And earth has to pay a smaller fee when it comes to the carbon footprint of this process. Adobe also made some interesting news this week. First of all, they are releasing an AI assistant for the Acrobat software that is the PDF creation and viewing software. It costs$4. 99 a month and it can help you read, understand, and summarize PDF documents. I must admit I don't see a huge benefit in using that because you can take PDF documents and upload them to Cloud or ChatGPT and do all these things over there, but if you are an Adobe user and already using a lot of PDF stuff, that might be worth paying for. The more interesting thing that Adobe released that became viral across every place, if you're consuming any kind of AI content, is AI video capabilities for Adobe Premiere. You can now use these capabilities to extend the video. So you have a video that ends at a certain point and you want to add a few seconds in order to align it with the music or your needs and so on. You can literally generate frames out of thin air that will Seamlessly stitched to the end of the segment you had before you can add and remove Objects from the video you can change the type of clothes people wear in specific scenes And you can generate video from text directly within premiere pro. They're also planning to integrate third party generation tools into premiere pro such as OpenAI Sora and movie generation tools like PikaLabs and Runway. So this entire ecosystem of Adobe Premiere that is probably the most commonly used post production tool in the world today is getting a lot of AI attention. If you haven't watched the videos that show these demos, definitely worth watching. And the last piece of news I'm going to dive into the imagination in action event comes from Microsoft. Microsoft just released a model they call Vasa one V. A. S. A. One, which is a model that can turn a single image into deep fake videos in a really quick turnaround time. Milliseconds. And these are still not very high resolution videos. There are 512 by 512 resolution, but they're running at 40 frames per second, and you can upload a face of a person and upload a audio file a person speaking, and it knows how to create the face motion and the lip syncing in a perfect way. It looks completely realistic. This is similar to tools like Haygen, Only it looks even more realistic than Haygen and it starts just with an image and not a video of you as the spokesperson. This is obviously a really impressive capability that has a lot of useful benefits if you are a content creator for any purpose, whether it's internal or external, but it also comes with very serious fears of generating deepfakes very easily and with literally just an image of a person. So now let's switch and talk about the summit at MIT. The summit was absolutely mind blowing. First of all, in the level and quality of the people that participated, whether in the audience and on stage, literally the smartest people on the planet and the most influential people on the planet when it comes to AI, and the Were there either in person or through a zoom call. I will name just a few Yan Lecun the chief ai scientist for meta was there. Arvind Srinivas, the CEO of Perplexity. ai was there. Hassani, the CEO of Liquid. ai and his team were there. Maheshwari, the CTO of GROQ, was there. That's GROQ with a Q, the chip manufacturer, not GROQ with a K. the AI model by X. AI, Steve Case was there, the founder of AOL was speaking there about the internet revolution and its similarities and differences, then what happens now with AI. Binod Khosla was there, one of the most successful venture capitalists in the tech world today, and the first external investor in open AI was speaking there, and many others, influential people in the AI space and beyond when I say beyond people from the financial world was there speaking about the impacts on the financial world. Michael Wise, the CTO of Universal Pictures was there talking about the impact of AI on entertainment and so on and so forth. Again, the most influential and smartest people on AI on the planet, the amount of AI and computer science PhDs that were in each room were probably around 50 percent of the people at any given point, literally incredible. So now let me share with you, first of all, my biggest stakes and then I'm going to share some of the specific things that some of the specific individuals have shared. So the first thing I want to jump into, which is maybe counterintuitive when we're coming to talk about AI implications is human relationships or the human aspect of the AI revolution. And I shared that in many cases before, and I talk about this a lot in the courses that I teach you when I do consulting, but human relationships, human collaboration and human innovation. are still the most important aspect of everything we do, including business, including the AI revolution. It was very clear that the achievements that are achieved today in AI are all built around collaborations between really smart people. And that anything you want to do in the future, the human relationship is going to be a very important factor of it because a lot of the other stuff, everybody will be able to do pretty damn well, because they will be using AI tools to do these things. So if most of what a business can do, everybody can do. the AI is actually pretty well, the differentiator between one business and the other is going to be the human relationships that these people have, and their ability to collaborate with other people. Another very important aspect of this is what skills are going to be the most relevant in this new future with AI. And in this particular case, I'm going to quote Vinod Khosla that said that the most important skill in this future is going to be adaptability or the ability to learn how to learn because the world is going to change significantly faster than we're used to. The only way to stay updated is to know how to continuously learn new skills and new capabilities so you can adapt to what the new norm is. And I agree with him 100%. This is obviously a very tough question for anybody who has kids on what do you push your kids to learn, whether it's at school or after school activities or in their college or postgrad degrees. And there isn't a clear answer. But again, the general The idea is we will need to learn how to learn and adapt a lot faster as humans. Next thing I want to talk about is the concern about the future of society and humanity with AI and its implications. The good news is that every one of the large companies have their own internal concerns about the potential risks of the things that they're developing. And they're investing a lot of resources and putting a lot of effort and try to reduce those risks that includes relevant teams that are reviewing everything that they're doing that includes developing technology that can safeguard what can and cannot be done with some of these tools. Etc. There's obviously a very big debate between the open source approach of let's open this to everybody so everybody can see what's going on So when people are making changes to the code We can see what they're doing and it's open to the public versus the company that are saying this is too dangerous and we should allow few companies to control this I'm not gonna state what my opinion on this is. all i'm going to say is that really smart people You Are on both sides of that fence. But as I mentioned the good news is Whether it's open source, like Llama 3 that just came out, when we were speaking to Jan Le Coon, he specifically said that they're putting a lot of efforts into making sure that it's used safely and pushing his belief that if it's open source, then everybody can see what's happening and that keeps us safe. Obviously, there are people there from OpenAI and Microsoft and so on who are doing it exactly the opposite way, and they're putting a lot of efforts in keeping those systems safe. We had an entire session with Alexander Alex Madry, the head of preparedness in OpenAI, and he was speaking about the amount of efforts that goes in OpenAI into keeping these models as safe as possible. I must admit that while I'm very happy that these organizations care about these things, the flip side of this is that they're moving very quickly to develop and deploy better and more capable models. And I'm sure that within that race, safety is not the top priority, even if it is a priority. So I truly hope that we're going to see more collaboration between these companies. On the safety aspect of this, and I truly hope that the government and governments all around the world will jump in order to enforce some rules and guidelines and regulations on what AI can and cannot do in order to try to really allow us to enjoy the amazing benefits, which I'm going to talk about shortly that AI can bring while reducing and maybe even eliminating the risks. There was a fascinating session about required breakthroughs in AI in order to make the next step in that the people in the panel were senior people in various aspects of this field, including Raj Agarwal, the general manager of Gen AI at AWS, Dina Katabi, a professor for that in MIT, Alan Chabra, the executive vice president, worldwide partners at MongoDB, And others. They shared some of the mutations and hence the breakthroughs that they're expecting to happen in the near future to overcome these issues. So one of the issues that they mentioned was that rag, which is the ability to extract information from a file. Documents and or information that you have still has limitations, and it's not good enough, even at the enterprise level, not to mention beyond that. There's still hallucinations, even when you're referencing specific data, and that obviously hinders the deployment of that in larger enterprises and across more safeguarded data where accuracy is important, and he's expecting those problems to be resolved in the near future. They spoke about the breakthrough in code generation, whether using copilots or by using natural language in order to code computers. This is something that was mentioned in several different aspects and in several different conversations. the concept that computers will adapt to humans versus the other way around. Meaning instead of us having to learn how to write code, computers will understand our universe through natural language, through voice, through vision capabilities that will be installed on wearable devices, and we will be able to program and interact with computers. Naturally, just like we interact with other people, and the computers will adapt to that. One of the benefits of this is obviously an explosion in innovation, because if any person on the planet can quote unquote create new software that does something specific, it will allow more people to be creative from a code generation perspective, way beyond the amount of people that are doing this today. There are several different companies who are working on solutions like that On basically allowing each and every one of us, any person in the world that has access to the internet to be able to write code and create software. Ash Keller, who is the CEO of Neurosity, mentioned that he's looking forward to a breakthrough that will allow foundation models to understand human brain activity in order to connect with us on a very basic level and in order to help us to create cures for existing brain damage, whether by, whether through age or through various injuries, raj from AWS mentioned something that again, if you've been in this field for a while, you would already know that data cleaning and data access is the biggest problem AI adoption today, data in most companies is messy and it's In multiple places in multiple formats and bringing it all together and cleaning it up in a way that a I can actually use it effectively is 80 percent of the work in most large AI projects. And he's looking for the breakthrough that will allow generative AI or other AI capabilities to do the AI data cleaning on their own, which will save a huge amount of work and will allow us to harness and mine significantly more information a lot faster. This is true, obviously, for society as a whole, but also for each individual company or enterprise that wants to implement AI solutions. They also mentioned a very important point when it comes to the struggle between the wish to run forward faster while keeping everything safe. And when I say everything safe from the model creation perspective, it's obviously keeping us safe from what the model can do. And from a company perspective, that's obviously keeping us safe safeguarding the data that they own, whether it's data or private data or customer data and aligning with regulations of their industry while they're doing so. And almost everybody and in almost every conversation, there was a discussion about the huge demand for compute on this path, the amount of energy that it requires and the amount of money that is spent on computer chips in order to achieve the next level of models and in order to get to AGI. Yann LeCun mentioned that Meta is going to invest$30 billion in buying additional NVIDIA chips in order to train their next huge billion parameter model and the implications is obviously not just money. It's also the amount of energy that these computers require, and hence the amount of carbon footprint that they leave on our planet. There has to be a breakthrough in that. And two different companies that were represented in the event are working in that direction. One of these companies is GROQ, again, G R O Q, and they have developed what they're called the Language Processing Unit, which is a completely new innovative chip design that can run between 10 to a thousand times more efficient the NVIDIA GPUs, which means they are running with significantly less compute to get to similar results. Now in addition to running significantly faster, GROQ also has some Huge benefits from a performance perspective. GROQ chips currently generate significantly more tokens per second than any other platform right now, by a very big spread. They're also generating the first token significantly faster than any other platform. So you're getting very short latencies, which will allow to have much more natural communications with large language models and other AI models that will support what I just said earlier, which is that seamless interaction with computers and AI systems. The other company that is at the tip of the spear of promoting a new kind of AI that will dramatically reduce the need for compute and will be a lot more efficient is Liquid AI. Their entire and leadership team was there in one of the panels. And what they are doing is they're developing a new kind of AI. Of a infrastructure, so they're not using the transformer model like everybody else is using. They've actually developed a new kind of model that they're calling Liquid AI. And Liquid AI is a new architecture that supports explainability and higher efficiency. What do I mean by explainability? You can actually understand what the model is doing and why, versus the black box of the current architecture which allows faster development of new models because you can actually understand what the model is doing. So it provides transparency and significantly higher efficiency. It requires 10 to 20 times less cost and energy to train a model and 10 to 1000 times higher efficiency when at Inference, which is the generation of tokens, basically the responses you're getting out of these models. What they've developed is a developer package and the core models that they're giving to companies in order to develop their own in house AI solutions because these models run significantly more efficiently, companies can run them in their own environment, even on premise computers, and that means that their data and the models themselves never have to leave the company's organization closed loops as it's running today, which obviously provides a significant benefit when it comes to data safety and security. And because it's a lighter model, it can also run on device in the future, including much smaller devices, which will allow AI to be significantly more distributed. This concept was raised by several different people on various panels on the need for balance between on device AI capabilities to cloud AI capabilities, and that the balance in the future is going to be a lot more towards On device capabilities versus cloud capabilities for obvious reasons, a data security, be latency, like a lot of the assistance that are coming are going to be running on our local devices, whether they're going to be wearables or our own home or company computers or our phones, et cetera. One of the more inspiring aspects of this summit was that there's a huge number of really smart people who are working to use AI to solve some of the biggest problems the human race and our planet has, such as curing cancer and other diseases, fighting viruses, supporting the elderly against diseases that they have, such as dementia and others. And they're doing it in incredible ways. To give you an example, jeffrey von Maltzan, who is a brilliant biology scientist who is betting on AI to solve a lot of problems. Him and his team are working on understanding how the actual structure of proteins is building various viruses and so on, so they can build drugs down to the molecular level so they can attack viruses in more advanced ways that we can do today. He gave some astonishing numbers. As an example, he mentioned that 90 to 95 percent of the time we fail to predict Which molecules are going to be a good drug that solves the problem is trying to solve and does not generate really high risks with the help of AI, we should be able to drive the failure number by something he's saying. Let's say we drive it down to 50 percent instead of 90 to 95%. We are changing medicine in the world completely. One of the interesting exercises that they did, and I'm sure I cannot explain it as good as he did, but they took a huge variety of existing drugs and used AI to try to help target the same locations and the same methods that these drugs are addressing now, but not in a way that will infringe on the IP of the companies that have created these existing drugs. And they were able to do this across all the drugs they tried it for. And that process would have taken them a decade to do in traditional methods and took them three months using AI. What he believes, which to me is the most inspiring thing that was said all day, is that in many cases, the AI based research will allow us not to only find new treatments, but to actually find cures to some major diseases that we have today. Surprisingly, there were not a lot of discussions about AGI. It was mentioned several times and actually two of the leading scientists and experts on this field, actually agree that the current models that we have probably cannot achieve AGI. One of them is Stephen Wolfern who has been in this field of computer communication and computer systems for the last 40 years, and he thinks we are not there yet. He believes that these models, while they're very good at guessing next words, are literally doing just that, and they don't really have an understanding of our world, and that language is actually, Easier than what we thought it's going to be and not as easy as understanding how the world around us works, so he thinks That there's a limitation to how large language model can progress and hence he believes that in their current shape We cannot achieve AGI. Another person who has been loud about the inability of large language models to achieve AGI is Yann LeCun. I've shared that with you several times before in the conversation with Yann. He's saying that the current large language models cannot achieve AGI because they do not understand how our world operates. They cannot reason, they cannot plan, and they do not have long term memory. And hence, at Meta, they are taking a new approach to how to teach the next version of models. They are trying to teach models the way babies learn about the world. They're calling this new approach VGEPA. JEPA stands for Joint Embedding Architecture. Doesn't matter what it stands for, but it's a way to allow these models to learn just like babies and animals learn about the world. And in their initial testing, it's achieving very positive results as far as actually understanding how the world around us works versus just Having a perception that they understand it because they can guess words. And he's saying that because these models will understand how the world works They will be able to make predictions about the world and hence will be able to plan Based on real world limitations and capabilities, which will allow us to achieve AGI and do a lot more with these models than what we can do right now. And because we're speaking of Yallacoon, I want to mention the fact that more or less everybody on every single panel there Even the people working for the closed source companies like OpenAI, Microsoft, and so on, have all shared that the open source world of AI is critical to the success of this world, partially because it's allowing significantly wider access to the models and a lot more brain power can go into development. and monitor what the next generation of models can do, but also because of cost. Several different people, including Arvind Srinivas, which is the, who is the CEO and co founder of Perplexity, said that currently the cost aspect of large language models leans very favorably in the For businesses to run open source models as the back end of their platform versus the closed source ones And what he's suggesting is something i'm doing with several of my clients is to test things out across multiple models And then pick the one that gives you the best cost efficiency combined with sufficient results, right? So they don't have to be the best results. They have to be results that are good enough for your needs and then optimized for cost efficiency, which in most cases will end up being an open source model. Another good piece of news is that the entrepreneurial spirit is alive and well, including in the AI world. This. Particular event had over a hundred startup founders. 80 of them started pitching their businesses at 7 a. m. in the morning in front of an audience of a few hundreds of people. And each and every one of them had 90 seconds to present their pitch. So people flew in from all around the country and some of them from overseas in order to give a 90 second pitch about their AI business. It was really inspiring to see and be in that room and learn about the huge Variety of ideas on how AI can be implemented across industries and across different aspects of our day to day life and business. In the conversation with Michael Wise, who is the senior VP and CTO of Universal Pictures, there was a very interesting conversation about the impact of AI tools on the future of entertainment, and specifically the movie industry. And Michael shared that they are using more and more of these AI tools in their work. He does not think it's going to have a significant impact on the future of movies as a whole. He thinks that there's very little risk that these tools can create movies in the quality of Hitchcock or Christopher Nolan. And that the key of the movie is the aspect of storytelling and not just the production of the movie. I must admit that while he obviously knows this industry a lot more than I do, I think he is either not sharing what he really thinks, because he's in charge of a lot of people in that industry, or he is Somewhat wrong. And I think that because of two different aspects. One is most of the movies that are being generated and watched. They are not produced by Hitchcock or Christopher Nolan or Steven Spielberg. They are day to day movies that are blockbusters, of course, multiple language in multiple genres that will be able to be completely generated by AI beginning to end, including Really touching or intriguing human stories completely generated by AI at a fraction of a cost of production of the movies that we see today. If you take it to the next step, the ability to personalize movies to people's specific needs, wishes, wants, atmosphere, moods is also going to be possible. It's obviously a movie that is recorded with an actor and shared on screen will stay more static. So that's aspect number one. And aspect number two, I do think that over time, these models will be able to create extremely touching and sophisticated movies, just like Hitchcock and Spielberg and so on. So my personal belief is that the movie industry as it is today is at a very serious risk, not to mention the fact that the democratization that's going on right now. Of this field will allow people that otherwise will never be known to potentially be the next Hitchcock or Christopher Nolan because they'll be able to produce these movies at home on their own and make huge blockbusters that maybe would never arrive to Hollywood or any other large movie production studio. And I want to end up on a few very interesting notes that came from the conversation with Vinod Khosla, again, one of the most successful venture capitalist investors in the tech world ever. So he mentioned several very interesting points. One, that computers will adapt to humans instead of humans adapting to computers. We talked about that and he's claiming that this will obviously drive a huge explosion in creativity because everybody will be able to create code. He also said that because Of this seamless interaction with computers, there will become just like electricity, meaning it's not something you think about when you turn on the light. You don't think about the fact that there are wire in the walls and pyre production, and power plant somewhere that are creating that power. It's a seamless something that we're doing every single day. And he believes that computers will become the same thing. It's not going to be something that we think about. It's just going to be something running in the background that we will use as infrastructure to everything we are going to do. The outcome of both of these is obviously an explosion in diversity in everything that we're going to do because of complete personalization and democratization of capabilities that currently exist only to the people who have the resources and the knowledge on how to interact effectively with computers. Another interesting point that he raised that many other people mentioned as well, that the vast majority of our interaction with the internet is going to be done through agents, meaning not people going to the internet directly, but actually people interacting with agents that will help them solve various problems and do a lot of work for them. including searching the internet, finding solutions, building plans and more or less everything that we'll need to help us live a better, faster, and more efficient and fulfilling life. He believes we'll get to a day that doctors and teachers and geologists and everything else we need in order to promote the human race will be in abundance, meaning it will be either free or almost free and available to all this generation. He's obviously really inspiring and I really hope we get to that day and he said that he is somewhat excited as well as afraid from the day that GPT version X will start training GPT version X plus one, meaning once these models can start training the next models. On their own, very interesting and scary things can happen all at once. So a quick summary for this unique episode. We are on a rocket that is flying faster and faster. The exact destination is very unclear, including to the people who are adding rocket fuel to this rocket. That being said, it's fascinating. It has some immense implications to everything we do. In life and in business, both positive and negative and hopefully jointly by learning and educating as many people as possible on the implication. we can jointly dramatically reduce the risks while increasing the amazing positive impacts that this technology can have on the human race and the world. And our planet, that's it for this week. I will be back on Tuesday with another fascinating interview that we'll dive into how to use AI in your business. If you enjoy this episode, and if you like this podcast in general, I would really appreciate if you will pull up your phone right now and rate us on your favorite podcasting platform, yes, do it right now. Pull up the phone. I know otherwise you'll forget because I forget to do things like that. So I'm not blaming you or anything, but put up your phone right now. Give us a review on your favorite platform. And if people that can benefit from this podcast, while you have the podcast open, click the share button, share it with them. This is our best way to reach additional people. And this is your way to drive additional AI education to more people. So we can do what I just mentioned earlier, potentially enjoy the benefits and reduce the risks. So that's your way to play a role. and until next time, have an amazing weekend. And I will see you back on Tuesday.