Learning Library

← Back to Library

AI Trends 2024: Reality Check

Key Points

  • 2024 is shaping up as the “reality‑check” year for generative AI, moving from hype‑driven buzz to more measured expectations and widespread integration of AI as co‑pilot features within existing software like Microsoft Office and Adobe Photoshop.
  • Multimodal AI is gaining traction, with models such as GPT‑4V and Google Gemini able to process text, images, and video together, enabling richer interactions like visual‑aided instructions and seamless language‑vision queries.
  • Energy and scalability concerns are prompting a shift toward smaller, more efficient models; while massive models consume electricity equivalent to thousands of households, newer open‑source LLMs are achieving strong performance with billions rather than trillions of parameters.
  • The convergence of these trends—realistic deployment, multimodal capabilities, and resource‑lean architectures—is expected to define how AI is embedded in everyday workflows by the end of 2024.

Full Transcript

# AI Trends 2024: Reality Check **Source:** [https://www.youtube.com/watch?v=sGZ6AlAnULc](https://www.youtube.com/watch?v=sGZ6AlAnULc) **Duration:** 00:09:26 ## Summary - 2024 is shaping up as the “reality‑check” year for generative AI, moving from hype‑driven buzz to more measured expectations and widespread integration of AI as co‑pilot features within existing software like Microsoft Office and Adobe Photoshop. - Multimodal AI is gaining traction, with models such as GPT‑4V and Google Gemini able to process text, images, and video together, enabling richer interactions like visual‑aided instructions and seamless language‑vision queries. - Energy and scalability concerns are prompting a shift toward smaller, more efficient models; while massive models consume electricity equivalent to thousands of households, newer open‑source LLMs are achieving strong performance with billions rather than trillions of parameters. - The convergence of these trends—realistic deployment, multimodal capabilities, and resource‑lean architectures—is expected to define how AI is embedded in everyday workflows by the end of 2024. ## Sections - [00:00:00](https://www.youtube.com/watch?v=sGZ6AlAnULc&t=0s) **2024 AI Reality Check** - The speaker explains that 2024 marks a shift from hype to realistic expectations as generative AI moves from standalone chatbots to integrated co‑pilot features and multimodal models like GPT‑4V and Google Gemini that blend language and vision capabilities. ## Full Transcript
0:00we're a little ways into 2024 now and 0:03the pace of AI certainly isn't slowing 0:04down but where will it be by the end of 0:07the year well we've put together nine 0:10trends that we expect to merge 0:12throughout the year some of them are 0:14Broad and high level some are a bit more 0:16technical so let's get into them oh and 0:19if you stumbled across this video in 0:212025 let us know how we did okay Trend 0:24number one this is the year of the 0:29reality 0:31check it is the year of more realistic 0:36expectations when generative AI first 0:38hit Mass awareness it was met with 0:41breathless news coverage everyone was 0:43messing around with chat GPT darly and 0:45the like and now the dust is settled 0:47we're starting to develop a more refined 0:49understanding of what AI powered 0:51Solutions can do now many generative AI 0:54tools are now being implemented as 0:55integrated elements rather than 0:58Standalone chatbots and like they 1:01enhance and complement existing tools 1:03rather than revolutionize or replace 1:06them so I think co-pilot features in 1:08Microsoft Office or generative fill in 1:10Adobe Photoshop and embedding AI into 1:13everyday workflows like these helps us 1:15to better understand what generative AI 1:18can and cannot do in its current form 1:22and one area generative AI is really 1:25extending its capabilities that is in 1:28multi 1:30model 1:33AI now ai multimodal models can take 1:37multiple layers of data as input and we 1:40already have interdisciplinary models 1:42today like open AI GPT 4V and Google 1:46Gemini that can move freely between 1:48natural language processing and computer 1:50vision tasks so users can for example 1:53like ask about an image and then receive 1:55a natural language answer or they could 1:58ask out loud for instructions to let say 2:00repair something and receive visual aids 2:02alongside step-by-step text instructions 2:06new models are also bringing video into 2:08the fold and where this really gets 2:10interesting is in how multimodal AI 2:13allows for models to process more 2:15diverse data inputs and that expands the 2:17information available for training and 2:19inference for example by ingesting data 2:22captured by video cameras for holistic 2:24learning so there's lots more to come 2:26this 2:27year now Trend three 2:30that relates to smaller 2:35models now massive models they jump 2:37started the generative AI age but 2:39they're not without drawbacks according 2:41to one estimate from the University of 2:43Washington training a single gpt3 size 2:46model requires the yearly electricity 2:48consumption of over a th000 households 2:51and you might be thinking sure that's 2:53training we know that's expensive but 2:55what about inference well a standard day 2:58of chat GPT queries Rivals the daily 3:01energy consumption of something like 3:0433,000 households smaller models 3:07meanwhile are far less resource 3:09intensive much of the ongoing innovation 3:11in llms has focused on yielding greater 3:14output from fewer parameters now GPT 4 3:18that is rumored to have around 3:211.76 trillion parameters but many 3:25open-source models have seen success 3:27with model sizes in the 327 billion 3:30parameter range so billions instead of 3:33trillions now in December last year 3:36mistol released mixol that is a mixture 3:39of experts or Ane model integrating 3:42eight neural networks each with 7 3:44billion parameters and mistol claims 3:46that mixol not only outperforms the 70 3:49billion parameter variant of llama 2 on 3:51most benchmarks at six times faster 3:53influence speeds no less but that it 3:56even matches or outperforms open AI far 3:59larger GPT 3.5 on most standard 4:02benchmarks smaller parameter models can 4:04be run at lower cost and run locally on 4:08many devices like personal laptops which 4:11conversely brings us to Trend number 4:13four which is 4:16GPU and Cloud 4:20costs the trend towards smaller models 4:23is p driven as much by necessity as it 4:26is by entrepreneurial Vigor the larger 4:29the model the higher the requirement on 4:31GPS for training and inference 4:33relatively few AI adopters maintain 4:35their own infrastructure so that puts 4:38upward pressure on cloud costs as 4:40providers update and optimize their own 4:42infrastructure to meet gen demand or 4:45while everybody is scrambling to obtain 4:47the necessary gpus to power the 4:50infrastructure if only these models were 4:52a bit more optimized they need less 4:55compute yes that is Trend number five 5:00that is model 5:03optimization now this past year we've 5:06already seen adoption of techniques for 5:08training tweaking and fine-tuning 5:10pre-train models like quantization you 5:13know how you can reduce the file size of 5:15an audio file or a video file just by 5:17lowering its bit rate well quantization 5:20lowers the Precision used to represent 5:22model data points for example from 16bit 5:25floating point to 8 bit integer to 5:27reduce memory usage and speed up 5:29inference 5:30also rather than directing directly 5:33fine-tuning billions of model parameters 5:36something called Laura or low rank 5:38adaptation entails freezing pre-train 5:41model weights and injecting trainable 5:43layers in each Transformer block and 5:46Laura reduces the number of parameters 5:48that need to be updated which in turn 5:49dramatically speeds up fine tuning and 5:52reduces the memory needed to stor model 5:54updates so expect to see more model 5:56optimization techniques emerge this year 6:00okay let's uh let's knock out a few more 6:03and the next one is all about custom 6:07local 6:09models open-source models afford the 6:12opportunity to develop powerful custom 6:15AI models that means trained on an 6:17organization's proprietary data and 6:19fine-tuned for their specific needs 6:22keeping AI training and inference local 6:25avoids the risk of proprietary data or 6:26sensitive personal information being 6:28used to train closed Source models or 6:31otherwise pass through to the hands of 6:32third parties and then using things like 6:35rag or retrieval augmented generation to 6:38access relevant information rather than 6:40storing all of that information directly 6:42within the llm itself that helps to 6:44reduce model 6:46size Trend number seven that is virtual 6:52agents now that goes beyond the 6:56straightforward customer experience 6:58chatbot because virtual agents relate to 7:02task automation where agents will get 7:04stuff done for you they'll they'll make 7:07reservations or they'll complete 7:09checklist tasks or they'll connect to 7:11other services so lots more to come 7:13there Trend number eight that is all 7:17about 7:19regulation now in December of last year 7:22the European Union reached provisional 7:24agreement on the artificial intelligence 7:26act also the the role of copyright 7:29material in the training of AI models 7:31used for Content generation remains a 7:33hotly contested issue so expect much 7:36more to come in the area of 7:38Regulation and finally we're at Trend 7:41number nine which is the continuance of 7:45something called Shadow 7:49AI what's that well it's The Unofficial 7:53personal use of AI in the workplace by 7:56employees it's about using gen AI 7:59without going through it for approval or 8:01oversight now in one study from Ernest 8:03and Young 90% of respondents said they 8:06used AI at work but without corporate AI 8:09policies in place and importantly 8:12policies that are observed this can lead 8:15to issues regarding security privacy 8:18compliance that sort of thing so for 8:20example an employee might unknowingly 8:22feed trade secrets to a public facing AI 8:25model that continually trains the model 8:27on user input or they might use 8:30copyright protected material to train a 8:32proprietary model and then that could 8:35expose the company to legal action the 8:38dangers of generative AI rise kind of 8:40almost in a linear line with its 8:42capabilities and that Line's going up 8:46with great power comes great 8:48responsibility so so there you have it 8:51nine important AI trends for this year 8:54but but but why nine don't these things 8:57almost always come in tens well yes yes 9:00they do and that's your job what is the 9:04one AI trend for 2024 that we haven't 9:08covered here the missing 10th Trend let 9:13us know in the 9:14comments if you have any questions 9:16please drop us a line below and if you 9:18want to see more videos like this in the 9:20future please like And subscribe thanks 9:23for watching