BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//Huntsville AI - ECPv6.8.3//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-WR-CALNAME:Huntsville AI
X-ORIGINAL-URL:https://www.hsv.ai
X-WR-CALDESC:Events for Huntsville AI
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/Chicago
BEGIN:DAYLIGHT
TZOFFSETFROM:-0600
TZOFFSETTO:-0500
TZNAME:CDT
DTSTART:20240310T080000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0500
TZOFFSETTO:-0600
TZNAME:CST
DTSTART:20241103T070000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240221T180000
DTEND;TZID=America/Chicago:20240221T190000
DTSTAMP:20260428T042830
CREATED:20240129T033417Z
LAST-MODIFIED:20240218T190012Z
UID:1548-1708538400-1708542000@www.hsv.ai
SUMMARY:AI Breakthroughs in Video
DESCRIPTION:This week we will be talking about some exciting breakthroughs in video generation and understanding. The main splash was with OpenAI’s announcement of Sora last week\, but there have also been similar releases and updates from Google\, Apple\, and UC Berkeley. There’s a description of each below if you want to read ahead. It will be impossible to cover all of these papers in an hour\, so let me know if you want to do a deeper dive into these at a later meetup. \nFebruary 15 – OpenAI Sora – “Sora can generate videos up to a minute long while maintaining visual quality and adherence to the user’s prompt.” There is no paper linked to the release\, but the post on OpenAI that covers the technical details is really good. \n\nhttps://openai.com/sora\nhttps://openai.com/research/video-generation-models-as-world-simulators\n\n\nFebruary 13 – UC Berkely – Large World Model – this is a multi-model (text and video) model capable of processing 1M tokens. It can generate images and videos as well as describe the content of videos or answer questions about what happened in a video.   \n\nhttps://largeworldmodel.github.io/\nhttps://huggingface.co/LargeWorldModel\nhttps://arxiv.org/abs/2402.08268\n\n\nFebruary 8 – Apple KeyFramer – “A design tool for animating static images (SVGs) with natural language”. This isn’t necessarily a video tool\, but generates CSS code to animate an SVG. \n\nhttps://arxiv.org/abs/2402.06071\nhttps://www.theverge.com/2024/2/14/24072917/apple-keyframer-ai-animation-tool-2d-images-text-descriptions\n\n\nJanuary 23 – Google Lumiere – Text to Video\, Image to Video\, and Stylized Generation. \n\nhttps://lumiere-video.github.io\nhttps://arxiv.org/abs/2401.12945\n\n\nOctober 31\, 2023 – SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Prediction – I’m not sure how to attribute this to a particular organization\, since it appears that several Chinese universities contributed. While not quite as impressive as the other releases above\, it shows a broader move to video generation. \n\nhttps://vchitect.github.io/SEINE-project/\nhttps://arxiv.org/abs/2310.20700\nhttps://github.com/Vchitect/SEINE\n\n\nDetails: \n\nDate – 02/21/2024\nTime – 6-7pm\nLocation – HudsonAlpha\nAddress –  601 Genome Way Northwest\, Huntsville\, AL 35806\nZoom –https://us02web.zoom.us/j/81284789520?pwd=SVk0bGFtR2UrYklaRzRtajNYbUtRZz09
URL:https://www.hsv.ai/event/topic-tbd-2/
LOCATION:HudsonAlpha\, 601 Genome Way Northwest\, Huntsville\, AL\, 35806
ATTACH;FMTTYPE=image/png:https://www.hsv.ai/wp-content/uploads/2024/01/AI-Breakthrough-in-Video.png
END:VEVENT
END:VCALENDAR