{"id":77335,"date":"2024-02-19T03:05:44","date_gmt":"2024-02-19T08:05:44","guid":{"rendered":"https:\/\/blog.cyberconservices.com\/?p=77335"},"modified":"2024-02-18T14:26:25","modified_gmt":"2024-02-18T19:26:25","slug":"when-a-i-can-make-a-movie-what-does-video-even-mean","status":"publish","type":"post","link":"https:\/\/blog.cyberconservices.com\/index.php\/2024\/02\/19\/when-a-i-can-make-a-movie-what-does-video-even-mean\/","title":{"rendered":"When A.I. Can Make a Movie, What Does \u201cVideo\u201d Even Mean?"},"content":{"rendered":"<p><span class=\"BaseWrap-sc-gjQpdd BaseText-ewhhUZ BylinePreamble-iJolpQ iUEiRd jslZfG gnILss byline__preamble\">By\u00a0<\/span><a class=\"BaseWrap-sc-gjQpdd BaseText-ewhhUZ BaseLink-eNWuiM BylineLink-gEnFiw iUEiRd ggMZaT cXqSTL eErqIx byline__name-link button\" href=\"https:\/\/www.newyorker.com\/contributors\/joshua-rothman\">Joshua Rothman<\/a> &#8211; For the past couple of weeks, I\u2019ve been making a home video on my phone, using Apple\u2019s iMovie software. The idea is to weave together clips of my family that I\u2019ve taken during the month of February; I plan to keep working on it until March. So far, the movie shows my five-month-old daughter cooing and waving her arms; my five-year-old son chasing me with a snowball; and a visit to the spooky, run-down amusement park in our town, among other things.<\/p>\n<p class=\"paywall\">I thought of my movie while absorbing the announcement, yesterday, of Sora, an astonishing new text-to-video system from\u00a0<a href=\"https:\/\/www.newyorker.com\/magazine\/2023\/12\/11\/the-inside-story-of-microsofts-partnership-with-openai\">OpenAI<\/a>, the makers of\u00a0<a href=\"https:\/\/www.newyorker.com\/tech\/annals-of-technology\/chatgpt-is-a-blurry-jpeg-of-the-web\">ChatGPT<\/a>. Sora can take prompts from users and produce detailed, inventive, and photorealistic one-minute-long videos. OpenAI\u2019s announcement featured many fantastical video clips: an astronaut seemingly marooned on a wintry planet, two pirate ships dueling in a cup of coffee, and \u201chistorical footage of California during the gold rush.\u201d But two other clips were more intimate, the sort of thing that an iPhone might capture. The first was generated by a prompt asking for \u201ca beautiful homemade video showing the people of Lagos, Nigeria in the year 2056.\u201d It \u201ccaptures,\u201d if that\u2019s the word, what seems to be a group of friends, or perhaps relatives, sitting at a table at an outdoor restaurant; the camera pans from a nearby open-air market to a cityscape, which is divided by highways sparkling with cars at dusk. The second shows \u201creflections in the window of a train traveling through the Tokyo suburbs.\u201d It looks like footage any of us might capture on a train; in the glass of the window, you can even see the silhouettes of passengers superimposed on passing buildings. Curiously, none of them seem to be filming.<\/p>\n<p>These videos have flaws. Many have a too-perfect, slightly cartoonish quality. But others seem to capture the texture of real life. The wizardry behind this is too complicated to easily describe; broadly speaking, it might be right to say that Sora does for video what ChatGPT does for writing. OpenAI claims that Sora \u201cunderstands not only what the user has asked for in the prompt, but also how those things exist in the physical world.\u201d In its statistical,\u00a0<a href=\"https:\/\/www.newyorker.com\/science\/annals-of-artificial-intelligence\/what-kind-of-mind-does-chatgpt-have\">mind-adjacent<\/a>, (probably) unconscious way, it grasps how different kinds of objects move in space and time and interact with one another. Sora \u201cmay not understand specific instances of cause and effect,\u201d the developers write\u2014\u201cfor example, a person might take a bite out of a cookie, but afterward, the cookie may not have a bite mark.\u201d And yet the A.I.\u2019s over-all comprehension of the objects and spaces it conjures means that it isn\u2019t just a system for generating video. It\u2019s a step \u201ctowards building general purpose simulators of the physical world.\u201d Sora performs its work not just by manipulating pixels but by conceptualizing three-dimensional scenes that unfold in time. Our own heads probably do something similar; when we picture scenes and places in our mind\u2019s eye, we\u2019re imagining not just how they look but what they are.\u00a0\u00a0<a href=\"https:\/\/www.google.com\/url?rct=j&amp;sa=t&amp;url=https:\/\/www.newyorker.com\/science\/annals-of-artificial-intelligence\/when-ai-can-make-a-movie-what-does-video-even-mean&amp;ct=ga&amp;cd=CAEYBCoUMTA2NDMzMzIwMzIyMTY0MjQzMTQyGjQ3N2UyZTUxNmVmMWQ0Nzk6Y29tOmVuOlVT&amp;usg=AOvVaw1_VgJgNOrwdQfFAoW8km7c\" target=\"_blank\" rel=\"noopener\">Read On:<\/a><\/p>\n<p>&nbsp;<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Sora, the new text-to-video system from OpenAI, doesn\u2019t make recordings\u2014it renders ideas.<\/p>\n <a class=\"more-link\" href=\"https:\/\/blog.cyberconservices.com\/index.php\/2024\/02\/19\/when-a-i-can-make-a-movie-what-does-video-even-mean\/\"><span class=\"more-msg\">Continue reading &rarr;<\/span><\/a>","protected":false},"author":1,"featured_media":77338,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_exactmetrics_skip_tracking":false,"_exactmetrics_sitenote_active":false,"_exactmetrics_sitenote_note":"","_exactmetrics_sitenote_category":0,"advanced_seo_description":"","jetpack_seo_html_title":"","jetpack_seo_noindex":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":true,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2}},"categories":[638,1224],"tags":[639,1225,1278],"class_list":["post-77335","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai","category-chatgpt","tag-ai","tag-chatgpt","tag-openai"],"jetpack_publicize_connections":[],"jetpack_featured_media_url":"https:\/\/blog.cyberconservices.com\/wp-content\/uploads\/2024\/02\/Computer-writing-a-film-script-while-2.png","jetpack-related-posts":[],"jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/posts\/77335","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/comments?post=77335"}],"version-history":[{"count":1,"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/posts\/77335\/revisions"}],"predecessor-version":[{"id":77339,"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/posts\/77335\/revisions\/77339"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/media\/77338"}],"wp:attachment":[{"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/media?parent=77335"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/categories?post=77335"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blog.cyberconservices.com\/index.php\/wp-json\/wp\/v2\/tags?post=77335"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}