Pages

Showing posts with label Gen-3. Show all posts
Showing posts with label Gen-3. Show all posts

Tuesday, August 13, 2024

In Flux


Been doing some thinking over the past two weeks. I spent much of July focused on the current state of AI video and was trying to determine its limitations and capabilities. That meant researching Gen-3, Dream Machine, and Kling. There are others but those are the main three so far. There is great potential. No doubt. But, to get the best results you need to use Image-to-Video, preferably with one or two images. I can see generating between a series of single images in a longer sequence as a thing. Having one at the start and one at the end of a 10-second clip is wicked cool, but I've heard of these tools being able to do 2-3 minute sequences. Imagine adding 120 single images and next thing you know you have a whole scene created without needing to be so tedious with these shorter clips. Not all the models are doing these beginning and end frames yet but they will soon. Oh, and FLUX is giving Midjourney a run for its money as far as realistic images. 

I love all of that. Imagining how it will progress is just as fascinating to me. It's amazing to watch. New use cases are rolling out every few weeks. For instance, you can shoot a video with your camera and take an image from that video to add a VFX sequence that can be edited into the live shot using editing software. After I saw that I realized I could go back and test on old short films I made years ago. I am discovering these brilliant techniques people are coming up with and testing them to determine possible use cases. 

Like many people who have realized over the past year that we have entered a new era-- one that seems likely to change society and the way we live our lives, I have been trying to determine how to pivot. While I am no longer a young man, I still have dreams and aspirations similar to those of that younger version of me who returned from California with a creative fire burning in his eyes 25 years ago. The stories are the key. They reveal the lessons learned along the way, and the possible futures based on the world as it is perceived.  

Give me an hour at a cafe with a cup of coffee, a good book, my phone, a notebook, and a pen and I couldn't be happier. It's a pocket of time when I am free to let my mind wander. While some of my best story ideas happen while I am out for a walk, so many of those ideas are fleshed out at a cafe. The constant change around me, people coming and going, as I sit there observing while looking inwards, making connections, recalling the journey, and trying to predict and plan for what comes next. 

I have given myself till the end of this week to assess AI Video models to see what I could learn and then determine how that might impact me creatively in the near future. I still have another week, and there are likely to be many new use cases to discover, but I feel I have made my mind up already.

I am not in this space to be the guy who is the first to discover new techniques with these GenAI tools. I do not feel obligated to post content every day to keep my engagement metrics up. No, not yet at least. I want to see how others are using these tools to learn from them so that I can tell my stories in new ways. I think we are in a new frontier-- creatively speaking, and I consider myself one of those pioneers. My goal is to create a multimedia company. Or a "media empire" as a friend recently joked after reading my most recent TV series bible. It lays out big plans for the TV series that involve a few new ways of interacting with the content. 

Simply put, the goal is to work with Gen AI tools to be able to do more. Two things came to mind last spring when I began to immerse myself in the Gen AI space: How do I use these tools to help me creatively? And, how do I use this technology to help others? 

I am by no means an altruistic saint who thinks only of helping his fellow man every second of every day. Far from it. We are a screwy species and it is often best to mind our own damn business. But I do come from a long line of educators so maybe genetically that is where it comes from. Anyway, an APP was one of the first things I thought of last spring after sitting down with GPT-4 for a month. I have been researching ever since. 

While I won't go into detail about the APP at this time, it is interesting that other than creating moving and still images to accompany my written words that I thought of creating an APP to help others. The idea just made sense. Even more so now. Not only can I help others with it but I can also help myself as well.

As I was assessing the current state of AI Video tools last week I realized something. 

If I am serious about starting a multimedia company, I can't expect AI video trailers/ short films, or graphic novels to fund the way forward... yet. AI Video has gotten a whole hell of a lot better than it was this time last year. However, it's not easy to tell a substantial story. And while the trailer I am working on means a lot to me, it cannot be my main focus. These tools need to get a lot better. Right now you need to be a patient and persistent puzzle master to piece together a worthwhile 2-minute trailer. You'll need to pay out your ears for all the tools needed to create something special. But it can be done. Within a few months, folks will start creating longer works where they have pieced together using the same methods from shorts to create something special. We'll learn their process and cringe at how difficult it was. And yet that will be the most difficult it will ever be. By this time next year, it will be so much easier to do all of this.

There is a window that has opened for AI Video creators and those like myself who are gradually learning more about it every day. The familiarity with these current tools and the proven results of using them may help during the big run to create content that will likely arise next year once AI Video takes its next big leap forward. That leap should be to provide the ability for these models to take a script from a scene, ask you questions about it to make sure it understands what you are wanting and then generate the scene. Once these models can communicate with us like LLMs do using chain of thought then we will see a massive explosion of AI-empowered storytelling. 

For the time being, AI Video is still too unstable, both with its outputs and the overall process. These tools have only become worth my time since June. Sora was in February, but that doesn't count because we still haven't gotten our hands on it. Again, I am not here to discover all the techniques and share those. The people doing that are amazing and I thank them for what they are doing. Their work will be a road map for all of us. They are the OG pioneers, charting the path forward for the rest of us. 

As far as the APP, I can help people with it while working in the background on the more creative side of things. I want to avoid going the clickbait route where I create disposable content to feed a metric. I prefer substance not only with my creative output but also with the APP. The goal is to provide a service that people need.  I want to create value for others and I fully plan on doing a free version of the APP, which may be all people ever need. And that is great! But, I also see charging a monthly fee for premium services for those who need more than the basic service. 

The decision to focus on the APP is not the one that I wanted to make. If I was calling all the shots, then I would have access to all AI video tools that are being held back for the election. That would mean I might be able to go full-steam ahead on making movies and TV with AI tools. Something that I may be able to do now with animation, which, as I have said before, has more room for error than the life-like AI content. But I am not in the animation mindset yet. Once I transition to the comic book series/ graphic novel then I might be more open to focusing on AI animation. Thinking about that now maybe I should focus on the comic book sooner rather than later. Food for thought. 

My evaluation with one week to go in my AI video assessment period is that with the publicly available tools you can make comics, illustrated novels, commercials, trailers, music videos, short live-action films, and longer animated projects that most people would never know are largely AI-generated. The VFX part of this can't be overlooked. That means those who have been filming live-action sequences but have been strapped financially, can actually do some amazing things right now with AI tools. That is all great but these are not my main creative focus. While this company I am creating will include illustrated novels and comics, these current capabilities are still short of where I would need them to be to create realistic AI TV series and movies. That said, in the meantime, I can focus on all the other things I can create using AI video and audio tools, which is a lot. For me, it is all training for TV and movies, though.  

If I had access to all the tools that are being worked on behind the scenes I would likely have a different take on things in this moment. I like to think I have some idea of what may be in the pipeline, but you never know for sure. The good thing is that it is highly likely that these tools will only get better, and fast. So, it makes sense to focus more of my attention on creating the APP for the next few months. Once the dust settles after the election, it will be the perfect time to shift my main focus back to AI Video. Not that I won't be working on AI Video at all between now and then. No, I just need to prioritize the APP for now to try an make some headway before Fall. 

This time last year, I was thinking that we would be right about where we are with video. A short scene is not a performance, though. Not yet, at least. Consistency and stability are nearly solved and performance will be the next big hurdle. Or at least I think it should be. I believe we may have an AI-animated movie out by the end of the year that will be indistinguishable from a traditionally animated movie.

While I want to be able to do all of these things, I am not attempting to be the first. I want to keep learning about all of it because my goals are more intricate than just trying to be the first to create a proper AI movie. That said, I have thought about what that might be like-- the first AI movie that most people cannot tell was created using only AI tools. It could be a hybrid that actually has some live-action. That seems likely to happen soon, which will raise a lot of eyebrows. And that may open the door for the first all-AI movie that generates enough buzz to create some acceptance and appreciation from the public. The Blair With Project always comes to mind when I think of this. 

That said, these AI tools will continue to improve with each passing week. While my main focus will be on the APP, I will keep working on the trailer for the TV series. I won't be sharing a lot of details about the APP until it is ready for public testing. My initial goal is to have it ready for initial testing by November. Rolling out the APP after the election is a good target. It seems likely that even better tools will become available then. This will allow me to adapt the APP based on any updates before final testing and release. 

I'll keep pushing with the trailer and the illustrated novel series in the background. If I am to create a proper multimedia company I need to have a lot of content. I am also open to doing more with these tools in ways that may not be top of mind at this moment. I may get adept enough with these tools that others may want my assistance with their projects. I could grow fond of creating commercials or fall in love with AI animation. Maybe I decide to create a video game. Who knows? 

The one thing I would stop everything to work on is a new form of storytelling entertainment. If the tools get good enough that I can do all I laid out in the bible for my recent TV series, then that will be my top focus. In reality, I am building towards that anyway. So it is best that I take this step-by-step approach toward the likely inevitability of a more immersive entertainment experience. 

It is a process. A process guided by imagination and fueled by rapid technological change. Embracing it was much easier than expected. Some are vehemently against any use of AI for anything involving creativity. Again, I get it. However, the dreams I have had for over 20 years were stifled in the pre-AI era. My creative visions had remained only partially realized through the written word. The chance to create more with these stories may allow me to fulfill the creative goals that I began setting out for myself when I returned from LA at the turn of the century.

I am an independent artist, and through the years I have grown to value my artistic freedom more than I felt a need to sacrifice it all for someone else's idea of success. I just love to create stories. And with AI I will be able to create all the worlds I've ever imagined while maintaining my artistic autonomy. And that's all that matters. Thanks for reading. 

Tuesday, July 23, 2024

All At Once





I love Time Bandits. A TV series based on the movie is coming out tomorrow on AppleTV, which I didn't realize until I started looking for the GIF. Hope it's not what caused AppleTV to essentially shut down. Anyway, I love the original. So, what does this have to do with anything? 

Today, I got up early and went to the store for a few things. That was when I saw an odd wreck. Some guy ran his car over a curb, between two poles, through a flower bed, and into the stone wall/ sign for an apartment complex. 


Went right between two poles and smashed right into the wall, which you can see in the picture cracked upon impact. The guy was talking to the cops. I'm guessing one of his flip-flops got stuck and he couldn't stop or he was intoxicated. A peculiar accident to see at the beginning of the day. 

When I got home I took out the trash. As I was putting a new liner in the can, a story idea came to me. Ideas sometimes develop over days, weeks, or months. This one came all at once. I normally don't Tweet about ideas, but this one was one of those amazing ones that comes almost fully developed. 


Is it too late to be thinking in terms of Classic Hollywood? 

For a decade, with the rise of streaming, I rarely thought in terms of movies, but stories just decide on their own what they want to be. This story is either an anthology movie, a mini-series, or a holodeck experience.

A what?

By the time I get around to actually writing this story Gen AI tools may make it possible. Nothing wrong with thinking ahead. If this is to be a new frontier we are entering, and it sure seems like it, then we need to think beyond what has been possible up until now. A holodeck or a truly amazing immersive video game-like XR headset experience are what seem the most likely next big steps.

As a storyteller of fictional worlds, I have long wondered what it would be like to tell stories in a gaming format. The worldbuilding on my recent TV series got me thinking about the creation of a gaming world. This was before the era of AI started last year. 

Again, I write a lot of anthologies. I do this because I have a lot of stories brewing at any given time. Most of which have been on the backburners for years. One has been sitting there on simmer for 2 decades. It too is an anthology tale but on the grandest of scales; it feels like the time has to be right, like I have to earn the privilege of writing that story. And I haven't yet. 

Many of the other stories that are waiting their turn upon the stage are stand-alone stories. As time passes while I am working on other projects, these stories will sometimes magically coalesce into something greater than their parts. An Anthology is born. What once could have been three to upwards of a dozen or more, otherwise stand-alone, stories come together like a pod of killer whales ready for whatever the ocean might bring. 

Good grief, I'm mixing metaphors here. What it means is there are certain tales within the fictional world of my current TV series that a gamer may want to explore, especially one of the three interwoven stories in the first season. The same could be said for this new idea I'm so excited about, but also the one that's been sitting on the back burner for 20 years. 

Today's new story harkens back to the early 2000s once again. An era my mind gravitates towards. So many of my beliefs about the world were carved during that period. It was a time when I started to think more like a writer and less like an actor. The new millennium began with so much drama, much more than any of us could have expected. This is likely why the first decade of the 21st century so deeply resonates with me. 

The reason this new story idea triggers my recollection of that time is because it has similarities to both Monarch and Psykosis. The Monarch similarities are because of the story Cipher, which actually echoes back to the 90s when I was studying philosophy and poetry. Therefore, this new story echoes back to those days when I was kicking around Hollywood and Los Feliz. 

Why do so many of these stories always seem to be anthology-type tales? I am not sure I can pin that down. With this current idea, could it be a movie? Yes, of course. Could it be a series? 100%. It really is a collection of stories that make up one story. Sound familiar? An anthology. 

The TV series I am creating a trailer for is the same. However, this new story might be better served as a movie. In fact, even though there are a ton of storylines it might be better to put it all into this one overarching story. I think people will relate to it. A full series might dilute the multi-story potency if not done right, but a mini-series might be another option. I've been surprised by how many movies I once wrote have evolved into miniseries. The specific nature of this new story's overall tale makes it perfect for a future holodeck expedition or a video game. It will be very personal.

The stories I have been comparing it to since this morning are all movies. If it were to be a series, I think it would be better as a limited series than one with multiple seasons. The multiple-story aspect would likely be better served in a more contained format (a movie or a handful of episodes) instead of a sprawling 8 to 14-episode season.

I will probably develop it as a movie and expand if needed depending on how it feels once completed. 

And that's kind of how these things develop. An Idea pops and the next thing you know you're making plans on how to write it. I keep wondering if I will have a normal story idea stir my soul like this one day. But then what is a normal story? A Norman Rockwell painting comes to mind. Hallmark movies and Romantic Comedies too. I'm not sure any of that is normal anymore nearly a quarter of the way into the 21st century.

Nowadays, everything seems to be perceived through a superhero filter, but that's not normal. Normalcy is no longer one view of the world. The most normal story I created over the past decade was not received well. My writing partner thought it was boring, and then I used up one of my last contacts from my time in California with it as well. I thought it was something that it wasn't or had a hard time making it come across. I can't help but think it was too much like an episode in a telenovela. Which was too bad because it was meant to tie into the Monarch universe. Oh well. We go again. 

Okay, it's time to get back to working on the Trailer for the TV series. Need to get some traction on that by the end of the week. I finished the script over the weekend and started working on visuals yesterday. 


I probably should have paid for unlimited generations when I signed up for Runway Gen-3 at the beginning of the month. That way I could just keep generating and generating without concerns for running out of tokens. However, by having spent most of the month focused on getting the script for the trailer right I have also been reviewing the strengths and weaknesses of the model based on the outputs of others. 

We'll see how this week and next week go. Apparently Kling will become available worldwide very soon. I have been reluctant to jump through hoops to get access. It's a Chinese company which gives me a bit of pause but it was the technical hoops I would have to jump through just to use it that has really kept me from using it so far. 

If I can achieve all I want with the trailer using Gen-3, Luma, and Hedra then I will. But Kling may be better than all of those models. We'll see in just over a week what I've got and go from there. And I can't help but wonder if all my new ideas will be geared toward a holodeck or a cinematic videogame. Thanks for reading. 

Monday, July 1, 2024

July




June was interesting. July is already intriguing and it's only the first day of the month. Buckle up! 

I spent much of June waiting for OpenAI and Google to release all of the features they had both pimped out to us in May. OpenAI did come out and say that the Voice model will gradually roll out and that most of us plebs won't see it until the fall. Whether that means after the US presidential election or not who knows. But that was not the only new feature. I also need to try the image creation capabilities they teased. Especially to help me create a graphic novel or illustrated novel. I prefer to use only a few tools to create everything I need for these image-heavy projects. I like Midjourney a lot more for image creation, but I keep hoping that OpenAI will improve either Dall-E 3 or provide a new image generation tool with better quality and more capabilities. Not sure where Google's updates are either. I especially wanted to try the video model Veo and Project Astra. Oh well. I guess this is yet another lesson in how patience is a virtue. 

While the big boys have been overpromising and underdelivering in a timely manner, we now have a few new AI video generation models to fawn over. I touched on this in my last post. However, I have had time to think on things since then. On Friday, Runway started to grant access to more people, namely those in the Creative Partners Program. While I did apply to this last week, I was too late to get access. Hopefully, I'll be allowed to join the CPP program at some point so I can get early access to future tools. After seeing what GEN-3 could do I was thrilled to see  GEN-3 Alpha rolled out to everyone today. I am all signed up and ready to start using these new tools. Perfect timing. Thanks, Runway. 

Over the past year, I was reluctant to use the existing AI video tools, something I have mentioned here several times. The quality was not good enough. My focus for part of the last year had been on AI images. Even my writing plans have been guided by the great quality of AI images and the ease with which they can be created. My main focus after the recent two-month query period for a TV series I created was meant to be on a two-part illustrated novel series and a graphic novel series. Having learned enough about creating AI images, I felt confident I could not only create illustrated novels but also graphic novels. However, with these AI video tools all dropping in the past few weeks, and more still to come, I have been forced to reconsider my immediate plans. 

Ever since last spring, I have had an eye on the AI video space with the thought of diving in once the quality reached a certain point. Sora had me dreaming, but its belated release had me focus on what I could do with AI images. If I had access to Sora in February, I would have created a trailer for the TV series to go with the pitch deck and the series bible I created for my query package. Oh well. 

I knew when I saw those early Sora videos that other companies would start to catch up. And when they did I would pivot some of my time and attention to AI video. While AI images are at a point where I can create what I need for the illustrated novel series and for the graphic novel series, I think those projects have become secondary for the next month. It is time to learn to use these AI video tools. I have been waiting so long to have this type of control over moving images once again.

It is one thing to write a story and have people read it. With a novel and illustrated novels, I still have control over what a reader sees. Whereas when screenwriting I have to rely on countless others to bring my vision to life. With AI video tools I have near total control. I say near because we are still early in the AI video space and these things are not perfect, even if they are incrementally better than what we were seeing before Sora. This reminds me of the kind of control I had back when I was making short films back in the day. Because of that, I will spend a big chunk of time in July focused on AI video and learning all I can about AI audio tools. 

The one thing I have not mentioned much about here is my desire to create an APP. I spoke with the people close to me over the past year about my desire to create it, but I wasn't sure if the APP was something that was needed because I saw others creating somewhat similar APPS or GPTs. However, I think I can make an APP that can help a lot of people and help me learn more about the process of creating an APP. I had considered making a GPT through OpenAI, but I think an APP is a better way to go, even though I will have to do a good bit of research. I think it can help more people in that format than as a GPT. 

GPTs seem to be quickly becoming a thing of the past. Microsoft is doing away with them and there are rumors that OpenAI is not as keen on them as they used to be. I want to keep learning about technology but I also want to create. I will likely lean on AI to help me build the APP, while also learning about the process. I am an artist not a martyr, so I don't mind leaning on AI for not only the image and video side of my new creative process but also some of the technical aspects of creating and launching an APP. I have learned a lot over the past year, but I cannot just sit down and crank out this APP without some guidance. 

So, I am making my main focus of July all about educating myself. Learning about AI video, AI Audio, and APP creation with AI assistance. We'll see if I can learn all that I need in one month. Maybe, maybe not, and it may be that I need to keep at it for another month or two. I'm up for the challenge. In whatever free time I have left, I'll also try and get some work done on the first book in the illustrated novel series and create a few panels in the graphic novel. Busy. Busy. Thanks for reading.