Jump to content

Welcome to FutureTimeline.forum
Register now to gain access to all of our features. Once registered and logged in, you will be able to create topics, post replies to existing threads, give reputation to your fellow members, get your own private messenger, post status updates, manage your profile and so much more. If you already have an account, login here - otherwise create an account for free today!
Photo

On Media Synthesis: An Essay on The Next 15 Years of Creative Automation

deep learning media synthesis deepfakes GANs artificial creativity AI style transfer encryption art artificial neural network

  • Please log in to reply
5 replies to this topic

#1
Yuli Ban

Yuli Ban

    Born Again Singularitarian

  • Moderators
  • PipPipPipPipPipPipPipPipPipPipPip
  • 20,704 posts
  • LocationNew Orleans, LA

One of my favorite childhood memories involves something that technically never happened. When I was ten years old, my waking life revolved around cartoons— flashy, colorful, quirky shows that I could find in convenient thirty-minute blocks on a host of cable channels. This love was so strong that I thought to myself one day, "I can create a cartoon." I'd been writing little nonsense stories and drawing (badly) for years by that point, so it was a no-brainer to my ten-year-old mind that I ought to make something similar to (but better than) what I saw on television.
The logical course of action, then, was to search "How to make a cartoon" on the internet. I saw nothing worth my time that I could easily understand, so I realized the trick to play— I would have to open a text file, type in my description of the cartoon, and then feed it into a Cartoon-a-Tron. Voilà! A 30-minute cartoon!
Now I must add that this was in 2005, which ought to communicate how successful my animation career was.
Two years later, I discovered an animation program at the local Wal-Mart and believed that I had finally found the program I had hitherto been unable to find. When I rode home, I felt triumphant in the knowledge that I was about to become a famous cartoonist. My only worry was whether the disk would have all the voices I wanted preloaded.
I used the program once and have never touched it since. Around that same time, I did research on how cartoons were made— though I was aware some required many drawings, I was not clear on the entire process until I read a fairly detailed book filled with technical industry jargon. The thought of drawing thousands of images of singular characters, let alone entire scenes, sounded excruciating. This did not begin to fully encapsulate what one needed to create a competent piece of animation— from brainstorming, storyboarding, and script editing all the way to vocal takes, music production, auditory standards, post-production editing, union rules, and more, the reality shattered every bit of naïveté I held prior about the 'ease' of creating a single 30-minute cartoon (let alone my cavalcade of concepts coming and going with the seasons).
 
In the most bizarre of twists, my ten and twelve-year-old selves may have been onto something; their only mistake was holding these ideas decades too soon.
In the early 2010s, progress in the field of machine learning began to accelerate exponentially as deep learning went from an obscure little breakthrough to the forefront of data science. Neural networks— once a nonstarter in the field of artificial intelligence— underwent a "grunge moment" and quickly ushered in a sweltering new AI summer which we are still in.
In very short form, neural networks are sequences of large matrix multiples with nonlinear functions used for machine learning, and machine learning is basically statistical gradient modeling. Deep learning involves massive layers of neural networks, parsing through a frankly stupid amount of data to optimize outputs.
As it turns out, deep learning is very competent at certain sub-cognitive tasks— things we recognize as language modeling, conceptual understanding, and image classification. In that regard, it was only a matter of time before we used this tool to generate media. Synthesize it, if you will.

Media synthesis is an umbrella term that includes deepfakes, style transfer, text synthesis, image synthesis, audio manipulation, video generation, text-to-speech, text-to-image, autoparaphrasing, and more.

AI has been used to generate media for quite some time— speech synthesis goes back to the 1950s, Markov chains have stitched together occasionally-quasi-coherent poems and short stories for decades, and Photoshop involves algorithmic changes to preexisting images. If you want to get very figurative, some mechanical automatons from centuries prior could write and do calligraphy.
It wasn't until roughly the 2010s that the nascent field of "media synthesis" truly began to grow thanks to the creation of generative-adversarial networks (first described in the 1990s by Jürgen Schmidhuber). Early successes in this area involved 'DeepDream', an incredibly psychedelic style of image synthesis that bears some resemblance to schizophrenic hallucinations— networks would hallucinate swirling patterns filled with disembodied eyes, doglike faces, and tentacles because they were trained on certain images.
When it came to generating more realistic images, GANs improved rapidly: in 2016, Google's image generation and classification system proved able to create a number of recognizable objects ranging from hotel rooms to pizzas. The next year, image synthesis improved to the point that GANs could create realistic high-definition images.

 

Neural networks weren't figuring out just images— in 2016, UK-based Google DeepMind unveiled WaveNet for the synthesis of realistic audio. Though it was meant for voices, synthesizing audio waves with such high precision means that you can synthesize any sound imaginable, including musical instruments.

And on Valentine's Day, 2019, OpenAI shocked the sci-tech world with the unveiling of GPT-2, a text-synthesis network with billions of parameters that is so powerful, it displays just a hint of some narrowly generalized intelligence— from text alone, it is capable of inferring location, distance, sequence, and more without any specialized programming. The text generated by GPT-2 ranges from typically incoherent all the way to humanlike, but the magical part is how consistently it can synthesize humanlike text (in the form of articles, poems, and short stories). GPT-2 crushes the competition on the Winograd Schema by over seven points— a barely believable leap forward in the state of the art made even more impressive by the fact GPT-2 is a single, rather simple network with no augmentation made by other algorithms. If given such performance enhancements, its score may reach as high as 75%. If the number of parameters for GPT-2 were increased 1,000x over, it very well could synthesize entire coherent novels— that is, stories that are at least 50,000 words in length.

This is more my area of expertise, and I know how difficult it can be to craft a novel or even an novella (which need only be roughly 20,000 words in length). But I am not afraid of my own obsolescence. Far from it. I fashion my identity more as a media creator who merely resorts to writing— drawing, music, animation, directing, etc. is certainly learnable, but I've dedicated myself to writing. My dream has always been to create "content", not necessarily "books" or any one specific form of media.
This is why I've been watching the progress in media synthesis so closely ever since I had an epiphany on the technology in December of 2017.

We speak of automation as following a fairly predictable path: computers get faster, algorithms get smarter, and we program robots to do drudgery— difficult blue-collar jobs that no one wants to do but someone has to do for society to function. In a better world, this would free workers to pursue more intellectual pursuits in the STEM field and entertainment, though there's the chance that this will merely lead to widespread unemployment and necessitate the implementation of a universal basic income. As more white-collar jobs are automated, humans take to creative jobs in greater numbers, bringing about a flourishing of the arts.

 

In truth, the progression of automation will likely unfold in the exact opposite pattern. Media synthesis requires no physical body. Art, objectively, requires a medium by which we can enjoy it— whether that's a canvass, a record, a screen, a marble block, or whathaveyou. The actual artistic labor is mental in nature; the physical labor involves transmitting that art through a medium. Thus, pure software can automate the creation of entertainment with humans needed only as peripheral agents to enjoy this art (or bring the medium to the software).

 

This is not the case with most other jobs. A garbageman does not use a medium of expression in order to pick up trash. Neither does an industrial worker. And while many of these jobs have indeed been automated, there is a limit to how automated they can be with current software. Automation works best when there are no variables. If something goes wrong on an assembly line, we send in a human to fix it because the machines are not able to handle errors or unexpected variables. What's more, physical jobs like this require a physical body— they require robotics. And anyone who has worked in the field of machine learning knows that there is a massive gap between what works heavenly in a simulation and what works in real life due to the exponentially increasing variables in reality that can't be modeled in computers even of the present.

 

To put it another way, in order for blue-collar automation to completely upend the labor market, we require both general-purpose robots (which we technically have) and general AI (which we don't). There will be increasing automation of the industrial and service sectors, sure, but it won't happen quite as quickly as some claim.

 

Conversely, "disembodied" jobs— the creatives and plenty of white-collar work— could be automated away within a decade. It makes sense that the economic elite would promote the opposite belief since this suggests they are the first on the chopping block of obsolescence, but when it comes to the entertainment industry, there is actually an element of danger in how stupendously close we are to great changes and yet how utterly unprepared we are to deal with them.

 

There are essentially two types of art: art for art's sake and art as career. Art for art's sake isn't going away anytime soon and never has been in danger of automation. This, pure expression, will survive. Art as career, however, is doomed. What's more, its doom is impending and imminent. If your plan in life is to make a career out of commissioned art, as a professional musician, voice actor, cover model, pop writer, or asset designer, your field has at most 15 years left. In 2017, I felt this was a liberal prediction and that art-as-career would die perhaps in the latter half of the 21st century. Now, just two years later, I'm beginning to believe I was conservative. We need not to create artificial general intelligence to effectively destroy most of the model, movie, and music industries.

 

Models, especially cover models, might find a dearth of work within a year.

 

Yes, a year. If the industry were technoprogressive, that is. In truth, it will take longer than that. But the technology to completely unemploy most models already exists in a rudimentary form. State-of-the-art image synthesis can generate photorealistic faces with ease—we're merely waiting on the rest of the body at this point. Parameters can be altered, allowing for customization and style transfer between an existing image and a desired style, further giving options to designers. In the very near future, it ought to be possible to feed an image of any clothing item and make someone in a photo "wear" those clothes.

In other words, if I wanted to put Adolf Hitler in a Japanese schoolgirl's clothes for whatever esoteric reason, it wouldn't be impossible for me to do this.

 

And here is where we shift gears for a moment to discuss the more fun side of media synthesis.

 

With sufficiently advanced tools which we might find next decade, it will be possible to take any song you want and remix it anyway you desire. My classic example is taking TLC's "Waterfalls" and turning it into a 1900s-style barbershop quartet. This would could only be accomplished via an algorithm that understood what barbershop music sounds like and knew to keep the lyrics and melody of the original song, swap the genders, transfer the vocal style to a new one, and subtract the original instrumentation. A similar example of mine is taking Witchfinder General's "Friends of Hell" and doing just two things: change the singer into a woman, preferably Coven's Jinx Dawson, and changing a few of the lyrics. No pitch change to the music, meaning everything else has to stay right where it is.
The only way to do this today is to actually cover the songs and hope you do a decent enough job. In the very near future, through a neural manipulation of the music, I could accomplish the same on my computer with just a few textual inputs and prompts. And if I can manipulate music to such a level, surely I needn't mention the potential to generate music through this method. Perhaps you'd love nothing more than to hear Foo Fighters but with Kurt Cobain as vocalist (or co-vocalist), or perhaps you'd love to hear an entirely new Foo Fighters album recorded in the style of the very first record.

 

Another example I like to use is the prospect of the first "computer-generated comic." Not to be confused with a comic using CGI art, the first computer-generated comic will be one created entirely by an algorithm. Or, at least, drawn by algorithm. The human will input text and descriptions, and the computer will do the rest. It could conceivably do so in any art style. I thought this would happen before the first AI-generated animation, but I was wrong— a neural network managed to synthesize short clips of the Flintstones in 2018. Not all of them were great, but they didn't have to be.

 

Very near in the future, I expect there to be "character creator: the game" utilizing a fully customizable GAN-based interface. We'll be able to generate any sort of character we desire in any sort of situation, any pose, any scene, in any style. From there, we'll be able to create any art scene we desire. If we want Byzantine art versions of modern comic books, for example, it will be possible. If you wanted your favorite artist to draw a particular scene they otherwise never would, you could see the result. And you could even overlay style transferring visuals over augmented reality, turning the entire world itself into your own little cartoon or abstract painting.

 

Ten years from now, I will be able to accomplish the very thing my ten-year-old self always wanted: I'll be able to download an auto-animation program and create entire cartoons from scratch. And I'll be able to synthesize the voices— any voice, whether I have a clip or not. I'll be able to synthesize the perfect soundtrack to match it. And the cartoon could be in any art style. It doesn't have to have choppy animation— if I wanted it to have fluidity beyond that of any Disney film, it could be done. And there won't be regulations to follow unless I chose to publicly release that cartoon. I won't have to pay anyone, let alone put down hundreds of thousands of dollars per episode. The worst problem I might have is if this technology isn't open-source (most media synthesizing tools are, via GitHub) and it turns out I have to pay hundreds of thousands of dollars for such tools anyway. This would only happen if the big studios of the entertainment industry bought out every AI researcher on the planet or shut down piracy & open source sites with extreme prejudice by then.

 

But it could also happen willingly in the case said AI researchers don't trust these tools to be used wisely, as OpenAI so controversially chose with GPT-2.
 
Surely you've heard of deepfakes. There is quite a bit of entertainment potential in them, and some are beginning to capitalize on this— who wouldn't want to star in a blockbuster movie or see their crush on a porn star's body? Except that last one isn't technically legal.
And this is just where the problems begin. Deepfakes exist as the tip of the warhead that will end our trust-based society. Despite the existence of image manipulation software, most isn't quite good enough to fool people— it's easier to simply mislabel something and present it as something else (e.g. a mob of Islamic terrorists being labeled American Muslims celebrating 9/11). This will change in the coming years when it becomes easy to recreate reality in your favor.

 

Imagine a phisher using style transferring algorithms to "steal" your mother's voice and then call you asking for your social security number. Someone will be the first. We have no widespread telephone encryption system in place to prevent such a thing because such a thing is so unthinkable to us at the present moment.

 

Deepfakes would be the best at subtly altering things, adding elements that weren't there and you didn't immediately notice at first. But it's also possible for all aspects of media synthesis to erode trust. If you wanted to create events in history, complete with all the "evidence" necessary, there is nothing stopping you. Most probably won't believe you, but some subset will, and that's all you need to start wreaking havoc. At some point, you could pick and choose your own reality. If I had a son and raised him believing that the Beatles were an all-female band— with all live performances and interviews showcasing a female Beatles and all online references referring to them as women— then the inverse, that the Beatles were an all-male band, might very well become "alternate history" to him because how can he confirm otherwise? Someone else might tell him that the Beatles were actually called Long John & the Beat Brothers because that's the reality they chose.

 

This total malleability of reality is a symbol of our increasingly advanced civilization, and it's on the verge of becoming the present. Yet outside of mentions of deepfakes, there has been little dialogue on the possibility in the mainstream. It's still a given that Hollywood will remain relatively unchanged even into the 2040s and 2050s besides "perhaps using robots & holographic actors". It's still a given that you could get rich writing shlocky romance novels on Amazon, become a top 40 pop star, or trust most images and videos because we expect (and, perhaps, we want) the future to be "the present with better gadgets" and not the utterly transformative, cyberdelic era of sturm und drang ahead of us.

 

All my ten-year-old self wants is his cartoon. I'll be happy to give it to him whenever I can.
 
If you want to see more, come visit the subreddit: https://www.reddit.c...MediaSynthesis/


  • Casey, Alislaws, starspawn0 and 1 other like this

And remember my friend, future events such as these will affect you in the future.


#2
funkervogt

funkervogt

    Member

  • Members
  • PipPipPipPipPipPip
  • 788 posts

 

Models, especially cover models, might find a dearth of work within a year.

 

Yes, a year. If the industry were technoprogressive, that is. In truth, it will take longer than that. But the technology to completely unemploy most models already exists in a rudimentary form. State-of-the-art image synthesis can generate photorealistic faces with ease—we're merely waiting on the rest of the body at this point. Parameters can be altered, allowing for customization and style transfer between an existing image and a desired style, further giving options to designers. In the very near future, it ought to be possible to feed an image of any clothing item and make someone in a photo "wear" those clothes.

Many things are technologically possible and would save us money, but we don't do them. I think the same will hold true for replacing all human models with digital images of models. Some jobs in the modeling industry will be lost to hyperrealistic artistry, but certainly not all and perhaps not even most. 

 

Whether it goes one way or another, it's of little consequence since the vast majority of "models" make very little or no money from doing gigs. Their careers are usually low-paid and short, and during them, they live in poverty (e.g. - four "models" sleeping on cots in a studio apartment in New York City) and have to do mundane side jobs (waitress, bartender) to pay their bills. If all modeling jobs were automated, the economy wouldn't even feel a ripple. 


  • Yuli Ban likes this

#3
Casey

Casey

    Member

  • Members
  • PipPipPipPipPipPip
  • 676 posts

Great post. Although media synthesis is something that I've been aware of for a few years now (2016-ish, maybe? I'm sure it's not something that I was aware of as a baby futurist back in 2012 or 2013), it's something that I've only been moderately excited about and has always been on the backburner for me, far as the futurist landscape goes. After reading this thread, though, I'll have to move it higher up the chain of command and start actively paying more attention to it. The malleability of reality is a very interesting concept, and goes well beyond simply allowing regular people to create complex forms of media by themselves (although that by itself is plenty exciting).

 

And yeah, the idea of media synthesis has certainly become more mainstream since 2017 or so. In fact, there's a currently airing anime called 'Carole & Tuesday' where media synthesis is one of the focal points of the story, focusing upon a future where almost all art is created by AIs. There's a few specific details that trigger my futurism autism (of the 'technology should have progressed more than it has' variety), and since only the first episode has aired it's too early to tell whether there will be a Luddite-ish "robots bad, humans good" moral, but I'm appreciative towards the series regardless for bringing the concept of media synthesis a few notches closer to mainstream status.

 

I could make a highly simplified 'obscurity to mainstream' chart that goes something like...

 

1. Technology that is still under wraps and hasn't been revealed to the outside public in any capacity

2. Futurists

3. Nerds (for lack of a better word) that aren't into futurism

4. Regular people. In other words, the technology has become mainstream enough - in terms of either awareness, or actual day to day use - that you can take for granted that, if you bring up the subject with anyone you pass by in the street or come across at the store, they will have heard of it.

5. People that have been stranded on a desert island for decades

 

At the moment, as with many technologies that futurists of the 2010s have been reading about throughout the decade, media synthesis is in a transitory period between 3 and 4. It was pretty common knowledge amongst futurists during the 2014-2016 region, but something that was rarely discussed outside the digital hallways of places like KurzweilAI or FutureTimeline. The concept started to become known amongst anime fans come 2017 thanks to websites that were able to randomly generate anime girls, and DeepFakes brought media synthesis into a sort of pseudo-mainstream status, with farmers in their 60s and 70s who haven't played a videogame since Pac-Man at least somewhat aware of the way things are changing thanks to warnings from Fox News or CNN regarding DeepFakes.

 

(Something that I've enjoyed seeing play out since I became a futurist the December of 2011 has been watching these ideas and subjects percolate throughout society. In 2010 and to a lesser degree 2011, some of the primary focuses of the technological and medical worlds - Virtual Reality, self-driving cars, gene therapy, Deep Learning, and so on - received relatively skimpy amounts of attention even among futurist circles. The main focus was still upon the mainstreaming of things that began their takeoff during the 2000s, like smartphones, streaming services, and GPSes that have been ubiquitous throughout the 2010s. The aforementioned things and a whole slew of others didn't really become a routine part of daily discussion on futurism forums until around 2012 and especially 2013. It was only around 2017 that awareness of these things began to expand beyond niche, underground circles, though it might be 3-5 years before you can take for granted that random men or women at grocery stories in the countryside are aware of them.)

 

starspawn said that GPT-2 might have the data scaled up by 100x to 1000x by the end of this year, and that would be really exciting to see. That's probably an understatement, actually - a coherent novel produced entirely by an AI would probably be the single biggest milestone in the field's entire history. If media synthesis was a nascent technology around 2014 and 2015, and a technology that was advanced enough to be useful in a limited amount of cases during 2017~2018, then a fully coherent novel or novella in 2019 would see the field begin to dip its toes into 'mature technology that has very broad uses' waters. It would be a fantastic coda for the 2010s to begin with AI that was no more impressive than SmarterChild - a dumb, goofy little toy that wasn't good for much more besides giving middle school and high school kids a laugh - and end with an AI-produced novel.


  • eacao, Yuli Ban and waitingforthe2020s like this

#4
starspawn0

starspawn0

    Member

  • Members
  • PipPipPipPipPipPipPip
  • 1,273 posts

I'd say what we will see in the not-too-distant future won't be so much the end of artist, musician, movie-maker, poet, writer, cartoonist, etc. as professions, but rather the end of the need of technical talent.  Top talent will still be appreciated; it's all that talent below the 99.9 percentile that won't be necessary. 

 

E.g. everyone will be able to craft a very high-quality poem to read at a poetry slam, that perfectly fits their experience and mood.  

 

People will want to know who the artist is, still.  They will look them up on social media, scrutinize every post like a stalker.  Completely talentless people who nontheless lead interesting lives will become their targets and become celebrities (who have no chance of becoming one at present).

 

I guess another way to look at it is:  IQ and talent will matter less and less, up to a point.  Grit, taste, and who a person is (the life they've lived, their perspective) will matter a lot more. 

 

It's already happening, in fact -- many celebrity musicians aren't very good natural singers (their songs are autotuned) and don't even write their own songs.  


  • Yuli Ban likes this

#5
Yuli Ban

Yuli Ban

    Born Again Singularitarian

  • Moderators
  • PipPipPipPipPipPipPipPipPipPipPip
  • 20,704 posts
  • LocationNew Orleans, LA

^ That's something I wanted to mention myself. But my rambling had dragged on for about 3,000 words by that point and I had to end it somewhere. 

 

Once more, I've said it somewhere else the exact same sentiment: when creative jobs begin going, it's not going to be the creators and publishers on the dole first. It's going to be the salaried programmer, the keyframe artist, the cover model, the technical writer, the sound mixer, some of the thousands of names you see scrolling on animated feature credits.

 

Take, for instance, the video game industry. If everything develops the way we think it will, then at least 60% of the people involved in creating a triple A video game won't be needed by 2029. There won't be need for so many coders, graphic artists, asset designers, voice actors, musical producers, and more. What you'll be left with is maybe the top 40% of necessary talent— the lead designers, the directors, some of the quality assurance, and maybe a few specialized personnel for tasks that even state-of-the-art AI cannot consistently handle.

 

Likewise, you also see this in animation. Take a bogstandard anime series. Anime, IIRC, costs around $100,000 per episode (3x cheaper than an episode of a Western cartoon), and a lot of that is due to all of the moneysaving techniques that now defines anime (e.g. the limited animation, stock expressions, cookie-cutter character designs, overuse of CGI, obscuring mouths when characters are talking, etc.) A lot of the actual animation process is outsourced to budget studios in the Philippines, and the limited budget means strict deadlines— stricter than you might see in Western shows. Thus, when shows first air on television, they're often very rough looking. It's not until months later with the Blu-Ray releases that they're effectively cleaned up. Animators on these shows are underpaid and forced to work very long hours, so they're rarely bringing their A-game. This is a result of Japan's "Lost Decade", which is when anime budgets really slumped and it turned out that audiences didn't care (factoid: it's not like Westerners were any better; Western cartoons had even less animation than anime back in the Hanna-Barbara Days of the 1960s and '70s). The Great Recession didn't make things any better. The move to digital animation in the 2000s wasn't done because creators loved computerized cartooning— they did it because it was cheaper, quicker, and less labor intensive, even though it removed a lot of the "soul" of the shows they were working on. Of course, salaries didn't increase in tandem. 

 

Now imagine what it might be like ten years hence. Algorithms needn't be paid at all— it's just a matter of upfront costs of the software (if there are any) and the costs of running the computers themselves. I can't be so excited as to say that they'll be able to generate entire shows in ten years, but I imagine they'll be able to automate so much of the process as to only require a skeleton crew for the same budget and quality of the product. 

 

There's one movie I saw recently (after learning about it from RebelTaxi) called Next Gen, and one of the most impressive aspects of it is that it is Pixar quality animation done in Blender, a free program you can download right now. When I learned this, one of the first things I thought of was "What if someone used Blender or an equivalent to automate the production of a movie?" You'd still need a lot of computing power to render all this, but consider that no one could do that 20 years ago even with top of the line computers. Nowadays, individuals with beefy rigs can put together shockingly good art pieces, though it still takes quite a bit of time. 

 

Budgets won't go down; more will be repurposed for marketing.


  • Casey likes this

And remember my friend, future events such as these will affect you in the future.


#6
Yuli Ban

Yuli Ban

    Born Again Singularitarian

  • Moderators
  • PipPipPipPipPipPipPipPipPipPipPip
  • 20,704 posts
  • LocationNew Orleans, LA

Great post. Although media synthesis is something that I've been aware of for a few years now (2016-ish, maybe? I'm sure it's not something that I was aware of as a baby futurist back in 2012 or 2013), it's something that I've only been moderately excited about and has always been on the backburner for me, far as the futurist landscape goes. After reading this thread, though, I'll have to move it higher up the chain of command and start actively paying more attention to it. The malleability of reality is a very interesting concept, and goes well beyond simply allowing regular people to create complex forms of media by themselves (although that by itself is plenty exciting).

 

And yeah, the idea of media synthesis has certainly become more mainstream since 2017 or so. In fact, there's a currently airing anime called 'Carole & Tuesday' where media synthesis is one of the focal points of the story, focusing upon a future where almost all art is created by AIs. There's a few specific details that trigger my futurism autism (of the 'technology should have progressed more than it has' variety), and since only the first episode has aired it's too early to tell whether there will be a Luddite-ish "robots bad, humans good" moral, but I'm appreciative towards the series regardless for bringing the concept of media synthesis a few notches closer to mainstream status.

 

I could make a highly simplified 'obscurity to mainstream' chart that goes something like...

 

1. Technology that is still under wraps and hasn't been revealed to the outside public in any capacity

2. Futurists

3. Nerds (for lack of a better word) that aren't into futurism

4. Regular people. In other words, the technology has become mainstream enough - in terms of either awareness, or actual day to day use - that you can take for granted that, if you bring up the subject with anyone you pass by in the street or come across at the store, they will have heard of it.

5. People that have been stranded on a desert island for decades

 

At the moment, as with many technologies that futurists of the 2010s have been reading about throughout the decade, media synthesis is in a transitory period between 3 and 4. It was pretty common knowledge amongst futurists during the 2014-2016 region, but something that was rarely discussed outside the digital hallways of places like KurzweilAI or FutureTimeline. The concept started to become known amongst anime fans come 2017 thanks to websites that were able to randomly generate anime girls, and DeepFakes brought media synthesis into a sort of pseudo-mainstream status, with farmers in their 60s and 70s who haven't played a videogame since Pac-Man at least somewhat aware of the way things are changing thanks to warnings from Fox News or CNN regarding DeepFakes.

 

(Something that I've enjoyed seeing play out since I became a futurist the December of 2011 has been watching these ideas and subjects percolate throughout society. In 2010 and to a lesser degree 2011, some of the primary focuses of the technological and medical worlds - Virtual Reality, self-driving cars, gene therapy, Deep Learning, and so on - received relatively skimpy amounts of attention even among futurist circles. The main focus was still upon the mainstreaming of things that began their takeoff during the 2000s, like smartphones, streaming services, and GPSes that have been ubiquitous throughout the 2010s. The aforementioned things and a whole slew of others didn't really become a routine part of daily discussion on futurism forums until around 2012 and especially 2013. It was only around 2017 that awareness of these things began to expand beyond niche, underground circles, though it might be 3-5 years before you can take for granted that random men or women at grocery stories in the countryside are aware of them.)

 

starspawn said that GPT-2 might have the data scaled up by 100x to 1000x by the end of this year, and that would be really exciting to see. That's probably an understatement, actually - a coherent novel produced entirely by an AI would probably be the single biggest milestone in the field's entire history. If media synthesis was a nascent technology around 2014 and 2015, and a technology that was advanced enough to be useful in a limited amount of cases during 2017~2018, then a fully coherent novel or novella in 2019 would see the field begin to dip its toes into 'mature technology that has very broad uses' waters. It would be a fantastic coda for the 2010s to begin with AI that was no more impressive than SmarterChild - a dumb, goofy little toy that wasn't good for much more besides giving middle school and high school kids a laugh - and end with an AI-produced novel.

1: I'm sure you heard of it before then. It was very early in the year when I created the dedicated thread on it, and I had already tossed the idea around since around 2013 by that point so I'm sure that at least mentioned it before then. 

 

2: Ah, the history & development of futurology— one of my favorite topics to discuss. It boggles the mind to realize how much has happened in the past ten years. Maybe on the KurzweilAI/MindX forums, you could see people taking a lot of these topics seriously, but even there you had such skewed perceptions of how long it would take before we would see so many technologies we casually discuss today. I remember going on the KAI forums back around 2012, 2013 or so and seeing quite a bit of pessimism that things were happening so "slowly" (though to be fair, that was probably mostly from MissKaioshin). In regards to media synthesis, I can definitely tell you that while I considered the possibility of it being a practical technology as far back as 2013, I definitely did not expect it to mature so quickly. Perhaps "mature" is the wrong word since it's still comparatively an embryo compared to where it will be. But it can't be overstated just how pathetic a lot of information technology was coming into the decade and how massively deep learning assisted it. I recall one particular graph that showed just how big of a leap a deep learning-based system was over another one when it came to image recognition, and that was in 2012.

fvUQK9J.png

 

3: I said it elsewhere: Joe Shmoe and Jane Doe live 10+ years behind futurists. If it's not beer, football, crime shows, or blockbuster movies, they don't really pay attention save in their areas of interest. To them, the world of 2019 really is no different than 1999 but with smartphones and better CGI. They hear "AI creating media" and their minds immediately add 981 years to the calendar. As a result, they're still in the mindset that their kids, grandkids, great-grandkids, and so on will be doing "respectable" jobs that they can understand. There may be some things arising like robot civil rights movements, jet-powered pencils, & space miners going on strike, but the world of tomorrow isn't that different from the world of today. 


  • Casey and waitingforthe2020s like this

And remember my friend, future events such as these will affect you in the future.






Also tagged with one or more of these keywords: deep learning, media synthesis, deepfakes, GANs, artificial creativity, AI, style transfer, encryption, art, artificial neural network

0 user(s) are reading this topic

0 members, 0 guests, 0 anonymous users