As usual for this sort of thing, the results vary all the way from “impressive” to “absolutely horrific.” I will say it seems to work quite well for many of the women in the thread.
I’m curious what the training set was for the model. I’m guessing perhaps a set of Japanese imageboard threads where artists draw manga interpretations of portraits. (These portraits tend to be of photogenic women, generally of east-Asian ethnicity, so it makes sense to see that most of the better examples in the thread seem to fall within that area.)
Bumped into this about a week ago and after quite a few tries it looked pretty good.
Figured to toss a few dollars to the person who created this via the paypal link at their site.
But then paypal told me: "Currently PayPal accounts in China are only able to send payments. This recipient is not eligible to receive funds."
Meh..
Emailed the author to inform them and they mused about using bitcoin, but that the threshold is too high for casual users.
"Sooooo many people here. Thank you for visiting :)
Gotta upgrade my servers.
GPU server is the bottle-neck, but my credict card billing is flying
Sigh ~"
And someone replied "put some ads on there", so the creator hastily slapped some ads on there to help with the server bills, and... this happened it seems.
Me too. It seems that the site is compromised and issues redirects to malware without user interaction. Maybe @dang can consider changing this to an archived version of the site until it’s fixed?
> The selfie dataset contains 46,836 selfie images annotated with 36 different attributes. We only use photos of females as training data and test data. The size of the training dataset is 3400,and that of the test dataset is 100, with the image size of 256 x 256. The size of the training dataset is 3400,and that of the test dataset is 100, with the image size of 256 x 256. For the anime dataset, we havefirstly retrieved 69,926 animation character images from Anime-Planet1. Among those images,27,023 face images are extracted by using an anime-face detector2.
These seem tiny, don't NNs need more samples to achieve decent quality?
The CycleGAN datasets [1] all have less than 10k images. The two largest have 10,345 images and 5,129 images, the rest (like the famous horses2zebras) have less than 3k.
One day people will have computers embedded into their optic and auditory nerves.
They will quite literally see a fantasy world, with filters and augmented reality changing everything they perceive.
Deep fakes and 3D transforms will mean any horny people can see anyone they want to naked. Other people will be able to live in a complete Lord of the Rings fantasy world with dragons, magic, elves and orcs.
...and all of that will be illegal hacks to the official form of the tech, which is supposed to show what the content pushers intended you to see.
The right to run ad-blockers, which would, in passing, enable content alteration like you suggest, would be something we'd actually have to fight for.
Most people wouldn't understand, and would be stuck in a world where they can no longer control what they see or hear, or how what they see or hear gets changed before their brain is allowed to perceive it.
There will be a person on the street corner crying out that the world isn't real... but only the people running hacks and filters will hear her.
I love those science fiction stories where everything is shiny and all the tech is amazing and just works. But, there's no future I can see that goes from here to there.
This concept is very entertaining. The PS3 game Haze and Black Mirror episode Men Against Fire have similar premises. Are there any other works of media exploring this?
One of the more famous ones was the classic game (now, series) Syndicate. The core premise was exactly as described above - a chip that would get wired into the brain to alter all sensory perception. The idea was that the world in the setting had become awful enough that this level of personal surrender was no longer outrageous - people were primarily just altering the perception to be "nicer".
The core plot of the game then assumed that once a critical percentage of people had these, it was a viable tool for criminal organizations to engage in massive campaigns of thought control and overthrow the governments of countries all over the world, essentially replacing states entirely.
Jack Vance's scifi/fantasy classic "The Eyes of the Overworld" starts with a quest which has the protagonist steal magical eye glasses which make the denizens of a village who literally live drenched in shit and mud and eat garbage believe they are lords and ladies living in palaces and eating the most delicious foods.
They know it's not the truth, but for all intents and purposes, while wearing the magical eyeglasses it becomes the truth.
The rest of the book is hilarious and I fully recommend it.
The following instructions are for Linux or macOS. It may work on Windows too, but I'm not very experienced with Windows. No special hardware is required (you can run it on a CPU; no GPU needed).
Install pyenv[0] and pyenv-virtualenv[1]. Clone the repo and set up an environment:
The public sidewalk has no privacy policy, either, and I show my face there all the time. Same goes for the airport, restaurants, every ATM, and every convenience store that takes videos of my face.
Hmm, I just tried this with a few pictures, including one of mine. I forgot about all the "your selfies are being data-mined by the bad guys!" stories.
Ever heard of browser fingerprinting? If this site is doing it, now it has my IP, my browser fingerprint, and my face (as well as some friends' faces...). Search for my face on Facebook and they have my name too. Sell to marketer.
The results: I wasn't impressed with the results, it seemed like I could have done better using some OpenCV and first principle algorithms. The examples they give are excellent, but when trying them they are awful.
The viruses: I didn't see all these viruses people were talking about, the software worked "as intended" for me without the redirects. There was no advertising on the page either... I run with uBlock Origin in Chromium.
The dataset: Where do these images come from? These seem like the sorts of pictures that might be uploaded to some older social media website, like Bebo or something.
2 cents: Personally I would like to try and do some work with a dataset that has a 1:1 mapping of person to character, but that would require artists and quite a bit of money. I was thinking that you could build out such a database for free by randomly matching anime characters and real pictures together and have a human rate closeness - then train a network based on these ratings.
One wonders if sites like this exist to link faces to IP addresses, though. Get someone to upload a selfie from their computer, then you can add it to your database. Share on Facebook and Twitter, and they know your socials. Next time you walk into a store (post-mask era anyway), they find you in the database, and can see if you've visited their website or not, who your friends are, what sort of things you're interested in, etc.
I'm almost surprised that this isn't more common, now that I think about it.
Edit to add: I had to flag this link. The site is now redirecting to malware. I guess people did in fact want that database!
I couldn't try because it has some of the most aggressive ads i've ever seen. After I upload a pic it says my phone needs cleaning and sends me to a cleaning app on the play store. No time to check the results.
oh that's interesting and it produces a little story.
I don't know in the back of my mind if I ever really get into that personal assistant thing and want a wrapper/face around it, this would be interesting. Sadly many of the voices are bad... but there are those services you can sample a voice and get a voice to use for TTS, that could be something. Then licensing issues but yeah and latency from API trips
Put some donation links (PayPal/bitcoin/patreon), try using teir 1 ad companies like Google AdSense and if they don't accept you, try something like native ads (outbrain)
I can understand how such GPU intensive website can be difficult to support for a creator.
I'm actually inclined to say that's less impressive because the model should be good at faces. On the other hand, the model doesn't seem to work when multiple people are in the picture. I'm wondering if somehow the gandhi face was recognized as not being a person in the picture, and was animeitized using a similar but different logic.
I'm guessing the author wanted this to be aimed at people who get the joke and/or say that the project shouldn't be taken too seriously by using a meme in the title.
https://twitter.com/tkasasagi/status/1250427941567094786