Ok, not really. But I can imagine that being a headline of some inflammatory “news” article. 🗞️
They’re working to remove implicit, societal gender bias from machine translations in Google Translate by changing the underlying architecture of the machine learning model they use. Basically, the model now produces a masculine and feminine version and then determines which is most likely needed. It appears that in some cases, like translating from the gender-neutral Turkish language, the system will return both versions. ✌️
This is after they announced that all gender pronouns will be removed from Gmail’s Smart Compose feature because it was showing biased tendencies with its recommendations. 📧
It’s early in the process but it appears that they are dedicated to this work and have big dreams. 🔮
This is just the first step toward addressing gender bias in machine-translation systems and reiterates Google’s commitment to fairness in machine learning. In the future, we plan to extend gender-specific translations to more languages and to address non-binary gender in translations.
Apparently researchers at Carnegie Mellon and Facebook’s Reality Lab decided there is nothing to worry about and the method for making them needed to be better. So they give us Recycle-GAN. ♻️
We introduce a data-driven approach for unsupervised video retargeting that translates content from one domain to another while preserving the style native to a domain, i.e., if contents of John Oliver’s speech were to be transferred to Stephen Colbert, then the generated content/speech should be in Stephen Colbert’s style.
Fantastic. Just what we need. A system that transfers content while maintaining stylistic integrity all while not needing a great deal of tweaking/input to make it happen. 😵
Also, why is Facebook helping to make fake content easier to create? Don’t they have enough problems on this front already? 🤔
Big Blue has rolled out a tool called Rosetta that can scan photos for text, extract the text it finds, and then “understand” that text. 👁️🗨️
This is huge as it means the platform can now increase accessibility by reading photos, it can pull out information from photos of menus and street signs, and it can monitor memes and images for destructive written content. And those are just a few examples I’m sure. ♾️
Personally, I’m interested to see how this impacts Facebook’s text content in ad image guidelines. It used to reject any ad that contained more than 20% text based on image size (but used a weird grid-based measuring system). Then it moved to an approach where the more text your image contain the narrower it’s delivery/reach. Facebook’s reason was always that “users preferred images with little to no text”, but I always figured it was more about their inability to automate filtering for content. Users don’t appear to have any issues with text overlays when it comes to organic content. 🖼️
Their post has a bunch of technical details if you want to nerd out. 🤓
I’m super interested in the world of NLP (natural language processing), so the news that performance increased dramatically with ELMo piqued my interest. 💡
The biggest benefit in my eyes is that this method doesn’t require labeled data, which means the world of written word is our oyster. 🐚
Yeah, yeah, word embeddings don’t require labeled data either. ELMo can also learn word meanings at a higher level, which I think means it will have far more impact and a wider range of applications. 📶
ELMo, for example, improves on word embeddings by incorporating more context, looking at language on a scale of sentences rather than words.
Our cuddly, red muppet still picks up the biases we embed in our writings though. So plenty more work to be done. 🛠
Time for Day 2 of Machine Learning Mastery’s 7 day course on Statistics for Machine Learning. The assignment? List three methods that can be used for each descriptive and inferential statistics. 3️⃣✖️2️⃣
Let’s start with descriptive; or, “methods for summarizing raw observations into information that we can understand and share”
Mean – an oldie but goodie. Thinking through examples from my day job: average time on page, average order amount, etc.
Standard deviation – the one that was hammered into me in college stats and econometrics.
Modality – this makes me think of gradient descent and the local vs. global maximum search.
Now on to inferential statistics; or, “methods that aid in quantifying properties of the domain or population from a smaller set of obtained observations called a sample”.
Chi-square – I was to say I used these first two in various econ courses.
I take issue when people seem to think that AI is only AI if it resembles what we’ve been shown in movies. I actually unsubscribed from a podcast when one of the hosts said that none of what is currently being touted as AI counts because it’s essentially not AGI or super AI. #petty 😒
That being said, I think this framework laid out by Ben Evans is pretty spot on: 👌
Indeed, I think one could propose a whole list of unhelpful ways of talking about current developments in machine learning. For example:
Data is the new oil
Google and China (or Facebook, or Amazon, or BAT) have all the data
AI will take all the jobs
And, of course, saying AI itself.
More useful things to talk about, perhaps, might be:
Enabling technology layers
Machine Learning’s current superpower is level of automation that seems almost magical. Of course this also means that products can claim to be AI/ML based but really just be a crazy automation stack. And maybe using some “new” terminology to talk about AI/ML/DL will help lead to constructive conversations and a better informed public instead of turning everything into a discussion about Terminator. 🤖
This might be my favorite part of the post: ❣️
Talking about ML does tend to be a hunt for metaphors, but I prefer the metaphor that this gives you infinite interns, or, perhaps, infinite ten year olds.
ML is amazing, but it isn’t omnipotent or truly intelligent in the way we would probably consider the meaning of that word (at least our limited, ego-driven meaning of it). Yeah, it can be like a superpower, but its superpower is that it’s the quietest assembly of unlimited 10 year olds you’ve ever (legally) put to work. ⚡
I was already excited about the potential or pairing machine learning and all kinds of devices, but this post from Pete has me even more excited. 😆
If you think about it, a lot of Internet of Things devices are pretty dumb. Dumb in that they aren’t very smart, not that they are a dumb product idea. Though there are some of those too. Machine learning can make these devices smart, and who knows what other devices could become possible. Pete has come pretty cool ideas. 💭
The quick “why” is: machine learning can be done on a CPU and use very little energy, this means that not a lot of electricity is required to power a microcontroller running a CPU-based deep learning model. Most of the current AI/ML focus is on the big stuff running on crazy compute clusters, but smart, tiny, cheap devices could be more revolutionary. At least in the short run. ✊
You can also watch his corresponding talk at the recent CogX conference. ▶️