Properly exciting times.
I remember when I was but a sprog, wide-eyed in wonder, sitting on my Dad's lap as we watched Next Gen. I don't think anybody back then would have imagined technology to be as advanced as it is now.
Seems like something Apple should have led the way on?
I'm amazed at how forward-thinking the Star Trek series are. It's literally like looking into the future.
I'm watching the Voyager and Enterprise series again now on Netflix. Never get bored of them.
That's getting to core of what we've imagined computers could be I think. There is a very long road ahead still, but progress is being made everyday."We are moving away from a world where people must understand computers to a world in which computers must understand us,"
For years the way has been led by airlines and other call center applications. That's where the money is. I think MS, Google and others picked this up out of a general interest in Deep Learning and even there the voice recognition is mostly icing on the cake-- what they're interested in is understanding the underlying sentence graph.Seems like something Apple should have led the way on?
Your comment has been added to the MR-forum wall of fame.You should meet my ex-wife.
I have what I would consider a very slight speech impediment. If you and I were to sit down and have coffee, within a few moments, you could understand 80% of what I said... Probably closer to 99%.
I've let to have a single word I have uttered properly received by any voice recognition equipment. Not a single word.
So I take all this with a dump truck load of salt.
However it will be a long time before computers can understand the real meaning of what's being said, he cautioned.
Apparently they have good microphones...?
Besides, wasn't Dragon Naturally at 1-2% ten years ago?
The quote I see under Accuracy says --
Bruh...You should meet my ex-wife.
It will be a long time before FELLOW HUMANS understand the real meaning of what’s being said.
Seems like something Apple should have led the way on?
Apparently they have good microphones...?
Besides, wasn't Dragon Naturally at 1-2% ten years ago?
I'm with you on this. I was hoping Siri for Mac would at least be able to eliminate the server requirement.I'm still trying to figure out why Siri has to have everything sent to Apple for decoding.
It took human beings hundreds of thousands of years to come up with written language, let alone to be able to transcribe speech into writing. Computers are outpacing us by a remarkable amount.It's amazing to me how long it has taken us to get computer voice recognition and synthesis to a workable level. We've been working on this for decades, and are just now barely getting passing marks. Image recognition, on the other hand, seems to be moving along much more quickly.
Apparently they have good microphones...?
Besides, wasn't Dragon Naturally at 1-2% ten years ago?
Apparently they have good microphones...?
Besides, wasn't Dragon Naturally at 1-2% ten years ago?
[doublepost=1476913848][/doublepost]That's fantastic! Now if they can do the same for Microsoft support staff that are based in.....other countries....wink, wink....
Researchers at Microsoft claim to have created a new speech recognition technology that transcribes conversational speech as well as a human does (via The Verge).
The system's word error rate is reportedly 5.9 percent, which is about equal to professional transcribers asked to work on the same recordings, according to Microsoft.
"We've reached human parity," said chief speech scientist Xuedong Huang in a statement, calling the milestone "an historic achievement".
To reach the milestone, the team used Microsoft's Computational Network Toolkit, a homegrown system for deep learning that the research team has made available on GitHub via an open source license. The system uses neural network technology that groups similar words together, which allows the models to generalize efficiently from word to word.
The neural networks draw on large amounts of data called training sets to teach the transcribing computers to recognize syntactical patterns in the sounds. Microsoft plans to use the technology in Cortana, its personal voice assistant in Windows and Xbox One, as well as in speech-to-text transcription software.
But the technology still has a long way to go before it can claim to master meaning (semantics) and contextual awareness - key characteristics of everyday language use that need to be grasped for Siri-like personal assistants to process requests and act upon them in a helpful way.
"We are moving away from a world where people must understand computers to a world in which computers must understand us," said Harry Shum, who heads the Microsoft AI Research group. However it will be a long time before computers can understand the real meaning of what's being said, he cautioned. "True artificial intelligence is still on the distant horizon."
Article Link: Microsoft Hails 'Historic Achievement' in Speech Recognition Technology
Great, so now Windows 10 is even better at harvesting your personal conversations and relaying them back home to Microsoft.
When should we expect the update to be pushed, with these Win 10 machines not giving the option to opt out of updates?
....
....and you can stop reading here. As both an Apple and MS customer, I never believe a word MS says on future products until it hits the market. And then it is usually 1/2 as good with 1/3 of the features as the promises.
I really don't think speech is an efficient way of interacting with our machines except in very narrow applications. I don't want to sit in an office of people dictating, I don't want to be on a subway full of people browsing the web by voice.
I don't think the computers are doing this by themselves...It took human beings hundreds of thousands of years to come up with written language, let alone to be able to transcribe speech into writing. Computers are outpacing us by a remarkable amount.
Image recognition is pretty amazing. To see a computer correctly identify a dog in a photo is really cool. But look closer and you'll see that the computer recognizes other objects as dogs that clearly aren't dogs. I see photos of cats, horses, pigs, etc., where the computer calls it a dog. Like speech recognition, image recognition is still a work in process.I don't think the computers are doing this by themselves...
My point though was about the relative rate of development between speech recognition and other technologies. Image recognition is an obvious comparison point, but look at control systems for drones, autonomous vehicles and deep space probes. Some kid can build a Lego machine in his garage to solve the Rubik's cube in seconds. We can find ourselves anywhere on the planet to within a couple feet using satellites.
But that most basic form of communication is still hard to crack.
Deep learning algorithms are now able to not only recognize dogs and flowers, but able to identify which breed of dog and species of flower:Image recognition is pretty amazing. To see a computer correctly identify a dog in a photo is really cool. But look closer and you'll see that the computer recognizes other objects as dogs that clearly aren't dogs. I see photos of cats, horses, pigs, etc., where the computer calls it a dog. Like speech recognition, image recognition is still a work in process.