No we aren’t. Lightning to 3.5mm adapter does exist.Are we going to get better codecs than AAC then finally? Apple has killed the headphone jack, so we are stuck with Bluetooth now.
Last edited:
No we aren’t. Lightning to 3.5mm adapter does exist.Are we going to get better codecs than AAC then finally? Apple has killed the headphone jack, so we are stuck with Bluetooth now.
For source, you want highest possible quality, preferably raw, so in case there’s a better format in the future, you can go back to the source master instead of transcoding a lossy version.Thanks, but I prefer to trust my own ears. I listen to streaming music on a headphone system consisting of a Meridian Prime Headphone Amplifier (with Meridian Prime power supply) and a pair of Sennheiser 800S headphones. I've certainly heard 96/24 files that sounded better than their 44.1/16 versions. I can't tell the difference between 192/24 and 96/24. And of course I've heard 44.1/16 that sound fantasic. My opinion, and it's one I'm sticking by, is that the difference between CD quality and 96/24 is noticeable with good equipment and material that was well recorded and mixed - particularly acoustic material. It's also my opinion that the difference isn't so great that a poorly mastered hi res audio file will necessarily sound better than a well mastered Redbook file. But at the end of the day, Apple and other streaming music services ought to strive to make sound quality as good as possible. And there's simply no reason to limit music streaming to mp3 quality, or even CD quality, when there's plenty of bandwidth and storage is cheap.
And if Apple didn't think that hi res audio sounded better, why would they ask the record lablels to supply them with hi res digital masters? They've been doing that since the "mastered for iTunes" program began. Apple knows that getting the best sounding mp3 file means starting with the best sounding master. But I don't think Apple can credibly claim that the mp3 files are indistinguishable from the master. Personally, I'd like to hear the master rather than an mp3 approximation.
I wouldn't be so sure. The lightning jack won't be around forever. I wouldn't be surprised if future iPhone models dispensed with connector ports entirely and went completely wireless. Much like Apple's stubborn commitment to inferior mp3 quality, Apple's refusal to embrace more advanced Bluetooth codecs like AptX HD and LDAC, or offer their own version, is frustrating. Perhaps someone will develop an ultra low power wifi chip that can go into headphones. All I know is that we can do a lot better than the status quo.No we aren’t. Lightning to 3.5mm adapter does exist.
Thanks, but I prefer to trust my own ears. I listen to streaming music on a headphone system consisting of a Meridian Prime Headphone Amplifier (with Meridian Prime power supply) and a pair of Sennheiser 800S headphones. I've certainly heard 96/24 files that sounded better than their 44.1/16 versions. I can't tell the difference between 192/24 and 96/24. And of course I've heard 44.1/16 that sound fantasic. My opinion, and it's one I'm sticking by, is that the difference between CD quality and 96/24 is noticeable with good equipment and material that was well recorded and mixed - particularly acoustic material. It's also my opinion that the difference isn't so great that a poorly mastered hi res audio file will necessarily sound better than a well mastered Redbook file. But at the end of the day, Apple and other streaming music services ought to strive to make sound quality as good as possible. And there's simply no reason to limit music streaming to mp3 quality, or even CD quality, when there's plenty of bandwidth and storage is cheap.
And if Apple didn't think that hi res audio sounded better, why would they ask the record lablels to supply them with hi res digital masters? They've been doing that since the "mastered for iTunes" program began. Apple knows that getting the best sounding mp3 file means starting with the best sounding master. But I don't think Apple can credibly claim that the mp3 files are indistinguishable from the master. Personally, I'd like to hear the master rather than an mp3 approximation.
At medium volumes agreed. Lossless really shines at loud volumes on high quality equipment. Lossy washes out at high volumes.Studies have proven you right time and again. Anyone can say they “can tell” a difference, but the testing shows most people can’t. It takes the right equipment in the right settings, with the right listener, and even then the numbers of people who can tell a difference are low.
I also have the Sennheiser HD 800S, connected to a Schiit Magnius amplifier connected to a Hegel HD12 DAC. Balanced cables between them all.
I love the headphones, and I decided to get the amp as the Hegel HD12 isn't quite powerful enough to drive the Sennheiser HD 800S, in my opinion. The only way to get fully balanced cables end-to-end is by connecting the Hegel HD12 to a balanced amplifier.
I have listened to 256kbps AAC, 1411kbps ALAC, Tidal MQA 24bit/96kHz and some music from Blu-Rays at 24bit/192kHz I would be hard-pressed to separate them in a blind test.
Going from running unbalanced using my Hegel HD12 to fully balanced by tossing the Schiit Magnius into the mix made for a much bigger change than going from 256kbps ALAC to 24bit/192kHz tracks.
And this is pretty much being backed up by almost every unbiased blind test out there. Even on extremely high-end and expensive equipment, next to no one can separate the source material when you go past 16bit/44kHz 200 kbps VBR if decent codecs and encoding have been done from decent masters.
And the funny thing about hearing is that it's degrading with age. By the time you reach a point in life where you have the money to spend on great equipment, your hearing will most likely have degraded to a point where you are far less likely to be capable of picking up the slight difference in audio quality. If you take this expensive equipment and put them on the ears of a 14-year-old, the person will have a hearing making them far more likely to tell the difference, but they would most likely not care to begin with.
We are all going to be different, of course. So your mileage will vary. But unless someone is capable of entering an unknown environment with high-end equipment they have no personal experience with and listen to a complete blind-test and prove to me they are actually able to hear and tell the difference, I'm going to have a tough time believing them.
Whatever differences you are telling yourself that you are hearing are most likely down to the differences between the masters that have been used for whatever encode you are listening to and the placebo. And it's nothing wrong with that. Placebo is a powerful thing, and if it gives you joy and you feel you get something out of it, there is nothing wrong with it.
With all that said. I see nothing wrong with Apple offering a "Hi-Fi" option. The closer the encoding is to its master, the less likely we will have any kind of theoretical degradation of the audio. And it's not like streaming in 256 kbps or 1411 kbps or whatever is much of an issue in 2021. So providing the option doesn't really hurt anyone, so it doesn't really matter if it's pointless or not. Why do we have to throttle the bitrate when we have the capacity in our networks for using higher bitrates.
This is what I believe too; I'm sure there are people with gifted ears and fantastic equipment that can hear the differences, but it seems so exceedingly rare that I'm highly skeptical of people who claim they can. That doubt is doubled when I hear flowery nostalgia for vinyl/analogue and qualitative descriptions of HiFi being warm, rich, full or natural.I also have the Sennheiser HD 800S, connected to a Schiit Magnius amplifier connected to a Hegel HD12 DAC. Balanced cables between them all.
I love the headphones, and I decided to get the amp as the Hegel HD12 isn't quite powerful enough to drive the Sennheiser HD 800S, in my opinion. The only way to get fully balanced cables end-to-end is by connecting the Hegel HD12 to a balanced amplifier.
I have listened to 256kbps AAC, 1411kbps ALAC, Tidal MQA 24bit/96kHz and some music from Blu-Rays at 24bit/192kHz I would be hard-pressed to separate them in a blind test.
Going from running unbalanced using my Hegel HD12 to fully balanced by tossing the Schiit Magnius into the mix made for a much bigger change than going from 256kbps ALAC to 24bit/192kHz tracks.
And this is pretty much being backed up by almost every unbiased blind test out there. Even on extremely high-end and expensive equipment, next to no one can separate the source material when you go past 16bit/44kHz 200 kbps VBR if decent codecs and encoding have been done from decent masters.
And the funny thing about hearing is that it's degrading with age. By the time you reach a point in life where you have the money to spend on great equipment, your hearing will most likely have degraded to a point where you are far less likely to be capable of picking up the slight difference in audio quality. If you take this expensive equipment and put them on the ears of a 14-year-old, the person will have a hearing making them far more likely to tell the difference, but they would most likely not care to begin with.
We are all going to be different, of course. So your mileage will vary. But unless someone is capable of entering an unknown environment with high-end equipment they have no personal experience with and listen to a complete blind-test and prove to me they are actually able to hear and tell the difference, I'm going to have a tough time believing them.
Whatever differences you are telling yourself that you are hearing are most likely down to the differences between the masters that have been used for whatever encode you are listening to and the placebo. And it's nothing wrong with that. Placebo is a powerful thing, and if it gives you joy and you feel you get something out of it, there is nothing wrong with it.
With all that said. I see nothing wrong with Apple offering a "Hi-Fi" option. The closer the encoding is to its master, the less likely we will have any kind of theoretical degradation of the audio. And it's not like streaming in 256 kbps or 1411 kbps or whatever is much of an issue in 2021. So providing the option doesn't really hurt anyone, so it doesn't really matter if it's pointless or not. Why do we have to throttle the bitrate when we have the capacity in our networks for using higher bitrates.
Let's see how long it will take for Hi-Fi music streaming to fail and be discontinuedSo they discontinue the Hi-Fi HomePod just before introducing Hi-Fi music streaming? 🤔
Aptx is Qualcomm, LDAC is Sony, Apple would probably come up with their own format before paying license fees to these. So forget about those.I wouldn't be so sure. The lightning jack won't be around forever. I wouldn't be surprised if future iPhone models dispensed with connector ports entirely and went completely wireless. Much like Apple's stubborn commitment to inferior mp3 quality, Apple's refusal to embrace more advanced Bluetooth codecs like AptX HD and LDAC, or offer their own version, is frustrating. Perhaps someone will develop an ultra low power wifi chip that can go into headphones. All I know is that we can do a lot better than the status quo.
For real, I bet you we see that tech comeback in another form tho. This isn’t the first time they canceled a high end speaker and the made a new one.So they discontinue the Hi-Fi HomePod just before introducing Hi-Fi music streaming? 🤔
That's what I was wondering!Wonder if there will be an upgraded iTunes Match fee?
that's for loseless audio.Hopefully they are airtags compatible
So it's plausible one could discern a difference if a remixer ****s things up, which is why bluetooth AAC vs AptX HD FLAC file or 3.5/USB-C comparisons may not be equivalent to directly comparing 326kbps and a FLAC file via wired direct connection.Not surprised at all they're doing a lossless Apple Music tier, in fact surprised they didn't beat Spotify to it. Clearly that's the future of streaming now we're no longer restrained by small storage and slow internet. No need for lossy audio anymore. And yes you can definitely tell the difference if you have good quality equipment to listen through. You can even do it through bluetooth with LDAC. I'm sure Apple will launch their own proprietary solution.
You beat me to it RE: Cymbals. lol.This is what I believe too; I'm sure there are people with gifted ears and fantastic equipment that can hear the differences, but it seems so exceedingly rare that I'm highly skeptical of people who claim they can. That doubt is doubled when I hear flowery nostalgia for vinyl/analogue and qualitative descriptions of HiFi being warm, rich, full or natural.
While those descriptions can be true, it wasn't until I ABX'd lossless against ~160kbps iTunes AAC VBR that I noticed how much of a challenge it could be to hear the differences. The only consistent give away were cymbals. Apple Music 256kbps AAC CBR should be enough for most people. The remaining audiophiles will then mostly be satisfied with access to CD lossless equivalents. It won't be until bandwidth/storage become so trivial that HiFi at 24+/96+ will become a mainstream focus. Maybe if the streaming services get into a pissing match about (essentially placebo) music quality?
The archivist in me loves that idea. But I don't think it particularly pragmatic just yet.