Posts from the ‘Uncategorized’ Category
When encoding video for home playback instead of streaming the goal is to achieve the best compression and quality in the fastest amount of time. When streaming games file size does not matter as long as a you do not saturate your network bandwidth. While for home use compression is king. I wanted to test what is the best format and encoder to meet these conditions. No AVC (h.264) was included as most video is already encoded in this format to begin with (AVC is commonly stored in the .mp4 container). This test uses handbrake 1.6.0 to test AV1 vs HEVC (h.265) to see if the AV1 is a good future proofing format.
For this testing a scene from the original Ghost in The Shell 1080p Bluray(Specifically the 25th Anniversary Edition) that I obtained from personal copy using Make MKV. For comparison of video quality I used the structural similarity index (SSIM) where 1 is an identical image and 0 is an unrelated image to the reference image. SSIM scores where taken from 4 different parts of the scene which will be shown below.




Below are the settings that where used to test each encoder using Handbrake.
Encoder | Preset | Encode Type |
x265 | Ultrafast | Constant Quality |
Faster | Constant Quality | |
Faster | Constant Quality + Tune:Animation | |
Faster | Constant Quality + Tune:Grain | |
Faster | Constant Quality + Tune:SSIM | |
Medium | Constant Quality | |
Slow | Constant Quality | |
Placebo | Constant Quality |
Encoder | Preset | Encode Type |
Apple Video Toolbox | Quality | Constant Quality |
Speed | Constant Quality |
Encoder | Preset | Encode Type |
NVENC | Fastest | Constant Quality |
Medium | Constant Quality | |
Slowest | Constant Quality |
Encoder | Preset | Encode Type |
AV1 | 12 | Constant Quality |
5 | Constant Quality | |
2 | Constant Quality |
Below you will see charts containing the averaged encoder SSIM values. The closer the value is to 1 the closer the image is to an exact copy of the original video.
Preset | SSIM |
Ultrafast | 0.953 |
Faster | 0.966 |
Faster + Tune:Animation | 0.960 |
Faster + Tune:Grain | 0.965 |
Faster + Tune:SSIM | 0.966 |
Medium | 0.967 |
Slow | 0.970 |
Placebo | 0.969 |
It is important to point out that moving the x265 preset up doesn’t cause massive changes in quality. Going from ultrafast to faster is quite a big jump in SSIM value but faster to medium has less range than adjusting the encoders tune. Notice that placebo does not beat Slow. People have this idea that increasing the preset increases the quality but it really just takes a longer time to search for the optimal settings which may not lead the a better quality.
Preset | SSIM |
Quality | 0.943 |
Speed | 0.943 |
Apple’s encoder appears to be broken in handbrake do not use it or buy a mac to use it.
Preset | SSIM |
Fastest | 0.966 |
Medium | 0.967 |
Slowest | 0.967 |
For our scene not a big difference between the presets for home use just use Medium or Fastest.
Preset | SSIM |
12 | 0.960 |
5 | 0.969 |
2 | 0.971 |
AV1 with its 2 settings showed the the best quality of the entire test. It however took forever as you will see the in the test below.
Preset | Avg. Encode Fps | Percentage of Original (%) |
Ultrafast | 174.4 | 12.0 |
Faster | 55.2 | 19.8 |
Faster + Tune:Animation | 52.6 | 13.1 |
Faster + Tune:Grain | 41.4 | 34.4 |
Faster + Tune:SSIM | 56.0 | 18.4 |
Medium | 29.7 | 22.0 |
Slow | 13.8 | 26.4 |
Placebo | 1.3 | 28.4 |
As you see adjusting the preset has a massive impact on encoding rate. Going from faster to medium does not drastically improve the encoding quality or compression but will take twice as long to encode.
Preset | Avg. Encode Fps | Percentage of Original (%) |
Quality | 114.8 | 3.1 |
Speed | 114.9 | 3.1 |
Once again do not use handbrake to encode using the M1 macs built in GPU. The files look bad probably because they have to low of a bitrate hence the High compression ,but I used constant quality encodes so this is annoying.
Preset | Avg. Encode Fps | Percentage of Original (%) |
Fastest | 304.1 | 31.4 |
Medium | 214.2 | 31.4 |
Slowest | 116.8 | 31.4 |
NVENC’s setting just adjust quality but do change compression. NVENC has to worst compression of the whole test. While the encoder is fast it does not provide much compression. In some extreme cases I have found that NVENC does not provide in compression at all. I generally avoid NVENC since my CPU is strong enough to encode but you should consider it if you have a weaker CPU. I would not recommend buying a new GPU for its encoder best to upgrade your CPU.
Preset | Avg. Encode Fps | Percentage of Original (%) |
12 | 261.9 | 22.0 |
5 | 6.5 | 21.2 |
2 | 0.4 | 24.2 |
The compression was middling but the encode times where some of the slowest in this whole test. Preset 5 had quality similar to x265 Slow ,but was half the speed and no where near faster which with tunes is not far off from Slowest. I would recommend AV1 as player support is poor and does not provide any benefits over the much more wildly supported H265 video codec.
Best performance
I created a top five list of best performing encoder. First the encoder setting had to perform above average and then second it must have the highest score consisting of the sum of milliSSIM per second and Megabytes saved per second. This tells us how much quality does the encode gain per second of encoding and how much file size reduction is achieved during the encode duration.
- x265 faster Constant Quality using the SSIM tune (Score: 13.17)
- x265 faster Constant Quality no tune (Score: 12.84)
- x265 faster Constant Quality using the Grain tune (Score: 8.67)
- x265 Medium Constant Quality no tune (Score: 6.80)
- AV1 (SVT-AV1) 5 Constant Quality (Score: 1.50)
Below is a side by side comparison of our original with the winning codec. I want to say since our metric is in fact SSIM using the SSIM tune is basically cheating however I generally recommend using faster as it gives you the best encode time and image quality.








I am not going to show the others as the SSIM scores are relatively similar and the quality is very good for all of them. Below I will show Apple Media Toolbox which had the worst quality to give an idea of the range.
Apple Video Toolbox








Conclusion
I was shocked at how poor AV1 did considering the hype around it. It was able to get the highest encode quality. However, its’ compression was generally around than of h.265 and encode time was much longer. Remember H265 came out in 2013 and AV1 in 2019. AV1 is not even competing with H265 but the new H266(VVC). The real big push for AV1 is that big companies do not have to pay any licensing fees to use the codec. However, the total license fees are capped for HEVC (H265). Also, the big push for that AV1 is better from an open source perspective is nonsense since both handbrake and x265 are open source and open souce encoders exist for h.266.
Notice that increasing the “quality setting” does not lead to drastic increases in SSIM score. This is due to the fact that the “Higher Quality” presets spend more time searching for more optimal ways to store the content which may or may not lead to better quality. I would suggest sticking with faster for x265 if your doing constant quality encode. Also, NVENC had the worse compression and middling quality. If you want to encode video for compression I would suggest a CPU upgrade before GPU. However, if you cannot get near real time on the x265 Ultrafast preset then you should use NVENC as it will give you better quality at a much faster speed. Apple’s video toolbox handbrake integration is bad do not use it. If you have to encode videos on a mac use Apple’s Compressor not Handbrake.
Raw Results
Here is a spreadsheet with my raw results below.
Computers
- MacBook Pro (13″,M1,2020) 16gb of ram and 512GB running MacOS Monterey
- Custom Built Desktop with a 3950x
Processor | AMD 3950x with additional 200Mhz allowed Boosted & Infinity Fabric running at 1867Mhz |
OS | Debian 11 Bullseye |
Graphics Card | Gigabyte GTX 1070 G1 |
Ram | Patriot Viper Gaming PC4-35200 4400MHz running at 3533Mhz |
SSD | WD – BLACK SN750 (WDBRPG0010BNC-WRSN) |
Motherboard | Asrock X570 Creator |
Cooler | Ice Gaint ProSiphon Elite |
Benchmarks
To compare how your computer would run these task I attaching a few benchmarks
Geekbench 5
Computer | Score |
Custom Desktop | |
-Single Core | 1418 |
-Multi-core | 16677 |
M1 Macbook Pro | |
-Single Core | 1063 |
-Multi-core | 4303 |
Phoronix Test Suite 10.8.4
Test | Result FPS |
SVT-AV1 Preset 4 Bosphorus 4K | 3.229 |
SVT-AV1 Preset 4 Bosphorus 1080p | 7.649 |
x265 Bosphorus 4K | 23.36 |
x265 Bosphorus 1080p | 43.87 |
Phoronix test suite is a free and open source cross platform application that can be used to benchmark computers.
Handbrake is a powerful video transcoding program than is greatly misunderstood. A lot of discussion goes into which preset to use for the best quality and fastest encode times. However, Handbrake’s presets are designed for best compatibility of devices regardless of quality. If you are planning to playback on VLC or Roku Media Player not to much to worry about in terms of comparability. However, if you have an Apple TV or a Bluray Player look at the support guide. The presets are basically useless and will lead to worse encode times when stringent format support is not required.
When encoding for home consumption your looking for
- Smallest File
- Quickest Encode Time
- Highest Quality
A lot of people after watching youtube videos think GPU encoding is perfect for this job, you would be wrong. Here are the issues with GPU encoding is that GPUs are geared towards live encoding for game streaming. The reason they can get these incredible encode time is they are doing a lot less optimized encoding. How video encoding gets great compression is by finding similarity between frames or reducing quality. Finding similarity between frames means you have to look at more frames before outputting a frame. Looking at more frames increases latency thus why it is generally avoided in GPU encoding as it has worse compression. Below I have video by YouTube Maraksot78 about compare file with different encoders. Intel quick sync encoding H.265 provided the smallest file of any of the hardware accelerated systems it still was 40% larger than the CPU encoded file and 1.6 times faster encode time.
Also, an annoying thing about handbrake is that custom presets are compatible across operating system or version. Which means you’re going to have remake them every time which is very annoying.
Here is what you need to compress video
First quick way to reduce encode time
- Copy content to internal drive if you have enough space. This way reading the file does not slow down encode time
- Make sure your using fast external storage (use a usb 3.0 drive or better yet an external drive enclosure)
- Use MediaInfo to learn more about the data you want to encode
Step 1 Learn more about your content
Is the video you want to encoded a rip from a Video Disc (DVD or Bluray or Maybe an HD DVD?)
Is the video progressive or Interlaced?
What format is it encoded in? as you may not get additional compression
Here are some examples from MediaInfo using the tree view
Example 1: A video from my digital camera

Here are the fields to pay attention to Complete name this is the name and location of the file. The format field tells you the container that is used to storage the video file. A container is basically like when you store a bunch different documents in a zip file so they are linked together. A container is always used because inside a video file is literally a separate audio, video and subtitle files. A container also limits what type of files maybe stored inside of it. The field format profile tells you what video profile your movie in this affect things like maximum quality that can be stored.

Now we will look at the video section. The fields you care are format, format info and Scan type. The format field tells you what codec this video is encoded in it currently encoded in AVC which stand for Advanced Video Codec as stated by the format info. AVC is also known as H.264 which is the name of the ITU standard for the video format. Then finally the Scan type field tell you if this progressively scan like modern display or interlaced. Knowing the scan type can help save encode time and do necessary processing.

Now finally onto the audio field. The fields of interest are Format, channel(s) and Compression mode. The format tells you what audio format is stored in. When encoding you never want to modify this format because increase computation and can hurt audio quality or even convert 5.1 to surround sound. Many will say you should convert it to save space but in the case of this file the audio only accounts for 2% of the total file size so your no saving anything.
Example 2 A rip of DVD movie

Notice that we have additional field in this such as text and menu along with additional audio fields. This has multiple audio tracks in this case English and Chinese. The text fields contain subtitles and the menu contains the DVD chapters.

So big difference is that the format field is Matroska. This is a MKV file and my preferred containers it allows a lot more file types to be stored and is an open format.

The fields are format, format version, frame rate and original frame rate. The format and format version tell you that this is encoded in MPEG-2 this is a very common format on DVD. The very important field is frame rate and original frame rate. DVDs can only store video at frame rates of 25 or 29.97 (this is known as 30 fps broadcast but it is not really 30fps because of reason that are beyond the scope of this article look up NTSC or watch the playlist by Technology Connections. But in short summary when your encoding video you need to set the frame rate to the frame rate field ,but handbrake will show the frame rate as Original Frame rate field.
The big difference is that this has surround sound with its six channels. The format is is AC-3 which has the commercial name of Dolby Digital. Look the field language this is the english track ,but Audio #2 is the Chinese track.

I am showing it for information consistency but you will not have to do much with it in handbrake.

This allow you to use scene selection.
Example 3 interlaced DVD

Here is the important fields to look out for is scan type, scan order and Standard. The scan type field shows that this video file is interlaced and that the first field is at the top of the screen. Film Maker IQ has a great video (The Joys and Sorrows of Interlacing) explaining what interlaced vs progressive video. However because this interlaced it will need to be de-interlaced to properly be displayed on modern display. The Standard will tell you what standard you need to set it and guess at the true framerate.

Here this is just to point out that DTS surround audio exist.
Now onto handbrake Settings. The goal is to show you how to encode a HEVC/H265 file passing thru all audio and subtitles into an MKV container.
Picture Top menu

Here is how we break Down the top menu
Open Source – This is what you click to open a video or folder of videos you want to encode
Add to Queue – If you imported a folder of videos this allows you to setup the same transcoding for all video in the imported folder. This allows you to set a bunch of settings on one video and apply it to all videos.
Start – will start the encoding of videos
Pause – pause allows you to pause encoding when you need more CPU powerful
Presets – Shows you list of available presets. If you are trying to encode for specific device this can be useful ,but they will not be used in this tutorial.
Preview – this allows you to preview your video encoding setting this very useful and can save you a lot of time compared to just rendering out a whole a video. This will also allow you to estimate how long an encode will take.
Queue – Allows you to see video that are in the encode queue and this opens up a new window
Activity – handbrake is basically a wrapper for a command line program and this just a cleaned up version of the command line output. Can be helpful for figure additional stats about encoding speed but generally you will will not need to interact with it.
Source – is the current video file that you are working on.
Angle – would allow you to selection another camera view if you have a file with multiple video streams
Range – allows you to select the region of the title you would like to encode. The selection range can be anything from chapters, seconds or frames
Preset – is the currently selected preset
Summary pane

This pane summarizes the encode settings
save as – controls the output file name
encode time – would tell you the progress of an ongoing encode
Dimensions Pane

Basically set for the bare minimum any cropping on downscaling will cause additional computation.
Orientation & Cropping
Unless the video has clear visible black bar around the edges you want everything in Orientation & Cropping off. Flipping should be unselected, Rotation should be at zero and cropping should be set to none. If your platform does not have a none option then should set it to custom and put zero in all the boxes.
Resolution & Scaling
Resolution Limit – None. This will allow you to reuse this as a template.
Anamorphic – Automatic
Optimal Size – Leave Checked
Allow Upscaling – leave unchecked. If checked this will drastically increase computation time depending on what your resolution limit is set to.
Borders
fill -none. You do not want to add a border to your video
Final dimensions
can be ignored let handbrake do it for you.
Filters Pane

Anything in this pane will be done with CPU and increase encode time
I mention the setting Detelecine (link wikipedia article). What happen is this processing is about undoing the process to turn convert film to television frame rate called telecine. Converting Cinema (shot on film) frame rate (24fps) to television frame rates 25 fps and 29.97 fps involves some interesting process but when converting to a progressive frame rate we want it to be at its original frame rate. Animated content is almost always rendered at 24fps because it is cheaper to do it this way. for reference for you save 60 frames every minutes when you compared to 25fps. You save 358 frames per minute compared to 29.97 fps. If you assume a film is approximately 90 minutes long that mean at 24fps saves you 32238 frames compared to rendering at 29.97fps and 5400 frames compared to 25 fps. In other terms renders in 29.97 compared to 24fps is like making your movie 22 minutes longer and almost 4 minutes longer at 25 fps. So assume it is at 24fps if it is animated or your regions equivalent.
If your content is progressive or the scan type is not provided then you want to set every filter to off. Playing with these filters will not really improve your encoding with out massive amounts of tweaking and then only marginally. If your frame rate differs original frame rate and is 24 fps or 23.97 or 23.976 enable the Detelecine this will help improve the quality of motion.
If your content is interlaced only turn on Interlace detection to default then set Decomb and choice the preset EEDI2 this will give you the best Deinterlace quality. If you know the content original came from film or is animated turn Detelecine to default.
Video Tab

Video Encoder – H.265 (x265) x265 that will encode your video
Frame Rate – the simplest and the least computational expensive option is same as source. However, if you video was originally from film set the frame rate to the appropriate frame rate for your region.
Make sure your have handbrake set to variable frame rate this means that only when a change on the screen this can drastically decrease your file size if you have a lot of static content. Make sure your quality is set to Constant Quality this will give you the best looking output by varying how large each frame needs to be to have maximum quality. HandBrake provide suggested values in the document “Adjusting Quality”. The smaller number the value the better the quality but do not go 2 less than suggested as this will just increase file size without really improving quality. Remember zero just render a whole saves each frame as an uncompressed image and reduce file size is about only storing the difference between frame.
Preset – set this to faster. A lot of people believe that this a quality setting including the handbrake team. It is not a quality setting but a compression vs quality. Basically increasing from faster to fast might reduce your file size, but it will definitely increase your encode time and remember since you are using constant quality encoding it will just increase the file size in specific regions to maintain quality so I would really only go from faster, medium & slow. I have not noticed visual quality drops or drastically bigger files using faster but have noticed the speed up in encode time. I have noticed quality drops using anything below faster.
Tune – If it was originally shot on film set to grain and if it is animated content set to animated. Otherwise set to none. This will also help reduce file size.
Profile – auto. Only the main and main still pictures exist so no reason to set it to anything other than auto.
Level – set to auto. (I am not even going to try to explain this. If your player says it only supports a specific level in h.265 set it to that level otherwise just leave it at auto)
Here is a link to the documentation for x265 the encoder used by handbrake to encode the files this better explains tunes and profile. You can learn a lot by reading this document it is very well written actually. A wikipedia page about the different types of frames used in encoding will help explain things like b-frames.
Audio

So here I am gonna to tell you to basically throw away everything that is already enabled away.
Click the button “Selection Behavior”
in the new window

in the new window
Track Selection Behavior – All Matching Selected Languages
Languages – Any. Or your preferred language. Since audio takes so little space compared to video no reason to really not keep all the tracks. Then the person who prefers subs is always comfortable.
Auto Passthru – Select all of the check boxes you do not want to waste cycles re-encoding anything. Then set Passthru fallback as Flac 16-bit so you preserve the maximum quality audio in a lossless format. In most cases it will not do anything but pass the audio.
Then click the minus button at the bottom and make sure the pane is completely empty.
Then add the following
Codec – Auto Passthru
Mixdown – 7.1
Samplerate – Auto
Bitrate – cannot be set as this is passed from the file itself
gain – 0
DRC – 0
NOW click ok.
In the Audio Pane click reload
(if you do not reload you will not pass thru all the changes.)
This will load every audio track including stereo files of the selected language because by selecting mix down 7.1 it will never actually mix down audio.
Subtitles Pane

then click selection behavior

Track Selection Behavior – All Matching Selected Languages
Languages – Any
Options – Add Closed Caption when available select this option. Add foreign audio search.
Burn-in Behavior – None and deselect DVD & Blu-ray subtitles.
Click OK
Then hit reload in the subtitle pane.
Now all the subtitles are added. Once again these are text files do not add much to the final output.
Chapter Pane

If you know or want to name the chapters you can in this pane.
Create New Preset

Click New preset
Resolution Limit – None
Audio – Repeat the selection from the audio pane it does not pull that from your current settings which is annoying.
Subtitles – Repeat the selection from the subtitle pane.
Now that you are done you will want to save a preset for different qualities based on resolution and maybe even different tunes.
Why I do not like handbrake’s presets
In general the presets basically never use H265 which will give you the maximum compression possible without loss in quality. Only streaming services and website do not use H265 because they do not want to have to pay a licensing fee, but as an end user who has devices that support the format it saves you a ton of space. Yes, the encode time is longer, but I have found with the faster preset my H265 files are shrunk by a 1/3 on average (big variance can occur depending on the source material) without visual drop in quality just by going to H265. You can fiddle with the knobs of h264 all you want your not gonna get that drop in file size. Many presets forces a frame rate which could cause you to have to waste time generate additional frames or lose frames.
Should you even do this?
A lot of YouTuber are pushing services like Plex and Jellyfin and show these encoding tutorials. However, the issue with home media servers is that ripping DVD and Blurays then transcoding video takes longer than watching the films themselves. A NAS and Libredrive compatible Bluray drive is going to cost more than just buying a nice Bluray player. Also, in getting an optimal video deinterlacing and upscaling in most cases you just need to pop in a disc into a Bluray/DVD Player. You can buy a 4k capable Bluray player for between $150-1000 which will get you some of the best Bluray player in the world. The cheapest two drive NAS that I could find was around $160 with no storage drives which means for a lower end Bluray player (which the Sony ones are actually quite good) you could be watching a bluray instead of ripping them. So, unless you have videos that raw video files or video downloads it is more worthwhile for you time and money to just play your disc on a player. You can read my article about optical storage media ,but it has a much longer life than common wisdom states. Optical media has already out lasted the other storage media at this time so backing up to SSD or hard drives is a waste of time.
An important question we need to ask is what is the purpose of the reviewer? Most people watch or read reviews to find out if something is worthwhile to purchase and to see how it compares to other in that price range. An important aspect of reviewers is to help their audience gain context about the current products they are reviewing, the available competition and that category in general. It is also, important to try to understand your audience that your targeting.
A review that upon first reading irked me upon further reading is actually a good review not flawless, but good. In the BuzzFeed’s review of the M1 iMac John Paczkowski talked about how the original iMac was so much better than his 1995 Compaq Presario. This seem like a strange reference to mention a nearly 3 decade only computer, but it gives the reader the context of why people cared about the original iMacs. He then goes on to mention his current computer a 2015 iMac and how it is much slower than the M1 and cannot have as many applications open. He then proceeds to say wirecutters top pick HP All-in-One 24-df1036xt is worse because it has worse performance, a worse screen and is not worthwhile even with its lower price. My complaint is that the reasoning behinds these opinions is not well supported. This review does not provide benchmarks or a quantitative Data besides he could have more tabs open. However, the reader is given actionable information that will help them make a purchase decision. I am now going to show you how reviewers can go wrong with quantitative numbers and be lead astray by numbers and specifications. I will be comparing the M1 Ultra Mac Studio with the Alienware R13 the media coverage.
(Many) Reviewers are Basically Useless
Both Youtube channels Linus Tech Tips and Gamer’s Nexus have done reviews of prebuilt computers. I am going to focus on Gamer’s Nexus Coverage of the Alienware R13 and compare it to other outlets. I am picking this computer because it is the flagship device from a major PC manufacturer. It also cost around the same as Apple’s Mac Studio another flagship device. We can see how these computers compare in benchmarks as well.

Gamer’s Nexus released 4 Videos about the computer and its components
- Crazy Bad $5000 Alienware Gaming PC: R13 Aurora Tear-Down (April 24 ,2022)
- The Worst Pre-Built We’ve Ever Reviewed: Alienware R13 $5000 gaming PC Benchmarks (May 1, 2022)
- “Fixing” the Alienware R13 Dumpster Fire (May 22, 2022)
- Sometimes Dell Actually Tries: Dell RTX 3090 Review, Tear-Down & Benchmarks (July 2021)
A quick summary of the problem that Gamer’s Nexus found was that the Alienware ran so hot and was so power limited that its CPU the i9-12900KF performed like an i7-12700K. A power limit means that the system did not provide enough power to the components it was due to software in Alienware’s case, but can also be a power supply issues as well.
Looking at Amazon on July4 , 2022 the i9 cost $549.97 and the i7 cost $377.98 (USD). That means that in the a flagship computer your CPU cannot perform to its full potential and performs like a CPU that cost more than $150 less than the one inside the computer. Also, the computer uses a R13 specific case, motherboard & power supply so you cannot reuse any of the components for a new build. Removing the front panel caused the internal temperature to drop by 10 degrees and during load the CPU was nearing the dangerous maximum temperature. The fans are loud. All the ram is green on an all black Motherboard & Graphics card PCB it does not match the aesthetic. The RAM that Alienware put inside the computer is Slow DDR5. The computer does not have a thunderbolt ports. The R13 is not meaningfully upgradeable as you are thermally and form factor limited. The front inputs are on the motherboard instead of a separate board. If your front inputs get damaged it requires you to replace your entire Motherboard not just the input board.

This is not the case with custom built computers and this is not the case with M1 Ultra Mac Studio as in seen in the Max Tech Teardown . Apple has more replaceable components so that if any port gets damaged you replace the breakout board not the whole motherboard. However, Apple locks it components to CPUs so enjoy trying to fix it but it can be physically be fixed. It also does not have 10gigabit which is used for High end connected storage and is standard on the Mac Studio. My custom built computer which is half the price of this computer has 10gigabit ethernet and thunderbolt so it is frankly unacceptable. When we visit the website OpenBenchmarking.org the i9-12900K (A variant that has same boost and clock but more enterprise features) it is in the 75th percentile of a site that allows dual core servers workstations. The i9-12900K in many benchmarks in in the 99%+ percentile so saying it is faster than any other system is a given. The real question is what value does Alienware bring to this components!
Other Media Coverage
Many of Many (An Australian Men’s LifeStyle site)
They mention the loud fans but are not as annoyed as Gamer Nexus. They never really compared it to other systems on the market or talked about thermal throttling. They never compared it to an external test bench. They did mention that boutique builds would use nicer material than was used on the Alienware. They also did not compare it to other offering.
Tech Targeted Media
PC Mag
When scrolling across the page they show similarly priced system but the prices are not as configured for the review but as the base system price which is misleading. They also mention that it does not feature glass side panel on the case that high end competitors have, but plastic. This computer cost $4.5k+ so maybe Alienware could put some nice glass on it I mean it is only a couple of months of rent in a two bedroom apartment in NYC (https://www.apartmentlist.com/renter-life/cost-of-living-in-new-york). They do not mention that it does not have comparable connectivity to other products in this price range and that the device thermal throttles. The Alienware system also performed below the test bench they used in the review for the i9-12900K they reviewed standalone which was not mentioned. System to system variance is real, but for $4.5k you can tune a system. Compared it to other offering and it was worse with a more powerful GPU. Also, the benchmarks never ran for a longtime so the thermal system was never really taxed. So, more benchmark and no real additional useful information.
Hot Hardware
Mentions fan noise of previous generations. They mention the wifi chip could be better I mean it is only $4.5K+ who needs better wifi. They did not experience any thermal throttling in testing and device even got the Editor’s Choice award even when it did not perform as well as on some of the benchmarks for the standalone i9-12900K that Hot Hardware reviewed. They also did not use almost any of the same benchmarks for the standalone beside Speedometer so it is very hard to show a drop in performance.
If you look at these website you would think that I am making up this thermal throttling issue, but I picked this device because I know it thermal throttles. So, the YouTuber JustIn Tech Tips made a video called “How to Mine Crypto on the Alienware Aurora R13 – Without Crashing!” he had to dissemble the GPU and apply new thermal paste on a New $4.5k+ computer. The GPU temperature dropped by almost 20 degrees Fahrenheit and allowed him to Crypto mine for 24 hours a day instead of instantly crashing! I am not a fan of crypto and recommend Nicholas Weaver lecture talking about its issues. JustIN Tech Tips also replaced the stock fans with a pair of after market fans (Noctua NF-A12x25) he found that it was quieter even at maximum speed. I am avoiding talking about his testing methodology because that is a whole another article and most channels do it wrong. He did see that when the where fans not at full blast they had worse temperatures. The difference in fan performance could be do to a poorly optimized fan curve or how quick it makes the fan spin for a given load. Remember Alienware only contribution to this system was the power supply, cooling, circuitboard and chassis design. The fact that both JustIn Tech Tips and Gamers Nexus could improve the system means that Alienware’s engineer where not trying very hard.
Now let’s see how the Alienware compares to the Mac Studio with M1 Ultra. First UFD Tech found that the Mac studio was very efficient and could mine and does not mention the device crashing when trying to mine on the device or the standalone PC he built. In the PC Mag review they mention even under load the Mac Studio fans do not get loud unlike the Alienware. They mention that the Mac studio is not user upgradeable ,but do not talk about how the Dell used not standard parts. In Linus Tech Tips review of the Mac Studio even with using unrealistic workloads they could not get the mac studio to thermal throttle and the fans never spun up past 50% of fan speed.
The Tech Journalist Pipeline
I could only find a couple, but Glassdoor says it all a Journalist Total pay is $85,580 ($52,551 Base Pay) and for an Electrical Engineer $102,303($81,649 Base Pay) so becoming a journalist will cause you to loses $15k+ in salary on Average. Making it hard to recruit any one from the industry with knowledge. PC Mag’s Glassdoor shows that a staff writer makes somewhere between $76K – $81K Base Pay. This is actually industry competitive base salary, but it does not show the difference in working environment.
A big incident of the poor culture in tech journalism is the infamous Verge PC Build which showed the incompetence of need for quick video for clicks by releasing sloppy videos. Check out Timmy Joe PC Tech video on the subject for a full context. Eric Jackson a former Contributor for Forbes wrote that “Why is there so little criticism of tech companies by mainstream business journalists today?”. Where he lamented that few journalist add their own opinion because they fear being wrong or losing access to companies. He points out the mainstream Business Press is not writing Long Form stories covering tech companies. Jackson says “Sure, you’ll get the occasionally bright blogger shining a spotlight on truth but the mainstream business press really needs to start pulling its weight.” YouTuber are the bright bloggers of the modern era.
Tech journalism has become tedious product journalism where printing the spec sheets for mass produced consumer products is celebrated as a great story and where there appears to be little understanding of bigger picture stories about how our digital technologies are transforming our industries, cities, and our societies, at a pace and scale that’s never been seen in our history.
Tom Foremski article in “The role of tech journalism in a post-technology world”
His words are still applicable today as very little context is given even in bench mark heavy reviews. For instance the technology outlets who benchmarked the Alienware R13 got burned because Alienware made sure that short benchmarks would run fine on the machine. As someone who works in a companies lab we do know what test your most likely going to run and we will see how we compare to thecompetition before hand.
Closing: Why does this matter?
The average consumer does not have the ability to do high level testing or try a large set of products that are available so they need journalist to provide them with context. The technology companies have basically spent over a decade with minimal journalistic oversight and it shows. Companies are savvy and know what is being tested and reviewers must update methodology to make sure the readers understand the benefits and downside of products.
People when backing up data like to focus on speed and apparent ease. This is why people mainly focus on portable hard drives or SSDs for backup solutions. While easy to copy and paste on they are easily damaged, very expensive and slower compared to desktop hard drives. Another disadvantage of the hard drive solution is determining the formatting. I simple will argue why you should consider blu-rays as your onsite backup medium.
Blu-ray Filesystem Advantage
While not thought of very often in the case of backing up data Filesystems are very important. Filesystems are how a computers understand the contents of storage drive. For instance Mac Formatted hard drives cannot be read without additional software on other operating systems. An NTFS hard drive is read only on MacOS. Some older portable drive formats also limit the file sizes that can be stored on them and have naming limitations. This means that you could have the storage space available to hold your Data ,but due to the name or file size it could not be backed up to your drive. Reformatting your drive causes all the data to be deleted off the drive so you have to transfer all the data off to put one file on. While optical storage including Blu-ray use the UDF (Universal Disk Format) which is vender independent and can support massive files and is natively supported in MacOS, Linux & Windows. Which makes it an excellent way to store your files platform independently.
Write Only and Storage Drive
Most of the blu-ray disk your going to find are Write only which for a hard drive is a pain, but for backup it isn’t. When you are backing up photos, home videos, music or movies you do not want them to be changed. When backing up this data the the whole point is to not lose it or have to be changed. Write only memory cannot be corrupted when the reading device loses power. Blu-rays are Impervious to powerful magnets so they beat hard drive storage and flash storage in this regard. Another big benefit of Blu-rays is that the storage media is separated form the playback device. If a hard drive’s read head dies the data could be fine but it impossible to read and SSDs have the same problem with a dead controller. This is a massive advantage because replacing a broken Blu-ray drive allows access to data while replacing the read head of a hard drive or SSD controller is expensive and difficult. An unforeseen advantage of the Read Only nature of the data is that it is immune encryption based ransomware since the data cannot be modified.
The format is backward compatible
A Blu-ray drive can read CD and DVDs which makes it more than a one trick pony. This seems trivial until you realize that electrical interfaces for internal storage have drastically changed just over my lifetime with following set of connectors IDE, SCSI, SATA (I,II &III), MSATA, SATA Express, M.2 & U.2. External interfaces have changed from USB mini, USB micro, eSATA, Firewire 400, Firewire 800, Thunderbolt 1, Thunderbolt 2 and Thunderbolt 3. While with a single drive you are able to access data from 1983 till the present.
Isn’t Optical Media Slow?
To put this to the test I created a Test set for transfer speed to compare real world performance. It was broken down into three groups pictures, sound and video. In the pictures I had the pictures and video from a trip I took during graduate school which totaled 8.7GB including mainly pictures and some short videos. In the sound category I have Led Zeppelin’s entire Discography in CD quality FLAC, A copy of the En Avant! Audio companion CDs and Finally Mike Dean’s 420 & 422 in HiRes FLAC. On the Video section I had DVD rip of the first season of The Batman, Rosewood in 1080p, The Recording of Underoath playing “LOST IN THE SOUND OF SEPARATION TRANSMISSION” and finally HDR test patterns. The Performance was carried out by copying files to and from the WD NVME drive which is assumed to be faster than the rest of the drives but can not be profiled as you cannot fairly copy to the drive itself. The test depended on the drives bottlenecking the WD NVME drive. The Access time was found by using Gnome’s built in drive benchmark tool using a file size of 954MiB to cause a cache overflow on the drives with known cache sizes. A cache overflow just means that if a drive has a fast memory cache on the device if the file transferring is larger than this it will have to access the slower drive. This is like fusion drive which had a mix of a hard drive and an SSD if you transfer a file larger than the fast SSD you would be limited to hard drive speeds.
I have benchmarked several storage disk below.
Model | Read (Megabit/s) | Write (Megabit/s) | Access Time (ms) | Estimated Cache Size (Megabyte) | GigaByte/(USD) |
WD_BLACK D10 8TB | 1886 | 1782 | 11.74 | 256 | 40 |
256GB PRO Elite | 2229 | 761 | 0.42 | Unknown | 4.97 |
16GB | 268 | 40 | 1.08 | Unknown | 0.94 |
WD_BLACK SN750 NVMe | ? (due to testing) | ? (due to testing) | 0.17 | Unknown | 6.67 |
CS900 (120GB) | 1344 | 2668 | 0.28 | Unknown | 4.29 |
USA Broadband | 134 | 19 | note* | n/a | n/a |
USA Mobile | 53 | 9 | note* | n/a | n/a |
WH16NS60 | 186 | 92 | 32.29 |
*The access time cloud =Time spent downloading the file +the disk access time
Make | Model | Model Numer | Type |
Western Digital | WD_BLACK D10 8TB | WDBA3P0080HBK-NESN | USB Hard Drive |
PNY | 256GB PRO Elite | P-FD256PRO-GE | USB 3.0 Flash Drive |
Infinitive | 16GB | BL16?425517B | USB 2.0 Flash Drive |
Western Digital | WD_BLACK SN750 NVMe | WDBRPG0010BNC-WRSN | Internal NVME SSD |
PNY | CS900 (120GB) | SSD7CS900-120-RB | Internal Sata SSD |
Median USA Fixed Broadband | Ookla Nov 2021 | ||
Median USA Mobile | Ookla Nov 2021 | ||
LG | WH16NS60 | 6x Dual Layer (50GB) | Dual Layer Ridata Blu-ray Disc |
Blu-ray is one of the slowest media of the bunch and has awful access time. However, blu-ray is fast enough for all the content and still faster to read and write than the median fixed broadband in the United States which is in the top 10 for speeds. Remember the broadband comparison is for if a cloud storage solution could use all of your internet bandwidth which no cloud storage solution is capable of doing. You may have 10TB of cloud storage but how are you going to download that quickly and reliability in case of a failure.
To put this in to perspective the highest bitrate content I had was a Video captured on my camera at rate of 2.375 MB/s or about 10 times slower than Blu-ray’s peak read speed. I had 4k Freaking test patterns in HDR unable to saturate the read speed. Yes, Blu-ray is slow but vary few files even have peak bitrates that can saturate Blu-rays speed. Very few storage alternatives can beat Blu-ray’s GB per dollar.
Bit Rot and Disk Failure
Many will argue that Blu-ray is a bad format because it is a dying media and can suffer from disk rot ,but hard drives fail all the time. What really matters is what is the quality and longevity of your media that you own and frankly it is impossible to know this fact. You can store them in temperature and humidity controlled environments and still have failures. Another argument against Blu-ray is you will need to make a directory to find all your files because they will span multiple disk. However, storage size and file size have been on the increase for a while it is more likely that you will need to span your data across multiple media anyways. Any storage media will need a distributed file directory so this is once again a moot point. Also, I wanted to avoid talking about high-end backup like LTO (Linear Tape-Open) ,but it will prove my point. This argument argument that hard drives will last for 20 years is a moot point because what are you going to read these files on? I am going to show that interface disappearance is the thing to worry about not so much the underlying media. LTO drives need to at minimum be two generations backwards compatible with reading up to LTO-7. The current LTO standard is 9 which means with two previous versions of support 8 which came out in 2017 which means a current drive can only support media at minimum up to three years ago which is not great. However, LTO-1 was released 20 years ago so that means any LTO-1,LTO-2 or LTO-3 drive should be able to read this media so first lets fine some of these drives. A quick search of Newegg and eBay shows that vast majority of these Drive use the SCSI interface which is not supported by modern computers and would require an adapter. The only real adapter I could find was the Ratoc USB 2.0 to Ultra SCSI Converter which is now discontinued. Unless you can find remaining adapters, the remaining drives and your operating system can read the partition format you are out of luck even reading your 20 year old drives. You can go out and purchase a USB blu-ray drive and read a Blu-ray,DVD or CD you burned on a 2003 Blu-ray drive ,but your new LTO drive will not read your 2003 LTO-3 drive. Disc rot is nowhere near as big as actually being able to read the media and is the same as data loss.
Conclusion
While arguing for Blu-rays I would like to point out that all is not well in burning software it is harder to find and they are relatively unreviewed. So, if you’re on a Mac Roxio’s toast from their site is one of the only software packages available (The Mac App Store one is super old). On Windows you have Roxio and Nero for Blu-ray burning. On linux you can use XFburn or K3b to burn disc. However, I would only recommend using them to create ISO files then burning with xorriso the command line tool (see tips for that). Also, unless large files need to be backed up I the initial cost of a blu-ray drive may break even. However, for the speed, cost per Gigabyte, immunity to power failure, immunity to EMP, immunity to encryption based ransomware attacks I believe Blu-rays are an excellent choice for most people’s Local backup needs.
Backup Tips
- Never store data in .zip, folders of compressed folder format as you will a program that can uncompress the folders after which may not exist
- Also, .zip was designed for documents so using it to compress audio, pictures or video is basically pointless. If lossless convert to a lossless format like Flac or Alac. Using CPU encoding you can see massive file size drops going from H.264 to HEVC.
- On linux the most reliable way to burn to use K3b to create a disk image and then use xorriso to burn the disk with command line. I recommend creating the disk image first the burning the disk image.
- xorrecord -v dev=/dev/sr0 blank=as_needed -eject padsize=300k ‘.iso’
- The above script has led to fast flawless burns
- Create a database of stored files on a writeable medium as you will need this to access the data.
- For backing up to CD or DVD just use the built in utilities on each OS they should be fine.
- Store on multiple formats
- Check your Media to make sure the contents still exist
- Create an List of media Storage
- Avoid Hard disk in enclosures buy separate disk and enclosures (Portable hard drives or SSDs)
- Only expect a decade out of your backup and then begin porting as drive and reading will start to become limited
- RAID storage is not backup it is additional redundancy.
See for Additional information
At the beginning of the pandemic I was considering getting an e-reader so I could read and not have to go the store to pick up books. Then once the pandemic ended I would have something portable to read on. After looking into e-readers I decide to just buy physical books.
Then in my news feed I found an article from Markus Riley called “The Advantages of E-Books And Other Forms of Electronic Reading” appear. I felt like some of the claim where deceptive. E-reader’s hype is enmeshed in the techno utopia replacement of legacy technology in other words DISRUPTION!!!! E-readers have more disadvantages than they appear. Here are some common arguments.
E-Books Are Cheaper
A common half truth about e-books is that they are cheaper than physical books. Your average e-book is most likely less expensive than that same book new. An obvious disadvantage for e-books is you cannot buy used e-books which can have a massive savings. However, if you are buying only new eBooks once you throw in the cost of an e-reader this savings becomes murky at best. For instance, I have been reading this light novel series Konosuba over this the pandemic and new volumes cost somewhere between $14 and $10.49 so let’s say on average it cost $12.30 for the physical book. The Konosuba’s ebook average price is $7.99. If we buy the least expensive e-reader (without ads for the kindle people) for $110 dollars we need to purchase 26 books. Konosuba only has 17 volumes so you can’t break even just buying this series. In Riley’s article he mentions textbooks can have massive savings when purchased as e-books.
If we look at the two most recent text books I purchased which where “Wavelets: A Concise Guide” & “Advanced Signal Processing: A Concise Guide” this story gets complex. The difference in the E-book versus physical for “Advanced Signal Processing: A Concise Guide” is $13.50 when looking on Amazon which had the cheapest price. However, when “Wavelets: A Concise Guide” is purchased from the publisher which had the lowest price there was no difference between the physical and electronic edition. The cheapest e-reader is not comparable to physical books but, I will get to that in another section. E-readers file formats are not always transferable and may require conversion software which is not always free. E-book are notorious for their strict DRM (Digital Rights Management) mechanism making it hard to leave the purchase ecosystem.
note: Most e-readers do not have a replaceable battery and have minimal resale value. This means minimal value can be extracted on resale
E-readers are More Portable
An e-reader is more portable than your entire physical library, but most people are reading only one or two books at a time. Looking at a single volume of the Konosuba light novel which weights about 6.9oz and my copy of “Wavelets: A Concise Guide” weights 15.5oz. A 2021 Kindle has a weights of 6.1 oz and an ONYX BOOX MAX Lumi is an e-reader with a 13.3” screen has a weight 20.1oz. If I carry one pleasure book and my text book this is approximately two ounces more than the large e-reader. If I only carry the pleasure book then it is a tenth of an ounce more than a kindle. The e-readers are small bookish size so your not really saving space. Both of the mentioned books are sub 1 inch so at best your saving less than half an inch of thickness going with the e-reader.


It should noted that I am 6ft tall. The Konosuba Light Novel is the book on the left and has dimensions of 8.25″x5.45″ and the book on the right is “Wavelets: A Concise Guide” with dimensions 9.25″x 6.125″.
E-Reader are Better Than Other Screens for Eyestrain
There is no evidence to back this claim up. It is mixed at best. In the study listed below people preferred the physical book anyways!
See “E-Readers and Visual Fatigue” by Simone Benedetto, Véronique Drai-Zerbib, Marco Pedrotti, Geoffrey Tissier & Thierry Baccino
Physical Book purist just care about feel and smell. They are technophobes not ready for the digital future!
This is an argument that is implied and not often directly stated . However, this critique is idiotic because, all the e-readers that I can find have displays smaller than US letter (13.9 inch diagonal) or A4 paper (14.3 inch diagonal) even BOOX MAX Lumi 13.3”. The standard sized e-reader are somewhere between 6-7 inches which is approximately the size of a sheet of paper folded twice or smaller….. so tiny. If you are lucky you will receive an e-reader that has a 300PPI (pixels per inch) which is the minimum density that most books are printed. Most e-readers have a PPI less than 300 and will actually be less crisp than a physical book. PPI (Pixel per inch) is approximately equal to DPI (dots per inch) where the dots are dots of ink.
Most e-reader displays do not even have an aspect ratio that matches the physical print so diagrams will not laid out properly. None of the book mentioned in this article are even near the size of a standard e-reader with “Wavelets: A Concise Guide” having a diagonal of 11.1” and Konosuba having a diagonal of 8.9” inches neither of these books is large. Comparing Ron Larson’s Calculus book (this is most likely the calculus book you used in school. It has the circle thing on the cover has a diagonal length of 14.1”). You would be forced to get the ONYX BOOX MAX Lumi which cost $879 to read content at its native scale but, not at its native DPI. This is not to meant to be a straw man argument against eBooks this literally the only option to get near actual scale! Even if you go for a 10.3” display your breakeven point is still tripled putting you closer to nearly 80 books (if you purchased a Onyx BOOX NOVA 3).
E-readers are supposed to emulate and expand on the book experience yet they are unable achieve this goal. When you tap a large book you get a nice thud. Books can have beautiful embossed covers and marbled pages. Book can have glossy art pages that shimmer in the sunlight. Books edges can be covered in beautiful gold pigments. Book are not merely an utilitarian way to shovel unstimulating text into your brain. Reading is an experience driven by the text, layout, paper choice, color and gray tones. Experience is an important part of a book and cannot be discounted!
The majority of e-readers do not have color and the ones that do have color displays only have low PPI in the color mode. E-readers are unable to model the different characteristics of papers like its whiteness, reflectivity, texture and composition. Books that use mixed paper types ,like Konosuba does -with its glossy color inserts, cannot be replicated on e-readers. E-readers whole purpose is to replicate text better than traditional emissive displays and they fail at this simple task.
Books Mentioned in this Article
Konosuba: God’s Blessing on This Wonderful World!, Vol. 9 (light novel) by Natsume Akatsuki
- ISBN-13: 9781975332334
- Dimensions (LxWxH): 8.2″ x 5.5″ x 0.85″ (208mm x 140mm x 22mm)
- Weight : 6.9oz (196g) (from amazon)
Wavelets: A Concise Guide by Amir-Homayoon Najmi
- ISBN-13: 9781421404967
- Dimensions (LxWxH): 9.25″ x 6.125″ x 0.75″ (235mm x 156mm x 19mm)
- Weight : 15.5oz (439g) (from kitche scale a Taylor model 5257576 )
Calculus of a Single Variable 8th Edition by Larson, Hostetler & Edwards
- ISBN-13: 9780883903469
- Dimensions (LxWxH): 10.5″ x 7.75″ x 1.25″ (267mm x 197mm x 32mm)
- Weight : 4.5lb (2 kg)
It should noted that I dispute the thickness of Konosuba Volume 9 because in real life it smaller than Wavelets. I personally measured the thickness of Wavelets the rest comes from amazon from all the books.
Additional Reading
- Electronic Books – A Bad Idea by Jakob Nielsen
- Kindle 2 Usability Review by Jakob Nielsen
- Do we really want vendor lock-in with e-book readers? by Peter Cohen
- 7 Reasons Why Ebook Sales Are Falling–and Print Book Sales Are Rising Again by Glenn Leibowitz
- Reasons for Declining Ebook Sales: My Update on the Ebook Industry, and Musings on My Participation in it by Harma-Mae Smit
In technology the same goals can be achieved in many different ways. There is a near infinite possible ways to communication between two devices. All you need to do is look to the past to see how many different keyboard and mouse connectors existed. We currently live in a unique time when you can just walk into almost any retailer buy a computer, a random computer keyboard, a random musical keyboard and a whatever computer mouse they are most likely going to work with each other. This capability to connect two random devices is the result of good and robust standards. In this article I am going to discuss two standard MIDI and DVI. MIDI is the undisputed greatest digital communication standard ever invented. DVI is a failure of a standard that even now is unable to live up to the goals it set out to achieve in its standardization.
What do these standards do?
Here is a snippet from “The Complete MIDI 1.0 Detailed Specification Revision 96.1” which explains the purpose of midi.
“The Musical Instrument Digital Interface (MIDI) protocol provides a standardized and efficient means of conveying musical performance information as electronic data.”
The Complete MIDI 1.0 Detailed Specification Revision 96.1
Musical performance information is a very broad term, because this information can include volume level, timing between equipment and even custom messages that can contain anything the user wants.

DVI is explained in a snippet from “Digital Visual Interface Revision 1.0” which was released 02 April 1999.
“The Digital Visual Interface (hereinafter DVI) specification provides a high-speed digital connection for visual data types that is display technology independent. The interface is primarily focused at providing a connection between a computer and its display device.”
Digital Visual Interface Revision 1.0
DVI supports analog video in some form factors so its own description is not quite right.
What did they set out to do?
To determine if a standard is successful we need to know what each standard was trying to achieve. To start we will look at what DVI was trying to achieve.
Another snippet explaining the goals from “Digital Visual Interface Revision 1.0”
The DVI interface enables:
- Content to remain in the lossless digital domain from creation to consumption
- Display technology independence
- Plug and play through hot plug detection, EDID and DDC2B
- Digital and Analog support in a single connector
Now for understanding the goals of MIDI we need to look at the document that inspired MIDI a paper presented at the 70th AES convention. The proposal written by Dave Smith and Chet Wood of Sequential Circuits for what was then dubbed “The ‘USI’ , Or Universal Synthesizer Interface” along with comments from others in the industry would evolve into MIDI.
“The Universal Synthesizer Interface is a specification designed to enable inter-connecting synthesizers, sequencers and home computers with an industry-wide standard interface.”
“Expandability – The interface is designed to be expandable. …. A status byte extension capability is provided.”
The ‘USI’ , Or Universal Synthesizer Interface
This means that you can make a new special code for a expanded functionality on MIDI. Also, a very important thing about MIDI is the device only responds to status bytes (messages types) it understands and ignores the rest. Which mean you can implement a new version of MIDI without worrying about causing glitches in older devices.
What makes a standard successful?
The success of a standard will be judged by the following criteria.
- Does the standard achieve the goals of its development?
- Does the standard have longevity?
- How long was in common usage?
- Are products still sold today using the standard ?
- Does it still exist in compatible revisions (eg. USB-C is not physically compatible with USB A but, is electrically compatible so that with a simple passive adapter you to connect a USB A to USB-C). This revision cannot be developed by an standards body that was independent of the original standards body.
- Does the standard provide easily accessible and affordable documentation?
When is a standard super seceded?
So lets quickly discuss why a standard gets super seceded even after achieving adoption and could be considered successful. The standard could lose market share, because it is implemented poorly, it is expensive to implement, it does meet design goals and finally it does not evolve with the application.
So lets start with MIDI does it meet design goals
- Enable inter-connecting synthesizers, sequencers and home computers with an industry-wide standard interface
- Expandability
We can see from the 8-bit guy’s video on the Roland MT-32 MIDI sound module that it was used to create music in early computer games and that computers where capable of sequencing of MIDI. Atari consider MIDI so important it was installed by default on their computers. These facts alone show that it achieved its first goal.

Now its second goal expandability the example I give of this is the first commercially successful digital synthesizer the Yamaha DX7. A video YouTuber—AudioPilz—in his Bad Gear series covered this synthesizer. It had a difficult programming interface, but could be controlled by using third-party editor programs. Literally the only feature that cannot be controlled by MIDI on the Yamaha DX7 is the power switch. I found this out by skimming the manual for the Yamaha DX7 on a Website called DX7 SYSEX. SYSEX is how the DX7 manages it audio presets and how MIDI expansion is implemented. SYSEX is stand for System Exclusive message which described in “The Complete MIDI 1.0 Detailed Specification Revision 96.1” as
“System Exclusive Messages System Exclusive messages may be used to send data such as patch parameters or sample data between MIDI devices. Manufacturers of MIDI equipment may define their own formats for System Exclusive data. Manufacturers are granted unique identification (ID) numbers by the MMA or the JMSC, and the manufacturer ID number is included as part of the System Exclusive message. The manufacturers ID is followed by any number of data bytes, and the data transmission is terminated with the EOX message. Manufacturers are required to publish the details of their System Exclusive data formats, and other manufacturers may freely utilize these formats, provided that they do not alter or utilize the format in a way which conflicts with the original manufacturers specifications.
Certain System Exclusive ID numbers are reserved for special protocols. Among these are the MIDI Sample Dump Standard, which is a System Exclusive data format defined in the MIDI specification for the transmission of sample data between MIDI devices, as well as MIDI Show Control and MIDI Machine Control.”
The Complete MIDI 1.0 Detailed Specification Revision 96.1
To Long; Didn’t Read (TL;DR)
- Manufacturers are allowed to send custom MIDI messages when using System Exclusive messages
- Manufacturers must use their Manufacturer ID number for their custom messages
- Manufacturers must publish their System Exclusive Messages it cannot be secret
- Manufacturers must allow other manufacturers to utilize these formats freely (without royalty) ,but they cannot alter their original manufacturers message format.
The beauty of SYSEX expansion method is that MIDI can be expanded without having to issue new revisions or modifying the baseline messaging format. Which means you can run multiple custom SYSEX message formats into a synthesizer that is unaware of these new format’s existence and experience no random misfiring or adverse affects. MIDI clearly met its design goals and has many features that are not discussed.
Does MIDI have longevity? MIDI is still in common usage considering it is nearly impossible to purchase a computer, tablet or cellphone that does not support MIDI suggest that it is in common usage (Supported by Windows, MacOS, Linux, Android and iOS). Considering that a whole industry of MIDI controller and keyboard exist it definitely still sold today meeting the second condition of a longevity. Also the standard has released its first ever revision of version 2.0 which was released January of 2020 expanding the capabilities while maintaining backwards compatibility. MIDI documentation is freely available to anyone who creates a registered account with the MIDI association. Clearly MIDI is a successful standard that is willing and able to evolve with user use cases and is easy enough to implement that people have been implementing it for decades.

Now Lets discuss DVI first does DVI achieve the goals of its development. First lets see its goals
The DVI interface enables:
“
- Content to remain in the lossless digital domain from creation to consumption
- Display technology independence
- Plug and play through hot plug detection, EDID and DDC2B
- Digital and Analog support in a single connector
“
DVI manages to achieve goals one through three flawlessly but, fails at the fourth goals which means it fails it objective. The reason it fails is because not every cable is able to achieve this performance. The DVI working group decided to spread these capabilities across five cables (there is only one midi cable).
DVI-A – it is a fully analog VGA compatible DVI cable. So it is a funny shaped VGA cable.
DVI-D Single Link – Fully digital but, half the bandwidth of a dual link DVI
DVI-D Dual Link – Fully digital and can achieve the maximum bandwidth of the DVI specification.
DVI-I Single Link – Analog and digital but, only supports half of the maximum bandwidth of the digital DVI section
DVI-I Dual Link – Analog and full digital bandwidth
There should of only ever been Dual link DVI-I cables this meets the design goals and is less confusing for consumers. This cable would have been more expensive but, cost is controlled by difficulty of manufacturing and number of units made. Less units are made when you have five different types of cables. We can see this in terms of cost on Monoprice for a 6ft VGA cable, a Dual link DVI-I and an HDMI 2.0 cable. The VGA cable was the cheapest at $4.09 USD . The HDMI 2.0 cable cost $6.99 USD and the dual link DVI-I was the most expensive at $9.15 USD. The HMDI 2.0 supports higher resolution and refresh rates than DVI but, cost less showing the benefits of economies of scale.
The longevity of a standard is how long it was in common usage. We can answer this question with an Intel press release which said AMD would remove DVI-I in 2015 and hinted at that Displayport would allow the features of the future. So lets say 1999-2015. Are products still being sold using the standard lets check the top 10 best selling products on Amazon & Bestbuy.

Abbreviation | Resolution |
FHD | 1920×1080 |
QHD | 2560×1440 |
UHD | 3840×2160 |
Example FHD@75HZ -1920×1080 at 75Hz
Amazon Best Seller | |
1 | Acer SB220Q FHD@75Hz (HDMI & VGA) |
2 | HP 24mh FHD@60Hz (Displayport,HDMI & VGA) |
3 | Sceptre E205W-16003R -FHD@60Hz (HDMI & VGA) |
4 | Sceptre C275W-1920RN FHD@75Hz (HDMI & VGA) |
5 | HP VH240a FHD@60Hz (VGA & HDMI) |
6 | Asus VG248QG FHD@165Hz (HDMI,DVI-D & Displayport) |
7 | LG 27GL83A-B 1 QHD@144Hz (Displayport & HDMI) |
8 | Dell SE2719H FHD@60Hz (HDMI & VGA) |
9 | AOC C24G1 FHD@144Hz (HDMI, Displayport, DVI & VGA) |
10 | Ben Q GW2480 FHD@60Hz (VGA,HDMI & Displayport) |
Bestbuy Best Seller | |
1 | HP 6XJ00AA#ABA FHD@60Hz (HDMI & VGA) |
2 | LG 24ML44B-B FHD@75Hz (HDMI & VGA) |
3 | Lenovo Q24i-10 FHD@75Hz (HDMI & VGA) |
4 | Dell S2319NX FHD@60Hz (HDMI & VGA) |
5 | HP X24ih FHD@144Hz (HDMI & Displayport) |
6 | HP V21 FHD@60Hz (HDMI & VGA) |
7 | Acer K242HQL FHD@60Hz (DVI,HDMI & VGA) |
8 | HP OMEN QHD@165Hz (Displayport & HDMI) |
9 | Samsung UR55 UHD@60Hz (HDMI & Displayport) |
10 | LG 27GN850-B QHD@144Hz (HDMI & Displayport) |
Seven of the twenty monitors would not have been able to hit their refresh rate targets using DVI. However, their is no resolution that DVI could not achieve in best selling monitor. The rest of the monitors normally would have a combination of HDMI and VGA which could of easily been driven from a single Dual link DVI-I port on the back of the monitor. For 65% of the best selling monitors on each website a single Dual link DVI-I port would work with no loss in capability. DVI also appeared on some of these monitors which seems pointless considering HDMI can handle a DVI-D signal. No revisions of DVI exist because, the standard body no longer exist and the website is not even up. In an HP white paper from March 2011 title “An Overview of Current Display Interfaces” the body was defunct and had not met since 2006. This meant that they could never adopt apple’s Mini DVI like Mini display port was made a standard. Also HDMI is not a continuation of DVI because, it was not released by the same standards body. According to the same HP white paper as quoted below.
“ After the DDWG failed to agree on a specification for a consumer version of the DVI specification, Silicon Image formed a new consortium to develop a digital interface specifically for the consumer TV market, this time with six leading CE companies. The result was the High Definition Multimedia Interface, or HDMI.”
“An Overview of Current Display Interfaces” by HP
As, the website is not up I am not able to tell the standards cost or distribution or any known funding model. DVI is a failure because, it did not accomplish its goal in most cases and did not adapt to the needs of the application.
Who is the standards body?
An important part of a successful standard is not just well thought out technical details, but ways to maintain and update standards. Standards like DVI where superseded because, they where never maintained or updated. To understand what makes a good standards body we must look at the difference between MIDI Manufacturers Association (MMA)/ MIDI Association and Digital Display Working Group (DDWG) .
We will start with the MIDI Association which has a clear financing model. Corporate members pay to support the organization and get to input into the future MIDI standards and branding. The MMA does not have different tiers of corporate membership everyone pays a fee based on yearly sales and is an equal member. The MIDI Associate also has an individual membership which is free and allows user to get standards for free but, does allow them to supply any input.

As far as I can tell the DDWG does not have a way for anyone to apply to join the organization. You were allowed to join the DDWG Implementer forum (DDWG-IF). The DDWG-IF has two membership tiers Promoters which are made up of the founding members and Adopter which is anyone who joins the DDWG-IF. Their is no way to become a promoter according to “DIGITAL VISUAL INTERFACE SPECIFICATION REVISION 1.0 ADOPTER’S AGREEMENT”. The agreement also needs the sign off from a promoter. Joining the DDWG-IF does not give you any right to influence or change the standard. The only people who could change the standard appear to be the Promoters. This is very different from the MIDI Associate which is probably the reason that only the initial version of the standard was released. Since adopters where not allowed to provide any input and where “Not Partners” as stated in the adopter agreements. The other problem with this approach is DDWG does not have any administrative authority.
“While the Promoters may select an entity to handle certain administrative tasks for them, no party is authorized to make any commitment on behalf of all or any of them.”
Except from DIGITAL VISUAL INTERFACE SPECIFICATION REVISION 1.0 ADOPTER’S AGREEMENT
So then is authorized to implement a revise DVI standard in the first place? Also, DDWG did not even process memberships they had Promoters (founding members) process memberships. As an adopter you do pay a membership fee, but what these funds do is very unclear.
Why you should care!
You should care, because bad standards do not die! You can still buy computer monitors with DVI connections even when a standards body doesn’t exist. As much as I trashed DVI when HDMI was in developed DVI was the basis. DVI was a standard with no user feedback and many of the problems with HDMI inevitably come from this problem. Standards get rolled into each other and become new standards.
DVI is by far not the worst standard in existence some successful standards are written to advantage and/or require you to use a specific companies products. DVD-Video is a prime example of biased standards. This fact can be demonstrated by the archive of “DVD FAQ” of videodiscovery.
“
A disc can have up to 8 audio tracks (streams). Each track can be in one of three formats:
- Dolby Digital (AC-3): 1 to 5.1 channels
- MPEG-2 audio: 1 to 5.1 or 7.1 channels
- PCM: 1 to 8 channels.
Two additional optional formats are supported: DTS and SDDS. Both require external decoders.
The “.1” refers to a low-frequency effects (LFE) channel that connects to a subwoofer.
…
Discs containing 525/60 (NTSC) video must use PCM or Dolby Digital on at least one track. Discs containing 625/50 (PAL) video must use PCM or MPEG-2 audio on at least one track. Additional tracks may be in any format. Many MPEG-2 discs include Dolby Digital.
“
For the case of the North American NTSC Disc one of the required tracks was Dolby Digital meaning to be compliant with the standard you had to include a Dolby Decoder while you did not have to include a decoder for the competing standards DTS or SDDS. It is also mentioned that many MPEG-2 Disc had Dolby Digital tracks as well. This gives Dolby a massive advantage over its competitors since it mandatory or effectively mandatory in most cases. Dolby’s competitors are now expensive questionable additions on a player.
If you are a product designer or a customer who want to have choice in what they buy then you care about standards. If you want new cutting edge features then you care about standards. Some features can not be added without violating standards. The future direction of technology depends on good standards that allow for growth and creative approaches to problem solving.
See Also
Sources
- The Complete MIDI 1.0 Detailed Specification version 96.1 third revision
- Digital Visual Interface DVI revision 1 released 02 april 1999
- “The ‘USI’ or Universal Synthesizer Interface” by Dave Smith and Chet Wood
- The Best Sound for MS-DOS Games – Roland MT-32
- Bad Gear – Yamaha DX7 – Synth Of Fear???
- Manual for DX7 with midi format documentation access from dxsysex
- MIDI Support by OS
- Details about MIDI 2.0™, MIDI-CI, Profiles and Property Exchange
- Monoprice access Jan2 2021 at 3:36Pm EST products numbers: 614, 85 & 3992
- Leading PC Companies Move to All Digital Display Technology, Phasing out Analog (an Intel press release)
- “An Overview of Current Display Interfaces” (an HP white paper) released March 2011
- About page on midi.org access jan 3 2021
- http://www.ddwg.org/jointheif.asp wayback archive from july 2012
- http://www.ddwg.org/lib/DDWG_Adopters_Agreement.pdf wayback archive from march 17 2012
- http://www.videodiscovery.com/vdyweb/dvd/dvdfaq.html wayback archive from march 29 1997
Image Sources
- https://www.sweetwater.com/insync/midi-interface-buying-guide/
- http://www.vintagesynth.com/yamaha/dx7.php
- https://www.amazon.com/dp/B07CVL2D2S?tag=gr-single-product-review-20&linkCode=ogi&th=1&psc=1
- https://www.midi.org/midi-articles/details-about-midi-2-0-midi-ci-profiles-and-property-exchange
- https://web.archive.org/web/20120717012946/http://www.ddwg.org/
In this this article all data will be referenced in bits using decimal prefix aka 1 kilobit (kb)= 1000 bits and 1 byte=8bits
In music you hear the debate compressed versus lossless music. Audiophiles will say that you can hear the difference between compressed and uncompressed music. Then someone runs a listening test such as ABX or others outlined in standards such as ITU P.800 (it is a free download), which quantitatively generate no difference to minimal difference in the metrics between compressed and uncompressed media. I am going to start with the basics what is compressed music?
Lesson One
There are three ways to store media content uncompressed, losslessly compressed and lossly compressed. First, uncompressed is the simplest way to store media. It is the exact captured content stored in a file unaltered. Some example of uncompressed audio format include WAV and AIFF. Second, lossless compression is a format that is reduced in sized but, when uncompressed the media is numerically identical and indistinguishable (aka each sample is exactly the same across the entire audio file). Examples of a lossless audio format are FLAC and ALAC. The third and final type is lossy compression. Lossy compression causes data to be lost and is not identical to the uncompressed audio format Examples of a lossy audio format is MP3 and Ogg.
Media Storage | Uncompressed | Lossless | Lossy |
Codec Name | WAVE & AIFF | FLAC & ALAC | Ogg,MP3 & AAC |
Common File Ending | .wav, .aiff or .aif | .flac & .m4a | .ogg, .mp3 & .m4a* |
*.m4a is a container and can hold an audio file encoded in losslessly or lossyly
Lesson Two
The quality of a lossy audio format is determined by a comparing how indistinguishable the sound or metrics of the compressed audio to the uncompressed audio in a listening test or listening test derived metric. It is important to know, this is relatively complex to truly achieve. Agata Rogowska investigated how using different musical instrument samples (stimuli) where distinguishable based on LAME-MP3 (the MP3 encoder used by audacity) encoding. He found the low bitrate (32kbps) or highly compressed MP3s of horns from the Sound Quality Assessment Material recordings for subjective tests (EBU-TECH 3253) was indistinguishable from the uncompressed horn sounds.
However, a viola sound could be distinguished almost 95% of the time [1]. Even when moving to the generally recommended bitrate of 128kbps [2] Viola was still distinguishable 65% of the time. This is where 50% distinguishable would be a random guess. Almost two years after this investigation, the creator of MP3 would recommend that people stop using MP3 [3].
Lesson Three
The question that is overlooked in this discussion —Why are we even using compressed audio formats? Internet speeds having drastically increased since the MP3 codec was introduced in 1993 . Storage cost have drastically decreased. For instance, according Matthew Komorowski article, “A History of Storage Cost (Update)” it cost $250 per Gigabits ($2000 per GB) and $0.00375 per Gigabits ($0.03 per GB) in 1993 and 2015 respectively. Let’s examine storage cost of a CD, a CD at max can hold 80 minutes of audio. It cost a little bit less than $0.03 dollars to store a CDs worth of audio and under a $0.01 to store a CD as a 128kbps mp3. Now, how much does it cost to transfer the content of a CD over the internet? That can be estimated two ways. The first way is by using the data cap for Xfinity home of 9.6 per Terabits (1.2 TB [5]); then find the minimum internet speed to be able to hit the datacap in an average length month. which is about 4 Mbps. The slowest internet speed that Xfinity provides is up to 25Mbps for $49.95 per month (retrieved August 30th 2020). That means at maximum it cost $0.0005 per Megabit to transfer the content of a CD over the internet. The second way to determine the cost per megabit is to determine the maximum data that could be possibly be transferred to a home in an average month, then determine based on the monthly cost how much it would cost. As you know, for an additional $30 per month you can remove the data cap [5]. The fastest plan Xfinity provides is a 2 Gbps plan for $299.95 (retrieved August 30th 2020) plus the additional $30 per month. Using this calculation, it comes out to $0.00000000006 per megabit to transfer the content of a CD over the internet. Continually focusing on cost, the average active Spotify user listened to 25 hours of music per month in 2017 [6] which would cost (internet transfer cost can be neglected) $0.86-0.07 per month depending on the quality settings in storage alone. If Spotify premium were to stream in CD quality uncompressed it would cost $3.81 in storage. The average active user only listen a short time relative to the hours in an average month. Therefore, dedicate storage per user can be reduced by the ratio average hours listened to per month since it can be reused across users. This brings the uncompressed audio cost down to $0.13 per month, which is 1 percent of monthly revenue.
Media content is also streamed on Youtube, Youtube has stopped reporting streams overtime. The most recent data for peak plays in a week is Despacito by Luis Fonsi featuring Daddy Yankee is 157.307 million views in the 28th week of upload (this was captured by Engauge graph digitizer from a drawing of the plot). By manipulating the numbers, you can find that the peak plays per minute were approximately 15,600. Using youtube-dl to estimate the size (I did not download and would never condone acts of piracy) of all the □.mp4 video resolution, I summed the storage space for all supported video only tracks” and added the □”.m4a audio for each supported resolution. Youtube pre-encodes all supported resolutions [7], so in total Youtube needs 398 Mb per viewer. The social media stats website Social Blade estimates at the time of writing (september 9th 2020 noon) 7 billion views generated between $3.5-$27.8 million dollars [8]. Taking the middle of these revenue generated estimates a view is estimated to be worth $0.002. Assuming that the plays per minute all count as views. The peak plays per minute generated is $3,120 per minute in revenue and it cost $23.28 per minute (using internet bandwidth and storage) to serve the videos. Performing the same analysis with Spotify and adding raw CD quality audio (this could be shrunk using a lossless codec like FLAC), it would only take 121 Mb at a cost of $7.08 to serve versus $2.11 without lossless support. Or simply stated, it is a cost difference of 0.15% more of the revenue generated during peak playback. After examing cost, I ask the question again, why are we even using compressed audio formats? I am simply asking Streaming Services to let customers choose to listen to lossless quality audio for less than 0.2% of revenue.
Why Should I the Consumer Care?
As a consumer you should care because, you are not getting a discount on the reduced bandwidth usage or not having to ship the product. For instance searching on Amazon for Luis Fonsi’s album Vida I find that the CD cost $10.99 and the MP3 download cost $9.49. In more extreme cases ScHoolboy Q’s album CrasH Talk the CD cost $7.58 and the MP3 Download cost $9.49. In the most extreme case presto classical was hosting a sale on High Resolution Music (lossless music recorded at a higher sample rate and higher bit depth than CD) where the 96kHz 24bit was less expensive than the 320kbps mp3. Remember the lossless audio requires over 14 times (a little less when encoded in FLAC 7-8 times [9]) more storage than the MP3. The price difference between lossless and lossy codecs is relatively arbitrary. In many cases, a consumer has to buy a CD to obtain lossless audio even when they would prefer a lossless download. Why don’t music retailers and labels provide this to customers who want lossless audio?


The consumer is still bearing many hidden costs for being forced to use compressed music with non-FOSS (Free and open-source software) every device that can play these propriety Codecs is required to pay a licensing fee which is then passed on to the consumer. Compressed audio also trades storage space for power usage, as it requires more computation power to playback compressed audio.
Another hidden cost that many streaming devices and Bluetooth headphones is the native wireless protocols use lossy compression. This means that your lossly compressed audio could have sounded fine with one stage of compression but, could be drastically affected by a second round of audio compression. Also, many of the artifacts heard on internet calls are caused by lossy audio compression. Using lossy codecs increases latency by having to wait for the audio to encode and decode. This encoding time changes depending on the content which can cause video sync issues. One of the biggest shames is that we by licensing fee or government grants are still developing new lossy codecs that in general we do not need or use. Notice that most music stores sell you mainly MP3 even when the creators of the MP3 prefers that you use their new lossy audio creation [3]. What other problems could be solved instead of research new audio compression techniques for minimal gains?
References
- “Audibility of lossy compressed musical” by Agata Rogowska
- “Which MP3 bit rate should I use?” by Nate Lanxon
- “The MP3 Is Officially Dead, According To Its Creators” by Andrew Flanagan
- Matthew Komorowski personal blog “a history of storage cost (update)”
- https://dataplan.xfinity.com/faq/ (retrieved August 30th 2020)
- “Average monthly time spent listening to Spotify content among monthly active users worldwide from 1st quarter 2015 to 4th quarter 2017” by Amy Watson
- “How YouTube Works – Computerphile” on the Computerphile youtube channel.
- Social Blades statistical information on Luisfonsivevo channel
- “FLAC compression level comparison” by Nathan Zachary
People say they prefer analog to digital recordings for many reasons. Analog’s assumed superiority comes from a misunderstanding of how digital audio works and how to improve the sound quality of digital audio. The most common analog format that is listed as superior is vinyl. Vinyl has a distinctive advantage as an analog format. It is conceptually easy to understand since the movement of the needle creates a movement of air (you can even hear the music just from the needle movements alone). Improving sound quality for vinyl is as simple as following the groove of the record better, improved vibration reduction and playback speed regulation. None of these benefits and tweaks are a panacea for analog audio. The case for digital audio will be made evident in this article. It will be clear why digital audio beats analog audio of any format in recording.
About Vinyl
Many enthusiast purport vinyl is closer to the artistic intent. Throughout vinyl’s history it has been purely a mass-produced consumer standard, not a professional standard. Vinyl’s use in movie soundtracks and musical recordings was extremely short. The Fox Corporation when developing sound for “talkies”, always used optical sound. Warner Brothers movies used vinyl-based sound in the mid 1920s but, by the mid 1930s, Warner Brothers was using optical sound [1]. According to the 1952 SMPTE’s Progress Report by the end of 1951 75% of original production recording for Hollywood was done on magnetic tape. Artists have not recorded directly on vinyl for a long time; thus, violating the artistic intent argument. The ideal music listening experience would be in the control room or a live performance listening to the artist. Magnetic recordings needed to be transferred to vinyl which caused a loss in quality. Vinyl also does not allow for a separation of playback and song composition. So, if you have a continuous piece that is longer than a single side, it has to be interrupted. Tracks are separated on vinyl records by periods of silence but, songs without silence separating them have become very common in the digital era. These fully integrated songs becomes inseparable on vinyl records because, their is no way to denote beginning or ending. This violates the artistic intent. Lossless digital downloads meet artistic intent requirements much better. Consider Mike Dean’s Album 4:20 which was mixed and mastered by the artist. The Album was delivered directly to the consumer at no point did the album need to change formats.
Vinyl has poor value for those who want to argue collecting it is valuable (ask millennials about their beanie baby collections). An example of this: if you bought the original 1973 Pink Floyd Darkside of the Moon (UK Edition) and never played it you could get between $1,700-$3,478.26 (USD). Other than considering that an unplayed record is a waste of money since you do not get to enjoy the music, but this is beside the point. Assuming this record cost $5 (an inflation adjusted version of cost of Tom Petty’s Hard Promises price dispute with his label MCA). looking at the Dow Jones in March 1973, it was at $951.01 while in March 2020 it was $21917.16. If you invested the money instead you would only get $115.23—in this case preserving the Pink Floyd album would have been a worthwhile investment. However, if you pick another popular album from the billboard charts in 1973 Houses Of The Holy (UK Edition) today it could net you between $100-$200. Houses Of The Holy is not a clear winner you can sell at the upper end and it would be worth it otherwise not so much. If you selected to hold on to Don’t Shoot Me I’m Only The Piano Player (UK Edition) you could get between $19-$35—in this case the stock market would have been a way better bet, sorry Elton. All prices were taken from Discog (accessed on April-May 2020—only mint or near mint where used). The only reason the number one album of 1973 The World Is a Ghetto by WAR was not used was because no mint or near mint version could be found for sale.
The following video illustrates some of the deficiencies of vinyl. Ironically, this video was about why vinyl so great, yet it is one of the best explainers of the format’s shortcomings.
TLDW (To Long Didn’t Watch)
- Requires lots of processing of the original recording before you can even start cutting
- To cut a vinyl record you need to limit the bandwidth that your music uses
- Limits the slew rate of your music to preserve the cutting head not your music quality
- “It limits you to what you actually can do”
- Forces you to put the bulk of your content in the midrange frequencies instead of having a full range composition
- Depending on your sonic decisions your playtime will be adversely affected (having more bass reduces your playback time and so does having a louder record)
- To fit on a record your music will need to take on the vinyl sound but not necessarily your mixed sound
Benefits of Digital Audio
Now finally: why should you use digital audio? The story of digital begins with the return to the direct vinyl recording in the late 60’s and early 70’s. Why, after forsaking recording to vinyl directly would direct disc mastering exist again? What problem does this solve? Direct disc recording fixes modulation noise! Modulation noise is unique to magnetic tape and it was very hard to measure.
Modulation Noise – A noise which exists only in the presence of a signal and is a function of the recorded amplitude of the signal. The signal undergoes modulation in both amplitude and frequency. These two types of modulation produce sidebands above and below the signal frequency.
Let’s Breakdown what this definition means. The first sentence means that modulation noise only occur sonce audio has been recorded to the magnetic tape. A tape that has no audio recorded will still have a slight hiss called surface noise, but then will gain additional noise from the modulation noise once a signal is recorded to the magnetic tape. A perfectly noiseless tape will-which doesn’t exist-still have modulation noise. This sentence also means that as the amplitude of the signal increases or decreases the characteristics of modulation noise changes. So, recording louder will reduce surface noise but, will cause an increase in modulation noise. The second sentence means that the frequency and amplitude of a signal is varied over time. In the case of our examples a pure tone with a constant frequency 440Hz that undergoes frequency modulation will now have a frequency that varies from 429Hz to 451Hz. In case with amplitude modulation the amplitude of the signal is varying over time from being multiplied by 1 and 0.98 (which is below human perception of 1.1 to 0.89 if the amplitude was 1). Amplitude modulation of two pure sine waves produce what are known as side bands which means now additional tones are produced at 380Hz (440Hz-60Hz) and 500Hz (440Hz+60Hz). These extreme examples are just with these parameters turned up. The modulation noise uses a combination of FM and AM modulation. The rate of frequency modulation mirrors a Mark Levinson No 515 turntable and the amplitude is modulated by randomly generated white noise to mimic uneven gain size.
Below are auditory examples for different types modulation
Track List
- Unmodulated Sinewave
- AM (Amplitude Modulated) Sine Wave
- FM (Frequency Modulated) Sine Wave
- A Sinewave with simulated Modulation Noise
- AM (Amplitude Modulated) Extreme Sine Wave
- FM (Frequency Modulated) Extreme Sine Wave
Frequency Modulation (FM) Example


Amplitude Modulation (AM) Example


Simulated Modulation Noise


AM Modulation Extreme Example

FM Modulation Extreme Example

The source of these errors came from a non-uniform distribution of iron oxide grain size on the magnetic tape. This is because, the iron powder cannot be ground to perfectly even sized grains. Magnetic tapes work by gluing a fine magnetic powder to the tape material. Powder application tends to be inconsistent causing level inconsistencies. This is very similar to gluing glitter to a sheet of paper. Some glitter always comes off. The frequency modulation comes from the imperfect speed regulation of the tape machine [3]. It is noise generated at some frequency multiple of the incoming signal. Modulation noise destroys your low frequency response.
Early Digital
Nippon Columbia, or Denon, wanted to improve their recordings’ quality. In fact, according to Audio Restoration expert Thomas Fine, Denon, when pursuing digital, had one goal: “To produce recordings that were not compromised by the weaknesses of magnetic tape” [4]. Early digital was impossible to edit so it had very limited use cases, but direct disc masters could not be edited either so most early digital recordings where done as backups to direct disc recordings. Denon’s using NHK’s (NHK is like Japanese BBC) early prototype can be heard on the Steve Marcus + Jiro Inagaki & Soul Media recording of “Something”. Honestly, the recording sounds fantastic, maybe a little limited on the high frequency but that’s the only fault I hear.
From what I can tell this was recorded on the prototype NHK system, so it had a sample rate of 30kHz and a nonlinear 13-bit depth, that was built in 1969. Sample rate is how often an incoming analog signal is measured per second. So a sample rate of 30kHz means that the incoming audio signal is measured 30,000 times per second. While bit depth is how many unique values the sampled amplitude can take on in this case 8192 unique values. However, unlike a CD the spacing between amplitude values is nonlinear by design in the NHK prototype. See additional reading for further explanation of digital sound.
Early digital audio was focused on providing the best audio to consumers. Soundstream–another early digital pioneer located in Salt Lake City, Utah–was used as a backup to a direct disc recording for a Virgil Fox session. Jerry Bruck, the recording engineer on the session, remembers receiving a call from Soundstream founder Thomas Stockham that the analog feed had a hum problem -80dB down from peak amplitude which was well into the noise floor of the best recorders of the late 70’s. The best tape recorders with signal to noise reduction could only produce 70dB of signal to noise so this hum was inaudible on tape. This shows you one of the unintended consequences of digital recordings was improved signal-to-noise ratios. As far as I can tell no one really thought that improving the signal-to-noise ratio was that important as modulation noise was already removed. To put this into perspective, using OSHA’s guidelines for noise exposure, no worker should be exposed to an average level over 85dBA for a whole day without hearing protection.
If we make your playback level 85dBA then subtract the 70dB signal to noise ratio (I know it is not A-weighted) this gives us the noise floor of 15dBA that is needed to hear this dynamic range during playback. This noise floor is quieter than a quiet study room (20dBA) and the North Rim of the Grand Canyon (30dBA) [6]. In most playback cases, vinyl records and magnetic tape are effectively noiseless. Soundstream’s test recordings were good but had limited high frequency content due to only sampling at 37.5kHz (the true maximum frequency was around 15kHz for anti-aliasing purposes) with 16bits. This early recording allowed Soundstream to be co-booked for many more recordings. Telarc Records was impressed by these early Soundstream recordings but found the high frequencies limited and asked for an improvement which Soundstream agreed to do. Soundstream bumped the sample rate up to 50kHz.Telarc Records and Soundstream worked with band director Fred Fennell to record an assorted set of pieces. Band director Fred Fennell who directed the live band could only say the words “wow” when he heard the digital recording of the session [4]. Telarc Records is famous for its bass drum sounds but, this is extremely difficult to cut into vinyl records but, trivial for digital recordings. If you own a copy of Telarc’s Stravinsky’s Firebird Suite & Borodin’s Polovtsian Dances on Vinyl and love its rich warm sound, then you agree with me that digital recordings sound excellent.
Digital Misconceptions
Now it’s time to discuss digital misconceptions. A common misconception – with sampling is you are losing information. This is not true. An example, let’s say we want to represent data from a single linear line and we need to pick a way to preserve information about the line. We can choose between two representations A or B which contains more information?
Representation A: A set of points
X | Y(x)=2x |
---|---|
0 | 0 |
1 | 2 |
2 | 4 |
3 | 6 |
4 | 8 |
5 | 10 |
6 | 12 |
7 | 14 |
8 | 16 |
9 | 18 |
10 | 20 |
11 | 22 |
12 | 24 |
13 | 26 |
14 | 28 |
15 | 30 |
16 | 32 |
17 | 34 |
18 | 36 |
19 | 38 |
20 | 40 |
21 | 42 |
22 | 44 |
23 | 46 |
Representation B: An Equation
Y(x)=2x for x equal 0 to 23 with a step size of 1
Neither representation contains more information about the line data. Each can be used to create the other’s representation. Even if we took smaller and smaller steps both would remain interchangeable. This is the same with sampled audio. There is a lot of discussion about how digital audio cannot represent the complex overtones of an instrument because, they surpass the audio bandwidth of even the 192kHz. Can analog recordings represent these higher frequencies? Let’s say you are working on capturing an ensemble containing a full orchestra–a Buchla Music Easel and Minimoog Voyager (both are fully analog synthesizers)–directly to vinyl. You use Earthworks QTC50 omni microphones for orchestra and room microphones. While you record the soloist with a Neumann TLM103 you connect everything to a fully analog SSL XL-Desk. First, the Earthworks QTC50 begins to roll off at 50kHz[7] and both analog synthesizers can produce harmonics past 20kHz but the Neumann TLM103 can only produce up to audible content up to 20kHz [8]. Your nice SSL XL-Desk [9] has an upper frequency response of 40kHz which is less than your QTC50 so you’re actually rolling off the high end of the microphone. While if you recorded this signal digitally at 192kHz all of this would be preserved. All physically realizable devices have to have finite decaying bandwidths or else they would require infinite amounts of energy to power (look at definition of energy in signal processing and divergent infinite series to gain an understanding of this point or Technology Connection’s Videos ).
Conclusion
Digital audio does not sound worse than analog. Digital has many advantages over analog from degradation-free copying to longer playback times. You will not gain any benefits from playing audio back using analog media–at best it will sound the same. The collectors’ value of analog media is minimal and speculative at best on average. Analog audio limits artistic expression and can conflict with artistic intent. In particular, vinyl has been a purely consumer audio standard for decades and has not factored in music capture, reproduction or production professionally in a long time. If you enjoy maintaining vinyl and the comradery you gain from being in the analog community that is fine. But you will gain nothing in sound quality from analog music playback compared to digital music playback. You will also lose many of the benefits that digital provides like tracks seeking, glitch free playback, error correction, easy archiving and organizing.
Additional Reading
- Xiph’s Digital Show and Tell Video
- Xiph’s 24/192 Music Downloads…and why they make no sense
- “Introduction to Signal Processing” by Sophocles J. Orfanidis
- Technology Connection’s “Nyquist-Shannon; The Backbone of Digital Sound”
- Technology Connection’s “Sound By Numbers: The Rise of Digital Sound”
Reference
1. “Motion Picture Sound Recording “ by John G. Frayne
2. “The 35mm Album Master Fad” by Thomas Fine presented at the 135th AES Convention
3. “Modulation Noise in Tape Recordings” by Robert Z. Langevin
4. “ The Dawn of Commercial Digital Recordings” by Thomas Fine
5. “An Historical Overview of the development of PCM/Digital Recording Technology at Denon” by Almon H. Clegg et al
6. “How loud is too loud?” by OSHA (https://www.osha.gov/SLTC/noisehearingconservation/loud.html) accessed June 7th 2020
7.Earthworks QTC50 specification (https://earthworksaudio.com/products/microphones/qtc-series/qtc50/) accessed June 7th 2020
8. Neumann TLM103 specifications (https://en-de.neumann.com/tlm-103#technical-data) accessed June 7th 2020
9. SSL XL-Desk Owner Manual accessed June 7th 2020
There is a general consensus among the left, right and environmentalist is that increasing the amount multifamily residency (which includes apartments, condominiums, townhouses and duplexes) is what we should be doing in the future. The argument on the left is less restrictive zoning, proper incentives for building new homes and making it difficult to hoard land. The argument on the right is that restrictive zoning law slow the develop of different types of housing and the the amount of housing. Both agreeing that single family homes and intense zoning laws are affecting the supply and affordability of housing option. The environmental case for increasing housing density is it will make public transportation easier to facilitate. The increased residential density will reduce commuting time and distance. These are all excellent reasons to support multifamily residences . However no one is addressing the downside that many multifamily residences have which is poor sound isolation which is essential to proper living conditions.
Residential Noise Disputes
I myself have worked on multiple noise disputes they are complicated and extremely personal. A 2013 survey performed by the company FindLaw found that 42% of Americans where involved in a dispute with their neighbor of the disputes almost 50% where related to noise. The next category was Pets and Animals which was almost 30 percent of the complaints. I can not tell you what percentage of these dispute involved sounds the pets made. In multifamily residences this is further complicated by the fact people live in close proximity.
Noise vs Sound
The first thing to get out of the away is avoiding the term noise. To measure the level of sound produced we must use a sound level meter. No noise level meters exist. What is and is not noise is completely subjective. The only thing that can be done is to reduce the sound level that is transferred between units.
Sound Transmission
Sound can travel in two fashions airborne and structurally. Airborne sound transmission is when you can hear your neighbor talking through the walls. Structural sound transmission is when you can hear your neighbors foot steps in a far away room. Airborne transmission is when the sound travels through the air. Structural transmission is when sound travels through solid objects and then radiates from solid objects into the air. Most sound has a combination of both transmission methods.
Regulatory
How we deal with unwanted sound at a local level is very interesting here is an example from the noise ordinance (§ 74-131. Prohibited conduct paragraphs a-c).in Atlanta,Georgia. I picked Atlanta because, finding its rules would be pretty easy and I am cold.
Atlanta’s restrictions vary during the time or day of the week here is a table explain those shifts.
Sunday-Thursday | Friday-Saturday | |
7am-11pm | 300ft | n/a |
11pm-7am | 100ft | n/a |
7am-12pm | n/a | 300ft |
12pm-7am | n/a | 100ft |
Atlanta’s Single Family Zoning Laws
- Restrictions of 300 feet for 7:00 a.m. through 11:00 p.m. Sunday through Thursday and 7:00 a.m. through 12:00 midnight on Friday and Saturday.
- Mechanical sound-making devices. It is unlawful for any person or persons to play, use, operate, or permit to be played, used, or operated any radio receiving device, television, stereo, musical instrument, phonograph sound amplifier or other machines or devices for the producing, reproducing or amplifying of sound and/or noise at such a volume and in such a manner so as to create, or cause to be created, any noises or sounds which are plainly audible at a distance of 300 feet or more from the building, structure or vehicle, or in the case of real property, beyond the property limits, in which it is located, whichever is farthest, between the hours of 7:00 a.m. and 11:00 p.m. Sunday through Thursday and between the hours of 7:00 a.m. and 12:00 midnight on Friday and Saturday.
- Human-produced sounds. It is unlawful for any person or persons to yell, shout, hoot, whistle, or sing on the public streets or sidewalks or on private property so as to create, or cause to be created, any noises or sounds which are plainly audible at a distance of 300 feet or more from the place, building, structure, or in the case of real property, beyond the property limits, in which the person is located, whichever is farthest, between the hours of 7:00 a.m. and 11:00 p.m. Sunday through Thursday and between the hours of 7:00 a.m. and 12:00 midnight on Friday and Saturday.
- Commercial advertising. (this section was omitted)
- Party noise. It is unlawful for any person or persons in charge of a party or other social event that occurs on any private property to allow that party or social event to produce noise in such a manner that such noise is plainly audible at a distance of 300 feet or more from the building or structure from which the noise is emanating or in the case of real property, beyond the property limits, on which the party or social event is located, whichever is farthest, between the hours of 7:00 a.m. and 11:00 p.m. Sunday through Thursday and between the hours of 7:00 a.m. and 12 midnight on Friday and Saturday. For the purposes of this subsection, a “person in charge of a party or other social event” shall mean any adult person who resides in or on the premises involved in such party or social event and is present at such party or social event. For the purposes of this subsection, “noise” shall mean the same sounds, or any combination thereof, as described in paragraphs a. or b. above.
Atlanta’s Zoning Code For Multifamily Residency
Restrictions for areas within apartments, condominiums, townhouses, duplexes, or other such residential dwelling units. Except for persons within commercial enterprises that have an adjoining property line or boundary with a residential dwelling unit, it is unlawful for any person to make, continue, or cause to be made or continued any noise in such a manner as to be plainly audible to any other person a distance of five feet beyond the adjoining property line wall or boundary of any apartment, condominium, townhouse, duplex, or other such residential dwelling units with adjoining points of contact. For the purposes of this subsection, “noise” shall mean human-produced sounds of yelling, shouting, hooting, whistling, singing, or mechanically-produced sounds made by radio-receiving device, television, stereo, musical instrument, phonograph sound amplifier or other machines or devices for the producing, reproducing, or amplifying of sound, or any combination thereof. For the purposes of this subsection, “property line or boundary” shall mean an imaginary line drawn through the points of contact of (1) adjoining apartments, condominiums, townhouses, duplexes or other such residential dwelling units with adjoining points owned, rented, or leased by different persons; or (2) adjoining common areas or adjoining exterior walls. Said property line or boundary includes all points of a plane formed by projecting the property line or boundary including the ceiling, the floor, and the walls.
Understanding Zoning Laws
It seems relatively reasonable to keep “ sound and/or noise at such a volume and in such a manner so as to create, or cause to be created, any noises or sounds which are plainly audible at a distance of 300 feet (100ft) or more from the building….”. What does plainly audible mean is poorly defined.
However, Paragraph c reference to multifamily residences is utterly ridiculous “…to be plainly audible to any other person a distance of five feet beyond the adjoining property line wall…”. To put this in to perspective the width between to two walls adjacent to my neighbors is 15’ 6” (a little under 5 meters). Which means I have an approximately five foot (~1.5 meters) strip for which I could complain about both of my neighbors sound level. This means if I can hear sounds from my neighbor’s dwelling on my toilet I can not complain (since most most toilets are not more than five feet from the adjacent wall). In a more close to home explain here is the section on noise from my residence agreement.
“Owners, guests, and lessees will be expected to reduce noise levels after 10:00 p.m. so neighbors are not disturbed. At no time are musical instruments, radios, televisions or other sources of noise to be so loud as to become a nuisance, the determination of which is left to the sole discretion of the Board of Directors.”
In this case do not disturb anyone after 10pm seems reasonable. In my residences rules the phase “The determination of which sound become a nuisance will be left to the board of directors.” How will nuisance be determined? When writing this article I could hear my neighbor hammering was this a nuisance?
It is important to understand that an assumption that both of these noise rules have is that you understand the sound levels effect on your neighbor. In Atlanta if you are having a party and the person 300ft (~91m) away from you is also having a party what is “ plainly audible” has now changed. In the case of my residence rules what is a nuisance changes depending on the make up of my Board of Directors. Also, in the case of my residency rules if my neighbors are not home after 10pm and I have not become deemed a nuisance by the current Board of Directors it is impossible for me to violate the noise rules.
It is unreasonable to assume that you know the affect of your sound level on your neighbors. What is needed to do an assessment? The first thing you would need to know is what is the sound pressure radiating on the surface of the wall. The second thing you would need to know is the attenuation provided by neighboring walls. Third,you would need to know what is plainly audible or disturbing to that neighboring person. It is not possible to estimate the sound attenuation of a wall by just by looking at it. It is very unlikely that your neighbor will let you setup a sound level meter inside their house to determine the attenuation. It is also nearly impossible to determine the contributions from inside the neighbors environment and the contribution your making to the sound level in their environment at all times. What is plainly audible or disturbing is not based on sound level they receive from you but, the environment the sound is received in. This is why I am not a fan of noise ordinances or quiet policies because they suffer from these same weaknesses.
Legal Rights
It seems relatively reasonable to keep “ sound and/or noise at such a volume and in such a manner so as to create, or cause to be created, any noises or sounds which are plainly audible at a distance of 300 feet (100ft) or more from the building….”. What does plainly audible mean is poorly defined.
However, Paragraph c reference to multifamily residences is utterly ridiculous “…to be plainly audible to any other person a distance of five feet beyond the adjoining property line wall…”. To put this in to perspective the width between to two walls adjacent to my neighbors is 15’ 6” (a little under 5 meters). Which means I have an approximately five foot (~1.5 meters) strip for which I could complain about both of my neighbors sound level. This means if I can hear sounds from my neighbor’s dwelling on my toilet I can not complain (since most most toilets are not more than five feet from the adjacent wall). In a more close to home explain here is the section on noise from my residence agreement.
“Owners, guests, and lessees will be expected to reduce noise levels after 10:00 p.m. so neighbors are not disturbed. At no time are musical instruments, radios, televisions or other sources of noise to be so loud as to become a nuisance, the determination of which is left to the sole discretion of the Board of Directors.”
In this case do not disturb anyone after 10pm seems reasonable. In my residences rules the phase “The determination of which sound become a nuisance will be left to the board of directors.” How will nuisance be determined? When writing this article I could hear my neighbor hammering was this a nuisance?
It is important to understand that an assumption that both of these noise rules have is that you understand the sound levels effect on your neighbor. In Atlanta if you are having a party and the person 300ft (~91m) away from you is also having a party what is “ plainly audible” has now changed. In the case of my residence rules what is a nuisance changes depending on the make up of my Board of Directors. Also, in the case of my residency rules if my neighbors are not home after 10pm and I have not become deemed a nuisance by the current Board of Directors it is impossible for me to violate the noise rules.
It is unreasonable to assume that you know the affect of your sound level on your neighbors. What is needed to do an assessment? The first thing you would need to know is what is the sound pressure radiating on the surface of the wall. The second thing you would need to know is the attenuation provided by neighboring walls. Third,you would need to know what is plainly audible or disturbing to that neighboring person. It is not possible to estimate the sound attenuation of a wall by just by looking at it. It is very unlikely that your neighbor will let you setup a sound level meter inside their house to determine the attenuation. It is also nearly impossible to determine the contributions from inside the neighbors environment and the contribution your making to the sound level in their environment at all times. What is plainly audible or disturbing is not based on sound level they receive from you but, the environment the sound is received in. This is why I am not a fan of noise ordinances or quiet policies because they suffer from these same weaknesses.
Residential Rights
The problem with current noise law it is ineffective or oppressive. Quiet times should be based at a community level using survey data updated relatively frequently. Just because a person can’t afford or does not want to live in a single family residence does not mean they should have to live in silence. You should be allowed to watch a movie with your surround sound blasting. You should be allowed to hammer a nail into a wall. You should be allowed to vacuum in the morning and night in your private residence. You should be allowed to say your credit card number over the phone without the fear that anyone can hear you. You should be able to sleep in on the weekend without hearing your neighbors footsteps or door closing. Noise law should be focused on making sure that both the public and private nuisances are protected. Public nuisances are threats health or well being of the community. A private nuisance is unreasonably or illegally interfering with some ones right to enjoy property. You should not annoy your neighbor but, you should not be harassed by your neighbor for any sound that you make. If you can not watch a movie or play an album loud on your own private property without the threat of action from your neighbors then your neighbor has become a nuisance. Below are solutions that I recommend.
Building Solutions
- Every adjacent wall at minimum can be STC (Sound Transmission Class) 55 and should really be STC 60 and above measured in place (I would prefer these in OITC but, not as much information on partition that meet OITC (Outdoor–Indoor Transmission Class) ratings is provided)
- All stairs and floors need to be acoustically isolated from surrounding structure. Preferred is that the walls between buildings dwelling be physically separate with each wall on a separate stubs.
- All heating and cooling should be installed by default or in the case of air conditioning ducted if not already installed
- All already built buildings should be forced to disclose an average STC rating with a raw 1/3 octave band going down to 80Hz for condo/apartment location (aka if you have 13 floors with the same floor plan take an average STC for rooms in that building position)
- Builders and architects should have an average residential building STC rating attached to their name.
- All signal wires should be in conduit to allow for easy replace and stop the drilling of holes in walls
- All Door need to be isolated to the same standards walls
- When multiple dwellings are in the same room they should all be soft closing door that are acoustically isolated (multiple room mates living in the same apartment)
- IIC should be a minimum of 60
Noise Regulation Recommendation
- The abolishment of all current noise ordinances.
- 30dB(A) Leq (30 decibels A-weighted equivalent continuous sound level) night time average level and 45db(A)Leq Peak during night time hours as determined by the average workers start and finish time collected by survey
- A peak daytime noise level of 60db(A) in neighboring apartments and an 45Ldn (45 decibel day-evening-night noise level) neighboring apartments.
- The rebuilding and refunding of the EPA’s ONAC (Office of Noise Abatement and Control) from the Noise Control act 1972.
- The definition of noise violation should always be inside living areas and not at property lines.
- For outside of structures but, inside the property line noise should not cause hearing loss in 24 hours average of 70dB(A).
Additional Reading
Ambient Noise is “The New Second Hand Smoke” by Daniel Fink published In Acoustics Today
City Noise Might Be Making You Sick by Kate Wagner published in The Atlantic
Sources
Legal Articles about noisy neighbor
Top 5 Neighbor Disputes and How to Resolve Them by Betty Wang, JD
Legal How-To: Dealing With Noisy Neighbors by Betty Wang, JD
Can I Sue My Noisy Neighbor? by Andrew Lu
Reduce Noise Complaints With a Quiet Hours Clause by Erin Eberlin
Housing Articles
Study: Increased housing density helps combat climate change by Jeff Wattenhofer
Zoning, Land-Use Planning, and Housing Affordability by Vanessa Brown Calder
Atlanta’s Code of Ordinances section 74-131. Prohibited conduct.

Determining audio quality is one of the most difficult questions I receive about consumer electronics. People will ask me “which headphones should I buy? What are the best speakers to get? Can you recommend a DAC (Digital-to-Analog Converter)?”. Honestly, I do not know how to answer these questions. Sound quality is determined by a combination of factors that are both objective and subjective. However, I am going to discuss the most under discussed parameter in sound reproduction that is Engineering Tolerance. More simply stated how consistent are the products that come off the assembly line?
Every audio product will provide you with a set of measurements that are descriptive of performance. None of these measurements matter without a defined tolerance! Here is a great hypothetical example.
You look and see a banner ad on the side of you favorite audio site with the following text
“Our speaker produces 60dBSPL at 1 meter at maximum continuous output! Each speaker is matched to each other with obsessive precision! Made from the finest old growth Italian hardwood! We have a passion for the highest fidelity sound reproduction on the market!”
All your friends go to the store and pick a pair up. There is a group text among friends to discuss the performance.


You go back to the store and wait in the line for the customer service department. You get to the front of the line and explain your problem. The technician whirls the speaker around looking for damage. The technician plugs the speaker in to see if they power on. The technician goes into the back room for a while and comes back and says
“Sorry,there is nothing I can do it is within manufacturer tolerance. The amplifier works, the output is within tolerance” – Technician.
“What is the tolerance? I can’t hear anything from the speaker” – You
“Plus or minus 80dB”-Technician
which means the speakers maximum output can be either above the threshold of pain (120dBSPL) or below the threshold of hearing (0dBSPL) . In this example your speakers could be functioning as designed and be completely inaudible to you.
Albeit, the hypothetical example was extreme but, notice that the speakers can be matched with each other and the consumer can still receives poor performance. Knowing the matching tolerance is very important to understanding the performance. People who purchase the same speaker have drastically different experiences. Depending on which set they had. This is an issues with recommending products without known tolerances. Information on tolerance is very useful for comparison shopping because you can see how a set of speakers compare to each other. Tolerances can tell you if speakers actually have the same performance or are very different. When tolerance is not provided by a manufacturer it is very expensive and time consuming to determine. I generally recommend staying away from products that do not reference their tolerances.

Here are some basic terms to look out for in speakers. Measurements can be tricky.
- Distortion (THD) – The ratio between a fundamental frequency and the harmonics amplitude generated by the devices. An example would be sending 1kHz into a speaker and getting 1kHz, 2kHz and 3kHz output for the speaker these harmonics amplitudes are used to calculate total harmonic distortion. Total Harmonic Distortion (THD) is a method of quantifying how much distortion a system has. THD is the most common method report distortion but, it is not the only way. Total harmonic distortion can be calculated including the fundamental or excluding the fundamental which gives slightly different numbers. The lower the better but, it does not tell you how a speaker is distorted so it is very difficult to compare two speakers without plots of the individual harmonics.
- Frequency Response – The range of reproducible frequencies within a specific tolerance range. The range is bounded by a specific roll off frequencies which are measured at specific reductions in peak amplitudes. So a speaker with a frequency range of 20-20kHz with the roll off defined at -3dB means that at 20Hz the level is 3dB below peak acoustic output. It is important to understand that to have an even frequency response your roll off regions should be outside the range of human hearing. This is due to the fact that humans can only detect a minimum differences in amplitude of 1dB so this roll off would be noticeable to the ear. Another important point to understand is that different roll off locations can not be converted to each other as they need to be re read off the graph. This means that a -6dB roll off cannot be convert -10dB roll off without reading from the frequency response plot! The frequency response of a speaker is roughly equivalent to the piston band or the region with flat acoustic output above the speakers unique resonance. (see John Eargle’s “Loudspeaker Handbook 2nd Edition” pg.10)
- Electrical Power Consumption – A useless measurement that should be ignored! Electrical power consumption is only useful if you know the speakers acoustic output efficiency or sensitivity. Aka Power consumption ≠ Acoustic power . Most electrical ratings are reporting total power consumption in watts. A smart speaker’s power consumption includes the WiFi and other microchips power consumption as well not just the audio portion!
- Acoustic or Sound Power – The mechanical work that is done on the air by a speaker each cycle which is also measured in watts. (See Marshall Long “Architectural Acoustics 2nd edition” pg.63)
- Sensitivity – The efficiency of converting electrical power into acoustic output at 1kHz or 250Hz for sub woofers. This efficiency is reported in decibels. It is the measured sound output generated with 1W of power input at 1m. An example sensitivity would be Speaker A has a sensitivity of 94dB at 1 meter. The 1W of input is always implied. A common reporting error is to report sensitivity a fixed voltage. A fixed voltage sensitivity is wrong because the wattage at input is being changed. For instance it is common to report sensitivity at 2.83V. If the speaker has an impedance of 8 ohms at 1kHz then sensitivity at 2.83V is a true efficiency. However, if a speaker has an impedance of 4 ohms at 1kHz the reported efficiency is double the true efficiency. The sensitivity is not constant across all frequencies. (see John Eargle’s “Loudspeaker Handbook 2nd Edition” pg.11-12)
- SPL (Sound Pressure Level)- this the is a decibel scale where 0dBSPL is the threshold of hearing at 1kHz. It is a measurement of acoustic output power. (See Marshall Long “Architectural Acoustics 2nd edition” pg.66)
Now using our understanding of tolerance we can compare speakers to each other.

The criteria for a speakers in this comparison is a single speaker that was rated most popular, highest priced and or lowest price on Zzounds an e-commerce website. The last speaker in this comparison is one that I own.
JBL LSR305 | Yamaha HS8 | Genelec 8050B | JBL 305P MkII | |
ZZounds Filter Categories | Lowest | Most Popular | Highest | I own |
Price (USD) | 99.99 | 369.99 | 1,895.00 | 149.00 |
Cutoff Location (dB) | ? | -10 | -6 | -10dB |
Frequency Response (Hz) | 43-24k | 38 – 30k | 32-25k | 43-20k |
Frequency Response Tolerance (dB) | ? | ? | 2 | 3 |
Max Peak SPL (dB) | 108* | ? | ≥120 | 108 |
Max Continuous SPL (dB) | ? | 101 | 94 | |
Distortion @ 90dB (THD (%)) | ? | ? | 0.5 | ? |
Combined maximum driver Power (Watts) | 82 | 120 | 270 | 82 |
*C-weighted
First thing to notice is that the Genelec’s frequency response is not directly comparable to any of the other speakers. Yamaha makes the most popular speaker but, doesn’t provide the user with much comparison data. This makes a comparison more difficult. With our data we can conclude the maximum output of the Genelec 8050B is higher than the JBL LSR305 and 305p MKII. Also the JBL 305P MkII is merely a refresh of the JBL LSR305 which explains why they have nearly identical measurements. We cannot assume however that the Yamaha has a higher output than both JBL models. Remember power consumption is not equal to acoustic output! Normalizing the acoustic output to 1 watt the JBL has a sensitivity 70dBSPL and the Genelec has a sensitivity 71dBSPL. In this example even if the speakers used the exact same amplifier putting the same amount of power out as the JBL they would have different max peak SPL. In fact the Genelec has approximately two decibels more output compared to the JBL. It should be noted that this comparison was done using peak acoustic output but, we do not know how much power each amplifier used at peak output or their efficiency at peak output.
In conclusion even if when we compare a $150 speaker to a $2000 speaker, the difference in output is explained by a 1dB difference in sensitivity and 1dB less variance in frequency response. Neither manufacturer has provided tolerances for their sensitivity so depending on the individual model they could easily have identical sensitivities. With out proper information on engineering tolerances it is almost impossible to say which speaker is the better performer and that is why engineering tolerance is the most under discussed acoustic parameter.