What use are DATASETS really?

A bunch of people are trying to use datasets to implement permissions and ACLS via the GUI.
I use a zsh shell via ssh and give normal debian commands and scripts and automations. You cannot automate hardly anything via a gui. That is the #1 reason I abandoned TERRA-MASTER OS.

However, on TrueNAS there is the dataset dilemma.

Willnx says about datasets on truenas:

If it helps, here’s how my box at home is set up (with the datasets and such):
Vol1
+Berkeley (Unix dataset)
–>So many exports…
+Redmond (CIFS dataset)
–>Two different shares for my only two Windows boxes. Unique permissions for each.
+FTP (Unix dataset)
–>For the crap I always have to re-download when I make a new VM/ build a new PC.

I don’t have any quotas set up on these datasets either. I don’t care how big any of them get, provided it doesn’t fill my NAS 100% and turn it into a big brick.
Instead, I set up a reserve space on Vol1.

To me, the real power/ benefit of multiple datasets are:
More granular snapshots (Clones too)
Different compression preferences
More granular Deduplication control

Using datasets (in my option) for permission control is kind of like using a big wrench to hammer a nail; it works fine, but not it’s intended use.



My thoughts:

So, what I am hearing above is snapshots and clones are a reason to use datasets. Plus, if you have a windows and want SMB (because NFS is hard to do on windows) you do not want files being accessed by NFS and SMB at the same time because SMB does collaborate on file locking.

I use NAS for backup of desktops. But I also want virtual machines and server like Plex and Home Assistant, etc, and I want them in their own dataset space.

Further thoughts: First thing I plan to do when creating the first NAS test bed is to see if the superuser can cp directories in dataset A to directories in dataset B.

I am used to different users having their own private file space in their home directory and see no need for anything else to separate them. Datasets seems like something that comes from non-nix operating systems. Nix has been great for 40+ years and there is no need to overcomplicate it. Debian is debian is debian.

Why use RAID-10 in a 4 BAY NAS?

Has to do with lowest stress and quickest resilver.
Could use 3 bay RAIDZ1 with hot spare, but that is more stressful on all drives.

My plan A.

System 1: The back end. (Lower performance CPU). A 4 bay terramaster F4-424 pro should use 2 mirrors in raid 10 format.

System 2: The front end.

A 6 bay terramaster F6-424 MAX (higher performance a bandwidth:
Should use 2 mirrors or if 2 extra drives are added use 3 mirrors, all in RAID-10 format.
Have a cold spare available. Back up system 2 to system 1. Back up both to external USB drives and rotate offsite. 1 to back, 1 to cloud.

Have a third system in another location for fault tolerance.

A Typical Developer Type Discusses 2024 Display Managers.

If you ever wondered why I spent ZERO time and effort playing with Windows EVER any time in my life, this guy explains my world.

Air Traffic Control, for example, uses X11 protocols for display management.

Military systems use linux servers, and you might find X11 in their development labs (which are air-gapped from the outside world, of course).

Gas Turbine Engine systems use this technology.

Cell tower communications developers use all linux as their operating system.

Process control systems (paper mills, chemical processing) use linux as both development and target embedded operating systems.

Missile launch systems (including space shuttle) use large parallel linux based scientific computing clusters.

So this guy is speaking the bread and butter of what is on my desk every day. Display server, display manager, window manager, compositor, etc, is a choice for every node a linux engineer installs.
This guy is a typical example of the kind of people I have always worked with on a daily basis. The way he approaches problems, the way he speaks, his openness to alternative solutions for diverse use cases – all of this is what a typical engineer does daily.

Windows never had any role in any of that world. Windows was a toy sold to its victims of millions of idiot consumers. His talk is fascinating. I personally prefer KDE Plasma and Mate as desktops for PC’s. And I am experimentally switching to Wayland.

About Y-DNA

This was a note to myself to remember a few basic facts about Y-DNA and what the ftdna website was telling me.

Here is their statement:

Men do not just have one Haplogroup. They carry a “trail” of SNPs that are cumulatively passed from fathers to their sons.

Each man has many thousands of SNPs/Variants.

As of 27 March 2024, each FTDNA Big Y-700 included 639,776 Variants for each Kit’s results. (Show All: +; – ;?)

When two Kits verify a Variants existence, the Variant becomes a named SNP.

There are 20 Haplotypes: A through T.

In your example, the Haplotype is “R”.

Haplotype + SNP = Haplogroup.

In your case, you see three Equivalent SNPs in one Box on the Block Tree: FT43614; FT283922; and FTF15721.

The three SNPs formed in 3 separate and distinct Paternal Ancestors in your Tree. The names of the SNPs differ based on when they were first observed.

“BY” SNPs were named when Big Y-500 was the current test in 2029.

“FT” SNPs began in 2020 with the introduction of Big Y-700.

Once the number of “FT” SNPs reached 469,999, the names started over at “FTA”.

When “FTA” SNPs reached 99,999 the naming began again at “FTB”.

And the system continues.

The most recent SNP included in the ISOGG SNP Index is FTG641.

As of the end of March, there were 673,791 Variants/SNPs on the FTDNA Haplotree.

Shadow Banning

This came from a recent discussion about blog sites not being able to be posted on facebook and seen in the facebook browser.

Susan Lambeau, Looking at the facebook post in a browser if you click on the picture it launches a new browser window (the default browser of your system) and goes to https://phys.org/news/2023-12-theory-einstein-gravity-quantum-mechanics.html which is the URL of the article at phys.org. But using the facebook app what happens is you cannot get to phys.org. Instead Facebook generates a new web page that looks like the page at the above URL and send you to there, but inside the facebook app.

The implication here is facebook can censor the content. If they don’t want someone reporting on adverse side effects, for example, they can ban that.
Legally they can do that because they own the content on their website. They do not own the content on outside websites.

Even if you put the URL “https://phys.org/news/2023-12-theory-einstein-gravity-quantum-mechanics.html” on your facebook post people using the app will not be able to see the URL. They filter it out and re-write the display that is rendered.

This is important. Why? Well, for example, there is a recent scientific paper published in Canada about contamination level of DNA fragments in Pfizer covid vaccines. Facebook will not allow discussion of the paper. You cannot post the paper. You cannot post a link to a website talking about the paper either. They just remove your post. Sometimes they warn you not to try doing it again. The information is suppressed. This is one of the reasons people blog – to avoid censorship.

One of the side effects I recently noticed was it is not possible to get to a certain blog on medium.com where there are book reviews of books on science. The owner puts the link on her facebook group but facebook app users cannot get to the site to give feedback to the reviewer.

Really old people who still have computers and know what a browser is are slightly better off. But the new gen-alpha audience is totally locked into phones. Even my 40 year old daughter in law does not know what a browser is. And her kids have never seen a computer.

I think there is a word for this. I have heard it called “shadow banning”. The idea is to prevent unauthorized information from being viral. Or being capable of being viral.

So if a research scientist or an MD wants to say something that is “forbidden” she may be subject to shadow banning. Was there much of that happening during the lockdowns? You tell me!

When I travel I usually do not access facebook from a computer, but use a phone. I miss being able to copy a URL to paste into a note for later retrieval.
for example, the book review site. I had to ask the owner to message me the URL.

I hope this helps explain some of my concern. I want an open world wide web, not one that is controlled by a coalition of big tech that chokes off information.

Thanks!

Home Cloud Server

GMKtec Mini PC Windows 11 Pro Intel N100 (Up to 3.4GHz) 4C/4T, Mini Desktop Computer Dual LAN 1000Mbps 12GB DDR5 1TB Hard Drive, Micro PC 4K, Triple Display, WiFi6, BT5.2, Energy Efficient Nucbox G2

I built a cloud server based on this box. First I tested the windows 11 and it works fine. Then came the home cloud server. Also known as NAS (or network area storage).

INTRODUCTION

The G2 has USB 3.2 ports. So, I added a 1 terabyte USB NVME-SSD from Lexar. Also temporarily plugged a portable DVD reader. The latter is pretty slow but it is just for installation. So I changed the bios of the G2 and booting from the DVD reader I installed openmediavault 6.5 onto the SSD. Then set the bios to boot openmediavault (aka OMV).

OMV comes with a debian version of linux as it’s based OS. And the OMV software is layered on top.

Then I added an 8 TB seagate disk on another USB port. This needed to be re-formatted for use with OMV and I chose EXT4 format because it is the most flexible. The disk ended up giving 7.28 TB of usable storage after formatting.

Next I installed CasaOS because it has really cool home applications and runs a Docker management system. More on that later.

Next, using the OMV panel I created an NFS server to make the G2 cloud appear to my linux desktop. And then an SMB server to make it appear to the W11 machine.

On W11 I copied all my audible books in AAX and MP3 format to the G2. This was 329 GB of data and it took an hour. The average data transfer rate was 50 MBytes per second. Thats 500 megabits. It gave my wifi a pretty good workout.

RESULT:

I can now share files between windows and linux via this shared cloud device.

Videos on Amazon page here: https://a.co/d/1QKNlmp

Here is my dashboard:


I am running CasaOS and here is a panel from it:

The CasaOS OpenSpeedTest app measures my upload speed of 484.8 Mbps exceeds the 405.6 Mbps download speed to the G2.

My plex account on CasaOS:

And of course when I went to make the screen snap I totally got lost in a episode of Doc Martin on live Plex TV. Cannot resist that Cornish countryside.

Resources:

OMV on raspberry PI: https://www.naturalborncoder.com/hardware/raspberry-pi/2022/11/21/adding-an-storage-drive-and-smb-share-to-openmediavault-on-raspberry-pi/




More on this later.


Why ICE is not dead.

Synthetic fuels’ biggest draw is that unlike fossil fuels, the C02 they release into the atmosphere when burned in an engine is virtually equal to the amount taken out of the atmosphere to produce the fuel thus making them CO2-neutral overall. To sweeten the deal, ICE vehicles do not require any modifications to run on e-fuels, which can also be transported via existing fossil fuel logistics networks. Further, synthetic fuels can be blended in fossil fuels or can completely replace them in existing ships, airplanes or industrial technologies.

Building 8852be Driver


I bought an AMD Ryzen 7 5825U 8 core laptop. The wifi chip is an RTL8852 chip. Linux does not yet have a driver for it in any distributions. So, here is how to build it from source code.

First you add the repository.


sudo add-apt-repository ppa:kelebek333/kablosuzsudo add-apt-repository ppa:kelebek333/kablosuz
sudo apt update
git clone https://github.com/HRex39/rtl8852be.git
clone https://github.com/HRex39/rtl8852be.git
cd rtl8852be
make
sudo make install
sudo modprobe 8852be

Now when you reboot and go into control panel and look at adding a wifi device you will see the new device present. Mine shows up as a device named wl0.

Wind Draughts

Primary Category: …Energy…

An investor writes:

“I wrote this article  for the Australian edition of the British magazine Spectator a couple of weeks back. In essence, academics are FINALLY starting to realise that wind droughts are an issue with intermittent systems and studying them. As the article notes some work has been done in the UK, where it is known, for example, that some years back the wind made no contribution to the UK grid for nine days, and there were serious deficits during another drought at the end of last year. These wind droughts are an extreme event like cyclones or rain droughts. I saw some material recently on wind droughts in the US but I seem to have mislaid it. Perhaps someone has access? As for Australia there has been limited work to suggest that wind droughts in a given year might last for up to 36 hours. But that’s just from one year of data. As noted in the article there is no way to store enough power to tide the grids over such long periods. Australia is building one water dam project called Snowy 2.0 (after the region) but a fully renewables network would need at least six of seven. In any case the blindness of policy makers to this issue to date is just extraordinary. “

This got me to thinking.


   
My take: Storage of transient energy remains an issue. Tesla’s power wall is based on lithium battery technology and what counts here is Mega-Joules/Kg. ie, energy density of the storage mechanism. Also the economics of the life cycle of mining all the way through waste disposal and the the cost of each step.


I recently mentioned a physicist who remarked on TV about the subject of chemical based “replaceable energy storage cells”, ie, battery units, for personal road vehicles. There is a physical limit to that energy density. This was in a conversation about Tesla, which uses lithium battery technology. I simply pointed out the existence of the physicist’s remarks. And was instantly set upon by a protagonist of the original poster who was “triggered” by the point. We never did get around to addressing the actual issue, mainly because I do not respond to off the wall aspersions and argumentum ad-hominum attacks directed at third party people. And there were plenty of those from this particular protagonist.

The physicist had a real point. There are physical limits to chemical energy density and there is no “magic technology” that will save chemical batteries. There are alternative replaceable storage cell designs based on non-chemical energy storage and these are in research phases. And one could discuss those. But for the time being Tesla as a current product is not based on any of these.

And we have to beware red-herring arguments and be careful to compare apple to apples (not apples to oranges).

There are possibilities for building personal electric vehicles if the energy density problem could be solved. It is not going to be Iron Man’s fusion battery strapped to your chest, however. Wouldn’t that be nice if it were?

Major issues in personal vehicles are:

  • How far can you drive before recharging?
  • How much time is required to recharge?
  • Availability of recharging equipment?
  • Ultimate energy source of the recharge. Where did it come from? Where was it stored?


The Beat Goes On

The public is being told that in the near future everyone will be driving a vehicle powered by electric motors which run off chemical based batteries and these in turn will be powered by wind power and solar panels which store the energy in an energy infrastructure that easily distributes to resupply the personal vehicles. That is the main drumbeat. And humanity will be saved from climate change. Problem is, this is not credible. The drumbeat also includes elements of “if you don’t believe the drumbeat you must be a trog who is against science.” That is a non-sequitor.

Of course there are other sources of energy. among these are:

  • Hydro
  • Geothermal
  • Nuclear
  • Fossil Fuels

Systems Analysis

As a physics major turned engineer I believe these issues require an approach of systems analysis. In other words they are problem sets in systems analysis. All aspects must be solved simultaneously for society to be able to utilize any given solution set. Systems engineering is one of the types of jobs that I do. This type of thinking is particularly important for policy makers. Unfortunately most public debate ignores systems analysis and focuses on just one aspect of the problem set. This is naive thinking. When someone demonstrates such thinking I usually refuse to speak with them because it becomes a waste of time.

Areas I am interested in:

Capacitive power cells powered by fuel cells. Why? Higher energy density. Higher energy discharge capability. Fueling is rapid and fueling stations can be made readily available.

Hydrogen power.

Something more exotic.

These are completely separate discussions than vehicles powered by lithium power cells.

The above might answer how to build personal vehicles. But neither of the above answer the question of where the initial power comes from or how is the energy stored and transferred for availability to vehicles.

My experience is that folks who are in love with electric cars tend to focus only on the one aspect they care about and ignore the other issues entirely. And they seem to resent any questions about those aspects.

Now, wind draughts are one tiny aspect of energy gathering systems. Wind power freezing over in Texas or Minnesota is another such topic. These systems tend to be under-engineered and fail. The overall energy grid needs to be able to deal with such transient effects.

I plan to say something about large stationary power storage systems … soon.


Meanwhile:

This is interesting commentary.

Undergraduate basics.

Lorraine and Corson is the standard E&M textbook in upper division physics at California State University, or at least it was for many years. It is the one I used for my undergraduate work. It is a core prerequisite for senior level physics courses. Generally you take 16 credits of physics per semester and add in one general ed easy course as the 5th to make 18, but this varies.

Lorraine and Corson was WONDERFUL as a course.


I was in my senior year of physics when I got married and moved across the country and well, life took a different turn.

Let me just say right now: I have never met a BIOLOGIST who took this E&M course. Doesn’t mean there aren’t some. I myself eventually became an engineer who shipped 30+ engineering products. But I also went on to study biology, microbiology, and biochemistry.

Let me quote a review from Amazon.com about Lorraine and Corson:

5.0 out of 5 stars Great reference work.Reviewed in the United States on October 20, 2001

This book is intended primarily for students of Physics or Electrical Engineering at the junior or senior levels, although some schools will prefer to use it with first-year
graduate students. The book should also be useful for scientists and engineers who wish to review the subject.
The aim of this book is to give the reader a working knowledge of the basic concepts of electromagnetism. Indeed, as Alfred North Whitehead stated, half a century
ago, “Education is the acquisition of the art of the utilization of knowledge.” This explains the relatively large number of examples and problems. It also explains why
we have covered fewer subjects more thoroughly. For instance, Laplace’s equation is solved in rectangular and in spherical coordinates, but not in cylindrical
coordinates.
CONTENTS
A chapter on vectors (Chapter 1), a discussion of Legendre’s differential equation (Section 4.5), an appendix on the technique that involves replacing cos wt by exp jwt,
and an appendix on wave propagation.
After the introductory chapter on vectors, Chapters 2, 3, and 4 describe electrostatic fields, both in a vacuum and in dielectrics. All of Chapter 4 is devoted to the
solution of Laplace’s and of Poisson’s equations.
Chapter 5 is a short exposition of the basic concepts of special relativity, with little reference to electric charges. It requires nothing more, in the way of mathematics,
than elementary differential calculus and the vector analysis of Chapter 1. Chapter 6 contains a demonstration of Maxwell’s equations that is based on Coulomb’s law
and on the Lorentz transformation and which is valid only for the case where the charges move at constant velocities.
Chapters 7 and 8 deal with the conventional approach to the magnetic fields associated with constant and with variable currents. Here, as elsewhere, references to
Chapter 6 may be disregarded.
Chapter 9 contains a discussion of magnetic materials that parallels, to a certain extent, that of Chapter 3 on dielectrics.
In Chapter 10, the Maxwell equation for the curl of B is rediscovered, without using relativity. This is followed by a discussion of the four Maxwell equations, as well
as of some of their more general implications. The point of view is different from that of Chapter 6, and there is essentially no repetition.
The last four chapters, 11 to 14, concern various applications of Maxwell’s equations: plane waves in infinite media in Chapter 11, reflection and refraction in Chapter
12, guided waves in Chapter 13, and radiation in Chapter 14. The only three media considered in Chapters 11 and 12 are perfect dielectrics, good conductors, and
low-pressure ionized gases. Similarly, Chapter 13 is limited to the two simplest types of guided wave, namely the TEM mode in coaxial lines and the TE1,0 mode in
rectangular guides. Chapter 14 discusses electric and magnetic dipoles and quadrupoles, as well as the essential ideas concerning the half-wave antenna, antenna arrays,
and the reciprocity theorem.
For a basic and relatively simple course on electromagnetism, one could study only Chapters 2, 3 (less Sections 3.3, 3.4, 3.8, 3.9, and 3.10), 4 (less Sections 4.4 and
4.5), 7, 8, 9 (less Section 9.3 but conserving the equation v – B = 0), and 10. For a rather advanced course, on the other hand, Chapters 2, 3, 4, 5, 7, 8, and 9 could be
reviewed briefly using the summaries at the end of each chapter. One would then start with Chapter 6, and then go on to Chapter 10 and the following chapters. There
are, of course, many other possibilities.
In Chapter 12, Sections 12.3 and 12.7 could be dispensed with. They involve the application of Fresnel’s equations to particular cases and are not essential for the
remaining chapters. Chapter 13 is instructive, both because of the insight it provides into the propagation of electromagnetic waves and because of its engineering
applications, but it is not required for understanding Chapter 14. Finally, Chapter 14 is based on Chapter 10 and on the first two sections of Chapter 11.