That’s awesome, but no, they made something far more useful, lol. I’m glad to see projects like that though; it’s a lost art!
That’s awesome, but no, they made something far more useful, lol. I’m glad to see projects like that though; it’s a lost art!
Years and years ago I built my own 16 bit computer from the nand gates up. ALU, etc, all built from scratch. Wrote the assembler, then wrote a compiler for a lightweight object oriented language. Built the OS, network stack, etc. At the end of the day I had a really neat, absolutely useless computer. The knowledge was what I wanted, not a usable computer.
Building something actually useful, and modern takes so much more work. I could never even make a dent in the hour, max, I have a day outside of work and family. Plus, I worked in technology for 25 years, ended as director of engineering before fully leaving tech behind and taking a leadership position.
I’ve done so much tech work. I’m ready to spend my down time in nature, and watching birds, and skiing.
The article says that steam showing a notice on snap installs that it isn’t an official package and to report errors to snap would be extreme. But that seems pretty reasonable to me, especially since the small package doesn’t include that in its own description. Is there any reason why that would be considered extreme, in the face of higher than normal error rates with the package, and lack of appropriate package description?
Ants are the OG cooperative agent algorithms. Simulating ants use of pheromones to implement stigmergy path finding is a classic computer science algorithm.
In that case, never mix business and family. 😂
Lol, yep. Twenty years give or take testing just about everything along the way. 😂
I have to admit I used FreeBSD as my daily driver years ago. But I’ve also used everything else in the list at one point or another.
I use a terminal whenever I’m doing work that I want to automate, is the only way to do something such as certain parameters being cli only, or when using a GUI would require additional software I don’t otherwise want.
I play games and generally do rec time in a GUI, but I do all my git and docker work from the cli.
Right. Which gets us full circle, to never give investment advice, lol. That being said, at some point someone may sincerely look too you for guidance and you need to make a call as to whether you want to take that risk, what advice you give, and are you sure it is good advice.
I used to mentor student employees years ago, and when they wanted advice I always told them to max out workplace matches first, and then after that if they can save more, put it in tax advantaged savings programs that let you buy into indexed funds and never sell. In those cases you usually can’t even sell unless certain conditions are met and you sign disclosures, unlike most brokerages. Now, students you are giving them advice for the rest of there life and they likely don’t have $40k to panic sell/buy/sell to zero.
Honestly in my other comment I said never give investing advice, but as far as it goes, recommending investment in indexed funds is probably there exception with the caveat that it is a multi-year investment and there are dips.
That’s a great generalization! Don’t give lay people knowledge they can use to harm themselves, and recognize when you are the lay person.
Yes, times a thousand. But I would go even further.
Never give investment advice. You might explain what investments you have made and why you made them, but never give advice and never urge or prompt someone to invest. You should also end every conversation with “but that’s not advice and I’m not an expert.” It is too easy for either the investment to not work out, or for them to do it wrong (wrong timing, panic sale, misunderstood the options, etc).
The last thing you want on your conscience is someone investing a life changing amount of money just for it to go down in flames. I might invest $1000 in something that I think might pay off, tell someone they should invest, and next thing you know they drop in $40k and panic sell on a dip in two weeks, when I was planning to hold for five years. You never know.
That’s a good point. One of the things people struggle with I think is understanding the full scope of what was considered Greek and over what period of time. That and the competing representations of figures and the timeline of events means it really is like taking in a series of vignettes, each with its own take.
Stephen Fry did an excellent job making an updated and streamlined version of the mythology, effectively choosing from the myths what to accept as cannon in his retelling. If you haven’t read his books I would recommend them as being a wonderful story. He also narrated them himself on audible, which were also excellent.
Right. It just seems worth pointing out specifically rather than just using the term ace like the thread OP, because lumping those together doesn’t seem fair to ace folks. And at least for people like me who have read a lot of the Greek mythology, her aromantic nature is at least, if not more prominent in her personality than her chaste nature.
Yes, but what about love? As in, the emotion? Because Artemis is supposed to be unaffected by love, hence one of the three the goddess of love, Aphrodite, had no power over. But being ace doesn’t preclude someone from being in love.
Maybe it’s splitting hairs, but I recall the “chaste goddess of the hunt” and one of the three goddesses whom Aphrodite had no power. Additionally, goddess of healing, midwifery, and children. So I don’t know if the contemporary understanding of Ace matches that or not, as she is unaffected by love or lust.
That’s a good takeaway. AWS is the ultimate Swiss army knife, but it is easy to misconfigure. Personally, when you are first learning AWS, I wouldn’t put more data in then you are willing to pay for on the most expensive tier. AWS also gives you options to set price alerts, so if you do start playing with it, spend the time to set cost alerts so you know when something is going awry.
Have a great day!
So you just asked the most confusing thing about AWS service names due to how names changed over time.
Before S3 had an archival tier, there existed a separate service that AWS named AWS Glacier Storage, and then renamed to AWS S3 Glacier.
Around 2012 AWS started adding tiers to S3 which made the standalone service redundant. I received you look at S3 proper unless you have something like a Synology that can directly integrate with the older job based API used by the original glacier service.
So, let’s say I have a 1TB archival file, single tarball, and I upload it to a brand new S3 bucket, without version, special features, etc, except it has a life cycle policy to move objects from S3 standard to S3 Glacier instant access after 0 days. So effectively, I upload the file and it moves to Glacier class storage.
The S3 standard is ~$24/tb/month, and lets say worst case scenario our data sits on standard for one whole day before moving.
$0.77+$0.005 (API cost of the put)
Then there is the lifecycle charge to move the data from standard to glacier, with one request per object each way. Since we only have one object the cost is
$0.004 out of standard
$0.02 into glacier
The cost of glacier instant tier is $4.1/tb/month. Since we would be there all but one day, the cost on the first bill would be:
$3.95
The second month onwards you would pay just the $4.1/month unless you are constantly adding or removing.
Let’s say six months later you download your 1tb archive file. That would incur a cost of up to $30.
Now I know that seems complicated and expensive. It is, because it is providing services to me in my former role as director of engineering, with complex needs and budgets to pay for stuff. It doesn’t make sense as a large-scale backup of personal data, unless you also want to leverage other AWS services, or you are truly just dumping the data away and will likely never need to retrieve it.
S3 is great for complying with HIPAA, feeding data into a cdn, and generally dumping data around in performant way. I’ve literally dropped a petabyte off data into S3 and it just took it and did its thing.
In my personal AWS account I use S3 as a place to dump cache contents built by lambda functions and served up by API gateway. Doing stuff like that is super cheap. I also use private git repos (code commit), private container registry (ecr), and container host (ECS), and it is nice have all of that stuff just click together.
For backing up my personal computer, I use iDrive personal and OneDrive, where I don’t have to worry about the cost per object, etc. iDrive (not an Apple service) let’s you backup multiple devices to their platform and keeps them versioned.
Anyway, happy to help answer questions. Have a great day.
Thanks for posting. I just deployed to my container host in AWS ECS and it’s working well in my testing. Very easy deployment with docker.
With coffee
all thingsheart palpations are possible. It took me about a year and a half between work and studies. Definitely not a day. 😀