I think you are overestimating my character.

  • 0 Posts
  • 18 Comments
Joined 3 years ago
cake
Cake day: June 16th, 2023

help-circle
  • What? Not everyone lives in or near a metro area. Some people live in valleys or mountains where an OTA signal doesn’t reach. I should know that is where I grew up.

    Plus if you read up on the history of cable TV you will find that it was invented for just those reasons.

    “At the outset, cable systems only served smaller communities without television stations of their own, and which could not easily receive signals from stations in cities because of distance or hilly terrain”

    It isn’t that hard to read up on it and understand the history. Instead I guess just downvote because you don’t like the answer.


  • I grew up in the 70’s and had cable TV as soon as long as I could remember and it was the exact same broadcasts people that lived in DC or Baltimore got, we just got them from cable since the mountains we lived in prevented any OTA from getting through. And I am pretty sure I wasn’t alone with my cable TV.

    In 1968, 6.4% of Americans had cable television. The number increased to 7.5% in 1978. By 1988, 52.8% of all households were using cable. The number further increased to 62.4% in 1994. (https://en.wikipedia.org/wiki/Cable_television)

    I think one could argue they invented cable TV so that more people could see ads, not to stop showing ads.






  • Tasks the Apple Neural Engine Takes Responsibility For

    It’s time to dive into just what sort of jobs the Neural Engine takes care of. As previously mentioned, every time you use Face ID to unlock your iPhone or iPad, your device uses the Neural Engine. When you send an animated Memoji message, the Neural Engine is interpreting your facial expressions.

    That’s just the beginning, though. Cupertino also employs its Neural Engine to help Siri better understand your voice. In the Photos app, when you search for images of a dog, your iPhone does so with ML (hence the Neural Engine.)

    Initially, the Neural Engine was off-limits to third-party developers. It couldn’t be used outside of Apple’s own software. In 2018, though, Cupertino released the CoreML API to developers in iOS 11. That’s when things got interesting.

    The CoreML API allowed developers to start taking advantage of the Neural Engine. Today, developers can use CoreML to analyze video or classify images and sounds. It’s even able to analyze and classify objects, actions and drawings.

    https://www.macobserver.com/tips/deep-dive/what-is-apple-neural-engine/








  • I did know but actually forgot about that. I already had the text size in my control center. Regardless I will stick with apps that give me that option in the app itself. Memmy and voyager have it and are my gotos.

    I should add that I forgot about it as well defined apps have it in their app settings and do t make me go looking for it elsewhere. And I get it. Maybe I’m not the demographic you are after. I wish you well with the app.