I think it’s been about a year? IIRC Intel only started using TSMC for their processors with Meteor Lake, which was released in late 2023.
I believe their discrete GPUs have been manufactured at TSMC for longer than that, though.
I think it’s been about a year? IIRC Intel only started using TSMC for their processors with Meteor Lake, which was released in late 2023.
I believe their discrete GPUs have been manufactured at TSMC for longer than that, though.
I use a lot of AI/DL-based tools in my personal life and hobbies. As a photographer, DL-based denoising means I can get better photos, especially in low light. DL-based deconvolution tools help to sharpen my astrophotos as well. The deep learning based subject tracking on my camera also helps me get more in focus shots of wildlife. As a birder, tools like Merlin BirdID’s audio recognition and image classification methods are helpful when I encounter a bird I don’t yet know how to identify.
I don’t typically use GenAI (LLMs, diffusion models) in my personal life, but Microsoft Copilot does help me write visualization scripts for my research. I can never remember the right methods for visualization libraries in Python, and Copilot/ChatGPT do a pretty good job at that.
There is no “artificial intelligence” so there are no use cases. None of the examples in this thread show any actual intelligence.
There certainly is (narrow) artificial intelligence. The examples in this thread are almost all deep learning models, which fall under ML, which in turn falls under the field of AI. They’re all artificial intelligence approaches, even if they aren’t artificial general intelligence, which more closely aligns with what a layperson thinks of when they say AI.
The problem with your characterization (showing “actual intelligence”) is that it’s super subjective. Historically, being able to play Go and to a lesser extent Chess at a professional level was considered to require intelligence. Now that algorithms can play these games, folks (even those in the field) no longer think they require intelligence and shift the goal posts. The same was said about many CV tasks like classification and segmentation until modern methods became very accurate.
GPU and overall firmware support is always better on x86 systems, so makes sense that you switched to that for your application. Performance is also usually better if you don’t explicitly need low power. In my use case I use the Orange Pi 5 Plus for running an astrophotography rig, so I needed something that was low power, could run Linux easily, had USB 3, reasonable single core performance, and preferably had the possibility of an upgradable A key WiFi card and a full speed NVMe E key slot for storage (preferably PCIe 3.0x4 or better). Having hardware serial ports was a plus too. x86 boxes would’ve been preferable but a lot of the cheaper stuff are older Intel mini PCs which have pretty poor battery life, and the newer power efficient stuff (N100 based) is more expensive and the cheaper ones I found tended to have onboard soldered WiFi cards unfortunately. Accordingly the Orange Pi 5 Plus ended up being my cheapest option that ticked all my boxes. If only software support was as good as x86!
Interesting to hear about the NPU. I work in CV and I’ve wondered how usable the NPU was. How did you integrate deep learning models with it? I presume there’s some conversion from runtime frameworks like ONNX to the NPU’s toolkit, but I’d love to learn more.
I’m also aware that Collabora has gotten the NPU drivers upstreamed, but I don’t know how NPUs are traditionally interfaced with on Linux.
A lot of the cheap tablet SoC vendors like Rockchip (whose SoCs end up in low cost SBCs) really only do the bare minimum when it comes to proper linux support. There’s usually next to no effort to upstreaming their patches so oftentimes you’re stuck on their vendor kernel. Luckily for the RK3588(S), Collabora has done a considerable amount of work on supporting the SoC and its peripherals upstream. I run my Orange Pi 5 Plus (RK3588) on a mainline kernel and it works for my needs.
This practice is a lot easier to defend for a low cost SoC compared to something as expensive as a Snapdragon Elite though…
Yeah I think Lemmy would actually work pretty reasonably. It reminds me of how lots of software and projects have Reddit communities. I agree that being able to share 1 account over many services, and especially not having to pay for infrastructure is something that drives discord use over forum-based platforms.
Personally, I’d prefer that projects use forums for community discussions rather than realtime chat platforms like Discord or Matrix. I think the bigger problem of projects using Discord is not that it’s closed source, but rather that it makes it difficult to search (since no indexing by search engines) and the format deprioritizes having discussion on a topic over a long period of time. Since Matrix is also intended for chat, it has these same issues (though at least you can preview a room without making an account).
Right, as someone in the field I do try to remind people of this. AI isn’t defined as this sentient general intelligence (frankly its definition is super vague), even if that’s what people colloquially think of when they hear the term. The popular definition of AI is much closer to AGI, as you mentioned.
I’ve been using FreeTube since Piped was very inconsistent for me, but I guess that’s just the nature of these services. I’ll have to check out Invidious again, last time I tried it was several years ago and I stopped using it after the main instance shut down. Is it still under active development? I remember its development status being unclear, partially because the language it uses is not super mainstream, but it’s probably changed since then.
Fortunately, Invidious, Piped, Libretube and Newpipe all exist and work flawlessly so there’s no excuse to use proprietary trash like that.
Isn’t the very point of this post that Invidious and Piped don’t work flawlessly?
Can’t you still modify and distribute Grayjay, just not commercially? I understand that still prevents the app from being considered open source, but their reasoning is valid IMO (to prevent people from making ad-infested clones on the play store, which has happened with NewPipe before).
It’s unfortunately super clear from their Steam charts. When they had creator events and whatnot, the player count spiked, but other than that they only have about 1000 players active and I seriously doubt many people spend money on the game since it’s already rather F2P friendly.
It’s a shame, the game was a lot of fun and I still play with friends.
Yeah we used to joke that if you wanted to sell a car with high-resolution LiDAR, the LiDAR sensor would cost as much as the car. I think others in this thread are conflating the price of other forms of LiDAR (usually sparse and low resolution, like that on 3D printers) with that of dense, high resolution LiDAR. However, the cost has definitely still come down.
I agree that perception models aren’t great at this task yet. IMO monodepth never produces reliable 3D point clouds, even though the depth maps and metrics look reasonable. MVS does better but is still prone to errors. I do wonder if any companies are considering depth completion with sparse LiDAR instead. The papers I’ve seen on this topic usually produce much more convincing pointclouds.