technologique | Неотсортированное

Telegram-канал technologique - Technologique

697

Deeply involved developers about various aspects, tendencies & conceptions of programming technologies, FLOSS, Linux, security, cloud infrastructures & DevOps practices, distributed systems, data warehousing & analysis, DL/ML, web3, etc. Author: @andrcmdr

Подписаться на канал

Technologique

The data storage engine projects we're all waiting for!

I was expecting data storage engines and data warehouse solutions, cloud native solutions for data lakes, will be made using Rust, as systems language, in Rust community.

Long awaited stuff, for the whole time since 2015, stabilized Rust v1.0 compiler and Rust 2015 standard.

https://github.com/RustFS/RustFS

#Rust
#RustLang
#RustFS

Читать полностью…

Technologique

AI is dangerously centralized.

Why building community aligned AI is really matter, and how web3 technologies can play the key role to resolving current situation with centralized AI, owned by tech giant companies, and instead help to create a community driven ecosystem for AI development.

https://x.com/oleg_golev/status/1944157582144246077

The podcast:
https://x.com/autonolas/status/1926675599172452539

#AI
#AGI
#OpenAGI

Читать полностью…

Technologique

AI and AGI should be fully open sourced and loyal to builders and community!

The most important thing I should say and add to Steve's blog post is that AI should be open (now we see opposite things - a big tech concentrated AI market), free (as in freedom), monetizable and loyal, for creators/builders/developers good and for community win. And this is OML principle. And target goal of Sentient Foundation, who makes truly open AGI future, and already developed Dobby model (and Dobby is already free! =), Sentient Chat, Sentient OpenDeepSearch, OML Fingerprinting library, Agent Framework and Enclaves Framework (proud to be a leading part of it!).
And all of these parts of groundbreaking product portfolio and breakthroughs are made just within less than a year!
More good things to come! Stay turned!

https://steveklabnik.com/writing/i-am-disappointed-in-the-ai-discourse/

https://www.sentient.xyz

#AI
#AGI
#OpenAGI

Читать полностью…

Technologique

Modular provides MAX platform - it is MAX inference backend (engine) and MAX inference server (MAX Serve).

Just look at this:

https://builds.modular.com/models/DeepSeek-R1-Distill-Llama/8B-Q6_K

https://builds.modular.com/models/Llama-3.3-Instruct/70B?tab=deploy

In terms of deployment it is fantastic! Just one (relatively) tiny container!
And in terms of programming - GPU programming and acceleration without CUDA, using Mojo language (statically LLVM compiled), which has capabilities of Rust (static memory safety), LLVM MLIR (Multi-Level Intermediate Representation) byte code compilation for amazing low level code optimization and acceleration, syntax of Python and Mojo integrates (embrace) the whole Python ecosystem. I'm playing with Mojo for quite a while already (and it is best of both worlds - Rust and Python), but MAX just used recently. And Llama.cpp not even in comparison with MAX!

#Mojo
#MAX
#AI
#AGI

Читать полностью…

Technologique

https://www.youtube.com/live/AyH7zoP-JOg

Great conversation!

The privacy and confidentiality should be a fundamental human right in the information and ubiquitous computations era.

Always think about how your data will be used, what you say, message and what you'll prompt to search engine or AI model, how it can be and will be used, especially against your interests.

#AI
#AGI
#privacy
#confidentiality
#confidential_computing
#CC
#security

Читать полностью…

Technologique

OpenAGI summit at ETH Denver event.

The updates from Sentient.xyz:

Himanshu Tyagi presented a loyal AI implementation, a Sentient Chat:

https://youtu.be/UsuMbk32i44

#AI
#AGI
#OpenAGI

Читать полностью…

Technologique

OpenAGI summit at ETH Denver event.

The updates from Sentient.xyz:

Sandeep Nailwal speech about loyal AI principles:

https://youtu.be/u2_-dUCb_Yk

#AI
#AGI
#OpenAGI

Читать полностью…

Technologique

Fish shell 4.0, completely rewritten in Rust, just recently officially released!

https://fishshell.com/blog/new-in-40/

https://github.com/fish-shell/fish-shell

https://github.com/fish-shell/fish-shell/releases/tag/4.0.0

#Fish
#Shell
#Bash
#Rust
#RustLang

Читать полностью…

Technologique

The final of the story.

How's it started:
https://lore.kernel.org/lkml/20250108122825.136021-1-abdiel.janulgue@gmail.com/

How's it going:
https://lore.kernel.org/lkml/20250224-configfs-v4-0-9af9b5e611f6@kernel.org/

How's it ended:
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/MAINTAINERS?id=815291c11acda54515f1af5ce6fe307490de9127

Moral:
Don't resist to positive valuable changes and don't be a brake for others and for progress.

Rust to the Linux! 🐧
Rust to the Moon! 🚀

#Rust
#RustLang
#Linux

Читать полностью…

Technologique

The drama around Rust drivers inclusion into Linux kernel became worser and even worser.

Old kernel maintainer against of Rust code inclusion into DMA for drivers.
Asahi Linux maintainer resigned from Linux kernel maintainers.

The question is - what is better, maintain two languages and handle this complexity, one of it, Rust, provided memory safety, when usafe blocks are strictly located and controlled, or, continue to maintain old (and new) error prone code in C language with all that memory safety issues, as it was before, for many years, without accepting new approaches to develop safe system software, drivers and kernel modules.

Overall, that's the problem of old minded people, who doesn't want to put the efforts to accept something new, which can be better than old approach.

I've got only one question - if there are so much resistance from old minded people, kernel maintainers, for what the Rust for Linux subsystem was even included into the kernel, what old kernel maintainers and Linus himself expected when they've done that, i.e. introduction of Rust subsystem into kernel? They want to keeping it in cage of just subsystem for drivers development, when the problem is the whole old approach to memory management?

If this situation will be progressing (and it will!) many new wave kernel maintainers and kernel developers will leave and resign from Linux kernel. And that's not good for Linux, as maintainers became older and older and it's became tremendously hard to involve new generation of system developers into kernel development and maintenance.
Many new wouldn't even join Linux kernel, because of such toxic environment in conversations, and the better option here for new wave system developeers is to consolidate common efforts on some alternative kernel, like RedoxOS microkernel for example, which is a Unix POSIX compatible kernel, fully written in Rust, and someday, in 10-20 years from now, when Linux gets old and cost of maintenance of old error prone C code will be high (same as with assembly language kernels in previous age), Redox will became a Linux alternative, same as BSD kernels and distros, but better.

https://lore.kernel.org/lkml/20250207-rm-maint-v1-1-10f069a24f3d@marcan.st/

https://www.theregister.com/2025/02/05/mixing_rust_and_c_linux/

https://www.theregister.com/2025/02/07/linus_torvalds_rust_driver/

#Rust
#RustLang
#Linux
#kernel

Читать полностью…

Technologique

I also got a question - from where DeepSeek mined and gathered all of the data for data-set to train R1 model? Especially for such a short period of time!

For Facebook/MetaAI with LLama, Microsoft/OpenAI with GPT4o/o1/o3, Google AI (+ teams from Google Brain and DeepMind) with Gemini (former Bard) - it's naturally sourced either from search engine index from web, or social networks user data.

Some independent players like Anthropic with Claude (made by former OpenAI engineers, where Amazon as a largest investor),
Mistral AI (made by former Google DeepMind and Meta AI engineers) with open sourced models Mistal and Mixtral (mixture of experts, MoE, architecture, similar to what DeepSeek use) - they've got their own data sources.

Just think about it - DeepSeek R1 has really good quality of responses and reasoning, this means the data-set and reinforcement learning processes they use (explained in the paper: https://arxiv.org/pdf/2501.12948, https://raw.githubusercontent.com/deepseek-ai/DeepSeek-R1/refs/heads/main/DeepSeek_R1.pdf) are prepared and tuned in an really outstanding level!

(And now OpenAI states that DeepSeek steals data from their API and then uses distilled training based on their data.)

Читать полностью…

Technologique

DeepSeek R1 model release - disrupting of corporate monopoly and "Star Gate" project plan.

https://www.youtube.com/watch?v=WEBiebbeNCA

The big deal is that model R1 from DeepSeek is open sourced.

While LLama 3.1 for example it's still mainly closed, not fully open sourced (only model tensor layers in portable standardized format, runner for inference, and partially datasets for fine-tuned special models of LLama) and FSF recently evaluated Meta's LLama community license as non-free:
https://www.fsf.org/blogs/licensing/llama-3-1-community-license-is-not-a-free-software-license

While Meta stated about given freedom of usage and modifications:
https://ai.meta.com/blog/meta-llama-3/

Obviously, non-ethic usage possible (in medicine, bioinformatics, military, and as cyber weapon, etc.).

But this also restricts builders with honest usage from creation of other products faster and gives control over progress in AI only to large corporations and big-tech.

While model from DeepSeek has been given for free, as in freedom.

This shifts balance in current monopoly and gives powerful tool to builders, everyone who wants to create services.

Also this revealing the recent half trillion investing into "Stargate" project plan as inefficient.

Overall it's a good step forward in #AI world.

Links:

https://github.com/deepseek-ai/DeepSeek-R1?tab=readme-ov-file#6-how-to-run-locally

https://github.com/deepseek-ai/DeepSeek-R1?tab=readme-ov-file#7-license

https://huggingface.co/deepseek-ai/DeepSeek-R1

https://huggingface.co/deepseek-ai/DeepSeek-R1-Zero

Читать полностью…

Technologique

https://fishshell.com/blog/rustport/

I love Fish, using it everyday - shell is amazing, on par with zsh, and scripting language is like natural extension to Bash scripts, with a lot of additional helpful functions (like random, math, etc.). Now it's on Rust. Love it even more now! 💜

If you want to move to alternative shell - the best moment is now!
You won't miss anything in Fish, always can run/call Bash or substitute it through CLI.
Transition will be natural, just one day to start using it efficiently for everyday work. Less friction than with zsh. Oh... And fish_update_completions (https://fishshell.com/docs/current/cmds/fish_update_completions.html) running right from shell is a killer feature along with command history and fuzzy search substitutions from existing command history.

#Fish
#Shell
#Bash
#Rust
#RustLang

Читать полностью…

Technologique

Rust+Linux

Aaron0928/linux-reaches-a-major-turning-point-rust-drivers-set-to-enter-the-core-kernel-project-lead-says-18e1cbe4cc3f" rel="nofollow">https://medium.com/@Aaron0928/linux-reaches-a-major-turning-point-rust-drivers-set-to-enter-the-core-kernel-project-lead-says-18e1cbe4cc3f

Greg Kroah-Hartman, a key maintainer of the Linux kernel, stated that with these changes, Linux 6.13 now provides stronger support for developing kernel drivers in Rust.

#Rust
#Linux

Читать полностью…

Technologique

https://hez2010.github.io/async-runtimes-benchmarks-2024/

Whoa! Rust's Tokio runtime doing really great! 👍 (Even greater with just for cycles for making new Futures, mapping them on pool of system threads and awaiting in an event loop).
And statically AoT compiled C# amazes me a lot!

#Rust

Читать полностью…

Technologique

The one technically great web calls service, written in Rust, using Actix and NATS:

https://videocall.rs

https://app.videocall.rs

https://github.com/security-union/videocall-rs

Читать полностью…

Technologique

Open, Monetizable, Loyal AGI Platform

https://www.sentient.xyz

#AI
#AGI
#OpenAGI

Читать полностью…

Technologique

Whoa! We need to update our kernels!

https://hoefler.dev/articles/vsock.html

https://security-tracker.debian.org/tracker/CVE-2025-21756

#kernel
#Linux
#VSock

Читать полностью…

Technologique

Amazing things has been released by Modular development team (Mojo language and Max inference backend):

https://www.modular.com/blog/max-25-2-unleash-the-power-of-your-h200s-without-cuda

#Mojo
#MAX
#AI
#AGI

Читать полностью…

Technologique

Making truly open source AI, trustful, community driven, using top-notch TEE and blockchain technologies. That's the mission of Sentient.

Super interesting and informative podcast:
https://www.youtube.com/live/-P6sFtQRbl8

And you should take Sentient into serious consideration.

Give it a try!

https://chat.sentient.xyz

https://sentient.xyz

https://github.com/sentient-agi

https://huggingface.co/SentientAGI

#AI
#AGI

Читать полностью…

Technologique

OpenAGI summit at ETH Denver event.

The updates from Sentient.xyz:

Oleg Golev presented Sentient Enclaves Framework for Confidential AI applications and Sentient Agents Framework for creating AI agents:

https://youtu.be/Ah5FGrmj81M

A big milestone for our team of Sentient Builders!
Much kudos for the shout-outs! 🙌

#AI
#AGI
#OpenAGI
#TEE
#Enclaves
#AWS
#NitroEnclaves

Читать полностью…

Technologique

That's huge!
3FS (Fire-Flyer File System), a high performance distributed FS for distributed training (ML/DML) and inference just released! 🙌

https://github.com/deepseek-ai/3FS

#AI
#AGI
#LLM
#FS
#CPP
#Rust
#RustLang
#Linux

Читать полностью…

Technologique

OpenAGI summit at ETH Denver event.

https://openagi.xyz/schedule

openAGISummit/streams" rel="nofollow">https://www.youtube.com/@openAGISummit/streams

Live stream:
https://www.youtube.com/watch?v=wgtPtFvG1zo

#AI
#AGI
#OpenAGI

Читать полностью…

Technologique

Release of Rust 1.85 and standardized Rust 2024 Edition release

https://blog.rust-lang.org/2025/02/20/Rust-1.85.0.html

https://doc.rust-lang.org/edition-guide/rust-2024/index.html

https://doc.rust-lang.org/edition-guide/editions/index.html

And finally positive news about resolving situation with Rust support in Linux kernel:

https://lore.kernel.org/lkml/CAHk-=wgLbz1Bm8QhmJ4dJGSmTuV5w_R0Gwvg5kHrYr4Ko9dUHQ@mail.gmail.com/

https://lore.kernel.org/all/2025021954-flaccid-pucker-f7d9@gregkh/

infradead.org/T/#ma01177a189330fa9656bed9ba8b49b8f5fe10dc6" rel="nofollow">https://lore.kernel.org/lkml/Z7SwcnUzjZYfuJ4-@infradead.org/T/#ma01177a189330fa9656bed9ba8b49b8f5fe10dc6

#Rust
#RustLang
#Linux

Читать полностью…

Technologique

Here we go again!

SEV-SNP is vulnerable, again.

New AMD SEV-SNP vulnerability:

https://github.com/google/security-research/security/advisories/GHSA-4xq7-4mgh-gp6w

Exploit:

https://github.com/google/security-research/tree/master/pocs/cpus/entrysign

Reports about two recent vulnerabilities in SEV-SNP memory encryption and isolation mechanism, on CPU pipeline, cache and branch prediction level:

https://www.amd.com/en/resources/product-security/bulletin/amd-sb-3019.html

https://www.amd.com/en/resources/product-security/bulletin/amd-sb-3010.html

AMD reported that previous approaches to Spectre class attacks will work to fix new vulnerabilities:

https://www.amd.com/content/dam/amd/en/documents/epyc-technical-docs/tuning-guides/software-techniques-for-managing-speculation.pdf

TLDR:

This is related to TEE technology for confidential VMs (cVMs) wide used for confidential apps and data handling on servers, AMD SEV-SNP. Intel has similar technology for cVMs, TDX. AWS Nitro Enclaves which is widely used based on (Firecracker VMs and) EC2 instances wtth SNP or TDX support, depending on instance class.

This is memory isolation and encryption technologies for VMs and apps in it.
Vulnerability affects sensitive data in enclaves memory via side-channel attack, on CPU instructions pipelining, branch prediction and TLB cache (translation of virtual addresses of memory to real addresses of RAM). Thus sensitive data can leak unencrypted through processor cache, while handling by some privileged process (kernel, KVM subsystem, hypervisor), which can be read by other processes in a system (with escalation of privileges).
Thus it's a Meltdown/Spectre class/level vulnerability for CPU superscalar architecture.
Only solution here is disabling branch prediction on a CPU microcode level, via UEFI BIOS patch and patching (Linux) kernel with disabling IPC mechanisms for caching context between processes, which led to more kernel space (privileges) versus user space context switching and decrease of performance tremendously.

But this particular vulnerability is related and affects the microcode updates uploading itself, cracking microcode signature verification attacker (gained access to a host machine with ring0 privilege level, as root user and via kernel space rootkit, for example) can upload malicious microcode locally, which will gather confidential data from cVM processes via side-channels, from CPU cache.

The root cause - is insecure hash function in CPU microcode for microcode signature verification before updating itself, so signature hash can be spoofed this way.

#cVM
#TEE
#SEV
#SNP
#SEV_SNP
#AMD

Читать полностью…

Technologique

Holly Spirit!

Have tried out DeepSeek R1 now!

And it's so much creepy scary! 😱

Especially the speed (tokens per second) and reasoning, it reasoning, i.e. thinking our loud, like a good old Unix programmer! 😱

I've feed it some sample of code I wrote recently for enclave's init system (just about 400 SLoC in Rust), I was rewrote it from C to Rust, and asking about the ways of improvements in handling Unix signals for processes in Rust in an idiomatic POSIX way but using Rust standard library. Overall it should be the fully fledged Init system for Linux residential processes executed and existed in enclaves.

And you know what?

OpenAI GPT4o cannot solve this - always thinking for too long, making output, but not precisely to my prompts. And then asks to pay for subscription ('cause time and tokens limits exceeded). 😂 Probably OpenAI do this intentionally to be more commercially efficient. Just a wasting of time and money.
But I subscribed out of curiosity.

OpenAI GPT o1 - already with subscription. Same things. It cannot solve prompt to the fullest. Just 400 SLoC to analyze from source code and it always stops, asking for refinements and in final not giving fullest results, just code snippets, that aren't helpful, more like a hallucination (do not use any substance and make code! 😂).

LLama 3.1 70B self-hosted - works good. Not reasoning perfectly but gives meaningful hints. Downsides - also always stops, asking for refinements and cycling, this dialogue never end and you've never reach the final meaningful complex result. Code snippets with examples are helpful. Can be use as fast search engine for code samples with context for current task. Overall helpful.

DeepSeek R1:
It's mind-blowing! 😱
One precise prompt.
Precise analysis, weak and good points, snippets, examples.
Precise reasoning, thinking out loud, as me talking with my computer science teacher in University.
And speed - it's blazing fast! Tokens per second performance is way much faster than others, even visually!
By one full run I've got all the answers for my questions.
Best pair programming session with AI overall!
GitHub Co-Pilot sucks in comparison to R1!

This shows us that even in corporate monopoly market small companies can make big shifts, bring big difference and value, can innovate, and outperform giants.

My thoughts:
We're all will be replaced by such AI creatures! 😱

Joking! We can collaborate and create beautiful things! World is definitely changing now! We can adapt and adopt these technologies, and use them for the great good! (And I'm still believe in bright future.)

Overall, LLMs as neural network has inputs and outputs, and as an input for now it requires operator, engineer, human. It cannot make goal-setting via prompting itself! (At least for now!)

It's and interesting case and pair programming is so good application for reasoning LLMs and SLMs!

Paper:
https://github.com/deepseek-ai/DeepSeek-R1/blob/main/DeepSeek_R1.pdf

Try it by yourself:

https://chat.deepseek.com

https://github.com/deepseek-ai/DeepSeek-R1?tab=readme-ov-file#6-how-to-run-locally

https://github.com/deepseek-ai/DeepSeek-V3

Weights in tensorlayers HF format:
https://huggingface.co/deepseek-ai/DeepSeek-R1

Inference runners:
https://github.com/vllm-project/vllm

https://docs.vllm.ai/en/latest/serving/distributed_serving.html

https://github.com/sgl-project/sglang


For comparison with accessible alternatives:

LLama 3.1 70B chat service:
https://duck.ai


https://chatgpt.com

https://claude.ai


And try it out to deploy by yourself LLama 3.1/3.3 70B/405B via self-hosted installation with some custom inference runner (llama.cpp for example, or its Rust bindings) or cloud deploy from HuggingFace, and compare:
https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct
https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct
https://huggingface.co/meta-llama/Llama-3.1-405B-Instruct

https://github.com/ggerganov/llama.cpp
https://github.com/mdrokz/rust-llama.cpp
https://github.com/utilityai/llama-cpp-rs


#AI
#AGI
#LLM

Читать полностью…

Technologique

Good scheme for retrieving encryption key from TPM2.0 device and bypass LUKS encryption for disk partitions:

https://oddlama.org/blog/bypassing-disk-encryption-with-tpm2-unlock/

#Linux
#LUKS
#TPM

Читать полностью…

Technologique

https://mjg59.dreamwidth.org/70954.html

Matthew Garrett (kernel developer, FSF 2013 award winner for work on secure boot, TPM support and UEFI BIOS support in Linux kernel and Debian distro) on GPU TEE, TPM and DRM.
Must read for system level developers.

Читать полностью…

Technologique

https://www.intel.com/content/www/us/en/newsroom/news/intel-ceo-news-dec-2024.html

Pat is tech visionary guy, one of the main engineers created x86 and i386 CPU architectures, he had a vision for Intel development as a tech company in current market. This man was one last hope for Intel.
Jim Keller also hop off from Intel.

Not a good trend for Intel.
Probably we won't see a GPU/TPU TEE from Intel on a wide market, in Gaudi3 TPUs and forthcoming GPUs/TPUs.

GPU Arc Alchemist/Arctic Sound/Flex and next GPU Arc Battlemage won't compare with NVidia GPUs by performance, drivers optimization, and AI acceleration support and GPU TEE support, i.e. completely missing the demanding of confidential computing for AI acceleration.

https://www.intel.com/content/www/us/en/developer/articles/technical/finetuning-llms-on-intel-gpus-using-bigdl-llm.html

https://www.intel.com/content/www/us/en/docs/graphics-for-linux/developer-reference/1-0/alchemist-arctic-sound-m.html

#Intel

Читать полностью…

Technologique

Rust is an engineering tool, a really good modern tool (but still just a tool, not an egregor), applied and adopted already in many critical areas (like cyber security, distributed systems, cryptography, decentralised financial systems, AI), a type safe language, which respects safety, and way much better designed than C++, or Go (inconsistent data structures implementation, where behaviour depends on implementation, implicitly satisfied interfaces, is a truly nightmare), or even C. There's no usable system level language with integrated non-optional type safety on a type system level (rather than having all that kind of external tools, like linting, fuzzying, sanitizers, etc.). Probably from modern languages only Mojo and ATS can be accompanying Rust in this group (a type safe language with static typing and static memory management).
And thus learning curve is pretty steep and entry barrier is pretty high for Rust ecosystem.
The C language is a basis, fundamental technology, it's pretty simple, sharp, and dangerous, where freedom == responsibility.
The only cool thing of C++ is SFINAE. But it's a dynamic implicit thing for types coercion, thus completely eliminated by Rust type system.

#Rust
#CPP
#C
#Notes
#Thoughts

Читать полностью…
Подписаться на канал