

To expand: Just configure whatever profile you’re using (dev
, release
, …) to have link time optimization (lto) enabled:
[profile.release]
lto = "fat"
To expand: Just configure whatever profile you’re using (dev
, release
, …) to have link time optimization (lto) enabled:
[profile.release]
lto = "fat"
I’ve had this exact same gripe and can thankfully report that running EarlyOOM has fixed this for me.
And please don’t understand this the wrong way.
Ibis seems like a really cool project but with it being roughly half a year old me and many other people here simply have never heard of it before.
Including even a single short sentence describing what Ibis is in this and future posts helps us find projects that we care about more easily.
And we obviously care about Rust projects, otherwise none of us would be here.
Ibis is a federated online encyclopedia similar to Wikipedia.
This should be the first sentence of the post body.
Cushy is an experimental Graphical User Interface (GUI) crate for the Rust programming language. It features a reactive data model and aims to enable easily creating responsive, efficient user interfaces. To enable easy cross-platform development, Cushy uses its own collection of consistently-styled Widgets.
You can use their online web-editor (similar to OverLeaf for LaTeX) or download the open-source engine and run it locally (there are extensions available for many text editors).
Compared to LaTeX I find it much more comfortable to work with. It comes with sane, modern defaults and doesn’t need any plugins just to generate a (localized) bibliography or include links.
Since Typst is very young compared to LaTeX I’m sure that there are numerous docs / workflows that can’t be reproduced at the moment but if you don’t need some special feature I’d recommend giving it a shot.
There are some experimental models made specifically for use with Home Assistant, for example home-llm.
Even though they are tiny 1-3B I’ve found them to work much better than even 14B general purpose models. Obviously they suck for general purpose questions just by their size alone.
That being said they’re still LLMs. I like to keep the “prefer handling commands locally” option turned on and only use the LLM as a fallback.