

Were those nonsensical Difference percentages obtained via an existence that claims intelligence by any chance?


Were those nonsensical Difference percentages obtained via an existence that claims intelligence by any chance?


One can use custom viewers via core.pager and interactive.diffFilter in git configuration, not to mention defining custom difftools directly.
I primarily use delta for this (sometimes packaged as git-delta), which itself is implemented in Rust too.
For example, save this as a script called delta-s somewhere in $PATH:
#!/bin/bash
delta -s \
--minus-style='syntax #400000' \
--plus-style='syntax #004000' \
--minus-emph-style='normal #a00000' \
--plus-emph-style='normal #00a000' \
--line-buffer-size=48 \
--max-line-distance=0.8 $@
Then, in ~/.gitconfig, add
[difftool "d-sbs"]
cmd = diff -u $LOCAL $REMOTE | delta-s
And the you can just
git difftool --tool d-sbs HEAD~
You can further create an alias for that too of course.
Didn’t read the whole thing because I had to stop at the right column at the start.
Federated is “decentralized”. The correct word the author is looking for is “distributed”. And even then, direct exclusive P2P is only one form of “distributed”. Hybrid/Multiple approaches are also wide-spread (torrents anyone!).
Not sure how a technical writer gets such a basic aspect wrong.
Also, beyond the closed source aspect, and being a closed up platform in general, Discord was always literal spyware. And pretending like open-source projects who chose to use it didn’t know what they were doing, and glossing over the actions that ranged from collective nagging to almost literal fights in some communities because of such choices, reeks of willful blindness.


It’s laughable before you even get to the code. You know, doing “eval bad” when all the build scripts are written in bash 🤣
There is also no protection for VCS sources (assuming no revision hash is specified) in makepkg (no “locking” with content hash stored). So, if an AUR package maintainer is malicious, they can push whatever they want from the source side. They actually can do that in any case obviously. But with VCS sources, they can do it at any moment transparently. In other words, your primary concern should be knowing the sources are coming from a trustable upstream (and hoping no xz-like fiasco is taking place). Checking if the PLGBUILD/install files are not fishy is the easier part (and should be done by a human). And if you’re using AUR packages to the extent where this is somehow a daunting time-consuming task, then there is something wrong with you in any case.
Edit: That is not to say the author of the tool wouldn’t just fit right in with security theater crowd. Hell, some of them even created whole businesses using not too dissimilar theater components.


So try_from(&**p) is not a code smell/poor form in Rust?
No. It’s how you (explicitly) go from ref to deref.
Here:
p is *p is PathBuf**p is Path (Deref)&**p is &Path.Since what you started with is a reference to a non-Copy value, you can’t do anything that would use/move *p or **p. Furthermore, Path is an unsized type (just like str and []), so you need to reference it (or Box it) in any case.
Another way to do this is:
let p: &Path = p.as_ref();
Some APIs use AsRef in signatures to allow passing references of different types directly (e.g. File::open()), but that doesn’t apply here.


Let’s do this incrementally, shall we?
First, let’s make get_files_in_dir() idiomatic. We will get back to errors later.
fn get_files_in_dir(dir: &str) -> Option<Vec<PathBuf>> {
fs::read_dir(dir)
.ok()?
.map(|res| res.map(|e| e.path()))
.collect::<Result<Vec<_>, _>>()
.ok()
}
Now, in read_parquet_dir(), if the unwraps stem from confidence that we will never get errors, then we can confidently ignore them (we will get back to the errors later).
fn read_parquet_dir(entries: &Vec<String>) -> impl Iterator<Item = record::Row> {
// ignore all errors
entries.iter()
.cloned()
.filter_map(|p| SerializedFileReader::try_from(p).ok())
.flat_map(|r| r.into_iter())
.filter_map(|r| r.ok())
}
Now, let’s go back to get_files_in_dir(), and not ignore errors.
fn get_files_in_dir(dir: &str) -> Result<Vec<PathBuf>, io::Error>
{
fs::read_dir(dir)?
.map(|res| res.map(|e| e.path()))
.collect::<Result<Vec<_>, _>>()
}
fn main() -> Result<(), io::Error> {
let args = Args::parse();
- let entries = match get_files_in_dir(&args.dir)
- {
- Some(entries) => entries,
- None => return Ok(())
- };
-
+ let entries = get_files_in_dir(&args.dir)?;
let mut wtr = WriterBuilder::new().from_writer(io::stdout());
for (idx, row) in read_parquet_dir(&entries.iter().map(|p| p.display().to_string()).collect()).enumerate() {
Now, SerializedFileReader::try_from() is implemented for &Path, and PathBuf derefs to &Path. So your dance of converting to display then to string (which is lossy btw) is not needed.
While we’re at it, let’s use a slice instead of &Vec<_> in the signature (clippy would tell you about this if you have it set up with rust-analyzer).
fn read_parquet_dir(entries: &[PathBuf]) -> impl Iterator<Item = record::Row> {
// ignore all errors
entries.iter()
.filter_map(|p| SerializedFileReader::try_from(&**p).ok())
.flat_map(|r| r.into_iter())
.filter_map(|r| r.ok())
}
let entries = get_files_in_dir(&args.dir)?;
let mut wtr = WriterBuilder::new().from_writer(io::stdout());
- for (idx, row) in read_parquet_dir(&entries.iter().map(|p| p.display().to_string()).collect()).enumerate() {
+ for (idx, row) in read_parquet_dir(&entries).enumerate() {
let values: Vec<String> = row.get_column_iter().map(|(_column, value)| value.to_string()).collect();
if idx == 0 {
wtr.serialize(row.get_column_iter().map(|(column, _value)| column.to_string()).collect::<Vec<String>>())?;
Now let’s see what we can do about not ignoring errors in read_parquet_dir().
This consumes all readers before getting further. So, it’s a behavioral change. The signature may also scare some people 😉
fn read_parquet_dir(entries: &Vec<PathBuf>) -> Result<impl Iterator<Item = Result<record::Row, ParquetError>>, ParquetError> {
Ok(entries
.iter()
.map(|p| SerializedFileReader::try_from(&**p))
.collect::<Result<Vec<_>, _>>()?
.into_iter()
.flat_map(|r| r.into_iter()))
}
How can we combine errors from readers with flat record results?
This is how.
enum ErrorOrRows {
Error(Option<ParquetError>),
Rows(record::reader::RowIter<'static>)
}
impl Iterator for ErrorOrRows {
type Item = Result<record::Row, ParquetError>;
fn next(&mut self) -> Option<Self::Item> {
match self {
Self::Error(e_opt) => e_opt.take().map(Err),
Self::Rows(row_iter) => row_iter.next(),
}
}
}
fn read_parquet_dir(entries: &[PathBuf]) -> impl Iterator<Item = Result<record::Row, ParquetError>>
{
entries
.iter()
.flat_map(|p| match SerializedFileReader::try_from(&**p) {
Err(e) => ErrorOrRows::Error(Some(e)),
Ok(sr) => ErrorOrRows::Rows(sr.into_iter()),
})
}
let mut wtr = WriterBuilder::new().from_writer(io::stdout());
for (idx, row) in read_parquet_dir(&entries).enumerate() {
+ let row = row?;
let values: Vec<String> = row.get_column_iter().map(|(_column, value)| value.to_string()).collect();
if idx == 0 {
wtr.serialize(row.get_column_iter().map(|(column, _value)| column.to_string()).collect::<Vec<String>>())?;
#![feature(gen_blocks)]fn read_parquet_dir(entries: &[PathBuf]) -> impl Iterator<Item = Result<record::Row, ParquetError>> {
gen move {
for p in entries {
match SerializedFileReader::try_from(&**p) {
Err(e) => yield Err(e),
Ok(sr) => for row_res in sr { yield row_res; }
}
}
}
}


NCDC (No Code, Don’t Care)


As with all ads, especially M$ ones…
No Code, Don’t Care
At least if the code was available, I would find out what they mean by “spoofed Mime” and how that attack vector works (Is the actual file “magic” header spoofed, but the file still manages to get parsed with its non-“spoofed” actual format none the less?!, How?).
Also, I would have figured out if this is a new use of “at scale” applied to purely client code, or if a service is actually involved.


dyn compatibility of the trait itself is another matter. In this case, an async method makes a trait not dyn-compatible because of the implicit -> impl Future opaque return type, as documented here.
But OP didn’t mention whether dyn is actually needed or not. For me, dyn is almost always a crutch (exceptions exist).


If I understand what you’re asking…
This leaves out some details/specifics out to simplify. But basically:
async fn foo() {}
// ^ this roughly desugars to
fn foo() -> impl Future<()> {}
This meant that you couldn’t just have (stable) async methods in traits, not because of async itself, but because you couldn’t use impl Trait in return positions in trait methods, in general.
Box<dyn Future> was an unideal workaround (not zero-cost, and other dyn drawbacks). async_trait was a proc macro solution that generated code with that workaround. so Box<dyn Future> was never a desugaring done by the language/compiler.
now that we have (stable) impl Trait in return positions in trait methods, all this dance is not strictly needed anymore, and hasn’t been needed for a while.
I was just referring to the fact that they are macros.
printf uses macros in its implementation.
int
__printf (const char *format, ...)
{
va_list arg;
int done;
va_start (arg, format);
done = __vfprintf_internal (stdout, format, arg, 0);
va_end (arg);
return done;
}
^ This is from glibc. Do you know what va_start and va_end are?
to get features that I normally achieve through regular code in other languages.
Derives expand to “regular code”. You can run cargo expand to see it. And I’m not sure how that’s an indication of “bare bone”-ness in any case.
Such derives are actually using a cool trick, which is the fact that proc macros and traits have separate namespaces. so #[derive(Debug)] is using the proc macro named Debug which happens to generate “regular code” that implements the Debug trait. The proc macro named Debug and implemented trait Debug don’t point to the same thing, and don’t have to match name-wise.
Not sure if you’re talking about the language, or the core/alloc/std libraries, or both/something in-between?
Can you provide specific examples, an which specific languages are you comparing against?


(didn’t read OP, didn’t keep up with chimera recently)
From the top of my head:
The init system. Usable FreeBSD utils instead of busybox overridable by gnu utils (which you will have to do because the former are bare-bones). Everything is built with LLVM (not gcc). Extra hardening (utilizing LLVM). And it doesn’t perform like shit in some multi-threaded allocator-heavy loads because they patch musl directly with mimalloc. It also doesn’t pretend to have a stable/release channel (only rolling).
So, the use of apk is not that relevant. “no GNU” is not really the case with Alpine. They do indeed have “musl” in common, but Chimera “fixes” one of the most relevant practical shortcomings of using it. And finally, I don’t think Chimera really targets fake “lightweight”-ness just for the sake of it.
'0'..'9' (characters in ASCII) are (0+48)..(9+48) when read as integer values.
For readability you can do:
unsigned char zero = '0';
int h = getchar() - zero;
int l = getchar() - zero;
And as I mentioned in another comment, if this was serious code, you would check that both h and l are between 0 and 9.
Note that one of the stupid quirks about C is that char is not guaranteed to be unsigned in certain implementations/architectures. So it’s better to be explicit about expecting unsigned values. This is also why man 3 getchar states:
fgetc() reads the next character from stream and returns it as an unsigned char cast to an int, or EOF on end of file or error.
getchar() is equivalent to fgetc(stdin).


How is this literal joke still getting so much engagement?
nice CLAUDE.md. got it on the contributor list too.


An actually serious project that is not at the “joke” stage. Zero LLM use too:
https://nihav.org/
For audio at least, people should be aware of:
https://github.com/pdeljanov/Symphonia
The majority of actual rustaceans don’t care about these polls (or any “official community” activity for that matter).
If you want actually relevant (and objective) stats, look here.