Hacker News new | past | comments | ask | show | jobs | submit | jdeaton's comments login

Its called recursion

> Doesn't for_tree(...) look a lot nicer and simpler and less error prone than needing to implement a recursive function for each operation you would want to do on a tree?

No it does not


Yeah i vote it should be rebranded “why”

command: why install htop, -> install start: because you wan' it -> done: you got it

Something like that you mean?


I would love if the way to respond "yes" to the CLI is "why not"

Why not.

Obvious joke is obvious... but it seems the "motivation" is to write a project in rust.

Perhaps _why?

0.061 standard deviations? Thats like almost nothing?

Pytorch???


Maybe shes wearing those noise canceling headphones because of her auditory processing condition and not the other way around??

This seems like basic speculative attribution error- no research here.


This would be my guess as it sounds similar to what I experience and have been experiencing for decades long before any king of noise cancelling headphones. I find it very difficult to process noises when there are too many occurring at once and adding voices into the mix will make it look like I am lagging trying to respond. I have no idea what the underlying condition is, but I consider this problem to be called "misophonia". I think it is about not being able to filter sounds well.

Thankfully I'm not regularly in an environment that force me to wear noise cancelling headphones.

Sometimes I believe this was caused by my childhood bedroom being soundproofed, but I still went to noisy public school so who knows.


Something nice about this guide is that it generally transfers to GPU directly thanks to JAX/XLA.


if you're using tpu why are you using pytorch


there is limited TPU support in pytorch via torch_xla


Sounds limited


The interesting thing about this comment is that JAX is actually higher-level even than pytorch generally. Since everything is compiled you just express a logcial program and let the compiler (XLA) worry about the rest.

Are you suggesting that XLA would be where this "lower level" approach would reside since it can do more automatic optimization?


I'm curious, what does paradigmatic JAX look like? Is there an equivalent of picoGPT [1] for JAX?

[1] https://github.com/jaymody/picoGPT/blob/main/gpt2.py


yeah it looks exactly like that file but replace "import numpy as np" with "import jax.numpy as np" :)


JAX has a sub-system called Pallas[1] with a Triton-like programming model and an example implementation of Flash Attention [2]. It is quite fast. On TPUs I've heard that the XLA compiler already emits a flash-attention-like computation graph for a regular JAX implementation of attention so there's no need to have some specialized kernel in that case.

1. https://jax.readthedocs.io/en/latest/pallas/index.html

2. https://github.com/jax-ml/jax/blob/main/jax/experimental/pal...


Why is the study specifically of Turkish students


The paper at https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4895486 says:

> One of the co-authors, Özge Kabakcı, a high school math teacher and former department chair of the math department at our partner Turkish high school, led the development of all session materials.


I guess it would be hard to find a high school with the sample size that you need (thousand) that will agree on collaborating. And in the US every county will have different rules and in terms of math they don't teach it in standard way.

But why Turkish not British or any other place is going to be a question no matter the ___location. But do you really think the results will be significantly different if it is done lets say on Vietnamese students?


Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: